Intelligent Computing and Optimization: Pandian Vasant Ivan Zelinka Gerhard-Wilhelm Weber Editors
Intelligent Computing and Optimization: Pandian Vasant Ivan Zelinka Gerhard-Wilhelm Weber Editors
Intelligent Computing and Optimization: Pandian Vasant Ivan Zelinka Gerhard-Wilhelm Weber Editors
Pandian Vasant
Ivan Zelinka
Gerhard-Wilhelm Weber Editors
Intelligent
Computing and
Optimization
Proceedings of the 3rd International
Conference on Intelligent Computing
and Optimization 2020 (ICO2020)
Advances in Intelligent Systems and Computing
Volume 1324
Series Editor
Janusz Kacprzyk, Systems Research Institute, Polish Academy of Sciences,
Warsaw, Poland
Advisory Editors
Nikhil R. Pal, Indian Statistical Institute, Kolkata, India
Rafael Bello Perez, Faculty of Mathematics, Physics and Computing,
Universidad Central de Las Villas, Santa Clara, Cuba
Emilio S. Corchado, University of Salamanca, Salamanca, Spain
Hani Hagras, School of Computer Science and Electronic Engineering,
University of Essex, Colchester, UK
László T. Kóczy, Department of Automation, Széchenyi István University,
Gyor, Hungary
Vladik Kreinovich, Department of Computer Science, University of Texas
at El Paso, El Paso, TX, USA
Chin-Teng Lin, Department of Electrical Engineering, National Chiao
Tung University, Hsinchu, Taiwan
Jie Lu, Faculty of Engineering and Information Technology,
University of Technology Sydney, Sydney, NSW, Australia
Patricia Melin, Graduate Program of Computer Science, Tijuana Institute
of Technology, Tijuana, Mexico
Nadia Nedjah, Department of Electronics Engineering, University of Rio de Janeiro,
Rio de Janeiro, Brazil
Ngoc Thanh Nguyen , Faculty of Computer Science and Management,
Wrocław University of Technology, Wrocław, Poland
Jun Wang, Department of Mechanical and Automation Engineering,
The Chinese University of Hong Kong, Shatin, Hong Kong
The series “Advances in Intelligent Systems and Computing” contains publications
on theory, applications, and design methods of Intelligent Systems and Intelligent
Computing. Virtually all disciplines such as engineering, natural sciences, computer
and information science, ICT, economics, business, e-commerce, environment,
healthcare, life science are covered. The list of topics spans all the areas of modern
intelligent systems and computing such as: computational intelligence, soft comput-
ing including neural networks, fuzzy systems, evolutionary computing and the fusion
of these paradigms, social intelligence, ambient intelligence, computational neuro-
science, artificial life, virtual worlds and society, cognitive science and systems,
Perception and Vision, DNA and immune based systems, self-organizing and
adaptive systems, e-Learning and teaching, human-centered and human-centric
computing, recommender systems, intelligent control, robotics and mechatronics
including human-machine teaming, knowledge-based paradigms, learning para-
digms, machine ethics, intelligent data analysis, knowledge management, intelligent
agents, intelligent decision making and support, intelligent network security, trust
management, interactive entertainment, Web intelligence and multimedia.
The publications within “Advances in Intelligent Systems and Computing” are
primarily proceedings of important conferences, symposia and congresses. They
cover significant recent developments in the field, both of a foundational and
applicable character. An important characteristic feature of the series is the short
publication time and world-wide distribution. This permits a rapid and broad
dissemination of research results.
Indexed by SCOPUS, DBLP, EI Compendex, INSPEC, WTI Frankfurt eG,
zbMATH, Japanese Science and Technology Agency (JST), SCImago.
All books published in the series are submitted for consideration in Web of
Science.
Gerhard-Wilhelm Weber
Editors
Intelligent Computing
and Optimization
Proceedings of the 3rd International
Conference on Intelligent Computing
and Optimization 2020 (ICO 2020)
123
Editors
Pandian Vasant Ivan Zelinka
Department of Fundamental Faculty of Electrical Engineering
and Applied Sciences and Computer Science
Universiti Teknologi Petronas VŠB TU Ostrava
Tronoh, Perak, Malaysia Ostrava, Czech Republic
Gerhard-Wilhelm Weber
Faculty of Engineering Management
Poznan University of Technology
Poznan, Poland
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
v
vi Preface
vii
viii Contents
xix
xx About the Editors
Abstract. The wind speed has a great impact on the overall performance of the
wind energy harvesting system. Due to variable nature of wind speed, the
system is controlled to work only in a specified range of wind speeds to protect
both the generator and turbine from damage. This article presents adaptive
neuro-fuzzy inference system-based control scheme for operation of the system
between the cut in and rated wind speed. By controlling the generator speed to
its optimum value, the generator power and speed fluctuation can be reduced.
A Matlab/Simulink tool is used for the simulation and analysis of the system.
The obtained results indicate that adaptive neuro-fuzzy inference system is an
effective method to control the rotor speed.
1 Introduction
Wind energy technology have shown rapid growth among renewable energy sources in
most parts of the world due to depletion of fossil fuel reserves, rising pollution levels
and worrying changes in the global climate created due to conventional energy sources
[1]. According to Global Wind Energy Council (GWEC) report, around 50 GW of
newly installed wind power was added in 2018 slightly less than that of 2017, bringing
the global total wind power generation capacity to 597 GW [2]. The report indicates
that 2018 was the second year in a row with growing number of new installations for
energy generation.
Wind Energy Harvesting System (WEHS) can operate in both fixed as well as
variable speed mode of operation. In fixed speed wind turbines, the generator rotates at
almost constant speed and frequency for which it is designed regardless of variation in
wind speed [1]. As turbines forced to operate at constant speed, it should be extremely
robust to withstand mechanical stress created due to the fluctuation of wind speeds. On
the other hand, in variable speed wind turbines, the rotor of the generator is allowed to
rotate freely at any speed over a wide range of wind speeds. The generator is directly
connected to the grid in fixed speed operation whereas connected with the help of
power electronic equipment in variable speed system [3].
Thus, it is possible to control the rotor speed by means of power electronics to
maintain the optimum tip speed ratio at all the times under fluctuation of wind speeds to
produce maximum power. Different types of AC generators can be used along with the
modern wind turbine. Researchers classified them into synchronous and asynchronous
(induction) generators. Some are Squirrel-Cage Rotor Induction Generator, Wound-
Rotor Induction Generator, Doubly-Fed Induction Generator, Synchronous Generator
(With external field excitation) and Permanent Magnet Synchronous Generator [3].
Variable Speed Wind Turbine with Permanent Magnet Synchronous Generator
(VSWT-PMSG) is an attractive choice for villages, settlements and remote areas found
at very far distance from grid [4]. VSWT-PMSG can operate close to optimal speed
using maximum power point tracking (MPPT) for various wind-speeds.
This paper has seven sections: the first section contains general introduction about
WEHS and the purpose of this work is mentioned in Sect. 2. Section 3 presents
literature of the control methods applied on WEHS. Section 4 describes the details of
WEHS. Section 5 describes MPPT controller’s used in our work. The model of the
system, simulation results and discussions are given in the Sect. 6 before the conclu-
sion and future scope which is the final part of this paper.
The control system is used in WEHS to produce maximum power and supply the load
with a constant voltage and frequency under change in wind speed and load. WEHS are
controlled to work only in a specified range of wind speeds limited by cut-in and cut-
out speeds. Outside of these limits, the turbine should be stopped to protect both
generator and turbine from damage [5, 6]. The recent advancements in power elec-
tronics and control strategies have made it possible to regulate the voltage of the PMSG
in many different ways. The main factor that affects the performance of the system is
the variation in wind speed which affects power system stability and power quality. The
machine-side a torque/speed controller is applied to improve the performance of the
system at variable wind speed between cut-in and rated speed. Grid-side inverter is also
controlled to keep the DC-link voltage at a constant value and the current injected to
the grid at unity power factor to achieve maximum power delivery to grid as a desired
operating condition. Conventional control methods such as proportional integral
(PI) have proposed by different researchers for the control of wind energy system [7].
However, these methods need exact mathematical knowledge of dynamics system,
which is very often difficult to derive for complex systems. Nowadays to overcome
such problem researchers are tending to use intelligent control methods like fuzzy logic
control (FLC), neural networks (NN), adaptive neuro fuzzy interfacing system (ANFIS)
and genetic algorithm (GA) to control wind speed fluctuations.
In this paper, ANFIS controller is proposed to control rotor speed of WEHS. The
objective of ANFIS controller is to maximize energy production and ensure a con-
tinuous supply of energy to the grid by regulating the turbine speed in such a way that
the optimal tip speed ratio is maintained [8]. Modelling and simulation of the system is
Adaptive Neuro-Fuzzy Inference Based Modeling 5
3 Related Work
The studies related to modelling of WEHS using Adaptive MPPT network is limited
due to complexity of the system and advancement of the control techniques.
Researchers applied different MPPT techniques for control applications. Boobalan et al.
[9] proposed Fuzzy-PI control strategy to model WEHS using PMSG to provide
optimum power to the gird. The proposed vector control provides the desired voltage at
the output of the generator side convertor so as to control the generator speed. With the
vector control scheme, the magnitude, phase voltage and the frequency of the generator
currents are controlled and enable acceptable results in steady state with variation in
wind speed. However, if the wind speed changes significantly from time to time, this
method may require a long searching time to locate the maximum power point. Sarvi
et al. [10] have proposed a maximum power point tracking control algorithm based on
particle swarm optimization (PSO) and Fuzzy logic control for variable speed PMSG
based WEHS. The PSO algorithm is used to determine optimum rotor speed of turbine
generator and its maximum power according to different wind speeds whereas FLC is
used to adjust the duty cycle of boost converter. Ahmed et al. [11] applied fuzzy-PID
controller based MPPT to track the maximum power available from the WEHS and
obtain AC voltage with constant amplitude and frequency. The error between actual
rotor speed of PMSG and estimated value of rotor speed which depends on values of
DC current and voltage at the terminal of boost-converter and parameters of PMSG are
used to obtain maximum power. The rotor speed error is given as input to the fuzzy-
PID controller that controls the duty cycle of pulse width modulation (PWM) generator
and its output is connected to the boost converter. Wafa Torki et al. [12] have applied
vector control strategy using PI controller to model and control the direct-drive WEHS
with PMSG. The actual speed of the generator is compared to its reference value which
is obtained by MPPT method using the optimal tip speed ratio control and gives the
reference current. Marmouh et al. [13] have applied two controllers to produce max-
imum active power of WEHS equipped by PMSG whose stator is connected to the grid
through a back-to-back AC-DC-AC converter. The first one is FLC based MPPT
control algorithm of stator side control by using a hysteresis control and an optimal
generator speed reference which is estimated from different wind speeds for the gen-
eration of maximum active power and the second FLC was applied to grid side control
to ensure a smooth DC voltage between the two converters to its reference value.
Aamer Bilal Asghar et al. [5] have proposed hybrid intelligent learning based ANFIS
for online estimation of effective wind speed from instantaneous values of wind turbine
TSR, rotor speed and mechanical power. The estimated wind speed is used to design
the optimal rotor speed estimator for MPPT of VSWT. Gencer [14] demonstrated fuzzy
logic control system of variable speed WEHS to investigate power flow efficiency of
6 T. Mitiku and M. S. Manshahia
PMSG. Fuzzy logic controller was structured for MPPT to drive the WEHS at the
optimum speed that relates to maximum power at any wind speed. Kesraoui et al. [15]
examined the FLC control of the aerodynamic power for variable-speed wind turbine at
high wind speeds to oversee overabundance power created during high wind speeds
dependent on PMSG associated with the grid through consecutive power converter.
Pitch angle was additionally controlled utilizing fluffy rationale to restrain the aero-
dynamic power and the power on the dc bus voltage. Nadoushan et al. [16] exhibited
optimal torque control (OTC) of stand-alone variable-speed small-scale wind turbine
outfitted with a PMSG and a switch-mode rectifier. Right now, control method is
utilized for variable speed WEHS to remove optimal power from wind and keep the
voltage at 400 V with 50 Hz frequency and produce most extreme power output. The
simulation model of the framework was created utilizing MATLAB/SIMULINK
framework.
The output voltage of the generator is not acceptable due to the variation in amplitude
and frequency because of fluctuation in wind speeds throughout the day. A 1.5 MW
wind turbine is considered in this paper which consists of a 1.5 MW wind turbine and
PMSG connected to AC/DC converter and DC/AC converter modelled by voltage
sources. AC/DC converter is used to convert AC voltage with variable amplitude and
frequency at the generator side to DC voltage at the DC-link voltage. DC-link voltage
should be constant for direct use, storage and for conversion from DC to AC by using
inverter. Thus the obtained DC voltage again converted to AC voltage with constant
amplitude and frequency at the load side for electrical utilization. The back-to-back
PWM converter based power electronic interface is a suitable option for wind-power
applications with PMSG.
1
Pm ¼ CP ðk; bÞqpR2 Vw3 ð1Þ
2
The mechanical torque generated from wind power by wind turbine that will be
transferred through the generator shaft to the rotor of the generator is given by
the tip speed-ratio k that depends on the wind velocity and the rotation speed of the
shaft, xm in rad/s, given by [20–22]
116 21
Cp ðk; bÞ ¼ 0:5716 0:4b 5 e ki þ 0:0068 ð3Þ
ki
with
1 1 0:035
¼ 2 ð4Þ
ki k þ 0:008 b þ 1
Rxm Rnp
k¼ ¼ ð5Þ
Vw 30Vw
where R is blade radius and n is wind turbine rotor speed in revolutions per minute
(rpm). Since there is no gearbox, the shaft speed of the rotor and mechanical generator
speed are the same, again mechanical torque transferred to the generator is also equal to
the aerodynamic torque. The maximum value of Cp is 0.59 which means that the power
extracted from the wind is at all times less than 59% (Betz’s limit), this is because of
the various aerodynamic losses depending on rotor construction [3]. For VSWT the
pitch angle is nearly 0, therefore at, b = 0° the maximum power coefficient is 0.4412.
8 T. Mitiku and M. S. Manshahia
To maintain the tip speed ratio at optimum value, xm changes with the wind speed.
Therefore, to extract the maximum power from the wind the tip speed ratio should be
maintained at optimum value at any wind speed [23]. Hence, the wind turbine can
produce maximum power when the turbine operates at optimum value of Cp which is
Cp-opt. So it is necessary to adjust the rotor speed at optimum value of the tip speed ratio
kopt as shown in Fig. 2 below.
Fig. 2. Cp (k, b) characteristic for different value of the pitch angle b [10]
d 1 Rs Lq
id ¼ Vd id þ Pxm iq ð6Þ
dt Ld Ld Ld
d 1 Rs Ld 1
iq ¼ Vq iq Pxm ð id þ wf Þ ð7Þ
dt Lq Lq Lq Lq
The generator produces an electrical torque, and the difference between the
mechanical torque and the electrical torque determines whether the mechanical system
accelerates, decelerates, or remains at constant speed. The electric torque produced by
the generator is given by [21, 25, 26],
3
T e ¼ P w f i q þ L d Lq i d i q ð8Þ
2
3
Te ¼ Pwf iq ð9Þ
2
The active and reactive powers of PMSG in steady state are given by
Ps ¼ Vd id þ Vq iq ð10Þ
Qs ¼ Vd id Vd iq ð11Þ
Since wind turbine and generator shafts are directly coupled with out the gear box,
there is only one state variable. The mechanical equation of PMSG and WT is given by
dxm 1
¼ ð T e Tm f x m Þ ð12Þ
dt J
d
xe ¼ Pxm ; h ¼ xm ð13Þ
dt
Currently fuzzy logic control has played significant role in the development and design
of many real-time control applications [4]. FLC consists of three important functional
blocks i.e., fuzzification which assign fuzzy variables to the crisp data using mem-
bership function; inference engine that creates fuzzy rules by mapping from input to
output through the use of knowledge base in it and defuzzification which is the reverse
process of fuzzification provides the final output to the plant to be controlled [18].
Different adaptive techniques can easily be implemented in it to increase the perfor-
mance of the network. The ability of the network depends on the quality of the signals
used for training and the performance of the training algorithms and their parameters.
ANFIS is one of the neuro fuzzy networks that give best result in control of wind
energy system. ANFIS is a fuzzy inference system (FIS) whose membership functions
10 T. Mitiku and M. S. Manshahia
and rule base are appropriately tuned (adjusted) by ANN. It takes advantage from both,
the learning capability of ANNs and human knowledge based decision making power
of FIS. It uses a combination of least squares estimation and back propagation for
membership function parameter estimation.
(a)
(b)
Fig. 3. (a) A first-order Sugeno fuzzy model with two–input and two rules; (b) Equivalent ANFIS
architecture [16]
Adaptive Neuro-Fuzzy Inference Based Modeling 11
The sugeno’s fuzzy logic model and the corresponding ANFIS architecture is
shown in Fig. 3.
Assume the FIS under consideration has two inputs x, y and one output fi as shown
in Fig. 3 above. Square node indicates adaptive node whereas circle node indicates
fixed (non-adaptive) nodes. For a first-order Sugeno fuzzy model, a common rule set
with two fuzzy if-then rule is:
Here Oni denotes the output of the ith node (neuron) in layer n, pi, qi, and ri are
consequent parameters of the first order polynomial which are updated during the
learning process in forward pass by least square method [27].
Layer 1. Every node i in this layer is an adoptive node and consists of two inputs as
variables with a node function
O1i ¼ lAi ðxÞ for i ¼ 1; 2 and O1i ¼ lBi2 ðyÞ for i ¼ 3; 4 ð16Þ
where Ai is the linguistic label like big, very small, large, etc. associated with this node
function. lAi(x) is the membership function of Ai and it specifies the degree to which
the given x satisfies the quantifier Ai. It is chosen to be triangular or bell-shaped
membership function. Bell-shaped membership function with maximum equal to 1 and
minimum equal to 0, such as the generalized bell shaped function is selected for our
study.
Layer 2. Every node in this layer is non-adaptive node labeled p which multiplies the
incoming signals and sends the product out to the next layer. For instance,
The obtained result represents the firing strength of a rule. Other T-norm operators
that perform generalized AND can be used as the node function in this layer.
Layer 3. The neurons in this layer is non adoptive labeled by circle node N and
compute the normalized firing strength which is the ratio of firing strength of a rule to
the sum of the firing strengths of all rules. They compute the activation level of each
rule as
wi
O3i ¼ w ¼ ; i ¼ 1; 2: ð18Þ
w1 þ w2
For convenience, outputs of this layer will be called normalized firing strengths.
12 T. Mitiku and M. S. Manshahia
Layer 4. Every node i in this layer is adaptive square node which multiplies the
normalized firing strength of a rule with corresponding first order polynomial function
which produces crisp output
where wi is the output of layer 3, and, pi, qi, and ri are is known as consequent
parameters. This layer is called as defuzzification layer.
Layer 5. The single node in this layer is a fixed node labeled R that computes the
overall output as the summation of all incoming signals and transforms the fuzzy
classification results into a single crisp output and it is called the output layer, i.e.,
X w1 w2
O5i ¼ f ¼ wi fi ¼ f1 þ f2 ð20Þ
w1 þ w2 w1 þ w2
Therefore, when the values of the premise parameters in layer are fixed, the overall
output can be expressed as a linear combination of the consequent parameters in layer 4.
Adaptive network functionally equivalent to a first-order Sugeno fuzzy model is
constructed this way. The ANFIS learning or training algorithm is used to change all
the adjustable parameters to compare ANFIS output with trained data and identify the
parameters in the network. Each training period of the network is divided into two
phases. In the first phase (forward pass), functional signals go forward until layer 4 and
the consequent parameters are adjusted with Least-Squares Method. In the second
phase (backward pass), the error rates propagate backwards and the premise parameters
are updated with gradient descent (back propagation) method [28]. If these parameters
are fixed, ANFIS output is expressed as the summation of all incoming signals to
produce a single crisp output. Thus, a combination of gradient descent and least squares
methods can easily define optimal values for the result parameters, pi, qi, and ri.
The ANFIS based MPPT controller computes the optimum speed for maximum power
point using information on the magnitude and direction of change in power output due
to the change in command speed.
control technique. The relation between blade angular velocity reference xref and wind
speed Vw for constant R and kopt is given by
kopt Vm
xmref ¼ ð21Þ
R
First the wind speed is approximated by the proposed ANFIS based MPPT con-
troller to generate reference speed xmref for the speed control loop of the rotor side
converter control to track maximum power points for the system by dynamically
changing the turbine torque to operate in kopt as shown in Fig. 4. The PI controller
controls the actual rotor speed to the desired value by varying the switching ratio of the
PWM inverter. In the speed control loop, the actual speed of the generator is compared
to its reference value xref that is obtained by the ANFIS based MPPT control kopt which
is defined above. Then, the speed controller will output the reference q-axis current iqref.
The control target of the inverter is the output power delivered to the load [30, 31].
The advantages of ANFIS over the two parts of this hybrid system are: ANFIS uses
the neural network’s ability to classify data and find patterns. Then it develops a fuzzy
expert system that is more transparent to the user and also less likely to produce
memorization errors than neural network. ANFIS removes (or at least reduces) the need
for an expert. Furthermore, ANFIS has the ability to divide the data in groups and adapt
these groups to arrange a best membership functions that clustering the data and
deducing the output desired with minimum epochs [32]. The learning mechanism fine-
tunes the underlying fuzzy inference system. Using a given input/output data set,
Adaptive Neuro-Fuzzy Inference Based Modeling 15
As wind speed changing throughout the day, variable speed based wind power gen-
eration is useful for optimizing power output of wind energy harvesting system using
MPPT methods. This paper presents modeling and simulation of variable speed based
PMSG using ANFIS network. The network is applied to control the speed of the rotor
to adjust it to its maximum value to give maximum power output. Results have shown
that the voltage is near to rated voltage of the generator and speed fluctuations are
reducing. The future work is to control the pitch angle for the speed more than the rated
speed to produce rated power.
16 T. Mitiku and M. S. Manshahia
Acknowledgments. The authors acknowledge Punjabi University Patiala for providing the
internet facilities and the necessary library resources.
References
1. Zhang, J., Xu, S.: Application of fuzzy logic control for grid-connected wind energy
conversion system. In: Dadios, E.P. (ed.) Fuzzy Logic-Tool for Getting Accurate Solutions.
pp. 50–77. IntechOpen (2015) DOI: https://doi.org/10.5772/59923
2. WWEA: Wind power capacity worldwide reaches 597 GW, 50,1 GW added in 2018. World
Wind Energy Association, Brazil (2019)
3. Husain, M.A., Tariq, A.: Modeling and study of a standalone PMSG wind generation system
using MATLAB/SIMULINK. Univ. J. Electr. Electron. Eng. 2(7), 270–277 (2014)
4. Ali, A., Moussa, A., Abdelatif, K., Eissa, M., Wasfy, S., Malik, O.P.: ANFIS based
controller for rectifier of PMSG wind energy conversion system energy conversion system.
In: Proceedings of Electrical Power and Energy Conference (EPEC), pp. 99–103. IEEE,
Calgary (2014)
5. Asghar, A.B., Liu, X.: Adaptive neuro-fuzzy algorithm to estimate effective wind speed and
optimal rotor speed for variable-speed wind turbine. Neurocomputing 272, 495–504 (2017)
6. Pindoriya, R.M., Usman, A., Rajpurohit, B.S., Srivastava, K.N.: PMSG based wind energy
generation system: energy maximization and its control. In: 7th International Conference on
Power Systems (ICPS), pp. 376–381, IEEE, Pune (2017)
7. Khaing, T.Z., Kyin, L.Z.: Control scheme of stand-alone wind power supply system with
battery energy storage system. Int. J. Electr. Electron. Data Commun. 3(2), 19–25 (2015)
8. El-Tamaly, H.H., Nassef, A.Y.: Tip speed ratio and pitch angle control based on ANN for
putting variable speed WTG on MPP. In: 18th International Middle-East Power Systems
Conference (MEPCON), pp. 625–632, IEEE Power & Energy Society, Cairo (2016)
9. Boobalan, M., Vijayalakshmi, S., Brindha, R.: A fuzzy-PI based power control of wind
energy conversion system PMSG. In: International Conference on Energy Efficient
Technologies for Sustainability, pp. 577–583, IEEE, Nagercoil (2013)
10. Sarvi, M., Abdi, S., Ahmadi, S.: A new method for rapid maximum power point tracking of
PMSG wind generator using PSO_fuzzy logic. Tech. J. Eng. Appl. Sci. 3(17), 1984–1995
(2013)
11. Ahmed, O.A., Ahmed, A.A.: Control of Wind Turbine for variable speed based on fuzzy-
PID controller. J. Eng. Comput. Sci. (JECS) 18(1), 40–51 (2017)
12. Torki, W., Grouz, F., Sbita, L.: Vector control of a PMSG direct-drive wind turbine. In:
International Conference on Green Energy Conversion Systems (GECS), pp. 1–6. IEEE,
Hammamet (2017)
13. Marmouh, S., Boutoubat, M., Mokrani, L.: MPPT fuzzy logic controller of a wind energy
conversion system based on a PMSG. In: 8th International Conference on Modelling,
Identification and Control, Algiers, pp. 296–302 (2016)
14. Gencer, A.: Modelling of operation PMSG based on fuzzy logic control under different load
conditions. In: 10th International Symposium on Advanced Topics in Electrical Engineering
(ATEE), pp. 736–739, IEEE, Bucharest (2017)
15. Kesraoui, M., Lagraf, S.A., Chaib, A.: Aerodynamic power control of wind turbine using
fuzzy logic. In: 3rd International Renewable and Sustainable Energy Conference (IRSEC),
Algeria, pp. 1–6 (2015)
Adaptive Neuro-Fuzzy Inference Based Modeling 17
16. Nadoushan, M.H.J., Akhbari, M.: Optimal torque control of PMSG-based stand-alone wind
turbine with energy storage system. J. Electr. Power Energy Convers. Syst. 1(2), 52–59
(2016)
17. Aymen, J., Ons, Z., Nejib, M.M.: Performance assessment of a wind turbine with variable
speed wind using artificial neural network and neuro-fuzzy controllers. Int. J. Syst. Appl.
Eng. Dev. 11(3), 167–172 (2017)
18. Sahoo, S., Subudhi, B., Panda, G.: Torque and pitch angle control of a wind turbine using
multiple adaptive neuro-fuzzy control. Wind Eng. 44(2), 125–141 (2019)
19. Mitiku, T., Manshahia, M.S.: Fuzzy logic controller for modeling of wind energy harvesting
system for remote areas. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.) Intelligent
Computing and Optimization, ICO 2019. Advances in Intelligent Systems and Computing,
vol. 1072. Springer, Cham (2019). https://doi-org-443.webvpn.fjmu.edu.cn/10.1007/978-3-
030-33585-4_4
20. Heidari, M.: Maximum wind energy extraction by using neural network estimation and
predictive control of boost converter. Int. J. Ind. Electron. Control Optim. 1(2), 115–120
(2018)
21. Slah, H., Mehdi, D., Lassaad, S.: Advanced control of a PMSG wind turbine. Int. J. Mod.
Nonlinear Theory Appl. 5, 1–10 (2016)
22. Medjber, A., Guessoumb, A., Belmili, H., Mellit, A.: New neural network and fuzzy logic
controllers to monitor maximum power for wind energy conversion system. Energy 106,
137–146 (2016)
23. Sahu, S., Panda, G., Yadav, S.P.: Dynamic modelling and control of PMSG based stand-
alone wind energy conversion system. In: Recent Advances on Engineering, Technology and
Computational Sciences (RAETCS), pp. 1–6, IEEE, Allahabad (2018)
24. Jain, A., Shankar, S., Vanitha, V.: Power generation using permanent magnet synchronous
generator (PMSG) based variable speed wind energy conversion system (WECS): an
overview. J. Green Eng. 7(4), 477–504 (2018)
25. Elbeji, O., Mouna, B.H., Lassaad, S.: Modeling and control of a variable speed wind turbine.
In: The Fifth International Renewable Energy Congress, pp. 425–429. IEEE, Hammamet
(2014)
26. Sagiraju, D.K.V., Obulesu, Y.P., Choppavarapu, S.B.: Dynamic performance improvement
of standalone battery integrated PMSG wind energy system using proportional resonant
controller. Int. J. Eng. Sci. Technol. 20(4), 1–3 (2017)
27. Petkovic, D., Shamshirband, S.: Soft methodology selection of wind turbine parameters to
large affect. Electr. Power Energy Syst. 69, 98–103 (2015)
28. Oguz, Y., Guney, I.: Adaptive neuro-fuzzy inference system to improve the power quality of
variable-speed wind power generation system. Turk. J. Electr. Eng. Comput. Sci. 18(4),
625–645 (2010)
29. Farh, H.M., Eltamaly, A.M.: Fuzzy logic control of wind energy conversion system.
J. Renew. Sustain. Energy 5(2), 1–3 (2013)
30. Sekhar, V.: Modified fuzzy logic based control strategy for grid connected wind energy
conversion system. J. Green Eng. 6(4), 369–384 (2016)
31. Gupta, J., Kumar, A.: Fixed pitch wind turbine-based permanent magnet synchronous
machine model for wind energy conversion systems. J. Eng. Technol. 2(1), 52–62 (2012)
32. Thongam, J.S., Bouchard, P., Ezzaidi, H., Ouhrouche, M.: Artificial neural network-based
maximum power point tracking control for variable speed wind energy conversion systems.
In: 18th IEEE International Conference on Control Applications, Saint Petersburg (2009)
33. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent Computing & Optimization. Conference
Proceedings ICO 2018. Springer, Cham (2018). ISBN 978-3-030-00978-6
18 T. Mitiku and M. S. Manshahia
34. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent computing & optimization. In: Proceedings
of the 2nd International Conference on Intelligent Computing and Optimization. Springer
(2019). ISBN 978-3-030-33585-4
35. Mitiku, T., Manshahia, M.S.: Artificial neural networks based green energy harvesting for
smart world. In: Somani, A., Shekhawat, R., Mundra, A., Srivastava, S., Verma, V. (eds.)
Smart Systems and IoT: Innovations in Computing. Smart Innovation, Systems and
Technologies, vol. 141. Springer, Singapore. https://doi.org/10.1007/978-981-13-8406-6_4
36. Mitiku, T., Manshahia, M.S.: Fuzzy inference based green energy harvesting for smart
world. In: IEEE International Conference on Computational Intelligence and Computing
Research (ICCIC), pp. 1–4. IEEE, Madurai (2018)
Features of Distributed Energy Integration
in Agriculture
1 Introduction
Usually, the sources of energy supply for all structural units of holdings for elec-
tricity are electric grids (that are on the balance of power grid companies) and for heat
are their own boiler houses and boiler rooms (that are on the balance of heat supply
companies) or in some cases centralized heat networks.
The main problems of energy supply to agricultural holdings as well as all agri-
cultural enterprises, especially with facilities in places far from developed infrastruc-
ture, are:
• low reliability of electricity supply including low reliability of electricity supply to
heat supply facilities (boiler houses) what leads to a decrease in the quality of
products, underproduction, livestock losses and heat supply interruptions of agri-
cultural processes in case of in electricity interruptions of boiler rooms [2–4];
• low energy efficiency of used boiler houses what leads to an increase in the cost of
holding production;
• the lack of sustainable communication channels for the organization of systems for
monitoring the activities of energy supply facilities, energy consumption. It does
not allow efficiently monitoring of energy consumption by facilities [5–7];
• the lack of qualified specialists who can effectively organize interaction with power
sales and power grid companies. First of all, the low qualification of electrical
engineering personnel and engineering workers affects the difficulty of forecasting
consumption, meeting the requirements of energy sales companies to the conditions
of using more favorable price categories, which, in turn, does not allow agricultural
enterprises to reduce the cost of electricity in the structure of agricultural holdings
and hence the cost products [8].
companies. Basically, the proposals concern only the SGF directly and do not consider
its effective integration into the infrastructure of the enterprise taking into account its
features. In combination with the absence of a developed system of SGF operational
services, this makes it difficult to decide on the use of distributed generation.
Significant investments are needed in pre-project preparation of SGF implemen-
tation in agroholdings with unguaranteed validity of the effectiveness of SGF use.
Understanding this, enterprises are reluctant to carry out such work. In addition, a very
small number of examples of successful application of SGF in the agricultural sector
also “scares” agroholdings from their use.
There is a problem of consistency of legislation on the SGF use in terms of
connecting them to general-use electrical networks and using them for power supply to
third-party consumers. The process of joining and determining tariffs are overly
bureaucratic. This imposes difficulties on the effective use of SGF generating capaci-
ties. It is necessary either to unreasonably overestimate the power of the generators or
to use the SGF for incomplete coverage of the capacities of the enterprise. The con-
struction of its own network for the transmission of electricity to third-party consumers
requires the creation of its own network and distribution company or to involve them
from outside, which increases the cost of electricity sold.
It is necessary to coordinate legislation and regulatory documents in terms of the
possibility of using renewable energy sources, installations for processing agricultural
waste.
There is instability in raw materials for use as a fuel for SGF. This is primarily the
case of the use of biogas plants, plants for the incineration of crop waste, the forest
industry, etc. The market dictates the need to adapt to it in terms of production,
therefore, the structure and volume of production wastes may change and become
insufficient for SGF. The use of biomass specifically cultivated for energy needs as a
raw material requires justification since it needs the occupation of farmland areas for
these purposes.
It should be noted that the options of the SGF can vary greatly depending on the type,
quantity and quality of raw materials as well as the needs of the customer. Consider the
options for the structure of the SGF on the example of biogas plants (BGP), as one of
the most promising options for SGF [12, 13].
Individual BGP. Individual BGP is designed to partially or fully meet the needs of
small agricultural enterprises for electricity, fertilizer, cheap heat and in some cases for
complete energy autonomy. The structural diagram of an individual BGP is presented
in Fig. 1.
22 A. V. Vinogradov et al.
Local BGP. Local BGP is biogas plant designed to fully or partially cover the required
capacities of several enterprises connected to a common energy network. The structural
diagram of the local BGP is presented in Fig. 2.
The end products of such BGP are: electricity, biogas, bio-fertilizers used by
agricultural enterprises for technological purposes.
Network BGP. Network BGP is biogas plant intended for the sale of energy resources
(gas/electricity/fertilizers/heat) to an external network. Supply of raw materials for the
network BGP is engaged in one or more agricultural enterprises, the final product of
such the BGP is gas (electricity). The block diagram of the network BGP is presented
in Fig. 3.
At the same time, the property issue of network BGP has 2 solutions:
• BGP is jointly owned by enterprises and supplies its products in the network in the
form of electricity, heat, fertilizer or gas;
• BGP is an independent enterprise purchasing raw materials from enterprises and
selling its products in the appropriate networks. BGP can be specialized and sell
only gas to the gas network what significantly reduces the necessary equipment for
BGP and, accordingly, reduces the cost of its products. Auxiliary product in this
case can be fertilizer.
In the case of an agricultural holding, BGP can also become one of its structures
aimed at solving the tasks of supplying the units with the products they need.
24 A. V. Vinogradov et al.
As for the whole range of SGF, not only BGP, here, first of all, it is necessary to
conduct research aimed at determining the scope of using different types of SGF. For
example, if there are enterprises engaged in pasture cattle breeding in the structure of an
agricultural holding, it is rational to use mobile technological centers (shepherds’
homes on wheels, milking points, haircuts, etc.) equipped with solar batteries as an
energy source. Pumping installations on pastures, or stationary in areas of greenhouse,
livestock farms can be equipped with both solar and wind power plants. There are
promising SGF options for enterprises in the presence of a natural gas network that
uses gas to generate heat and electricity. But all these options require careful scientific
study with the development of practical recommendations on the required SGF
capacities, their choice in different climatic zones and with different specializations of
farms and other solutions.
Various promising options for the use of individual and local SGF in agroholdings
are listed in Table 1.
Table 1. Options for the use of individual and local SGF in agroholdings
SGF type Scope of perspective application Expected key results
in agroholding structure
Wind power plant • Power supply and • Reducing the cost of pumping
(or) mechanization of pumping water;
stations; • Reducing the cost of power
• Power supply of individual supply to individual objects
objects remote from the (including by reducing the cost
infrastructure of connecting to the network),
for example, at stationary points
on remote pastures
Biogas plant • Gas supply to facilities of • Reducing the cost of gas
different purposes; supply;
• Heating of livestock facilities • Reducing heating costs;
located near buildings of • Reducing the cost of power
various purposes, supply to individual objects
• Recycling of enterprise waste; (including by reducing the cost
• In some cases, the power supply of connecting to the networks);
of individual objects; • Reducing the cost of fertilizers;
• Effective use of fallow lands
(when growing energy plants)
or use of land for raw materials
of BSU in crop rotation;
• Reduction of fines for
environmental impact
(continued)
Features of Distributed Energy Integration in Agriculture 25
Table 1. (continued)
SGF type Scope of perspective application Expected key results
in agroholding structure
Solar power plant • Power supply to pumping • Reducing the cost of pumping
stations; water;
• Power supply (electricity and • Reducing the cost of electricity
heat supply) to individual supply to individual objects
objects remote from the (including by reducing the cost
infrastructure; of connecting to the network),
• Power supply to mobile for example, at stationary points
technological points (housing on remote pastures;
for shepherds on wheels, mobile • Improving the comfort of
milking stations, etc.); workplaces of mobile points
• Power supply to individual • Reducing the cost of creating
loads in enterprises (emergency backup power lines;
lighting, back-up power for the • Reducing the cost of thermal
responsible electrical receivers, energy
etc.);
• Drying of grain, energy supply
to storage facilities
Power plant on • Heating of the enterprise • Reducing the cost of electricity
natural gas (gas facilities located near buildings supply to individual objects
turbine, gas piston, for various purposes, (including by reducing the cost
etc.) • Power supply to the enterprise of connecting to the network);
facilities; • Reducing the cost of thermal
• The use of heat in technological energy
processes
Heat generator on • Heating of livestock facilities; • Reducing the cost of thermal
the enterprise waste • The use of heat in technological energy
processes
In addition to those indicated in Table 1 there can be other SGF as well as the
effects of their use. The final assessment in any case should be carried out taking into
account the characteristics of a particular farm. In the countries as a whole, it is rational
to formulate a program to investigate the potential of SGF use in the structure of
agriculture. It should provide for the solution of legislative and regulatory issues as well
as technical solutions, solutions for the SGF service infrastructure, organization of
advisory services and other aspects of SGF, for example, the use of the principles of
intelligent electrical networks and microgrids described in [14–16] when creating
projects of distributed generation.
4 Conclusions
1. Agricultural holdings are objectively interested in the implementation of distributed
generation projects in the case of obtaining results from these projects, which
consist in reducing the cost of production.
26 A. V. Vinogradov et al.
References
1. Epshtein, D., Hahlbrock, K., Wandel, J.: Why are agroholdings so pervasive in Russia’s
Belgorod oblast’? Evidence from case studies and farm-level data. Post-Communist Econ.
25(1), 59–81 (2013)
2. Vinogradov, A., Vinogradova, A., Bolshev, V., Psarev, A.I.: Sectionalizing and redundancy
of the 0.38 kV ring electrical network: mathematical modeling schematic solutions. Int.
J. Energy Optim. Eng. (IJEOE) 8(4), 15–38 (2019)
3. Vinogradov, A., Vasiliev, A., Bolshev, V., Vinogradova, A., Kudinova, T., Sorokin, N.,
Hruntovich, N.: Methods of reducing the power supply outage time of rural consumers. In:
Kharchenko, V., Vasant, P. (eds.) Renewable Energy and Power Supply Challenges for
Rural Regions, pp. 370–392. IGI Global (2019)
4. Tikhomirov, D.A., Kopylov, S.I.: An energy-efficient electric plant for hot steam and water
supply of agricultural enterprises. Russ. Electr. Eng. 89(7), 437–440 (2018)
5. Bolshev, V.E., Vinogradov, A.V.: Obzor zarubezhnyh istochnikov po infrastrukture
intellektual'nyh schyotchikov [Overview of foreign sources on the infrastructure of smart
meters]. Bull. South Ural State Univ. Ser. Energy 18(3), 5–13 (2018)
6. Sharma, K., Saini, L.M.: Performance analysis of smart metering for smart grid: an
overview. Renew. Sustain. Energy Rev. 49, 720–735 (2015)
7. Kabalci, Y.: A survey on smart metering and smart grid communication. Renew. Sustain.
Energy Rev. 57, 302–318 (2016)
8. Vinogradov, A.V., Anikutina, A.V.: Features in calculations for electric energy of
consumers with a maximum power over 670 kW and a computer program for selecting
the optimal price category [Osobennosti v raschetah za elektroenergiyu potrebitelej s
maksimal’noj moshchnost’yu svyshe 670 kVt i komp’yuternaya programma dlya vybora
optimal’noj cenovoj kategorii]. Innov. Agric. 2, 161–169 (2016)
9. Litti, Y., Kovalev, D., Kovalev, A., Katraeva, I., Russkova, Y., Nozhevnikova, A.:
Increasing the efficiency of organic waste conversion into biogas by mechanical pretreatment
in an electromagnetic mill. In: Journal of Physics: Conference Series, vol. 1111, no. 1 (2018)
10. Panchenko, V., Kharchenko, V., Vasant, P.: Modeling of solar photovoltaic thermal
modules. In: Vasant, P., Zelinka, I., Weber, GW. (eds.) Intelligent Computing &
Optimization. ICO 2018. Advances in Intelligent Systems and Computing, vol. 866.
pp. 108–116. Springer, Cham (2019)
11. Daus, Y., Kharchenko, V.V., Yudaev, I.V.: Managing Spatial Orientation of Photovoltaic
Module to Obtain the Maximum of Electric Power Generation at Preset Point of Time. Appl.
Solar Energy 54(6), 400–405 (2018)
Features of Distributed Energy Integration in Agriculture 27
12. Gladchenko, M.A., Kovalev, D.A., Kovalev, A.A., Litti, Y.V., Nozhevnikova, A.N.:
Methane production by anaerobic digestion of organic waste from vegetable processing
facilities. Appl. Biochem. Microbiol. 53(2), 242–249 (2017)
13. Kovalev, A., Kovalev, D., Panchenko, V., Kharchenko, V., Vasant, P.: System of
optimization of the combustion process of biogas for the biogas plant heat supply. In:
Vasant, P., Zelinka, I., Weber, G.W. (eds.) Intelligent Computing and Optimization. ICO
2019. Advances in Intelligent Systems and Computing, vol. 1072. Springer, Cham (2019)
14. Kharchenko, V., Gusarov, V., Bolshev, V.: Reliable electricity generation in RES-based
microgrids. In: Alhelou, H.H., Hayek, G. (eds.) Handbook of Research on Smart Power
System Operation and Control, pp. 162–187. IGI Global (2019)
15. Vinogradov, A., Bolshev, V., Vinogradova, A., Kudinova, T., Borodin, M., Selesneva, A.,
Sorokin, N.: A system for monitoring the number and duration of power outages and power
quality in 0.38 kV electrical networks. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.)
Intelligent Computing & Optimization. ICO 2018. Advances in Intelligent Systems and
Computing, vol. 866, p. 10. Springer, Cham (2019)
16. Vinogradov, A., Vasiliev, A., Bolshev, V., Semenov, A., Borodin, M.: Time factor for
determination of power supply system efficiency of rural consumers. In: Kharchenko, V.,
Vasant, P. (eds.) Handbook of Research on Renewable Energy and Electric Resources for
Sustainable Rural Development, pp. 394–420. IGI Global (2018)
Concept of Multi-contact Switching System
1 Introduction
At present, the development of “smart power grids” is a global trend related to the fact
that existing 0.4 kV distribution power grids are characterized by low reliability, sig-
nificant energy losses and insufficiently high power quality. The main reasons for this
are the oversized length of power lines, the insufficient degree of network automation
as well as the construction of networks using radial schemes which does not allow for
backup power supply to consumers. The most effective way to increase the power
supply system reliability is to connect distributed generation, for example, biogas
plants, solar panels, etc. [1–3], enabling these networks to work as microgrids [4]. The
joint work of distributed generation sources within the microgrid requires complex
automation. It consists in equipping the electric network with intelligent devices that
allow analyzing the operating modes and automatically reconfiguring the network to
localize the place of damage and restore power to consumers connected to undamaged
network sections.
Distributed automation of 0.4 kV electric networks should be carried out taking
into account the fact that these networks can have a large number of outgoing power
lines from the main transmission lines and there is a possibility to connect small
generation sources directly to 0.4 kV networks. In accordance with this, the concept of
the development of electric networks using multi-contact switching systems (MCSS) is
proposed in this paper.
2 MCSS
supply in terms of the number of power outages and damage to network ele-
ments; monitoring the power quality; monitoring the energy efficiency of the
network (energy losses, other energy efficiency indicators); monitoring other
parameters of the network and the overall power supply system depending on
the capabilities. The monitoring system can be combined with control and
accounting systems.
– The accounting system performs the functions: accounting for the amount of
electricity, adjusting the electricity cost depending on its quality or reliability
indicators of the power supply (for example, if the contractual terms for relia-
bility are violated, the cost decreases); accounting for the number and duration
of power supply interruptions; accounting of the number of the network
equipment work; accounting of other data. The accounting system can be
combined with control and monitoring systems.
– The control system manages all network elements depending on its operating
modes, specified switching requirements, received data from control, monitoring
and accounting systems in order to improve the energy efficiency of the network,
increase the power supply reliability, etc.
• Creation of database systems and information processing comprising:
– Databases of consumers connected to the network with their characteristics,
parameters of operating modes;
– Database of equipment installed on the network with the characteristics and
parameters of the modes of operation;
– Database of power supply reliability indicating the characteristics of power
supply interruptions, equipment damageability;
– Database of power quality with an indication of the quality parameters at dif-
ferent points of the network;
– Database of technical connections with an indication of the connection char-
acteristics, terms of implementation, etc.
The information collected in the database should serve as the basis for making
forecasts of electricity consumption, accident rates, etc. It is also used by control
systems that change the network configuration in various modes, switching, shutting
down equipment, etc.
• Ensuring the possibility of using various energy sources including renewable ones
(RES) and energy storage devices both in parallel with centralized systems and
without being switched on for parallel operation.
• Providing the ability to automatically change the network scheme in the event of
various situations.
• The possibility of organizing microgrids with different energy sources, operating in
parallel or in isolation from each other for different consumers.
Figure 1 shows an example of a distribution smart grid containing a transformer
substation (TS) and several renewable energy sources (RES), various types of switching
equipment. It is shown that the action of each device installed in the grid should be
monitored by monitoring, accounting and control systems. Each unit must be remotely
Concept of Multi-contact Switching System 31
controlled. Communication channels in the network can be organized with different data
transfer technology (JPS, JPRS, radio frequencies, PLC modems, etc. [8–10]).
The following monitoring, accounting and control systems are installed at trans-
former substations and renewable energy sources:
• Systems (sets of equipment) of determining the damage places in power trans-
mission lines and transformer substations – DDP [11];
• Systems (sets of equipment) for regulating the power quality indexes (including
means of adaptive automatic voltage regulation, for example [12]) - RPQI;
• Systems (sets of equipment) of substation automation (Automatic load transfer,
Automatic reclosing, Automatic frequency load shedding, relay protection, etc.) -
SAU (substation automation unit) [13];
• Systems (sets of equipment) of automatic reactive power compensation (regulation)
– RPC [14, 15];
• Advanced metering infrastructure - AMI. The AMI TS takes into account the
operation of the automation systems of the TS, operating modes, transformer
loading as well as carry out electricity metering for each outgoing line, at the inputs
of the transformer on the high and low sides [16–18];
• Other systems as they are developed.
Information from all specified and prospective systems should be transmitted to
information processing and control units having a communication channel between
themselves as well as to a control center (it can be the dispatching office of the electric
grid companies). This allows remote control of TS, RES and, if necessary, remote
control.
The inputs of all consumers are equipped with a AMI system allowing to determine
the values of electric power losses, interruptions in power supply, monitor the operation
of protective equipment and automation of the consumer network, detect unauthorized
electrical equipment (for example, a welding transformer), detect unauthorized con-
nection of generators without compliance safety rules, etc. In addition, consumers AMI
allows for electricity cost adjustment depending on its quality, on the number and
duration of interruptions in power supply and other factors [19]. It must be possible to
prepay electricity and other promising functions.
Power lines regardless of their performance (cable or air) are equipped with remote
monitoring systems for technical condition (insulation condition, overgrowth, incli-
nation of supports, etc.) transmitting data on the power transmission line element state
to IPCU and, respectively, to the SCC.
In the event of an emergency at one of the transmission line sections, automatic
control of the network switching equipment is based on the data received from the
control, monitoring and accounting systems. That is, the network configuration changes
due to switching contacts of the MCSS, sectionalizing points (SP), universal section-
alizing points (USP), points of automatic switching on the reserve (ASR) in such a way
as to isolate the damaged area and provide power to consumers in the backup areas.
The example shown in Fig. 1 does not exhaust all the possibilities of using the
presented systems depending on the required reliability parameters. The principles of
construction of the network and the types of switching devices presented in the Fig. 1
can also be used in networks of other voltage classes.
32 A. V. Vinogradov et al.
Fig. 1. An example of the use of multi-contact switching systems for the development of
microgrids containing renewable energy sources.
3 Discussion
The intelligent electrical network based on the use of the multi-contact switching
system allows to perform the following basic functions:
• To implement switching algorithms and automatic commands for switching in
networks (to perform network configuration changes, highlighting damaged areas,
etc.);
Concept of Multi-contact Switching System 33
• To “see” electricity loss in real time with the ranking of technological, commercial
while for power transformers to “see” the loss of idling and short circuit;
• To develop recommendations for reducing losses, optimal network configuration;
• To automatically change the electricity cost depending on its quality, power supply
interruption time and other parameters;
• To develop recommendations for changing the technological connection conditions
(change in capacity, change in the scheme, etc.);
• To obtain the calculated parameters of power lines and equipment, build mathe-
matical models of the network;
• To receive aggregate load schedules daily, monthly, annual …;
• To develop recommendations on the volumes and parameters of electricity storage,
modes of operation of balancing power plants;
• To obtain reliability indicators of the network and equipment installed in it with an
analysis of the most/least reliable network elements, brands and types of equipment,
issuing recommendations for selection, maintenance and repair;
• To develop recommendations for the selection and configuration of protection and
automation;
• To receive a mapping of the state of switching devices, recommendations for the
development and installation of new devices;
• To receive diagnostic parameters of the equipment and recommendations on the
timing of maintenance, repair, replacement, forecast accidents;
• To get other results based on the system capabilities.
The implementation of such electrical network intellectualization systems will
allow using technical and economic mechanisms to improve the efficiency of power
networks and power supply systems in general. It is possible to automate the imple-
mentation of justifications for the use of microgrids and networks with varying degrees
of automation, form multi-year databases on all parameters of power supply system
functioning, predict the operation modes of electrical networks, generating facilities,
current trends in the development of networks, electricity market. All this allows
integrating power supply systems, intelligent electrical networks into the digital
economy, creating new markets for equipment, communications equipment and soft-
ware for intelligent electrical networks and microgrids, and markets for electricity
services. The creation of new Internet equipment and Internet services based on the
capabilities of managing network equipment, regulating power quality, power supply
reliability, electricity storage, and the use of renewable energy sources are also ele-
ments of the digital economy involving the capabilities of intelligent electrical
networks.
One of the main advantages of such a concept of building autonomous networks
and networks connected to centralized systems is the possibility of flexible changes in
the power supply scheme. This means that if there is a shortage of power from one of
the energy sources or if there is damage to one of the network sections, it is possible to
automatically switch to another power source. It is also possible to transfer surplus to
the centralized network solutions for coding situations in the electrical network with the
peaks of energy production on renewable energy sources [4]. Solutions have been
34 A. V. Vinogradov et al.
developed for partitioning and backing up electrical networks [20], for improving the
quality of electricity and regulating the voltage in electrical networks [21].
In addition, the implementation of these systems will significantly improve elec-
trical safety in the operation of electrical networks. Receiving information about the
state of the power supply system and emergency conditions in it allows to quickly and
remotely localize the accident site or a section of the network where there is a danger of
electric shock, for example, when a wire break.
4 Conclusions
The developed concept of building intelligent electrical networks based on the use of
multi-contact switching systems allows to automatically change the network configu-
ration when using different energy sources, monitor the network situation including the
parameters of the equipment technical condition, the parameters of the network
operating modes, highlighting accident modes, manage switching equipment. This
makes it possible to use several power sources in the network, including renewable
ones which can work both in parallel and in isolation from each other.
References
1. Litti, Y., Kovalev, D., Kovalev, A., Katraeva, I., Russkova, Y., Nozhevnikova, A.:
Increasing the efficiency of organic waste conversion into biogas by mechanical pretreatment
in an electromagnetic mill. In: Journal of Physics: Conference Series, vol. 1111, no. 1 (2018)
2. Panchenko, V., Kharchenko, V., Vasant, P.: Modeling of solar photovoltaic thermal
modules. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.) Intelligent Computing &
Optimization. ICO 2018. Advances in Intelligent Systems and Computing, vol. 866.
pp. 108–116. Springer, Cham (2019)
3. Daus, Y., Kharchenko, V.V., Yudaev, I.V.: Managing spatial orientation of photovoltaic
module to obtain the maximum of electric power generation at preset point of time. Appl.
Solar Energy 54(6), 400–405 (2018)
4. Kharchenko, V., Gusarov, V., Bolshev, V.: Reliable electricity generation in RES-based
microgrids. In: Alhelou, H.H., Hayek, G. (eds.) Handbook of Research on Smart Power
System Operation and Control, pp. 162–187. IGI Global (2019)
5. Vinogradov, A.V., Vinogradova, A.V., Bolshev, V.Ye., Lansberg, A.A.: Sposob
kodirovaniya situacij v elektricheskoj seti, soderzhashchej mul'tikontaktnye kommuta-
cionnye sistemy i vozobnovlyaemye istochniki energii [A way of coding situations in an
electric network containing multi-contact switching systems and renewable energy sources].
Bull. Agric. Sci. Don 2(46), 68–76 (2019)
6. Vinogradov, A.V., Vinogradova, A.V., Marin, A.A.: Primenenie mul’tikontaktnyh kom-
mutacionnyh sistem s mostovoj skhemoj i chetyr’mya vyvodami v skhemah elek-
trosnabzheniya potrebitelej i kodirovanie voznikayushchih pri etom situacij [Application
of multi-contact switching systems with a bridge circuit and four outputs in consumer power
supply circuits and coding of situations arising from this]. Bull. NIIEI 3(94), 41–50 (2019)
Concept of Multi-contact Switching System 35
7. Vinogradov, A., Vasiliev, A., Bolshev, V., Vinogradova, A., Kudinova, T., Sorokin, N.,
Hruntovich, N.: Methods of reducing the power supply outage time of rural consumers. In:
Kharchenko, V., Vasant, P. (eds.) Renewable Energy and Power Supply Challenges for
Rural Regions, pp. 370–392. IGI Global (2019)
8. Bolshev, V.E., Vinogradov, A.V.: Perspektivnye kommunikacionnye tekhnologii dlya
avtomatizacii setej elektrosnabzheniya [Promising communication technologies for automa-
tion of power supply networks]. Bull. Kazan State Power Eng. Univ. 11(2), 65–82 (2019)
9. Ancillotti, E., Bruno, R., Conti, M.: The role of communication systems in smart grids:
architectures, technical solutions and research challenges. Commun. Technol. 36, 1665–
1697 (2013)
10. Khan, R.H., Khan, J.Y.: A comprehensive review of the application characteristics and traffic
requirements of a smart grid communications network. Comput. Netw. 57, 825–845 (2013)
11. Vinogradov, A., Bolshev, V., Vinogradova, A., Kudinova, T., Borodin, M., Selesneva, A.,
Sorokin, N.: A system for monitoring the number and duration of power outages and power
quality in 0.38 kV electrical networks. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.)
Intelligent Computing & Optimization. ICO 2018. Advances in Intelligent Systems and
Computing, vol. 866, p. 10. Springer, Cham (2019)
12. Vinogradov, A., Vasiliev, A., Bolshev, V., Semenov, A., Borodin, M.: Time factor for
determination of power supply system efficiency of rural consumers. In: Kharchenko, V.,
Vasant, P. (eds.) Handbook of Research on Renewable Energy and Electric Resources for
Sustainable Rural Development, pp. 394–420. IGI Global (2018)
13. Abiri-Jahromi, A., Fotuhi-Firuzabad, M., Parvania, M., Mosleh, M.: Optimized sectional-
izing switch placement strategy in distribution systems. IEEE Trans. Power Deliv. 27(1),
362–370 (2011)
14. Singh, B., Saha, R., Chandra, A., Al-Haddad, K.: Static synchronous compensators
(STATCOM): a review. IET Power Electron. 2(4), 297–324 (2009)
15. Haque, M.H.: Compensation of distribution system voltage sag by DVR and D-STATCOM.
In: 2001 IEEE Porto Power Tech Proceedings (Cat. No. 01EX502), vol. 1. IEEE (2001)
16. Bolshev, V.E., Vinogradov, A.V.: Obzor zarubezhnyh istochnikov po infrastrukture
intellektual'nyh schyotchikov [Overview of foreign sources on the infrastructure of smart
meters]. Bull. South Ural State Univ. Ser. Energy 18(3), 5–13 (2018)
17. Sharma, K., Saini, L.M.: Performance analysis of smart metering for smart grid: an
overview. Renew. Sustain. Energy Rev. 49, 720–735 (2015)
18. Kabalci, Y.: A survey on smart metering and smart grid communication. Renew. Sustain.
Energy Rev. 57, 302–318 (2016)
19. Vinogradov, A., Borodin, M., Bolshev, V., Makhiyanova, N., Hruntovich, N.: Improving the
power quality of rural consumers by means of electricity cost adjustment. In: Kharchenko,
V., Vasant, P. (eds.) Renewable Energy and Power Supply Challenges for Rural Regions,
pp. 31–341. IGI Global (2019)
20. Vinogradov, A., Vinogradova, A., Bolshev, V., Psarev, A.I.: Sectionalizing and redundancy
of the 0.38 kV ring electrical network: mathematical modeling schematic solutions. Int.
J. Energy Optim. Eng. (IJEOE) 8(4), 15–38 (2019)
21. Vinogradov, A., Vinogradova, A., Golikov, I., Bolshev, V.: Adaptive automatic voltage
regulation in rural 0:38 kV electrical networks. Int. J. Emerg. Electric Power Syst. 20(3)
(2019)
The Design of Optimum Modes of Grain
Drying in Microwave–Convective Effect
Dmitry Budnikov(&)
1 Introduction
The high energy intensity of post-harvest grain processing dictates the need to develop
new equipment that reduces energy costs for post-harvest processing [2, 5, 7]. Many
researchers note the possibility of reducing these energy costs due to the use of elec-
trophysical factors [1, 3, 4, 6, 8, 9]. These factors include ultrasound, ozone, aeroions,
infrared field, microwave exposure and others. In order to reduce costs while ensuring
the quality indicators, it is also necessary to develop modes and equipment for optimal
management of post-harvest processing plants. At the same time, it is worth consid-
ering the possibility of managing according to the criteria of minimum energy con-
sumption and minimum processing time.
2 Main Part
2.1 Research Method
A number of experimental studies are required to obtain the desired data. In this work,
the research was carried out on installations of two types. In the first case, the instal-
lation contained one microwave power source (a 900 W magnetron) acting on a sta-
tionary grain layer with a volume of 0.015 m3. In the second case, the installation
contained six similar sources located in three levels of height with two opposite
magnetrons at the level. In this case, the grain layer was moved vertically, and the
volume of simultaneously processed material was 0.1 m3.
At this stage, the planning of the screening experiment is aimed at searching for
energy-efficient grain drying modes using electrophysical factors, such as microwave,
for different states of the grain layer, taking into account the form of moisture coupling
in the grain. It is obvious that natural drying modes will have the lowest energy
consumption, but in this case the final material may be damaged due to limited time of
safe storage. The mode of constant and pulsed microwave exposure is used. Table 1
shows the factors of the screening experiment and their variation levels. The response
function is the cost of electrical energy for the drying process.
Since the screening experiment was carried out on an installation that allows
processing a small volume of grain mass, it makes sense only to dependence of the
relative deviations of energy intensity from the assumed initial one.
According to the results of the screening experiment, the mode with a thin layer of
grain mass and constant microwave exposure has the lowest energy consumption. At
the same time, the highest energy intensity is observed in the mode with pulsed
microwave exposure to the suspended layer that is 2.4 times higher than the cost in the
least energy-intensive mode. Despite this, it is worth considering that a significant
amount of energy (about 50%) was spent on creating a fluidized layer.
Thus, in the subsequent stages, it is worth dividing the total energy spent on the
process and the energy spent on the impact factor. In addition, the volume of the
processed material must be increased while maintaining the uniformity of the field
distribution in the processed material.
The measurement of grain material humidity by rapid assessment devices in the
process of microwave exposure gives false results associated with changes in the shape
of the moisture coupling and the properties of the material. Since that, the measurement
of humidity reduction is carried out by weighing the processed volume of the material.
38 D. Budnikov
The drying curves were taken to get the required data. After the data is approxi-
mated by a standard parametric model or a user-defined model, the approximation
quality can be evaluated both graphically and using various approximation fitness
criteria: SSE (least squares), R-square (R-square criterion), Adjusted R-square (ad-
justed R-square), RSME (root mean square error). In addition, one can calculate
confidence intervals for the found values of model parameters that correspond to
different probability levels, and confidence bands for approximation and data that also
correspond to different probability levels. The dependence of grain moisture change, W,
%, on the drying time, t, min, can be represented by the following equation:
W ¼ a ebt þ c; ð1Þ
The data logged during the experiments were processed directly in the SCADA
system. In addition, data downloaded from log files can be loaded into application
software packages such as Statistica, Matlab, Excel, etc., for further statistical and
regression processing. Table 3 partially presents the results of experimental studies of
the efficiency of a laboratory installation containing a single source of microwave
power according to the selected optimality criteria.
Table 3. (continued)
# Culture Moisture, Moisture q, MJ/kg of Optimality
W, % removal, DW, evaporated criterion
% moisture
5 Barley 16 1.5 7.6 Maximum
6 20 2 4.8 performance
7 16 1 6.3 Minimum energy
8 20 1 3.7 consumption
9 Sunflower 9 1.5 5.8 Maximum
10 12 2 3.9 performance
11 9 1 5.2 Minimum energy
12 12 1 3.6 consumption
2.2 Experiment
The modes presented in Table 4 was considered in the research work of the microwave-
convective module. The operating mode of sources of microwave power are: 0 -
without using the microwave; ½ - sources of microwave power work intermittently
(this mode was implemented as 10 s with microwave treating and 10 s without the
one); ¼ - sources of microwave power work intermittently (this mode was implemented
as 5 s with microwave treating and 15 s without the one).
In the course of experimental studies, wheat drying curves were taken from the
initial humidity of 20% [10, 11]. Further, according to these data, the dependence of the
rate of moisture removal in the drying process depending on the test mode of operation
of the unit was studied. Figure 1 shows the drying curves obtained as a result of
experimental studies. The current rate of moisture removal was also calculated from the
current humidity of wheat from the test mode of operation of the plant. As a result,
dependences of energy consumption for drying (evaporation of 1 kg of moisture) of
wheat were obtained depending on the implemented mode of heat treatment.
40 D. Budnikov
22
21
1 2 3 4
20
19 5 6 7 8
Moisture, W, %
18
17 9 10
16
15
14
13
0 20 40 60 80 100 120 140 160
Drying time, τ, min
Both the rate of moisture removal and energy costs significantly depend on both the
operating mode of the equipment and the current humidity.
These results allow to refine the developed models of equipment management and
implement processing control according to the specified optimality criteria.
3 Conclusions
The following conclusions can be drawn from the results of the experiment:
1. The highest energy intensity is observed in the drying mode with the pulse effect of
the microwave field on the suspended grain layer and is 2.4 times higher than the
cost in the least energy-intensive mode.
2. It should be taken into account that the application of the drying mode in the
fluidized layer has a very limited application in the processing of cereals, and
significant energy (about 50%) are spent on creating a fluidized layer.
3. When considering the energy intensity of drying from the point of view of prof-
itability, it is necessary to take into account the cost of energy carriers and sources
of generation (conversion) of energy.
4. The use of microwave fields allows to intensify the drying process in areas of
humidity close to the standard by 3–4 times.
5. It is advisable to apply this technology for drying grain at moisture levels close to
the standard one (from 17 to 14% for wheat).
References
1. Agrawal, S., Raigar, R.K., Mishra, H.N.: Effect of combined microwave, hot air, and
vacuum treatments on cooking characteristics of rice. J. Food Process Eng. e13038 (2019).
https://doi.org/10.1111/jfpe.13038
2. Ames, N., Storsley, J., Thandapilly, S.J.: Functionality of beta-glucan from oat and barley
and its relation with human health. In: Beta, T., Camire, M.E. (eds.) Cereal grain-based
functional foods, pp. 141–166. Royal Society of Chemistry, Cambridge (2019)
3. Basak, T., Bhattacharya, M., Panda, S.: A generalized approach on microwave processing
for the lateral and radial irradiations of various groups of food materials. Innov. Food Sci.
Emerg. Technol. 33, 333–347 (2016)
4. Dueck, C., Cenkowski, S., Izydorczyk, M.S.: Effects of drying methods (hot air, microwave,
and superheated steam) on physicochemical and nutritional properties of bulgur prepared
from high-amylose and waxy hull-less barley. Cereal Chem. 97, 483–495 (2020). https://doi.
org/10.1002/cche.10263
5. Izydorczyk, M.S.: Dietary arabinoxylans in grains and grain products. In: Beta, T., Camire,
M.E. (eds.) Cereal Grain-Based Functional Foods, pp. 167–203. Royal Society of
Chemistry, Cambridge (2019)
6. Pallai-Varsányi, E., Neményi, M., Kovács, A.J., Szijjártó, E.: Selective heating of different
grain parts of wheat by microwave energy. In: Advances in Microwave and Radio Frequency
Processing, pp. 312–320 (2007)
7. Ranjbaran, M., Zare, D.: Simulation of energetic-and exergetic performance of microwave-
assisted fluidized bed drying of soybeans. Energy 59, 484–493 (2013). https://doi.org/10.
1016/j.energy.2013.06.057
42 D. Budnikov
8. Smith, D.L., Atungulu, G.G., Sadaka, S., Rogers, S.: Implications of microwave drying
using 915 MHz frequency on rice physicochemical properties. Cereal Chem. 95, 211–225
(2018). https://doi.org/10.1002/cche.10012
9. Intelligent Computing and Optimization. Proceedings of the 2nd International Conference on
Intelligent Computing and Optimization 2019 (ICO 2019). Springer (2019). ISBN 978-3-
030-33585-4
10. Vasilev, A.N., Budnikov, D.A., Ospanov, A.B., Karmanov, D.K., Karmanova, G.K.,
Shalginbayev, D.B., Vasilev, A.A.: Controlling reactions of biological objects of agricultural
production with the use of electrotechnology. Int. J. Pharm. Technol. (IJPT). 8(N4), 26855–
26869 (2016)
11. Vasiliev, A.N., Goryachkina, V.P., Budnikov, D.: Research methodology for microwave-
convective processing of grain. Int. J. Energy Optim. Eng. (IJEOE) 9(2), 11 (2020). Article:
1. https://doi.org/10.4018/IJEOE.2020040101
Isolated Agroecosystems as a Way to Solve
the Problems of Feed, Ecology and Energy
Supply of Livestock Farming
1 Introduction
study has considered the efficiency of energy consumption increase and the possible
ways to reduce them. There have been analyzed these current issues and there have
been considered the variants of production organization.
When assessing the energy efficiency of production and determining the directions of
its improvement, it is expedient to use the methodology of energy evaluation of
products. It is expedient to apply a universal physical unit, whose exchange rate is
constant, solid and clear. Many researchers agree that this is an energy unit. Energy is
the only objective and universal measure of the value of any type of products, man-
ufactured not only by man, but also by nature.
This measure depends on neither supply and demand, nor on price. The energy
approach to the study of natural and production processes has a number of significant
advantages and wider opportunities compared to other methods [2]. The approach is
based on the necessity to take into account the objective laws of energy conversion
within the system, and at that, energy acts as a universal measure allowing us to
estimate both the value of material objects and the efficiency of production processes.
Without the use of energy analysis, effective production management is thought
impossible [3, 4]. One of the ways to implement this methodology is the environmental
and energy analysis of farm functioning. The issues of energy analysis of the
geosystems’ functioning have been studied since the beginning of the new century. In
particular, there has been developed an algorithm to conduct the study economic
activities of agroecosystems according to their energy assessment [5].
There are other researchers who carried out thorough work in this direction [6],
among them N.P. Mishurov [7].
In F.S. Sibagatullin’s work [8] there has been presented a methodology for cal-
culating the energy content of each element that is included in livestock production.
The energy content (E) was calculated on the basis of consumption in physical
units, taking into account its energy equivalent according to the formula [8]:
E ¼ Rn In ; ð1Þ
where R_n is the consumption of the resource in natural units (kg, hwt of feed units,
person ∙ h, kWt ∙ h); I_n is an energy equivalent of a resource [8].
The example of the conducted calculation is given in Table 1.
From the above presented data, it’s clear that the largest amount of energy has been
consumed on feed production.
Table 1. The energy consumption to produce a unit of product [8].
Isolated Agroecosystems as a Way to Solve the Problems of Feed, Ecology 45
If the share of energy consumption to produce various types of products in the total
amount of energy consumption is analyzed, it’s clear that 64.5% of the total amount of
energy consumption is required to produce milk. According to [9], the energy con-
sumed in the process is about 24% of total amount of energy consumption.
To produce cattle live weight, the share of energy consumption is 71.2% of the total
amount. In the production of pig live weight, this share rises up to 76.6%. At the same
time, direct consumption of electricity and fuel are 5.8% (of the total energy con-
sumption) for pig weight increase and 15.1% for cattle weight increase. Feed pro-
duction is the main consumer of energy.
In the works of E T. Sch. Fuzella [10], the energy efficiency is indicated as the ratio
of the energy flow at the system output to the energy flow at the system input. So using
this approach [10], there was made an estimation of the energy consumption by the
technological processes to produce various products for the SPK “Nelyubino”. Some
data of the conducted calculation are presented in Table 2. The energy flow at the
system output has been determined on the basis of the energy equivalent. As in Table 1,
the energy consumption structure has shifted to feed production with 72.5% of the total
energy.
Table 2. The estimation results of energy efficiency of livestock farming on the example of SPK
“Nelyubino”, TJ (1010 J).
Type of consumption Initial data Output
Year Type of products Year
1991 1999 2004 1991 1999 2004
Feed 60.2 45.5 42.6 Milk 8.1 6.6 6.5
Infrastructure 8.1 2.1 8.8 Beef 4.8 3.2 2.7
Electroenergy 12.3 11.3 12.3
Agromachinery 2.0 1.2 1.4
Fuel 0.5 0.6 0.7
Total 83.1 60.7 65.8 Total 12.9 9.8 9.2
46 A. N. Vasiliev et al.
In feed production the main share of energy is consumed on the production of crop
products, which is automatically transferred to livestock products [11].
According to the data presented in the table, in 1991 there were 0.16 units of a
product in energy equivalent per unit of consumed energy; in 1999 there were 0.16
units, and in 2004 there were 0.14 units. The calculation results have shown that the
highest energy efficiency of livestock production was 16%. For beef production energy
efficiency was not higher than 6%. That has shown that only 6% of the energy used for
beef production was consumed efficiently and 94% of the used energy were lost. The
problem of the lost energy is considered much more keen and urgent than a simple
calculation of energy consumption figures. Consuming 16% of energy to produce
livestock products, the farmer uses almost 84% of energy inefficiently. The energy
conservation principle must be taken into account in all sectors of agricultural activity,
and it’s necessary to find out what this energy is wasted on.
At this stage in the development of technology, this part of energy begins to
transform natural environment into which it has fallen, and the issue of energy inef-
ficiency in production automatically becomes an environmental concern for people
[12]. The issue is becoming keen for global agricultural production management.
Among the requirements of nature to biological objects, the provision of maximum
stability is of great importance. The most stable community to environmental devia-
tions is a biological community with a maximum variety of individual characteristics.
This requirement is practically inconsistent with the tasks of maximum productivity.
Maximum productivity leads to a minimum diversity. The best way of maximum
productivity is to cultivate a monoculture.
However, if a monoculture is cultivated, diversity is at a zero level, and a mono-
culture is absolutely unstable. To ensure the stability of a monoculture, people must
make additional efforts, i.e. to increase energy consumption.
The natural development and modification of biological objects in nature results in
an improvement of their diversity, that in its turn results in energy dissipation. The
organized use of a biological community, the purpose of which is to seize the obtained
product, reduces dissipation, and hence the stability of a biological system. Thus we
can conclude that any human activity that damages a balance of diversity conflicts with
the laws of nature and results in environmental problems. The development of
agroecosystems [13] has turned to be one of the way to solve this global contradiction
between humanity and the biological community.
An agroecosystem is referred to as the territory organized by man, where there is to
be ensured a balance to obtain products and return them to the fields to provide an
organized circulation of organic substances. Ration-ally organized agroecosystems are
necessarily to include pastures, livestock farming complexes, arable lands, though even
a perfectly organized agroecosystem cannot provide a complete cycle of substances.
This is due to the fact that part of the mineral substances is consumed by the yield. The
imbalance is eliminated due to the application of necessary fertilizers. Due to such
approach, agroecosystems are extremely unstable and incapable of self-regulation. The
additional energy resources introduced by a person must ensure the stability of the
agroecosystem. The experience has shown that agroecosystems, in which grain crops
prevail, cannot exist more than one year. In the case of a monoculture or perennial
grasses, the agroecosystem can break down after 3‒4 years.
Isolated Agroecosystems as a Way to Solve the Problems of Feed, Ecology 47
There are several ways to organize livestock agroecosystems. When free ranging of
cattle, an extensive option is implemented when the consumption of anthropogenic
energy is at a minimum level. The energy is consumed for support of cattlemen and the
primary processing of livestock products. An intensive way to implement an agroe-
cosystem involves the production of products at large complexes, while feed is obtained
due to the high energy input in the fields of the agroecosystem. Part of the manure is
brought to the fields, but before application it should be processed, that requires addi-
tional energy. The amount of obtained manure can be larger than is necessary to be laid
into the soil, and the environmental problem remains unsolved again.
Results and discussion.
By the example of livestock farming there has been considered the management of
such system, the effect of the components of the management system on environmental
friendliness of production [14, 15]. Figure 1 has presented a structural diagram that
allows discussing this issue in detail.
The main object of management is the livestock complex. The species of animals
are not important, since the control scheme will have the same structure. There should
be taken several parameters as controlled ones, namely volume of output (P), given in
tons; cost of production (C) in dollar per t; volume of production given in energy
equivalent (Ee), J; production waste (W), given in units of energy, J.
The use of the energy equivalent Ee as a controlled quantity is logical, since the
study of energy flows in ecosystems is one of the main research methods in ecology.
To implement technological production processes (Fig. 1), these are used com-
plexes that need general energy supply (Et) and energy supply for feed production (Ef).
The sources of energy I provide common energy supply, which is divided into indirect
(Ei) and direct (Ed) sources. Indirect energy supply is the main energy given in units of
energy consumed on buildings and structures, equipment, machinery.
48 A. N. Vasiliev et al.
Direct energy consumption includes the cost of fuel and electricity, used directly in
technological processes to produce finished products. The sources of energy II are
represented by feed (Efe), which are obtained in the second element of the agroe-
cosystem, i.e. plant growing. Initially the amount of livestock production fully con-
forms to the existing arable land and is provided with balanced crop rotation. The
system of mechanisms changes an energy flow (Efl) to implement technological pro-
cesses in the livestock complex.
Production results are evaluated using a monitoring system. The obtained data are
compared with a given indicator (Pp). Depending on the obtained comparison result
(ΔP) the control system management is to be corrected and updated. Deviations of the
controlled values from the given values can occur due to such disturbing effects as
changes in climate parameters, changes in feeding rations, changes in animal age and
weight, and other factors affecting the physiological state and productivity of animals.
The livestock agroecosystem presented in the form of a structural diagram is also
unstable, like any agroecosystem. From the point of view of the automatic control
theory, a system is considered unstable if, due to such disturbing effects, it deviates
from the given value and cannot return to its original state by itself. We have con-
sidered the results of stability loss of a given system through each controlled value.
As for such controlled value as volume of production (P), the stability loss for this
value means an irrevocable loss in animal productivity. This problem can occur due to
a significant influence of disturbing effects (e.g. an epidemic), as well as a lack of
regulatory measures (Efl), i.e. during the epidemic there were not provided appropriate
sanitary measures and vaccinations.
To ensure sustainability of such controlled value as cost price (C), there are
required restrictions on energy consumption. According to the data in Table 1 and
Table 2 it is necessary to reduce the energy consumption on feed production (Efe).
However, with the nowadays technologies, this can result in a production volume
decrease (P). Therefore, at present it is only possible to introduce restrictions on the
amount of consumed energy [16]. It has turned out that the stability of any controlled
effect is fully ensured by the energy supply of technological processes. Even for such
controlled value as volume of production, this thesis is not correct.
The current study has paid attention to the analysis of the system through equiv-
alent energy consumption.
As for equivalent energy consumption (Ee), the energy efficiency of the system
should strive for 1. The control system must reduce the energy flows Rt and Ee, while
simultaneously increasing the volume of output and its energy “costs”. In this case, the
“energy cycle” occurs without its significant, uncontrolled accumulation in production
waste. Those innovative plans, which are being actively discussed and adopted, have
got the exact opposite position. Indirect energy consumption is ready to be significantly
increased to produce products (e.g. to develop and use satellite systems for precise
farming, to use robotic systems, etc.). Moreover, the increase in such expenses can be
justified by only an improved product quality, with a reduced content of chemicals in
the final product, and a lower content of impurities. At the same time, direct energy
consumption is expected to decrease.
As a result, there is created another contradiction: to solve environmental problems
the indirect costs of production are increased and the energy efficiency of production is
Isolated Agroecosystems as a Way to Solve the Problems of Feed, Ecology 49
reduced, and it results in the amount of energy waste increase, that makes the envi-
ronmental problems more urgent.
Thus, on the one hand, the problem of the sustainability of the livestock agro-
ecological system is being solved by increasing energy regulatory effects, on the other
hand, these are simultaneously being created serious prerequisites for an environmental
disaster. One of the ways to solve the concern is to use waste as energy carriers. There
is so much energy in the agricultural waste that if it is obtained in the form of heat, fuel
and electricity, it will be more than enough for several cycles of animal production
[17]. Energy obtaining from organic and solid domestic waste is currently being
successfully implemented in various types of bioenergy mechanisms (biogas machines,
generators operating on mixed fuels obtained from organic waste, pyrolysis mecha-
nisms, etc.).
In the world, this issue has been successfully studied by a wide range of researchers
[18]. Nowadays technologies and equipment are being improved, which allows stating
that bioenergy is able to ensure energy independence of the agro-industrial sector.
Bioenergy is commonly considered as an element of renewable energy sources, but its
main task is to prevent an environmental disaster from energy stored in waste. It should
be noted that the simple waste burning with thermal energy release does not improve
the environmental situation. This energy remains unused, only its form changes, and it
does not return to the livestock production process.
The use of energy stored in waste reduces the share of hydrocarbons and gas in
direct energy consumption to produce livestock products. This increases the stability of
the livestock agroecosystem due to its cost price. However, the energy efficiency of the
system can only increase on 10‒15%. Still, more than 80% of the energy consumed to
produce livestock products is wasted. The necessity to increase livestock production
will require new energy costs for feed, which will result in an even greater decrease of
energy efficiency, and will lead to instability of the system according to cost price. The
issue could be solved by the use of non-standard solutions in the energy supply of
livestock.
Another way to increase the energy efficiency of livestock farming is to change the
structure of agroecosystems. Analyzing the data presented in Fig. 1 “the sources of
energy I” and “the sources of energy II” are the external ones to the agroecosystem
‘livestock farming’. Livestock farming largely (more than 75%) depends on the ‘ex-
ternal’ energy of crop production. The current trends in agriculture (precise farming)
show that the share of indirect energy consumption for feed production will grow
constantly and significantly. It is in this place that it is necessary to break the energy
cycle. It is necessary to reduce the share of energy consumption for feed production.
Therefore, there should not be developed the technologies to provide feed for the
livestock farming agroecosystem from the plant-farming agroecosystem. This is pos-
sible due to establishing so-called “isolated” agroecosystems. In this case, external
energy influx into the agroecosystem should be minimal. The animal husbandry should
provide the maximum feed requirements by its own resources. Most of necessary
technologies already exist and are undergoing serious production testing.
One such technology is the production and use of microalgae feed [19]. Microalgae
are actively used as feed additives in livestock and poultry farming, as it increases
animal immunity, weight, fertility and survival of young animals. In poultry, it results
50 A. N. Vasiliev et al.
in an increase in egg production and egg size. To implement this technology, the farms
specializing in cattle and poultry farming provide themselves with algal water pools
where livestock waste is utilized. As a result, 40% of the nitrogen from the wastewater
again enters the algae biomass and is eaten by animals.
Another technology that realizes the potential of livestock farming to produce feed
is the technology of growing fly larvae on organic waste, as they quickly multiply.
Larvae’s weight increases by 300‒500 times for a week. Bio-mass from a pair of flies
and their progeny with the full realization of the genetic potential will be more than 87
tons by the end of a year, i.e. it will be equal to the weight of six elephants [20].
Biomass of housefly larvae is a complete protein feed for pigs, calves, poultry, fur
animals, and fish. It contains 48‒52% of protein, 7‒14% of fat, 7‒10% of fiber, 7% of
BEV, 11‒17% of ash, as well as such biologically active substances as vitamins,
ecdysone, etc.
The processing of organic waste which is environmentally unfriendly and unsuit-
able for use in crop cultivation technologies can produce fertile humus. Its application
as a soil for growing feed in a protected ground will meet the needs of livestock
farming. Though organic waste of livestock farming should not be considered as
energy raw materials. The use of waste as an energy carrier will reduce the energy
intensity of livestock farming and increase its energy efficiency by no more than 15%.
The issue could be radically solved by using organic materials to produce highly
nutritious feed in closed systems, combined into an isolated agroecological system.
In this case, only one restriction must be fulfilled. The volume of the livestock and
its productivity (manufactured products and production waste) should maximize the
necessary amount of feed for the livestock. Ideally, the chain ‘feed ! animals ! a
technological process of production ! waste ! feed’ should have minimal energy
supply from the outside. The renewable energy sources acquire a special role as the
sources of direct energy supply (Ed) as shown in Fig. 1. Their use can increase the
share of capital costs, but significantly reduce the environmental problems of energy
supply.
3 Conclusions
livestock waste can be efficiently applied to produce high-protein animal feed and
reduce energy consumption by its products.
It is reasonable to use renewable energy sources to compensate direct energy costs
in the production of livestock products.
References
1. Mindrin, A.S.: Energoekonomicheskaya ocenka sel'skohozyajstvennoj produkcii [Energy-
economic assessment of agricultural products], 187 p (1987)
2. Bulatkin, G.A.: Analiz potokov energii v agroekosistemah [Analysis of energy flows in
agroecosystems]. Vestnik Rossijskoj Akademii Nauk tom 82(8), 1–9 (2012)
3. Perez-Neira, D., Soler-Montiel, M., Gutierrez-Pena, R., et al.: Energy assessment pastoral
dairy goat husbandry from an agroecological economics perspective. A case study in
Andalusia (Spain). Sustainability 10 (2018). Article number: 2838.
4. Guzman, G.I., de Gonzalez, M.G.: Energy efficiency in agrarian systems from an
agroecological perspective. Agroecology Sustain. Food Syst. 39, 924–952 (2015)
5. Pozdnyakov, A.V., Semenova, K.A., Fuzella, T.Sh.: Energeticheskij analiz funkcionirova-
niya agroekosistem v usloviyah estestvennogo nasyshcheniya-pervye rezul'taty [Energy
analysis of the functioning of agroecosystems in conditions of natural saturation - first
results]. Uspekhi sovremennogo estestvoznaniya (2), 124‒128 (2018)
6. da Silva, N.F., da Costa, A.O., Henriques, R.M., Pereira, M.G., Vasconcelos, M.A.F.: Energy
planning: Brazilian potential of generation of electric power from urban solid wastes—under
“waste production liturgy” point of view. Energy Power Eng. 7(5), 193 (2015)
7. Mishurov, N.P.: Metodologicheskie osnovy energeticheskoj ocenki proizvodstva moloka
[Methodological foundations of energy assessment of milk production]. Tekhnika i
oborudovanie dlya sela (5), 16‒19 (2017)
8. Sibagatullin, F.S., Sharafutdinov, G.S., Shajdullin, R.R., Moskvichyova, A.B.: Bioener-
geticheskaya ocenka i osnovnye puti snizheniya energoemkosti proizvodstva produkcii
zhivotnovodstva [Bioenergy assessment and main ways to reduce energy intensity of
livestock production]. Uchenye zapiski Kazanskoj gosudarstvennoj akademii veterinarnoj
mediciny im. N.E. Baumana T. 216, 295–302 (2013)
9. Rajaniemi, M., Jokiniemi, T., Alakukku, L., et al.: Electric energy consumption of milking
process on some Finnish dairy farms. Agric. Food Sci. 26, 160–172 (2017)
10. Fuzella, T.Sh.: Energeticheskaya ocenka funkcionirovaniya agroekosistemy (na primere
SPK «Nelyubino») [Energy assessment of the functioning of the agroecosystem (on the
example of the SEC ``Ne-Lyubino'')]. Vestnik Tomskogo gosudarstvennogo universiteta
vypusk (326), 203‒207 (2009)
11. Stavi, I., Lal, R.: Agriculture and greenhouse gases, a common tragedy. A review. . Agron.
Sustain. Dev. 33, 275–289 (2013)
12. Ghosh, S., Das, T.K., Sharma, D., Gupta, K., et al.: Potential of conservation agriculture for
ecosystem services. Indian J. Agric. Sci. 89, 1572–1579 (2019)
13. Marks-Bielska, R., Marks, M., Babuchowska, K., et al.: Influence of progress in sciences and
technology on agroecosystems. In: Conference: Geographic Information Systems Confer-
ence and Exhibition (GIS Odyssey), Trento, Italy, 04–08 September 2017, pp. 254‒263
(2017)
14. Lachuga, Yu.F., Vasilyev, A.N.: Napravleniya issledovanij v bioenergetike [Research areas
in bioenergy]. Vestnik Rossijskoj sel'skohozyajstvennoj nauki (2), 4‒7 (2015)
52 A. N. Vasiliev et al.
1 Introduction
One of the most progressive distributed technology today - Blockchain, The blockchain
is a decentralized digital database to store inside any nodes in the network called
Distributed Ledger Technology (DLT). The most popular example which uses
blockchain as a back-end operation system is Bitcoin. Bitcoin is an innovative payment
network and a new kind of money [1]. Moreover, blockchain can apply to many
aspects such as the hospital which uses blockchain with the information record and
payment [2]. In the financial system, blockchain is applied to make payments with the
transactions across the border. Blockchain also has more applications such as supply
chain, voting, and energy supply [3].
The decentralized application (DApp) plays an important role to apply with
blockchain technology. DApp overcomes the limitation of the locally running program
which is the performance limitation and cannot respond to the requirements of many
applications [4]. DApp could have four categories - open source, internal cryptocurrency
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 53–62, 2021.
https://doi.org/10.1007/978-3-030-68154-8_6
54 P. Thupphae and W. Ongsakul
support, decentralized consensus, and no central point of failure [4]. This study repre-
sents implement of DApp using the Ethereum open-source platform in Sect. 2. In
Sect. 3 shows implement of ERC 20 token. Section 4 shows the implementation of
energy trading and Sect. 5 shows the deployment and experimental result.
The proposed energy trading was implemented using the Ethereum blockchain
framework [5]. It includes the following three steps: (1) set up the Ethereum envi-
ronment; (2) set up the React and (3) set up identities with Ganache and wallets with
Metamask for participants.
async loadBlockchainData() {
const web3 = new Web3(Web3.givenProvider || 'http://localhost:7545')
}
}
After doing the step above, check the web3 loaded inside the console. By ($con-
sole.log(“web3”,web3)). The result would shown in Fig. 2.
async loadBlockchainData() {
const web3 = new Web3(Web3.givenProvider || 'http://localhost:7545')
const network = await web3.eth.net.getNetworkType()
const accounts = await web3.eth.getAccount()
const networkId = await web3.eth.net.getId()
const token = web3.eth.Contract(Token.abi, Token.networks[networkId]. address)
}
}
Fig. 3. A command-line use to detect the network, account, network ID and import the token.
async loadBlockchainData() {
const web3 = new Web3(Web3.givenProvider || 'http://localhost:7545')
const network = await web3.eth.net.getNetworkType()
const accounts = await web3.eth.getAccount()
const networkId = await web3.eth.net.getId()
const token = web3.eth.Contract(Token.abi, Token.networks[networkId]. address)
const totalSupply = token.methods.totalSupply().call()
}
}
Then the result should return the total supply which we defined = 10^18 Wei in
Fig. 6.
The implement of ERC20 Token has standard API to use with smart contracts [21]. The
function of ERC20 Token consists of name, symbol, decimals, totalSupply, balanceOf,
transfer, transferFrom, approve, and allowance. In this study focus on implementing these
functions on smart contract to DApp. The flowchart in Fig. 7. Represent the flow of smart
contract to create ERC20 token and use them to transfer. To make the ERC20 token, First
need to set name, symbol, decimal, and totalSupply of the token. In this study set
Name = Energy Token, Symbol = ET, Decimals = 18 (1 Ether = 1018 Wei) and
totalSupply = 1,000,000. Then set the balance of deployer equal to totalSupply to test the
transfer function. To execute the transfer function, the return values must return true to
execute this function. This function checks the balance of the deployer must more than the
transfer value. Then check the invalid receiver. If these return true, the transfer function is
executed. The balance of the deployer is decreased and the balance of receiver is increase
with the transfer value. The last of this function is to emit information from this transfer.
Start
BalanceOf;
Set Deployer :1,000,000 tokens
Receiver : 0 Tokens
Transfer;
Check1 : !balancceOfDeployer >= _value
Check2 : !invalidRecipient
True
balancceOfDeployer = balancceOfDeployer - _value
balancceOfReceiver = balancceOfReceiver + _value
False
End
export const makeBuyOrder = (dispatch, exchange, token, web3, order, account) => {
const tokenGet = token.options.address
const amountGet = web3.utils.toWei(order.amount, ‘ether’)
const tokenGive = ETHER_ADDRESS
const amountGive = web3.utils.toWei((order.amount * order.price).toString(), ‘ether’)
export const makeSellOrder = (dispatch, exchange, token, web3, order, account) => {
const tokenGet = ETHER_ADDRESS
const amountGet = web3.utils.toWei((order.amount * order.price).toString(), ‘ether’)
const tokenGive = token.options.address
const amountGive = web3.utils.toWei(order.amount, ‘ether’)
4.2 Trading
In a smart contract, the trading function which is internal function lives inside fillOrder
function. This time when the _tokenGet fetch the msg.sender balances which is the
person who called the filling order set equal to his balances minus the amountGet.
While the _tokenGet for the user is set to the user balances added the amountGet.
The _user is the person who creates the order. For the tokenGive use the balances
of the user who creates the order minus with the amountGive. This amountGive is
added to the msg.sender which is the person who fills the order. Moreover, the fee
amount is going to be paid by the person who fills the order and takes out of the
amountGet this case is the msg.sender. In this work, fee percent is set to 10% and add
60 P. Thupphae and W. Ongsakul
to the amountGet of the msg.sender. The feeAccount was created to update the balance
to the feeAmount (see Fig. 9).
Function _trade(uint256 _orderId, address _user, address _tokenGet, uint256 _amountGet, address _tokenGive,
uint256 _amountGive) internal {
uint256 _feeAmount = _amountGive.mul(feePercent).div(100);
tokens[_tokenGet][msg.sender] = tokens[_tokenGet][msg.sender].sub(_amountGet.add(_feeAmount));
tokens[_tokenGet][_user] = tokens[_tokenGet][_user].add(_amountGet);
tokens[_tokenGet][feeAccount] = tokens[_tokenGet][feeAccount].add(_feeAmount);
tokens[_tokenGive][_user] = tokens[_tokenGive][_user].sub(_amountGive);
tokens[_tokenGive][msg.sender] = tokens[_tokenGive][msg.sender].add(_amountGive);
The front-end of Solar-Energy Trading has been designed as Fig. 10. This user
interface has been deployed on the localhost:3000. In this study, suppose user1 has
defined to make buy order with the total payment amount is 1 ether and send to user 2.
Figure 11 represents transferring the total payment amount from user1 to user 2 in the
terminal interface (by setting 1 Ether = 11018 tokens). Figure 12 shows the trans-
action between user 1 and 2 on Ganache personal Ethereum blockchain. User 1 and
user 2 have the address 0xb877dCcB80F27b83E4f863c41f050f18FfAEcb9b and
0x12e622A7c90CE-fF482Fc79ADe96a3AcD17C9F282 respectively. The exchange
contact address is 0x0FDA0BA4c75c3552A42B9877c9b48fC6fddc022D.
6 Conclusion
The increase of solar PV and the progress of blockchain technology play an important
role in the energy section. The future market will become peer to peer market in the
microgrid. The excess energy is going to make a profit to prosumer. This paper presents
a decentralized application based on the blockchain technology framework. The
description is provided in technical details on setting up the ethereum blockchain, make
the token by using ERC 20 token, useful dependencies to set up DApp, and some part
of trading with smart contract.
References
1. B. Project. “Bitcoin is an innovative payment network and a new kind of money,” MIT
license (2009–2020). https://bitcoin.org/en/
2. Tasatanattakool, P.: Blockchain: challenges and applications. In: International Conference
on Information Networking (ICOIN), Thailand (2018)
3. Xu, X.: Architecture for Blockchain Applications. Springer (2019)
4. Wei Cai, Z.W.: Decentralized applications: the blockchain-empowered software system.
IEEE Access 6, 53019–53033 (2018)
5. Ethdotorg: Ethereum.org, Stiftung Ethereum. https://ethereum.org/en/
6. Cleverflare: Learn to Code Blockchain DApps By Building Simple Games, Loom. https://
CryptoZombies.io
7. Santander, A.: Hello Ethernaut, OpenZeppelin. https://ethernaut.openzeppelin.com
8. Remix - Ethereum IDE. https://remix.ethereum.org
9. Team, B.: Bootstrap, MIT. https://getbootstrap.com/docs/4.0/components/card
10. Team, B.: Bootstrap, MIT. https://getbootstrap.com/docs/4.0/components/navs
11. W3Schools: W3schools.com, W3.CSS. https://www.w3schools.com/Css/css3_flexbox.asp
12. Sphinx: web3.js, Core team (2016). https://web3js.readthedocs.io/en/v1.2.11/
62 P. Thupphae and W. Ongsakul
Abstract. The paper considers the design of the solar photovoltaic air cooling
module with a paraboloid type concentrator. Photovoltaic converters are located
on the surface of the concentrator, which ensures their cooling using a metal
radiator, the functions of which are performed by the solar concentrator itself. At
the same time, the profile of the solar radiation concentrator provides uniform
illumination of photovoltaic cells, which favorably affects their efficiency. It is
advisable to use high-voltage matrix silicon photovoltaic solar cells as photo-
voltaic converters, which have high electrical efficiency and maintain it at a high
concentration of solar radiation and heating with concentrated solar radiation.
1 Introduction
The use of solar radiation concentrators can reduce the number of photovoltaic con-
verters, which favorably affects the cost of the installation and the electricity received
with its help [1–3]. However, when photovoltaic cells convert concentrated solar
radiation, their heating occurs significantly, as a result of which their electrical effi-
ciency decreases [4], which indicates the need for their cooling [5–10]. When photo-
voltaic converters operate in a concentrated solar stream, their current-voltage
characteristic acquires a triangular shape, which indicates a significant decrease in their
efficiency [11]. All planar solar cells have this property, but matrix high-voltage silicon
photovoltaic converters do not lose electrical efficiency in concentrated solar flux due
to their structure [12, 13]. The use of a paraboloid type concentrator in the form of an
air radiator, in the focal region of which such matrix photovoltaic converters are located
simultaneously on its surface, eliminates the need for an active cooling system design.
However, for their stable operation, it is necessary to ensure uniform illumination of the
entire surface of the photovoltaic converters, which requires a special geometric
approach and design methods [14] for the profile of the solar radiation concentrator.
Fig. 1. The design of the solar photovoltaic module, where the concentrator is also an air-cooled
radiator for photovoltaic converters
Thanks to the use of silicon high-voltage matrix photovoltaic converters (Fig. 2), it
appears possible to obtain a high-voltage direct current (1000 V or more) at the output
of one solar module, as well as an increase in the electrical conversion efficiency of
solar radiation and, accordingly, a reduction in cost as module (specific electric power)
and generated electric energy. Due to the use of sealing technology of photovoltaic
converters with two-component polysiloxane compound, the period of the rated power
of the photovoltaic converters also increases.
Solar Module with Photoreceiver Combined with Concentrator 65
Fig. 2. Silicon matrix high-voltage photovoltaic converters with a voltage of more than 1000 V
Fig. 3. Principle of operation and the scheme of the concentrator photovoltaic solar module
66 V. Panchenko and A. Kovalev
X2
Y¼ ; ð2Þ
4f
ð2 R X Þ2 ¼ 4 f Y; ð3Þ
ð2 R Xc Þ ¼ ð2 R Xb Þ þ d0 sin a0 ; ð4Þ
( 12 )
Y0
Xb ¼ 2 f tg b0 1þ tg b0
2
1 ; ð5Þ
f
Xb2
Yb ¼ ; ð6Þ
4f
Xc ¼ Xb d0 sin a0 ; ð7Þ
Yc ¼ Yb h; ð8Þ
h ¼ d0 sin a0 ; ð9Þ
d0 ðYc f Þ
¼ : ð10Þ
sin a sin d sin u0
where b0 is the angle between the sunbeam, reflected from the surface of the
concentrator profile in the zone a – b at the point with coordinates Xa and Ya, coming
into the focal region of the concentrator (on the surface of the photovoltaic photore-
ceiver) at the point with coordinates (2 R – Xb), Yb and parabola focal length f;
Solar Module with Photoreceiver Combined with Concentrator 67
the angle b ¼ b0 n=N varies from 0º to b0, the values of the coefficient n are
selected from a series of integers n ¼ 0; 1; 2. . . N;
c0 is the angle between the sunbeam reflected from the surface of the concentrator
profile in the zone a – b at the point with coordinates Xb and Yb, coming into the focal
region of the concentrator (on the surface of the photovoltaic photoreceiver) at the point
with coordinates ð2 R Xc Þ, Yc and parabola focal length f;
a0 is the angle of inclination of the working surface of the profile of the concen-
trator in zone b – c;
the maximum radius of the solar concentrator is R;
the angle d is determined by the expression d ¼ p=2 þ a0 b;
angles a0 and b0 are selected in accordance with the boundary conditions.
The distribution of the concentration of solar radiation Kab on the surface of the
photovoltaic photoreceiver, reflected from the upper surface of the working profile of
the solar radiation concentrator (zone a – b) is calculated according to the formulas:
2
Rn þ 1 R2n
Kab ¼ ; ð12Þ
Ddn Rbcn þ 1 þ Rbcn
Ddn ¼ dn þ 1 dn ; ð13Þ
Rn ¼ Xn R; ð14Þ
The values of the X and Y coordinates in the zone of the working profile of the
concentrator c – d are determined using the system of equations:
Yc
tgam1 Yb
tgam ðXb Xc Þ
Ym ¼ ; ð16Þ
1
tgam1 tgam
ð Y c Ym Þ
Xm ¼ Xc ; ð17Þ
tga0m
ðYb Ym1 Þ
tgam1 ¼ ; ð18Þ
Xm1 ð2 R Xc Þ
ðYb Ym Þ
tgam ¼ ; ð19Þ
Xm ð2 R Xc Þ
ð2 R Xc Þ þ ðYc Ym Þ
Xm ¼ ; ð20Þ
tgam1
a0m ¼ a0 þ am ; ð21Þ
68 V. Panchenko and A. Kovalev
( 12 )
Y0
Xd ¼ 2 f tg b0 1þ 1 ; ð22Þ
f tg2 b0
Xd2
Yh ¼ Yd h Yd ¼ ; ð23Þ
4f
where a0m is the angle of inclination of the working surface of the profile of the
concentrator in zone c – d;
am is the angle between the sunbeam, reflected from the surface of the profile of the
concentrator in the zone c – d at the point with coordinates Xm and Ym coming to the
focal region of the concentrator (on the surface of the photovoltaic receiver) at the point
with coordinates Xb and Yb and the level of the coordinate Ym;
values of the coefficient m are selected from a series of integers m ¼ 0. . . M.
Distribution of the concentration of solar radiation Kcd on the surface of the pho-
tovoltaic receiver is calculated according to the formulas:
X
M X
M R2mn R2mðn þ 1Þ
Kcd ¼ Km ¼ ; ð24Þ
m¼0 m¼0 Ddn Rbcðn þ 1Þ þ Rbcn
Rm ¼ Xm R; ð26Þ
Using the formulas presented above, the coordinates of the working profile of the
paraboloid type solar concentrator are calculated, which ensures uniform illumination
of the focal region (silicon matrix high-voltage photovoltaic converters) with con-
centrated solar radiation. Figure 4 shows the profile of the reflecting surface of the
paraboloid type solar radiation concentrator with a maximum diameter of about
400 mm.
Three different zones of the concentrator profile can be distinguished on the profile
of the solar radiation concentrator, one of which has a flat shape and represents the
landing surface of the photovoltaic converters. This kind of photovoltaic solar modules
can be installed together on one frame with a system for constant tracking the position
Solar Module with Photoreceiver Combined with Concentrator 69
of the Sun. For a better layout and optimal filling of the frame space, the concentrators
of solar modules can be squared and closely mounted to each other.
Figure 5 on the left shows a three-dimensional model of the solar module with the
profile of the solar concentrator, calculated according to the developed method and
providing uniform distribution of concentrated solar radiation over the surface of
photoreceiver (silicon matrix high-voltage photovoltaic converters) in the focal region,
which is presented in the Fig. 5 on the right.
Fig. 5. Three-dimensional model of the solar module (on the left) and distribution of
concentrated solar radiation over the surface of photoreceiver (on the right)
uniform concentrated solar radiation, due to which the output electric power of the solar
module will be at a nominal level.
Thanks to the calculated profile of paraboloid type solar concentrator using
computer-aided design system, it becomes possible to create their three-dimensional
solid-state model. Figure 6 shows such concentrator of solar radiation of paraboloid
type and its single component (petal), into which it can be divided for subsequent
manufacture from metallic reflective material.
Fig. 6. Three-dimensional model of a paraboloid type solar concentrator and its single
component made of reflective metal
4 Conclusion
Acknowledgment. The research was carried out on the basis of financing the state assignments
of the All-Russian Institute of the Electrification of Agriculture and the Federal Scientific
Agroengineering Center VIM, on the basis of funding of the grant “Young lecturer of RUT” of
the Russian University of Transport, on the basis of funding of the Scholarship of the President of
the Russian Federation for young scientists and graduate students engaged in advanced research
and development in priority areas of modernization of the Russian economy, direction of
modernization: “Energy efficiency and energy saving, including the development of new fuels”,
subject of scientific research: “Development and research of solar photovoltaic thermal modules
of planar and concentrator structures for stationary and mobile power generation”.
References
1. Rosell, J.I., Vallverdu, X., Lechon, M.A., Ibanez, M.: Design and simulation of a low
concentrating photovoltaic/thermal system. Energy Convers. Manage. 46, 3034–3046 (2005)
2. Nesterenkov, P., Kharchenko, V.: Thermo physical principles of cogeneration technology
with concentration of solar radiation. Adv. Intell. Syst. Comput. 866, 117–128 (2019).
https://doi.org/10.1007/978-3-030-00979-3_12
3. Kemmoku, Y., Araki, K., Oke, S.: Long-term performance estimation of a 500X
concentrator photovoltaic system. In: 30th ISES Biennial Solar World Congress 2011,
pp. 710–716 (2011)
4. Kharchenko, V., Nikitin, B., Tikhonov, P., Panchenko, V., Vasant, P.: Evaluation of the
silicon solar cell modules. In: Intelligent Computing & Optimization. Advances in Intelligent
Systems and Computing, vol. 866, pp. 328–336 (2019). https://doi.org/10.1007/978-3-030-
00979-3_34
5. Sevela, P., Olesen, B.W.: Development and benefits of using PVT compared to PV. Sustain.
Build. Technol. 90–97 (2013)
6. Ibrahim, A., Othman, M.Y., Ruslan, M.H., Mat, S., Sopian, K.: Recent advances in flat plate
photovoltaic/thermal (PV/T) solar collectors. Renew. Sustain. Energy Rev. 15, 352–365
(2011)
7. Kharchenko, V., Nikitin, B., Tikhonov, P., Gusarov, V.: Investigation of experimental flat
PV thermal module parameters in natural conditions. In: Proceedings of 5th International
Conference TAE 2013, pp. 309–313 (2013)
8. Hosseini, R., Hosseini, N., Khorasanizadeh, H.: An Experimental study of combining a
Photovoltaic System with a heating System. In: World Renewable Energy Congress,
pp. 2993–3000 (2011)
72 V. Panchenko and A. Kovalev
9. Panchenko, V., Kharchenko V., Vasant, P.: Modeling of solar photovoltaic thermal modules.
In: Intelligent Computing & Optimization. Advances in Intelligent Systems and Computing,
vol. 866, pp. 108–116 (2019). https://doi.org/10.1007/978-3-030-00979-3_11
10. Panchenko, V.A.: Solar roof panels for electric and thermal generation. Appl. Solar Energy
54(5), 350–353 (2018). https://doi.org/10.3103/S0003701X18050146
11. Kharchenko, V., Panchenko, V., Tikhonov, P., Vasant, P.: Cogenerative PV thermal
modules of different design for autonomous heat and electricity supply. In: Handbook of
Research on Renewable Energy and Electric Resources for Sustainable Rural Development,
pp. 86–119 (2018). https://doi.org/10.4018/978-1-5225-3867-7.ch004
12. Panchenko, V.: Photovoltaic solar modules for autonomous heat and power supply. IOP
Conf. Ser. Earth Environ. Sci. 317 (2019). 9 p. https://doi.org/10.1088/1755-1315/317/1/
012002
13. Panchenko, V., Izmailov, A., Kharchenko, V., Lobachevskiy, Y.: Photovoltaic solar
modules of different types and designs for energy supply. Int. J. Energy Optim. Eng. 9(2),
74–94 (2020). https://doi.org/10.4018/IJEOE.2020040106
14. Sinitsyn, S., Panchenko, V., Kharchenko, V., Vasant, P.: Optimization of Parquetting of the
concentrator of photovoltaic thermal module. In: Intelligent Computing & Optimization.
Advances in Intelligent Systems and Computing, vol. 1072, pp. 160–169 (2020). https://doi.
org/10.1007/978-3-030-33585-4_16
Modeling of Bilateral Photoreceiver
of the Concentrator Photovoltaic Thermal
Module
Abstract. The paper describes the modeling of the thermal state and visual-
ization of the operating mode of the bilateral photoreceiver of the concentrator
photovoltaic thermal solar module. Based on the results obtained during the
simulation, the geometric parameters of the components of the photoreceiver of
the solar module are substantiated and selected. The developed design of the
solar module will provide high electrical and thermal efficiency of the solar
module. These types of concentrator photovoltaic thermal solar modules can
operate autonomously or in parallel with the existing power grid to power
consumers.
1 Introduction
Recently solar energy converters have been developing at a pace that outstrips the
development of converters of other renewable energy sources [1–6]. Along with
increasing the efficiency of photovoltaic converters, when the technology of their
manufacture is being improved [7–9], the cogeneration method is also used to increase
the overall efficiency of the solar module – when, along with electricity, the consumer
receives thermal energy. Such solar modules can be either a planar design [10] or a
concentrator design [11] when various concentrators and reflectors of solar energy are
used. Such devices must be manufactured with high accuracy, and their design is an
important and difficult task [12]. Concentrator thermal photovoltaic solar plants gen-
erate electrical and thermal energy, while the temperature of the coolant can reach
Fig. 1. Components of solar thermal photovoltaic modules of various designs and the model of
the bilateral photoreceiver “Model 4”
Modeling of Bilateral Photoreceiver of the Concentrator Photovoltaic Thermal Module 75
For calculation in a finite element analysis system, the photoreceiver model was
divided into finite elements, as well as broken down into components representing var-
ious physical bodies. When modeling the thermal state in the finite element analysis
system, the following parameters and module characteristics were adopted: module
length 600 mm; thermophysical properties of substances: silicon: thermal conductivity:
120 W/m K, heat capacity: 800 J/kg K, density: 2300 kg/m3; glass: thermal con-
ductivity: 0,46 W/m K, heat capacity: 800 J/kg K, density: 2400 kg/m3; rubber:
thermal conductivity: 0,15 W/m K, heat capacity: 1800 J/kg K, density: 1200 kg/m3;
nitrogen: atmospheric pressure: 101300 Pa, 20 ºC, dynamic viscosity: 1710–6 Pa s),
thermal conductivity: 0,026 W/m K; heat capacity at constant pressure: 1041 J/kg K;
density: 1,182 kg/m3.
At the next stage of modeling, the following parameters were adopted: specific heat
release on the photovoltaic converter: 2 W/cm2, gas temperature at the inlet: 15 ºC; gas
inlet speed: 10 m/s; gas pressure: 10 atm; total gas flow: mass: 0,066 kg/s, volumetric
at operating pressure and temperature: 5,50 l/s, volumetric at atmospheric pressure and
temperature 20 ºC: 56,12 l/s. The thermal state of the photovoltaic converter is shown
in the Fig. 3 on the left, the temperature of the coolant and other structural components
is shown in the Fig. 3 on the right.
Fig. 3. Thermal state of the photovoltaic converter (on the left) and temperature of the coolant
and other structural components (on the right)
as the concentration value (specific heat release on the photovoltaic converter), heat
carrier velocity and temperature of the heat carrier at the inlet, the necessary temper-
ature values of the photovoltaic converter and heat carrier can be obtained.
Thanks to the developed modeling method [14], a study was made of the bilateral
photoreceiver of the solar photovoltaic thermal module (“Model 4”), as a result of which
temperature distributions, flows, and coolant velocities were obtained, which made it
possible to analyze and optimize the layer-by-layer structure of the model. High-voltage
bilateral matrix photovoltaic converters are located in the focal region of the concen-
trator of the module. Bilateral photovoltaic thermal photoreceiver located in the focus of
a solar radiation concentrator was subjected to simulation. Bilateral photoreceiver of the
solar photovoltaic thermal module consists of 7 different components (Fig. 1), the
layered structure of which is presented in the Table 1. Modeling the thermal state in a
finite element analysis system with obtaining an array of temperatures allows getting a
more detailed three-dimensional picture of the thermal state of the module, in contrast to
two-dimensional analytical modeling, which allows comparing the thermal state of all
components and each separately for different parameters.
Table 1. Parameters of the components of the bilateral photoreceiver of the solar photovoltaic
thermal module (“Model 4”)
Component Thick., Comp. mat. Therm. Dens., Kin. Dyn. Heat Coef. of
mm cond., kg/m3 visc., visc., cap., therm. exp.,
W/mK Pas m2/s J/kgK 1/K
1. Bilateral 0,2 Silicon 148 2330 – – 714 2,54 10–6
electrogenerating
2. Transparent 0,2 Polysilo- 0,167 950 – – 1175 100 10–6
sealing xane
3. Transparent 0,1 Polyethy- 0,14 1330 – – 1030 60 10–6
insulating lene
4. Transparent 5 Water 0,569 1000 1788 1,78 4182 –
heat removal 10–6 10–6
5. Transparent 0,3 Polyethy- 0,14 1330 – – 1030 60 10–6
insulating heat lene
sink
6. Transparent 3; 5; 7 Air 0,0244 1,293 17,2 13,210– 1005 –
heat-insulating 10–6 6
Since the model of the photoreceiver is symmetric, the axis of symmetry will be the
middle of the bilateral electrogenerating component (silicon matrix high-voltage pho-
toelectric converter) (Fig. 4 above) in longitudinal section (Fig. 4 below). The thick-
nesses of the components in the drawing are taken preliminary.
78 V. Panchenko et al.
When modeling the thermal state of the model of the photoreceiver, the influence of
the concentration of solar radiation (3; 6 and 9 times), transparent heat-insulating
component (the air gap) (3; 5 and 7 mm), as well as the flow rate of the coolant (0,5; 5
and 50 g/s) on the thermal state of the coolant is considered. The temperature of the
bilateral electrogenerating component (silicon matrix high-voltage photoelectric con-
verter) should be below 60 ºC, since when the photovoltaic converters are heated, their
electrical efficiency gradually decreases.
As a preliminary calculation, a part of the bilateral electrogenerating component is
considered: a strip 10 mm wide, conventionally located in the middle part of the
photovoltaic thermal photoreceiver, the length of this part being equal to the length of
the entire module. Since the module is symmetrical with respect to the average hori-
zontal plane, only its upper half is considered.
The flow of the cooling agent is considered uniform throughout the cross section of
the cooling cavity. Three variants of the module design are considered, differing in the
height of the transparent heat-insulating component (air layer thickness) – 3 mm, 5 mm
and 7 mm (Fig. 5).
Fig. 5. Three variants of the module design (height of the transparent heat-insulating component
3 mm, 5 mm and 7 mm)
The influence of the concentration of solar radiation on the thermal state of the
module was studied – three cases were considered: a concentration of 3, 6 and 9 times
for each side of the module. Taking into account the fraction of solar radiation con-
verted into electrical energy, the heat flux is taken to be 2400 W/m2, 4800 W/m2 and
7200 W/m2.
Modeling of Bilateral Photoreceiver of the Concentrator Photovoltaic Thermal Module 79
For the case of water cooling with the temperature of 293 K, the thermal state was
simulated for mass flow rates of 0,05 kg/s, 0,005 kg/s and 0,0005 kg/s. It was noted
that at the mass flow rate of water of 0,05 kg/s, the water layer (transparent heat
removal component) does not warm up to the entire thickness, therefore, the calculation
for this flow rate was performed only for a design variant with a heat insulator thickness
of 3 mm.
To assess the effect of non-uniformity in the supply of the cooling agent (trans-
parent heat removal component), a solid-state model of a quarter of the photoreceiver is
constructed, shown in the Fig. 6. A design variant with a thickness of the transparent
heat-insulating component (air layer thickness) equal to 3 mm is considered. The layers
of transparent heat-insulating component and transparent heat removal component are
divided by thickness into 10 layers of elements (Fig. 6). The components of the model
with a small thickness are divided by thickness into 3 layers of elements.
The symmetry planes of the module are the average vertical and horizontal planes.
In the Fig. 7 on the left, the symmetry planes are marked in green. The coolant enters
and exits through the central holes marked in green in the Fig. 7 on the right. The flow
rate of the cooling agent is set at 0,205 kg/s and 0,0205 kg/s, which is equivalent to a
flow rate of 0,05 kg/s and 0,005 kg/s for a part of the module with a width of 10 mm.
Fig. 7. Symmetry planes of the model (on the left) and coolant inlet (on the right)
Fig. 8. Temperature distribution of the components of the model from the side of the inlet and
outlet of the coolant (above) and the velocity of the coolant from the side of the inlet and outlet of
the coolant (below)
Fig. 9. Temperature distribution of the components of the model along the entire length – top
view (above) and coolant velocity along the entire length – top view (below)
Optimization of the air gap and the flow rate of the coolant were carried out
according to the criterion for changing the temperature of the coolant (priority of the
first category). The remaining secondary criteria by which optimization did not occur
were of secondary importance, but also had priority depending on their contribution to
the heating of the coolant (the second category – coefficient 0,9, the third – 0,8, the
fourth category – 0,7 and the fifth – 0,6) (Table 2). Also, for changes in the criteria,
Table 2 shows the achievement of their desired values in terms of finding the optimum
– minimum or maximum, which is also taken into account when summing up the
secondary criteria, that is, the smaller the change, the better the heating of the coolant.
82 V. Panchenko et al.
If a change in the value of a secondary criterion reduces the main criterion (the priority
of the first category – heating of the coolant, relative to the value of which a decision is
made on the appropriateness of a particular module design), then when summing up the
magnitude of the change of this criterion is given with a negative sign. With the same
change in the main criterion (coolant temperature), the change in the total criterion is
calculated, which includes changes in all secondary criteria and preference is given to
the maximum value of the total criterion:
X
K ¼1 K1 0; 9 K2 0; 8 K3 0; 7 K4 0; 6 K5
The total criterion reflects a change not only in the parameter of the main criterion
(change in the temperature of the coolant), but also takes into account changes in other
criteria (change in the temperatures of the remaining components), which plays an
important role in the optimization process in view of the level of their contribution to
the thermal state of the module and heating of the coolant in particular.
4 Conclusion
As a result of the developed method for designing solar photovoltaic thermal modules
the designer can develop solar cogeneration modules of various designs, which can
then be tested in the finite element analysis system using the developed method for
modeling and visualization of thermal processes.
As a result of optimization of the geometric parameters of the bilateral photore-
ceiver of a concentrator solar photovoltaic thermal module, a module design is pro-
posed that will allow obtaining high values of the overall efficiency of the solar module
when the photovoltaic converters do not overheat and the coolant heats up to high
values.
Solar photovoltaic thermal modules of this kind can serve as cogeneration modules
for the simultaneous generation of electric and thermal energy for the consumers own
needs, which can operate both offline and in parallel with existing energy networks.
Acknowledgment. The research was carried out on the basis of financing the state assignments
of the All-Russian Institute of the Electrification of Agriculture and the Federal Scientific
Agroengineering Center VIM, on the basis of funding of the grant “Young lecturer of RUT” of
the Russian University of Transport, on the basis of funding of the Scholarship of the President of
the Russian Federation for young scientists and graduate students engaged in advanced research
and development in priority areas of modernization of the Russian economy, direction of
modernization: “Energy efficiency and energy saving, including the development of new fuels”,
subject of scientific research: “Development and research of solar photovoltaic thermal modules
of planar and concentrator structures for stationary and mobile power generation”.
Modeling of Bilateral Photoreceiver of the Concentrator Photovoltaic Thermal Module 83
References
1. Buonomano, A., Calise, F., Vicidimini, M.: Design, simulation and experimental
investigation of a solar system based on PV panels and PVT collectors. Energies 9, 497
(2016)
2. Ibrahim, A., Othman, M.Y., Ruslan, M.H., Mat, S., Sopian, K.: Recent advances in flat plate
photovoltaic/thermal (PV/T) solar collectors. Renew. Sustain. Energy Rev. 15, 352–365
(2011)
3. Kemmoku, Y., Araki, K., Oke, S.: Long-term perfomance estimation of a 500X concentrator
photovoltaic system. In: 30th ISES Biennial Solar World Congress 2011, 710–716 (2011)
4. Nesterenkov, P., Kharchenko, V.: Thermo physical principles of cogeneration technology
with concentration of solar radiation. Adv. Intell. Syst. Comput. 866, 117–128 (2019).
https://doi.org/10.1007/978-3-030-00979-3_12
5. Rawat, P., Debbarma, M., Mehrotra, S., et al.: Design, development and experimental
investigation of solar photovoltaic/thermal (PV/T) water collector system. Int. J. Sci.
Environ. Technol. 3(3), 1173–1183 (2014)
6. Sevela, P., Olesen, B.W.: Development and benefits of using PVT compared to PV.
Sunstain. Build. Technol. 4, 90–97 (2013)
7. Kharchenko, V., Nikitin, B., Tikhonov, P., Panchenko, V., Vasant, P.: Evaluation of the
silicon solar cell modules. In: Intelligent Computing & Optimization. Advances in Intelligent
Systems and Computing, vol. 866, pp. 328–336 (2019). https://doi.org/10.1007/978-3-030-
00979-3_34
8. Panchenko, V.: Photovoltaic solar modules for autonomous heat and power supply. IOP
Conf. Ser. Earth Environ. Sci. 317 9 p. https://doi.org/10.1088/1755-1315/317/1/012002.
9. Panchenko, V., Izmailov, A., Kharchenko, V., Lobachevskiy, Y.: Photovoltaic solar
modules of different types and designs for energy supply. Int. J. Energy Optim. Eng. 9(2),
74–94. https://doi.org/10.4018/IJEOE.2020040106.
10. Panchenko, V.A.: Solar roof panels for electric and thermal generation. Appl. Solar Energy
54(5), 350–353 (2018). https://doi.org/10.3103/S0003701X18050146
11. Kharchenko, V., Panchenko, V., Tikhonov, P., Vasant, P.: Cogenerative PV thermal
modules of different design for autonomous heat and electricity supply. In: Handbook of
Research on Renewable Energy and Electric Resources for Sustainable Rural Development,
pp. 86–119. https://doi.org/10.4018/978-1-5225-3867-7.ch004
12. Sinitsyn, S., Panchenko, V., Kharchenko, V., Vasant, P.: Optimization of parquetting of the
concentrator of photovoltaic thermal module. In: Intelligent Computing & Optimization.
Advances in Intelligent Systems and Computing, vol. 1072, pp. 160–169 (2020). https://doi.
org/10.1007/978-3-030-33585-4_16
13. Panchenko, V., Kharchenko, V., Vasant, P.: Modeling of solar photovoltaic thermal
modules. In: Intelligent Computing & Optimization. Advances in Intelligent Systems and
Computing, vol. 866, pp. 108–116 (2019). https://doi.org/10.1007/978-3-030-00979-3_11
14. Panchenko, V., Chirskiy, S., Kharchenko, V.V.: Application of the software system of finite
element analysis for the simulation and design optimization of solar photovoltaic thermal
modules. In: Handbook of Research on Smart Computing for Renewable Energy and Agro-
Engineering, pp. 106–131 https://doi.org/10.4018/978-1-7998-1216-6.ch005
Formation of Surface of the Paraboloid Type
Concentrator of Solar Radiation
by the Method of Orthogonal Parquetting
Abstract. The paper discusses the geometric aspects of the surface design of
solar radiation concentrators by adjusting its surface with single elements. The
proposed method of orthogonal parquetting allows optimizing the accuracy of
manufacturing a solar radiation concentrator and the smoothness of its working
profile. The geometric characteristics of paraboloid type concentrators are
considered. The article also discusses various methods of manufacturing solar
radiation concentrators, including the fan-surface parquetting method, which is
related to the orthogonal method of parking. The optimization of the number of
elements used in the developed method occurs either by the least number of
curved components of the concentrator or by the maximum number of flat
elements. As an example of the implementation of the developed method two
solar radiation concentrators designed for photovoltaic and photovoltaic thermal
photoreceivers are presented. Concentrations of solar radiation provide uniform
illumination in focal areas.
1 Introduction
Solar plants are the fastest growing stations based on renewable energy converters
[1–5]. Most of solar stations consist of silicon planar photovoltaic modules that gen-
erate exclusively electrical energy. To increase the overall efficiency of solar modules,
reduce their nominal power and accelerate the payback period, photovoltaic thermal
solar modules are relevant, which, along with electrical energy, also generate thermal
energy in the form of a heated coolant. Such photovoltaic thermal solar modules are
divided according to the type of construction used: planar [6–9] and concentrator (with
different concentrators of solar radiation) [10–12]. Since concentrator photovoltaic
thermal modules can work with bilateral photovoltaic converters, including silicon
matrix high-voltage [4, 5, 12] (which reduces the number of used photovoltaic con-
verters), as well as receive a higher temperature coolant at the output, which increases
the overall efficiency of a solar installation, concentrators of such installations require
more complex calculations related both to the development of a profile for uniform
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 84–94, 2021.
https://doi.org/10.1007/978-3-030-68154-8_9
Formation of Surface of the Paraboloid Type Concentrator of Solar Radiation 85
illumination of the focal region where the photovoltaic converters are located and to
methods for its manufacture according to geometrically specified requirements.
Methods for calculating the thermal state of concentrator solar photovoltaic thermal
modules are also necessary [13], since as the temperature of the photovoltaic converters
increases, their electrical efficiency decreases [14], which implies the need for their
effective cooling. The main types of concentrators for use in concentrator photovoltaic
thermal installations are paraboloid type solar concentrators, however, methods for
calculating their working profile and its manufacture are also relevant and important for
research.
Ideal parabolic concentrator of solar radiation (Fig. 1) focuses parallel sunrays to the
point, which corresponds to an infinite degree of concentration [15], which does not
allow evaluating the capabilities of the concentrator of solar radiation, since the Sun has
finite dimensions.
Fig. 1. Scheme of the formation of the focal region of the paraboloid concentrator of solar
radiation
Figure 1 shows the diagram of the formation of the focal spot of the paraboloid
concentrator of solar radiation: an elementary sunray with an angular size (2
0,004654 rad) is reflected from the surface of the concentrator of solar radiation and
falls on the focal plane, where the trace of this ray is an elementary ellipse with half
shafts: a ¼ p u0 =ð1 þ cos uÞ cos u; b ¼ p u0 =ð1 þ cos uÞ, where p = 2 f –
the focal parameter of the parabola; f – focal length [15].
86 V. Panchenko and S. Sinitsyn
From different radial zones of the concentrator of solar radiation (with different
angles U), the ellipses have different sizes, which, overlapping each other, form the
density of the focal radiation. An approximate estimate of the maximum radiation
density in focus is the calculation by the formula [15]:
1
EF ¼ q sin2 Um E0 ; ð1Þ
u20
ET ¼ Emax ecr ;
2
ð2Þ
2
180
Emax ¼ E0 q h2 sin2 u; ð3Þ
p
2
h
c ¼ 3;283 103 ð1 þ cos uÞ2 ; ð4Þ
p
where r – the radius in the focal plane; h – the measure of the accuracy of the
concentrator of solar radiation.
In the calculations, the focal length f and half-opening angle Um are considered
known, from where, using the representation of the parabola equation in the polar
coordinate system, the diameter of the concentrator of solar radiation is obtained [15]:
4 f sin um
D¼ : ð5Þ
1 þ cos um
1:2
Efid ¼ Rs sin2 um E0 ð6Þ
u20
where Rs – the integral transmittance of the system; u0 – the angular radius of the Sun
equal to 0,004654 rad.; E0 – the density of direct solar radiation.
The curves of the distribution of concentration of solar radiation in the focal plane
for four concentrators of solar radiation (spherical, quasiparabolic and parabolotoric)
with the same diameters and focal lengths (500 mm each) are shown in the Fig. 2 [18].
Formation of Surface of the Paraboloid Type Concentrator of Solar Radiation 87
Fig. 3. Paraboloid type solar radiation concentrators, the base of which is made of reed, guide
ribs, satellite dishes and using the fan-shaped method of parqueting of surface
88 V. Panchenko and S. Sinitsyn
As the initial data for parqueting, a planned projection of the surface of the paraboloid
is set (Fig. 4). The method for solving the task is related to the breakdown into
projection elements with the subsequent finding on the surface of the third coordinates
corresponding to the planned breakdown.
Fig. 4. The scheme of orthogonal parqueting of the surface of the paraboloid of the concentrator
When solving it is taken into account that the mathematical model of the inner
surface is given by the equation x2 þ y2 ¼ 2 p z.
The surface of the paraboloid S is dissected by one-parameter families of projecting
planes (Fig. 5). In this case, the planned projection is divided by a network with the
shape of the cell, depending on the relative position of the projecting planes. It is most
advisable to split the surface into elements with a planned shape of rectangles or
arbitrary quadrangles.
Fig. 5. The sectional diagram of the surface of the paraboloid by families of projecting planes
The cell parameters of the surface partition network are determined taking into
account the accuracy of the surface approximation ðDn D nÞ, where the numerical
value of the limiting error parameter D n ¼ 0;15%.
If the surface of a paraboloid is represented by an approximated set of elements
having the shape of equal squares in plan (Fig. 4) Si, then their angles on the surface
S correspond to the set of points of the skeleton, approximately taken as regular.
To determine the side of the square of the unit cell of the partition, in the circle of
the trace of the paraboloid on the plane X0Y, will enter a square with a side b (Fig. 4).
The parameter b is calculated by the area of a circle of radius r:
rffiffiffiffiffiffiffi
Scir
b ¼ 1;4 : ð7Þ
p
The constructed square S is a trace of the surface part of the paraboloid area.
90 V. Panchenko and S. Sinitsyn
where Nsm = 1 – the smoothness order of the approximating contour of the surface
composed of parquet elements; D n ¼ 0;15%.; CNsm – statistical parameter.
A point frame of dimension N = 680 is placed in the grid nodes of a square with a
side b, including segments of its sides, so the total number of unit cells of the partition
is determined by the formula:
pffiffiffiffi 2
N ¼ N 1 ¼ 625 ð10Þ
So, for example, for an area S = 250 m2 of a piece of the surface of a paraboloid
having a square trace on a plane X0Y, one cell has:
S 250
Si ¼ ¼ ¼ 0;4ðm2 Þ;
N 625
therefore, the square cell of the partition of the planned projection has a side li equal
to:
pffiffiffiffi pffiffiffiffiffiffiffi
li ¼ Si ¼ 0;4 ¼ 0;63:
li ¼ 0; 9 li ¼ 0;57:
where Kj is the number of elements in the i row (Fig. 4); n is the number of rows; (XN
and XK) are designations X1 of the coordinates of the boundaries of the planned
projection of the parquet element.
So, for example, the selected element
of the plan projection Si (Fig. 4) is located in
the i row with borders yimin ; yimax and in the j column with the X1 coordinates of the
borders (XNj,i and XKj,i).
Given the equal width of all belts yj+1 = yj + li, without taking into account the
gaps between the elements of the parquet, the matrix (11) is simplified and takes the
form:
The developed method for calculating the surface of a paraboloid profile using
orthogonal parquetting can be applicable for the manufacture of the working surface of
paraboloid type solar radiation concentrators. Similar solar radiation concentrators are
used in solar modules, where photovoltaic converters (including high-voltage matrix
ones), thermal photoreceivers and also combined photovoltaic thermal photoreceivers
can be used as photoreceivers of concentrated solar radiation. In concentrator solar
photovoltaic thermal modules, along with electrical energy, the consumer receives
thermal energy in the form of a heated coolant at the output.
The profiles of such paraboloid type solar radiation concentrators, providing uni-
form illumination in the focal region on the surface of the photoreceivers, are presented
in the Fig. 6 on the left, where the solar radiation concentrator is intended for photo-
voltaic converters located on the solar concentrator itself, which is an air-cooled
radiator for photovoltaic converters. Figure 6 below shows the profile of a paraboloid
type solar radiation concentrator that provides uniform illumination in the focal region
of the photovoltaic thermal cylindrical photoreceiver, in which the side surface con-
verts concentrated solar radiation into electrical energy and the upper end surface
converts concentrated solar radiation into thermal energy.
92 V. Panchenko and S. Sinitsyn
Fig. 6. Profiles of working surfaces of paraboloid type solar radiation concentrators (on the left),
partition of their surfaces into unit elements (in the middle) and individual elements of solar
radiation concentrators themselves (on the right)
5 Conclusion
Thus, it should be noted that the planned development of solar plants will occur at a
steadily increasing pace, and the share of use of hybrid photovoltaic thermal installa-
tions will increase, which will significantly increase the need for the design, manu-
facture and study of solar radiation concentrators and in particular the paraboloid type.
The presented method of orthogonal parquetting of the surface of paraboloid type
solar concentrators will allow optimizing the manufacturing process of concentrators
already at the design stage and will make it possible to obtain almost any working
profile of paraboloid type concentrators. The orthogonal scheme of parquetting surfaces
of rotation, in particular a paraboloid, is easy to implement data preparation. Also
presented method will help to control and set the manufacturing error at the design
stage of this kind of parabolic concentrators in order to ensure the expected charac-
teristics of the focal region of the concentrator. The developed method of parquetting
surfaces allows designing and manufacturing surfaces of rotation, including solar
radiation concentrators, with specified differential-geometric requirements, which will
ensure the expected distribution of the solar radiation illumination in the focal area.
Such solar radiation concentrators provide uniform illumination in the focal region of
various types of photoreceivers (photovoltaic, thermal, photovoltaic thermal), which
will positively affect the efficiency of solar modules and their specific power.
Formation of Surface of the Paraboloid Type Concentrator of Solar Radiation 93
Acknowledgment. The research was carried out on the basis of financing the state assignments
of the All-Russian Institute of the Electrification of Agriculture and the Federal Scientific
Agroengineering Center VIM, on the basis of funding of the grant “Young lecturer of RUT” of
the Russian University of Transport, on the basis of funding of the Scholarship of the President of
the Russian Federation for young scientists and graduate students engaged in advanced research
and development in priority areas of modernization of the Russian economy, direction of
modernization: “Energy efficiency and energy saving, including the development of new fuels”,
subject of scientific research: “Development and research of solar photovoltaic thermal modules
of planar and concentrator structures for stationary and mobile power generation”.
References
1. Ibrahim, A., Othman, M.Y., Ruslan, M.H., Mat, S., Sopian, K.: Recent advances in flat plate
photovoltaic/thermal (PV/T) solar collectors. Renew. Sustain. Energy Rev. 15, 352–365
(2011)
2. Nesterenkov, P., Kharchenko, V.: Thermo physical principles of cogeneration technology
with concentration of solar radiation. Adv. Intell. Syst. Comput. 866, 117–128 (2019).
https://doi.org/10.1007/978-3-030-00979-3_12
3. Buonomano, A., Calise, F., Vicidimini, M.: Design, simulation and experimental
investigation of a solar system based on PV panels and PVT collectors. Energies 9, 497
(2016)
4. Panchenko, V.: Photovoltaic solar modules for autonomous heat and power supply. IOP
Conf. Ser. Earth Environ. Sci. 317 (2017). 9 p. https://doi.org/10.1088/1755-1315/317/1/
012002.
5. Panchenko, V., Izmailov, A., Kharchenko, V., Lobachevskiy, Y.: Photovoltaic Solar
Modules of Different Types and Designs for Energy Supply. Int. J. Energy Optim. Eng. 9(2),
74–94 (2020). https://doi.org/10.4018/IJEOE.2020040106
6. Rawat, P., Debbarma, M., Mehrotra, S., et al.: Design, development and experimental
investigation of solar photovoltaic/thermal (PV/T) water collector system. Int. J. Sci.
Environ. Technol. 3(3), 1173–1183 (2014)
7. Sevela, P., Olesen, B.W.: Development and benefits of using PVT compared to PV. Sustain.
Build. Technol. 90–97 (2013)
8. Kharchenko, V., Nikitin, B., Tikhonov, P., Gusarov, V.: Investigation of experimental flat
PV thermal module parameters in natural conditions. In: Proceedings of 5th International
Conference TAE 2013, pp. 309–313 (2013)
9. Hosseini, R., Hosseini, N., Khorasanizadeh, H.: An Experimental study of combining a
Photovoltaic System with a heating System. World Renew. Energy Congress 2993–3000
(2011)
10. Kemmoku, Y., Araki, K., Oke, S.: Long-term performance estimation of a 500X
concentrator photovoltaic system. In: 30th ISES Biennial Solar World Congress 2011,
pp. 710–716 (2011)
11. Rosell, J.I., Vallverdu, X., Lechon, M.A., Ibanez, M.: Design and simulation of a low
concentrating photovoltaic/thermal system. Energy Convers. Manage. 46, 3034–3046 (2005)
12. Kharchenko, V., Panchenko, V., Tikhonov, P.V., Vasant, P.: Cogenerative PV thermal
modules of different design for autonomous heat and electricity supply. In: Handbook of
Research on Renewable Energy and Electric Resources for Sustainable Rural Development,
pp. 86–119 (2018). https://doi.org/10.4018/978-1-5225-3867-7.ch004
94 V. Panchenko and S. Sinitsyn
13. Panchenko, V., Kharchenko, V., Vasant, P.: Modeling of solar photovoltaic thermal
modules. In: Intelligent Computing & Optimization. Advances in Intelligent Systems and
Computing, vol. 866, pp. 108–116 (2019). https://doi.org/10.1007/978-3-030-00979-3_11
14. Kharchenko, V., Nikitin, B., Tikhonov, P., Panchenko V., Vasant, P.: Evaluation of the
silicon solar cell modules. In: Intelligent Computing & Optimization. Advances in Intelligent
Systems and Computing, vol. 866, pp. 328–336 (2019). https://doi.org/10.1007/978-3-030-
00979-3_34
15. Strebkov, D.S., Tveryanovich, E.V.: Koncentratory solnechnogo izlucheniya (Concentrators
of the Solar Radiation), Moskva, GNU VIESKH [Moscow, GNU VIESH], 12–30 (2007). (in
Russian)
16. Andreev, V.M., Grilikhes, V.A., Rumyantsev, V.D.: Fotoehlektricheskoe preobrazovanie
koncentrirovannogo solnechnogo izlucheniya (Photoelectric conversion of concentrated
solar radiation), Leningrad, Nauka (Leningrad, Science) (1989). 310 p. (in Russian)
17. Zahidov, R.A., Umarov, G.Y., Weiner, A.A.: Teoriya i raschyot geliotekhnicheskih
koncentriruyushchih system (Theory and calculation of solar concentrating systems),
Tashkent, FAN (Tashkent, Science) (1977). 144 p. (in Russian)
18. Alimov, A.K., Alavutdinov, J.N., et al.: Opyt sozdaniya koncentratorov dlya modul'nyh
fotoehlektricheskih ustanovok (Experience in creating concentrators for modular photo-
voltaic plants), Koncentratory solnechnogo izlucheniya dlya fotoehlektricheskih ustanovok
(Concentrators of solar radiation for photovoltaic plants) 17–18 (1986). (in Russian)
19. Hassen, A.A., Amibe, D.A.: Design, manufacture and experimental investigation of low cost
parabolic solar cooker. ISES Solar World Congress 28 August–2 September 2011 (2011).
12 p. https://doi.org/10.18086/swc.2011.19.16
20. Chandak, A., Somani, S., Chandak, A.: Development Prince-40 solar concentrator as do it
yourself (DIY) kit. In: ISES Solar World Congress August–2 Sept. 2011 (2011). 8 p. https://
doi.org/10.18086/swc.2011.23.02
21. Diz-Bugarin, J.: Design and construction of a low cost offset parabolic solar concentrator for
solar cooking in rural areas. In: ISES Solar World Congress August–2 Septe 2011 (2011).
8 p. https://doi.org/10.18086/swc.2011.30.05
22. Sinitsyn, S., Panchenko, V., Kharchenko, V., Vasant, P.: Optimization of parquetting of the
concentrator of photovoltaic thermal module. In: Intelligent Computing & Optimization.
Advances in Intelligent Systems and Computing, vol. 1072, pp. 160–169 (2020). https://doi.
org/10.1007/978-3-030-33585-4_16
Determination of the Efficiency of Photovoltaic
Converters Adequate to Solar Radiation
by Using Their Spectral Characteristics
1 Introduction
A prerequisite for the planned analysis is the adjustment of the equipment used to
ensure the equality of the integral (without the use of light filters) short circuit currents
of the reference converter from standard solar radiation and from a laboratory light
source. A comparison of the spectral (with the help of light filters) current photore-
sponses of the reference converter from two compared light sources gives an idea of the
values of the spectral averaged (over the bandwidth of the selected filter) power of the
monochromatic line of the used laboratory light source. Figure 1 shows the spectral
dependences of the current photoresponses of a reference silicon photoconverter when
illuminating with standard solar radiation AM 1.5 (according to the rating data for
calibration of this converter) and when illuminated from a laboratory light source
(halogen incandescent lamp).
Fig. 1. Comparison of the spectral densities of current photoresponses of the reference silicon
photovoltaic converter when exposed to standard solar radiation AM 1.5 (sunlight exposure) and
a laboratory light source (halogen lamp with a water filter h = 42 mm.) for equal integral short
circuit currents
From the analysis of these dependencies it follows that the spectrum of the labo-
ratory light source is depleted in the intensity of light fluxes in its short-wave part, but
differs in increased intensity in the long-wave part. The areas bounded by these
dependencies are equal in size, since they have the integral values of short circuit
currents equal.
Determination of the Efficiency of Photovoltaic Converters Adequate 97
Figure 2 presents the averaged spectral densities of power of the light bands (50 nm
wide in accordance with the calibration of the reference converter) of the photoactive
part (for Si) of the standard solar radiation AM 1.5 spectrum. It also shows the
numerical values of the power of the light flux of each such band. Figure 2 shows the
DP0kiAM1;5 values of standard solar radiation AM 1.5 obtained by dividing its pho-
toactive part of the spectrum for silicon into bands with a width of 50 nm.
Fig. 2. Averaged spectral densities of power of the bandwidths of 50 nm wide of the standard
solar radiation AM 1.5 of the photoactive part of the spectrum for silicon (numerical spectral
densities of power of the light bands are shown above the graph and have a dimension of W/m2)
The data in Fig. 1 and Fig. 2 make it possible to determine the power of the light
bands of the used laboratory light source for each selected wavelength.
Figure 3 shows the experimentally recorded spectral dependences of the current pho-
toresponses of two different photoconverters (reference photoconverter and the studied
one) at the same level and illumination spectrum from a laboratory light source. From
the analysis of the figure it follows that, due to the individual design features and the
technology of manufacturing photoconverters, these dependencies are not identical.
98 V. Kharchenko et al.
Comparison of the photoresponses of the converters allows to determine for the studied
photoconverter sample the current photoresponse per unit of the power of the light
current for a given spectrum wavelength.
Fig. 3. Comparison of the spectral densities of the current photoresponses of the silicon
reference and investigated photoconverters under exposure from a laboratory light source
It should be noted that the nature (form) of the volt-ampere characteristics of the
converter should not change if the light fluxes of some lines are mentally replaced by
equivalent ones according to the current photoresponse of other lines, and therefore, the
position of the operating point on the volt-ampere characteristics will not change. This
circumstance is due to the fact that mobile charge carriers (primarily electrons),
according to the kinetic energy obtained by the absorption of high-energy photons,
almost instantly thermolize according to the thesis described in [3]. This circumstance
makes these mobile charge carriers indistinguishable by the history of their origin.
In accordance with the foregoing, the power of monochromatic radiation with a
wavelength ki, generating the total short circuit current I0, is determined by the
expression:
P P ð1Þ
where: DP0kiAM1;5 is the density of power of the light fluxes of standard solar
radiation AM 1.5 with a wavelength ki and bandwidth of 50 nm;
SPVC is the area of the investigated photovoltaic converter;
i0kilab:ref : is the spectral density of the short circuit current of the reference photo-
voltaic converter from the laboratory light source at a wavelength ki;
i0kiAM1;5cal: is the spectral density of the short circuit current of the reference
transducer from the solar radiation AM 1.5 at a wavelength ki;
I0 is the short circuit current of the investigated photovoltaic converter according to
the volt-ampere characteristics;
i0kilab:PVC is the spectral density of the short circuit current of the investigated
photovoltaic converter under illumination from a laboratory light source at a wave-
length ki;
Δkcal. is the step of the calibration of the reference photovoltaic converter (usually
50 nm).
The spectral values of the coefficient of performance of the investigated photo-
voltaic converter will be determined according to the expression:
Pmax PVC
COPki ¼ ; ð2Þ
Pki
Figure 4 shows the dependence of the spectral values of efficiency of the investi-
gated serial silicon photovoltaic converter [12, 13] versus wavelength. From the
analysis of the obtained dependence, it follows that the COPki gradually increases from
zero (at ki = 0.4 lm) to 40% with ki = 0.95 lm with increasing wavelength. An
almost zero value of the COPki is observed under conditions when photons of the
corresponding (small) wavelength are completely absorbed by the doped layer of the
converter without p-n separation – by the transition of electron-hole pairs.
Fig. 4. The characteristic form of the spectral values of the efficiency of silicon photovoltaic
converters
where COPki are spectral values of the efficiency of the investigated photoconverter
obtained according to expression 2;
Kki is the fraction of the luminous flux power of the wavelength ki in the total
luminous flux power of a given spectrum.
Determination of the Efficiency of Photovoltaic Converters Adequate 101
6 Conclusion
Acknowledgment. The research was carried out on the basis of financing the state assignments
of the All-Russian Institute of the Electrification of Agriculture and the Federal Scientific
Agroengineering Center VIM, on the basis of funding of the grant “Young lecturer of RUT” of
the Russian University of Transport, on the basis of funding of the Scholarship of the President of
the Russian Federation for young scientists and graduate students engaged in advanced research
and development in priority areas of modernization of the Russian economy, direction of
modernization: “Energy efficiency and energy saving, including the development of new fuels”,
subject of scientific research: “Development and research of solar photovoltaic thermal modules
of planar and concentrator structures for stationary and mobile power generation”.
References
1. Bird, R.E., Hulstrom, R.L., Lewis, L.J.: Terrestrial solar spectral, data sets. Sol. Energy 30
(6), 563–573 (1983)
2. Kharchenko, V., Nikitin, B., Tikhonov, P., Panchenko, V., Vasant, P.: Evaluation of the
silicon solar cell modules. In: Intelligent Computing & Optimization. Advances in Intelligent
Systems and Computing, vol. 866, pp. 328–336 (2019). https://doi.org/10.1007/978-3-030-
00979-3_34
3. Arbuzov, Y.D., Yevdokimov, V.M.: Osnovy fotoelektrichestva (Fundamentals of Photo-
electricity), Moscva, GNU VIESKH (Moscow, GNU VIESH) (2007). 292 p. (in Russian)
4. Kharchenko, V.V., Nikitin, B.A., Tikhonov, P.V.: Theoretical method of estimation and
prediction of PV cells parameters. Int. Sci. J. Alt. Energy Ecol. 4(108), 74–78 (2012)
5. Kharchenko, V.V., Nikitin, B.A., Tikhonov, P.V.: Estimation and forecasting of PV cells
and modules parameters on the basis of the analysis of interaction of a sunlight with a solar
cell material. In: Conference Proceeding - 4th International Conference, TAE 2010, pp. 307–
310 (2010)
6. Nikitin, B.A., Gusarov, V.A.: Analiz standartnogo spektra nazemnogo solnechnogo
izlucheniya intensivnost'yu 1000 Vt/m2 i ocenka na ego osnove ozhidaemyh harakteristik
kremnievyh fotoelektricheskih preobrazovatelej (Analysis of the standard spectrum of
ground-based solar radiation with an intensity of 1000 W/m2 and assessment based on it of
the expected characteristics of silicon photovoltaic converters). Avtonomnaya energetika:
tekhnicheskij progress i ekonomika (Auton. Power Eng. Tech. Progress Econ. no. 24–25,
50–60 (2009). [in Russian]
102 V. Kharchenko et al.
7. Strebkov, D.S., Nikitin, B.A., Gusarov V.A.: K voprosu ocenki effektivnosti raboty
fotopreobrazovatelya pri malyh i povyshennyh urovnyah osveshchennosti (On the issue of
evaluating the efficiency of the photoconverter at low and high levels of illumination).
Vestnik VIESKH (Bulletin VIESH), 119 (2012). (in Russian)
8. Nikitin, B.A., Mayorov, V.A., Kharchenko, V.V.: Issledovanie spektral'nyh harakteristik
solnechnogo izlucheniya dlya razlichnyh velichin atmosfernyh mass (Investigation of the
spectral characteristics of solar radiation for various atmospheric masses). Vestnik VIESKH
[Bulletin of VIESH], 4(21), 95–105 (2015). (in Russian)
9. Nikitin, B.A., Mayorov, V.A., Kharchenko V.V.: Vliyanie velichiny atmosfernoj massy na
spektral'nuyu intensivnost' solnechnogo izlucheniya (The effect of atmospheric mass on the
spectral intensity of solar radiation). Energetika i avtomatika (Energy Autom.) 4(26), pp. 54–
65 (2015). (in Russian)
10. Kharchenko, V., Nikitin, B., Tikhonov, P., Adomavicius, V.: Utmost efficiency coefficient of
solar cells versus forbidden gap of used semiconductor. In; Proceedings of the 5th
International Conference on Electrical and Control Technologies ECT-2010, pp. 289–294
(2010)
11. Strebkov, D.S., Nikitin, B.A., Kharchenko, V.V., Arbuzov, Yu.D., Yevdokimov, V.M.,
Gusarov V.A., Tikhonov, P.V.: Metodika analiza spektra issleduemogo istochnika sveta
posredstvom tarirovannogo fotopreobrazovatelya i komplekta svetofil'trov (Method of
spectrum analysis of the studied light source by means of a calibrated photoconverter and a
set of light filters). Vestnik VIESKH (Bulletin of VIESH), 4(9), 54–57 (2012). (in Russian)
12. Panchenko, V.: Photovoltaic solar modules for autonomous heat and power supply. IOP
Conf. Ser. Earth Environ. Sci. 317, 9 p. (2019). https://doi.org/10.1088/1755-1315/317/1/
012002
13. Panchenko, V., Izmailov, A., Kharchenko, V., Lobachevskiy, Ya.: Photovoltaic solar
modules of different types and designs for energy supply. Int. J. Energy Optim. Eng. 9(2),
74–94 (2020). https://doi.org/10.4018/IJEOE.2020040106
Modeling of the Thermal State of Systems
of Solar-Thermal Regeneration of Adsorbents
1 Introduction
The leadership of Uzbekistan has set tasks to reduce the energy and resource intensity
of the economy, the widespread introduction of energy-saving technologies in pro-
duction, and the expansion of the use of renewable energy sources. The implementation
of these provisions, including increasing the efficiency of the use of solar energy in
thermotechnological processes of fruit and vegetable storages, is considered one of the
most important tasks.
Based on our studies, we developed a solar air-heating system for thermal regen-
eration of adsorbents and active ventilation of fruit and vegetable chambers [1–3].
Studies have shown that the thermal regime of thermal regeneration of adsorbents
depends on the intensity of convective heat transfer between the surface of the
adsorbent and the hot washer air. The main heat engineering parameters of convective
heat transfer is the heat transfer coefficient, which depends on many factors [4–8].
The aim of this work is to study the heat transfer and thermal regime of the developed
system of solar-thermal adsorbent regeneration in non-stationary mode. Convective
heat transfer during forced air movement through a fixed adsorbent layer in the solar
regeneration mode has a pronounced unsteady character, that is, air temperatures and
adsorbent temperatures change both in time and in space. In order to calculate the cycle
of adsorption plants, determine the duration of the solar regeneration regime, and select
the optimal thermotechnical parameters for their implementation, it is necessary to
calculate the temperature field and calculate the change in the temperature of the
adsorbent along the length of the adsorber at the studied time.
3 Research Results
The heat transfer coefficient in the granular layer of activated carbon in the processes of
thermal regeneration is determined by the method of modeling convective heat transfer
using criteria-based similarity equations [9, 10]. The experimental results were pro-
cessed using the following criteria similarity equations, that is, according to the formula
of V.N. Timofeeva:
The results of studies to determine the heat transfer coefficient from air to adsorbent
are shown in the Table 1.
Table 1. The results of studies of heat transfer in a granular adsorbent layer during thermal
regeneration by atmospheric air
№ W, m/s Reliq Nuliq a, W/m2 K
1 0,2 141,2 14,95 35
2 0,3 211,7 22 51,15
3 0,4 282,3 26,7 62,1
4 0,5 352,9 31 72
5 0,6 423,5 35,1 81,6
6 1,0 705,9 49,4 114,6
The process of unsteady heat transfer in the main apparatus of the adsorption unit –
the adsorber – allows presenting the physical model of the process as follows. The
process is one-dimensional, there is no heat exchange with the environment through the
side surface of the adsorber, at any time the temperature of the adsorbent particle can be
considered constant throughout its volume, convective heat transfer between the air
Modeling of the Thermal State of Systems 105
flow and the adsorbent layer is decisive, heat transfer is heat conduction through the air
and the layer in the axial direction and the heat transfer by radiation is small and can be
neglected, the air pressure during movement through the layer remains unchanged, the
mass flow rate of air is constant, all the thermal properties of the air and layer are
considered constant and independent of temperature [11–13].
When considering non-stationary temperature fields in adsorbers in heat transfer
processes, in addition to the above assumptions, it is assumed that the adsorption
properties of the layer do not affect the heat transfer process – the specific mass air flow
in any section remains constant, and the thermal effects associated with adsorption and
desorption are insignificant, and they can be neglected.
To derive the basic equations for air and particles enclosed in the elementary
volume of the layer, heat balance equations are compiled. For air: the change in the
enthalpy of air over the considered period of time plus the heat introduced by the flow
is equal to the amount of heat transferred to the layer during convective heat transfer.
For particles of the layer: the change in the enthalpy of particles over the considered
period of time is equal to the amount of heat transferred to the layer during convective
heat transfer [14–18].
Fig. 1. Design scheme of unsteady heat transfer during heating of the adsorbent layer
The desired functions are the air temperature ta and the temperature of the particles
of the layer tl, these functions are functions of two independent variables (Fig. 1).
The temperature of the adsorbent layer is determined by the following data, among
which mad – the mass of the adsorbent, kg; cl – specific heat capacity of the adsorbent J/
(kg °C); t0 – initial temperature of the adsorbent layer, °C; tl = t (s) – temperature of
the adsorbent at the time s (s 0).
106 G. Uzakov et al.
The amount of accumulated heat in the adsorbent grain layer at time s is determined
by:
and at time s = 0:
Q0 ¼ cl mad t0 : ð4Þ
Within the time ds, the amount of heat received by the adsorbent layer will increase
by:
This amount of heat is transferred to the adsorbent by air at the constant temper-
ature during convective heat transfer between the air and the adsorbent:
where a – the heat transfer coefficient, W/m2 ºC; ta – air temperature, ºC.
Equating (5) and (6), the following heat balance equation is obtained:
dt a
¼ ds: ð8Þ
ta tl cl mad
determined by:
Z
dt a ~1
¼ sþN ð10Þ
ta tl cl mad
ta tl ¼ x; ð11Þ
Modeling of the Thermal State of Systems 107
Taking into account expressions (11), (12) and the presence of the integration
constant C1 in Eq. (10), after integration we obtain:
Z
dx
¼ ln x: ð13Þ
x
Replace the left side of Eq. (10), taking into account (12):
a ~1
ln x ¼ s þ ln N ð14Þ
cl mad
The resulting Eq. (13) after the conversion, taking into account (10), has the fol-
lowing form:
1 a
s
t ¼ ta ecl mad : ð15Þ
1
1 0 1 1
t0 ¼ ta e ¼ ta or ¼ ta t0 : ð16Þ
~1
N ~1
N ~1
N
from (15) taking into account (17) we obtain the final equation:
The obtained Eq. (18) allows us to determine the temperature of heating of the
adsorbent in the solar air-heating installation, taking into account the environmental
parameters and thermal characteristics of the adsorbent itself (a, cl, mad).
From the analysis of the obtained dependence, we can conclude that the heating
temperature of the adsorbent in the adsorber is determined by its mass and heat
capacity, initial temperature, heat transfer coefficient, and duration of heating. The
resulting Eq. (18) will solve the following problems:
108 G. Uzakov et al.
1) Given the maximum temperature for heating the adsorbent, it is possible to deter-
mine the maximum duration of its heat treatment with air through a solar air heater
(the duration of the thermal regeneration of adsorbents).
2) Given the maximum duration of heat treatment of the adsorbent, we determine the
possible temperature of its heating.
The resulting mathematical Eq. (18) is convenient for practical engineering cal-
culations of thermal regeneration systems of adsorbents and does not require a large
amount of initial and experimental data.
We will calculate based on the following source data:
W = 0,2 m/s; a = 35 W/(m2 °C); cl = 0,84 kJ/(kg°C); mad = 28,8 kg; ta = 60 °C;
t0 = 18 °C;
a 35
b¼ ¼ ¼ 1;44 103
cl mad 0;84 103 28;8
Based on the calculation results, a graph of the temperature change of heating the
adsorbent in the adsorber was constructed (Fig. 2).
Fig. 2. The graph of the temperature of heating the adsorbent (activated carbon) in the adsorber
of a solar air-heating installation
4 Conclusion
1. The proposed mathematical model of the temperature field of the adsorbent layer
during solar heating allows one to determine the change in the temperature of the
adsorbent along the length of the adsorber and in time and also takes into account
the influence of the thermal characteristics of the adsorbent itself.
2. At the maximum heating temperature of the adsorbent, it is possible to determine
the duration of heat treatment with hot air heated in a solar air-heating installation.
Modeling of the Thermal State of Systems 109
3. As can be seen from Fig. 2 with increasing heat transfer coefficient (a) from air to
the adsorbent, the heating intensity of the adsorbent layer increases.
4. Thus, the obtained results of the study of the thermal regime of solar-thermal
adsorbent regeneration and the mathematical model of the temperature field of the
adsorbent layer make it possible to qualitatively control the process of solar-thermal
regeneration of the adsorbent and choose the optimal technological parameters of
the gas medium regulation systems in fruit storage.
References
1. Khuzhakulov, S.M., Uzakov, G.N., Vardiyashvili, A.B.: Modelirovanie i issledovanie
teplomasso- i gazoobmennyh processov v uglublennyh plodoovoshchekhranilishchah
(Modeling and investigation of heat and mass and gas exchange processes in in-depth
fruit and vegetable storages). Problemy informatiki i energetiki (Prob. Inf. Energy), no. 6,
52–57 (2010). (in Russian)
2. Khujakulov, S.M., Uzakov, G.N., Vardiyashvili, A.B.: Effectiveness of solar heating
systems for the regeneration of adsorbents in recessed fruit and vegetable storages. Appl.
Solar Energy 49(4), 257–260 (2013)
3. Khujakulov, S.M., Uzakov, G.N.: Research of thermo moisten mode in underground
vegetable storehouses in the conditions of hot-arid climate. Eur. Sci. Rev. 11–12, 164–166
(2017)
4. Uzakov, G.N., Khuzhakulov, S.M.: Geliovozduhonagrevatel'naya ustanovka s solnechno-
termicheskoj regeneraciej adsorbentov (Helio-air-heating installation with solar-thermal
regeneration of adsorbents). Tekhnika. Tekhnologii. Inzheneriya [Technics. Technol. Eng.],
no. 2, 7–10 (2016). https://moluch.ru/th/8/archive/40/1339/. (in Russian)
5. Uzakov, G.N., Khuzhakulov, S.M.: Issledovanie teploobmennyh processov v sistemah
solnechno-termicheskoj regeneracii adsorbentov (Study of heat exchange processes in
systems of solar-thermal regeneration of adsorbents). Tekhnika. Tekhnologii. Inzheneriya
(Tech. Technol. Eng.) (2), 10–13 (2016). https://moluch.ru/th/8/archive/40/1340/. (in
Russian)
6. Uzakov, G.N., Khuzhakulov, S.M.: Issledovanie temperaturnyh rezhimov geliovozduhona-
grevatel'noj ustanovki dlya sistem termicheskoj regeneracii adsorbentov (Study of the
temperature conditions of a solar air heating installation for thermal regeneration of
adsorbents). Geliotekhnika (Solar Eng.) (1), 40–43 (2017). (in Russian)
7. Abbasov, E.S., Umurzakova, M.A., Boltaboeva, M.P.: Effektivnost’ solnechnyh voz-
duhonagrevatelej (Efficiency of solar air heaters). Geliotekhnika (Solar Eng.) (2), 13–16
(2016). (in Russian)
8. Klychev, Sh.I., Bakhramov, S.A., Ismanzhanov, A.I.: Raspredelennaya nestacionarnaya
teplovaya model’ dvuhkanal’nogo solnechnogo vozduhonagrevatelya (Distributed non-
stationary thermal model of a two-channel solar air heater). Geliotekhnika (Solar Eng.) (3),
77–79 (2011). (in Russian)
9. Akulich, P.V.: Raschety sushil'nyh i teploobmennyh ustanovok (Calculations of drying and
heat exchange plants). Minsk, Belarus. navuka (Minsk, Belarus. Navuka) (2010). 443 p. (in
Russian)
10. Mikheev, M.A., Mikheeva, I.M.: Osnovy teploperedachi (Fundamentals of heat transfer).
Moskva, Energiya (Moscow, Energy) (1977). 320 p. (in Russian)
110 G. Uzakov et al.
11. Yanyuk, V.Ya., Bondarev, V.I.: Holodil'nye kamery dlya hraneniya fruktov i ovoshchej v
reguliruemoj gazovoj srede (Refrigerators for storing fruits and vegetables in a controlled gas
environment). Moskva, Legkaya i pishchevaya promyshlennost’ (Moscow, Light and food
industry) (1984). 128 p. (in Russian)
12. Kharitonov, V.P.: Adsorbciya v kondicionirovanii na holodil'nikah dlya plodov i ovoshchej
(Adsorption in conditioning on refrigerators for fruits and vegetables). Moskva, Pishchevaya
promyshlennost’ (Moscow, Food industry) (1978). 192 p. (in Russian)
13. Chabane, F.: Design, developing and testing of a solar air collector experimental and review
the system with longitudinal fins. Int. J. Environ. Eng. Res. 2(I. 1), 18–26 (2013)
14. Henden, L., Rekstad, J., Meir, M.: Thermal performance of combined solar systems with
different collector efficiencies. Sol. Energy 72(4), 299–305 (2002)
15. Kolb, A., Winter, E.R.F., Viskanta, R.: Experimental studies on a solar air collector with
metal matrix absorber. Sol. Energy 65(2), 91–98 (1999)
16. Kurtas, I., Turgut, E.: Experimental investigation of solar air heater with free and fixed fins:
efficiency and exergy loss. Int. J. Sci. Technol. 1(1), 75–82 (2006)
17. Garg, H.P., Choundghury, C., Datta, G.: Theoretical analysis of a new finned type solar
collector. Energy 16, 1231–1238 (1991)
18. Kartashov, A.L., Safonov, E.F., Kartashova, M.A.: Issledovanie skhem, konstrukcij,
tekhnicheskih reshenij ploskih solnechnyh termal’nyh kollektorov (Study of circuits,
structures, technical solutions of flat solar thermal collectors). Vestnik YUzhno-Ural’skogo
gosudarstvennogo universiteta (Bull. South Ural State Univ.), (16), 4–10 (2012). (in
Russian)
Economic Aspects and Factors of Solar Energy
Development in Ukraine
Renewable energy sources (RES) are one of the priorities of energy policy and
instruments to reduce carbon emissions. Efforts by countries to address this issue under
the Kyoto Protocol are not yielding the expected effect. The first steps of the world
community to solve this problem began at the 18th Conference of the Parties to the UN
Framework Convention and the 8th meeting of the Parties to the Kyoto Protocol, which
took place from November 26 to December 7, 2012 in Doha (Qatar) [1]. Under the
second period of the Kyoto Protocol (2013–2020), Ukraine has committed itself to
reducing greenhouse gas emissions by 20% (from 1990 levels) and has announced a
long-term goal by 2050 - to reduce emissions by 50% compared to 1990.
Why is the number of RES use projects growing in Ukraine, despite the fact that
tariffs are falling?
It turns out that at the same time there were several positive factors that determine
the development of alternative energy. The first factor is regulatory. In addition to
reducing green energy tariffs, the state guarantees that support will last long enough to
recoup investment in energy facilities.
As in some European countries, where the tariff decreases as the EU targets are met
(20% of energy is obtained from renewable sources by 2020), Ukraine also has
mechanisms in place to reduce it. The green tariff will be reduced from the base level in
2009 by 20% in 2020 and another 30% in 2025, and abolished in 2030 as a temporary
incentive for the development of RES.
The second factor in stimulating the construction of stations is that the conditions
for the construction of new facilities have been liberalized. In particular, the Verkhovna
Rada changed the rule on the mandatory local component in the equipment. Instead, an
incentive mechanism was introduced: the more domestic components in the station, the
higher the rate of tariff increase.
Another important factor is the rapid decline in technology in the world. In recent
years, capital expenditures for the construction of SES have decreased significantly, as
the cost of equipment has decreased. As of 2017, investments in 1 MW of capacity in
Ukraine fluctuate at the level of 0.75–1.05 million euros. Payback periods of projects
are 6–7 years. China intends to put into operation 15–20 GW of solar capacity
annually. By 2020, their total number should triple. Such a plan envisages investments
in China’s energy - $368 billion, this amount of investment (one of the main manu-
facturers of solar energy station (SES) components) will help reduce the cost of SES.
A factor in reducing the cost of SES, too, is to increase the efficiency of the
elements of SES (photovoltaic panels (FEP), batteries, etc.). To date, the production of
most commercial modules of solar cells is based on crystalline Si (I generation FE) and
amorphous thin-film SE with a large area of FE with a value of η * 5–8% (II-
generation FE). The concept of III-generation is the use of nano- and microstructures
(microwires). The main characteristic of FEP is the efficiency of photoelectric con-
version or efficiency (efficiency), which for the currently available technological
industrial FEP is in the range from 7% to 18% [1–7], and in laboratory developments
reaches 39–43% [4, 8].
Thus, the improvement of technologies and efficiency of SES elements has made
solar generation one of the leaders in terms of capacity growth not only in Ukraine but
also in the world.
One of the main elements of SES are batteries for electricity storage. Electricity
storage technologies are also developing rapidly - lithium-ion batteries, hydrogen
energy storage technologies, supercapacitors. These innovative developments have
higher productivity and lower cost.
Unfortunately, the current electricity storage technologies in Ukraine have a high
cost. One kW of storage capacity costs from $ 500 to $ 3,000. It is expected that within
3–5 years the price will drop by a third or more.
The factor that negatively affects the market of the domestic segment of RES is the
low technical condition of electrical networks. For example, due to the high density of
buildings in the Kyiv region, private homes have problems due to insufficient capacity
of electrical networks.
Importantly, the capacity of solar panels increased almost sevenfold: from 2.2 MW
(2015) to 16.7 MW at the end of 2016. That is, households began to install more
powerful panels. For all those who installed solar panels in 2017, there will be a fairly
high green tariff of 19 cents per 1 kWh. Gradually, this tariff will be reduced on the
same principle that works for industrial stations - up to 14 cents by 2030.
An important factor that encourages private individuals to install panels is the
reduction in the cost of small SES technologies. Ekotechnik Ukraine Group forecasts a
Economic Aspects and Factors of Solar Energy Development in Ukraine 113
5–10% drop in prices for solar power plant equipment for private households. In 2017,
the average price of the entire set of equipment is about 900–950 euros per 1 kW of
SES.
Ukrainian startup SolarGaps has offered a new solution - the world’s first sun blinds
have appeared. The project has already attracted the first investments and is preparing
to enter various markets around the world. The equipment of a usual window blinds in
the apartment, according to the startup, will cost $300. SolarGaps blinds in a three-
room apartment with windows facing south will be able to produce up to 600 W h, or
about 4 kW h per day (100 kW h hour per month).
Thus, solar energy is gradually becoming the cheapest source of energy in many
countries in the long run. Ukraine is no exception. Solar activity in Ukraine is enough
to ensure a return on investment for 6–7 years using the green tariff and 13–15 years
without it. This term can be compared with the payback of a classic thermal power
plant.
Ukraine has all the natural and regulatory prerequisites for the development of
RES. The share of green generation while maintaining favorable factors may reach
30% by 2035 due to the construction of 15 GW of new SES and wind farms.
Investments in RES and Ukraine’s economy will create tens of thousands of jobs –
which is an important social factor.
It is important in determining the conditions for the development of SES to
establish the degree of influence of various factors. In Fig. 1 presents a functional
diagram of the relationship of factors for two types of SES with a capacity of
P > 30 kW and P < 30 kW (private sector).
Analytical relationship between the function of SES development and factors:
ð1Þ
ð2Þ
Social factor X7
Logistics X8
Climatic factor X9
Fig. 1. Functional diagram of the relationship between the factors of development of solar
electricity.
Economic Aspects and Factors of Solar Energy Development in Ukraine 115
Since electricity is a commodity, and the cost of its generation and transportation
changes over time, it is advisable to consider a new principle - the formation of the
tariff for each consumer in the dynamics (during the period Dt, when the components of
the tariff are constant). This tariff can be called - dynamic.
Dynamic tariff (DT) is an integrated indicator of the current cost of electricity at the
input of the consumer, which is formed on the basis of the actual cost of electricity
from sources in the MicroGrid system (see Fig. 1), the cost of electricity losses during
transportation, taxes, planned profits and a number of functional factors, which
determine the management of the balance of generation and consumption of electricity,
the impact of consumers and the MicroGrid system on the quality of electricity.
The scheme of Fig. 2 shows the main elements of the MicroGrid system: con-
sumers – various objects of the settlement (utility and industrial spheres); SES - solar
power plants (can be with electricity storage); reclosers – switching devices with
automation and remote control; MicroGrid system control center – a server with a
computer center and an intelligent system control algorithm; information transmission
system from each element of the MicroGrid system to the control center (elements –
116 V. Kozyrsky et al.
SES, electricity storage, consumer meters, reclosers – equipped with means of trans-
mitting and receiving information).
During the operation of the MicroGrid system during the day the electrical load of
consumers changes and as the loads in the nodes of the circuit Fig. 2 will be controlled
on part of the reclosers, and the other part of the shutdown – the circuit configuration
will be optimized by the minimum total power loss in the MicroGrid system RDPmin .
Thus, the scheme of electricity transportation from SES can change and the consumer
can receive electricity from different sources during the day.
Functionality of determination of the dynamic tariff on input of the consumer of the
electric power:
ТД(Ci, Ki) = (Т1(С1)+Т2(С2) +Т3(С3) +Т4(С4) – Т5(С5)) · К1(S) · К2(I1) · К3(І2) ð3Þ
where the components of the electricity tariff: T1(C1) – the dependence of the tariff
for a particular consumer on the cost of electricity produced; T2(C2) – dependence of
the tariff on the cost of electricity losses for transportation; T3(C3) – the dependence of
the tariff on the amount of taxes; T4(C4) – a component determined by the planned
profit of the MicroGrid system; T5(C5) – a component determined by the sale of
electricity to the centralized power supply system at a “green tariff”; K1(S) – coefficient
that takes into account the management of the balance of generation and consumption
of electricity; K2(I1) – coefficient that takes into account the negative impact of the
consumer on the quality of electricity; K3(I2) – is a factor that takes into account the
negative impact of the MicroGrid system on the quality of electricity.
The coefficient K1(S) is determined by the total electrical load S of the MicroGrid
system and motivates consumers to use electricity during hours of minimum electrical
load of the system and can take values within n K1(S) m, for example – n > 0, and
m < 2 and can be determined on a contractual basis between electricity consumers, if they
are co-owners of the MicroGrid system, or consumers and the owner of the system.
The coefficients K2(U1) and K3(U2) are determined by the deviation of electricity
quality parameters from the norms and determine the motivation of the consumer and
the MicroGrid system not to reduce the quality of electricity.
Component T1(C1) determines the dependence of the tariff on the cost of electricity
produced by the generating installation, from which the consumer receives electricity
during the time Dt. In the case of a closed mode of operation of electrical networks
(see Fig. 2) of the MicroGrid system, the daily change of electrical loads of consumers
will be a switching change of the network. In this case, in order to optimize the mode of
operation of the network by the criterion RDPMIH (minimum total power losses during
the time period Dt) there is a change in the configuration of the circuit and the con-
sumer can receive electricity from one or more generating units. Since the cost of
electricity produced from different generating units will be different, the dynamics of
this component of the tariff for a particular consumer will change.
Component T2(C2) determines the dependence of the tariff on the cost of electricity
losses for its transportation to the consumer. Based on the conditions of formation in
the MicroGrid system of component T1(C1), and periodic changes in the configuration
time of the electrical network, it can be concluded that the component of tariff T2(C2)
should be calculated for the consumer for each time period Dt.
Economic Aspects and Factors of Solar Energy Development in Ukraine 117
Component T3(C3) determines the dependence of the tariff on the amount of taxes
and is a constant value at constant tax rates.
Component T4(C4) is determined by the planned profit of the MicroGrid system
and can be determined on a contractual basis by consumers as owners, or consumers
and owners of the MicroGrid system.
Component T5(C5) is determined by the sale of electricity to the centralized power
supply system at a “green tariff” and can be included in the formation of the tariff on a
contractual basis by consumers, owners of the MicroGrid system.
1 Conclusions
1. The most promising concept for the development of human energy supply systems
is the direction of Smart Grid technology, and for remote areas, such as rural areas -
the creation of MicroGrid systems.
2. Modern technical means allow to create a system of online accounting of the cost of
electricity input to the consumer (in real time), which requires the solution of a
number of economic problems.
3. Introduction of the concept of dynamic tariff - an integral indicator of the current
cost of electricity at the input of the consumer will reduce electricity losses for its
transportation, which will significantly affect the cost of electricity consumed by
consumers.
References
1. Introduction to Microgrids – What is a Microgrid. ABB. https://new.abb.com/distributed-
energy-microgrids/introduction-to-microgrids
2. International Energy Agency. https://www.iea.org/
3. European Commission. https://ec.europa.eu/
4. Bondarenko, S.A., Zerkina, O.O.: Smart Grid as a basis for innovative transformations in the
electricity market. BUSINESSINFORM, no. 4, pp. 105–111 (2019)
5. Cheremisin, M.M., Cherkashina, V.V., Popadchenko, S.A.: Features of introduction of smart
grid technologies in the electric power industry of Ukraine. Sci. J. Sci. Rise /4/2(9), pp. 27–
31 (2015)
6. Kozyrsky, V.V., Guy, O.V.: SMART GRID technologies in energy supply systems:
monograph. Comprint, Kyiv (2015). 336 p.
7. LAW of Ukraine on Voluntary Association of Territorial Communities (Vidomosti
Verkhovnoi Rady (VVR)), no. 13, p. 9 (2015)
8. Forst, M.: Germany’s module industry poised for growth. SUN Wind Energy 5, 256–263
(2011)
9. Bekirov, E.A., Khimich, A.P.: Computer modeling of complex power systems with solar
energy concentrators. Renew. Energy 1(24), 74–81 (2011)
10. Solar energy: industry review: based on materials from Nitol Solar Limited. - World Wide
Web access mode: https://nitolsolar.com/rusolarenergy/
11. Eckart von Malsen. Opportunities for large-scale projects. SUN Wind Energy 5, 254–255
(2011)
12. Solar energy/Wikipedia. [Free Internet Encyclopedia]. – World Wide Web access mode:
https://en.wikipedia.org/wiki/Solar_energy
A Method for Ensuring Technical Feasibility
of Distributed Balancing in Power Systems,
Considering Peer-to-Peer Balancing
Energy Trade
Mariusz Drabecki(&)
Abstract. In this paper a method for ensuring network feasibility of power flow
(in terms of all network constraints), when energy in power system is balanced
via peer-to-peer contracts, is proposed and analyzed. The method considers
subjective benefits (utility functions) of market participants. The method is
based on two optimization problems originated from Optimal Power Flow
standard formulations, which can be solved by system’s operator. It possibly can
be used in power systems with high penetration of distributed energy resources
(DERs) to give incentive to build and actively control those resources by market
participants. The method was tested on 9-bus test system, under three different
preference scenarios.
1 Introduction
aiming to maximize the social welfare function, which usually means minimizing the
overall generation cost, subject to all power flow technical constraints. Although being
least costly for the aggregated set of energy consumers, this approach completely
neglects individual preferences of market participants and of bilateral, peer-to-peer,
agreements made between them. I argue that allowing them in the balancing process
would add more freedom to the market itself.
Such a liberated way of energy balancing is especially interesting when considering
market mechanisms for power systems integrating many distributed energy resources
(DERs). This energy balancing scheme is important to systems with high DERs pen-
etration, as it directly gives incentives to market participants to both build DERs and
actively control their power output, to maximize participants’ subjective profits. Dis-
tributed power systems, where energy is traded on a peer-to-peer basis are gaining
recognition both theoretically and practically in multiple integration trials performed
worldwide [2].
Yet, as described in the following section, such a distributed self-balancing and unit
self-commitment may cause problems in terms of technical feasibility of power flow,
resulting from so-dispatched units. Some other problems may also arise while trying to
optimally balance demand and supply of energy in the system. Addressing these
technical issues lay in the scope of this paper and form its contribution.
Some interesting research in this field has already been conducted. Authors of [2]
outlined that depending on whether network constraints are taken into account or not,
and whether individual actions of peers are controlled centrally multiple ways of peer-
to-peer transactions can be identified, with their associated control models. One more
approach towards optimizing distributed p2p trade on the market was given in [3]. Yet,
authors of the latter neglected all network constraint and arising technical problems.
These were addressed in [4], yet limited to active power only. However, both of these
papers, together with [5] addressed the issue of maximizing subjective benefits of
market participants. Authors of [6] have pointed that p2p trading might be even more
beneficial when consumers optimize their power consumption and later trade nega-
watts. Some other examples of related research on relevant models of trade opti-
mization are given in [7–9]. However, the role of system’s operator in these works
remains unclear and possibly suppressed.
According to authors of [2, 7, 10], the technical feasibility issues (network con-
straints) might be addressed by special control systems at generation/load bus level –
Energy Management Systems. They will need to have capabilities of limiting possible
generation/demand of a given market participant to ensure network feasibility of the
dispatch. However, these systems would require correct integration and control
knowing the full picture of current situation in the grid. Therefore, it is reasonable to
assume that, at least in the period of transition from centralized to fully distributed
systems architecture, the system’s operator is a relevant entity to guard security and
stability of power supply to its customers, as it is the operator who knows all technical
issues of the grid.
This paper addresses the problem of possible infeasibility of power flow (both
active and reactive) when energy balancing in the system is accomplished through
peer-to-peer trading. In other words, the paper proposes a method for finding feasible
generating units dispatch when peer-to-peer balancing energy contracts are made.
120 M. Drabecki
The main goal behind any electrical power system is to safely and securely provide the
demanded amounts of electrical power to its consumers at each time instant. This
consists of two main tasks, namely generation of the power in system’s generation
buses and its transmission to the load buses.
For the first task it is necessary to produce the amount of apparent power which
equals exactly the demanded amount of power plus transmission losses. Transmission
of the power through the system is referred to as the power flow. It results from
combination of generating units setpoints (both for active and reactive power), demand
attached to every load bus and from current technical parameters of the power grid
itself. Normally, the power flow is estimated by numerically solving a set of nonlinear
equations as shown in [11].
The power flow can only be technically attainable (in this paper referred as feasible)
if all the variables, i.e. generating units setpoints, branch power flows, nodal voltages
and nodal voltage angles fall within their technical limits. As it was shown in [12],
network feasibility of the power flow, considering both active and reactive flow,
depends highly on grid model used for determining the power dispatch of generating
units. Thus, depending on the model of dispatching the units, despite the fact that all
generation setpoints lay within generating units capabilities, a non-feasible power flow
may be obtained.
3 Proposed Method
In this section the proposed method for ensuring technical feasibility of power flow,
when generating units dispatch is obtained via p2p balancing energy trade is described
in detail. It is assumed that the operator is only responsible for checking if the gen-
erating units dispatch resulting from so-agreed contracts yields a feasible power flow
and not for any a priori dispatch of generating units. This check is to be performed at
the moment of accepting bilateral contracts. If resulting power flow is not feasible, the
operator proposes a direction in which the market participants should be adjusting their
contractual position, so to get as close to the power flow feasible value as possible. This
is obtained by applying an additional constraint on accepted contracts. We can thus say,
that the operator serves only as a feasibility guard of the power flow without imposing
anything on the market participants for as long as possible. What is more, it is assumed
that only active power balance is subject of trade between peers.
The method is based on specially designed optimization problems. As contracts are
signed directly between market players basing on their subjective goals, overall gen-
eration cost (social welfare function) is unknown and is of little interest to the operator.
Ensuring feasibility of power flow 121
The optimization goal is thus to minimize the violations of agreed balancing con-
tractual positions, subject to all power flow technical constraints.
We shall consider two network flow optimization sub-problems formulated further
in this paper. The first sub-problem (Formulation (2)) allows for identifying the unit,
whose total amount of contracted power should be adjusted the most in order to make
the power flow feasible. This unit shall be considered as most problematic in terms of
taken contracts and should be the first to change its contractual position, while in search
of network-feasible power dispatch.
The second formulation (Formulation (3)) however, arbitrarily imposes changes on
generating units contractual positions, to obtain a feasible power flow while maxi-
mizing the use bilateral contracts between suppliers and consumers. I propose, the
operator to use this formulation when market participants could not come to a network-
feasible dispatch through negotiations in a considerable amount of time/rounds when
(2) was applied. Formulation (3) was first proposed in [13] and is cited in this paper as
part of the now-proposed method.
Both proposed formulations are the restrictions of the standard Optimal Power
Flow (OPF) problem [14]. Thus, any feasible solution of one of these two sub-
problems would yield a network feasible power flow in terms of satisfying all grid
technical constraints.
The method can be summarized in the following generic steps:
1. Accept contracts between suppliers and consumers that lay within all technical
limits of generating units, i.e. which do not exceed their technical maxima/minima.
Architectural design of an appropriate IT platform is beyond the scope of this article
and will not be elaborated.
2. Check the network feasibility of power flow that results from accepted contracts
between suppliers and consumers. In case of infeasibility, apply Formulation (2) to
identify the most problematic unit and to estimate transmission losses.
3. Calculate new constraint to be added.
4. Return information on most problematic unit and on the new constraint to market
participants.
5. Enforce market participants to adjust their market positions through bilateral
negotiations following the information issued in step 3 of this method, respecting
the additional constraint. After they agree on new contracts distribution go to step 0.
6. Use Formulation (3) to impose changes on market positions of the participants, if no
feasible solution is found through negotiations in a considerable amount of time (or
rounds of negotiations).
constraints. However, different other cost functions can be also used, such as: mini-
mization of transmission losses or re-dispatch of reactive power for enhancing the level
of system’s stability such as in [15].
Below, in (1), I cite a simplified formulation of the OPF problem as was given in
[14], with standard cost function i.e. minimization of the overall generation costs:
min fP ð1aÞ
subject to:
Pinj
i Pi þ Pi ¼ 0
D
8i 2 N ð1bÞ
Qinj
i Qi þ Qi ¼ 0
D
8i 2 N ð1cÞ
Pmin
i Pi Pmax
i 8i 2 N G ð1dÞ
Qmin
i Qi Qmax
i 8i 2 N G ð1eÞ
U min
i U i U max
i 8i 2 N ð1fÞ
Hmin
i Hi Hmax
i 8i 2 N ð1gÞ
0 Sl Smax
l 8l 2 N f ð1hÞ
where: f P – the total cost of generation and transmission, N- set of indices of all buses
in the system, N G - set of indices of all generating units, N f – set of indices of all
inj
branches in the system, Pinj
i =Qi - active/reactive power injection at bus i calculated
using standard, highly nonlinear power flow equations [11], PDi =Qi - active/reactive
D
min=max min=max
power demand at bus i, Pi =Qi – active/reactive output of unit i, Qi =Pi –
min=max
generation limits of unit i, U i – voltage magnitude at bus i, Ui - limits on voltage
min=max
magnitude of bus i, Hi – voltage angle at bus i, Hi - limits on voltage angles of
bus i, Sl – apparent power flow through line l, Smax
l – maximum value of apparent
power flow through line l.
The above standard OPF problem formulation provides basis for the proposed
optimization sub-problem Formulations (2) and (3) described below.
presented problem is a restriction of the standard OPF’s feasible set shown in (1), once
a feasible solution to (2) is found, the resulting power flow is guaranteed to be tech-
nically feasible.
X G= þ
min c1 T þ c2 i2N G
ðsP=Q;i Þ; where c1 c2 ð2aÞ
subject to:
P;i T
sG 8i 2 N G ð2bÞ
þ
P;i T
sG 8i 2 N G ð2cÞ
X X
þ
k2CN;i
PkC;i sG
P;i Pi k2CN;i
PkC;i þ sG
P;i 8i 2 N G ð2dÞ
Pmin
i Pi Pmax
i 8i 2 N G ð2eÞ
G= þ
sP;i 0 8i 2 N G ð2fÞ
G þ =
where c1 ; c2 – arbitrarily chosen positive costs of contracts violation, sP;i - slack
variable for making violation of limits possible for active power, CN; i – set of con-
tracts signed with generating unit i, PkC;i – contracted volume of active/reactive power
for generating unit i with contract k.
This formulation allows for identification of most problematic generating unit.
Once it is identified, one or more new constraints shall be formulated to limit the set of
acceptable contracts by the operator, as proposed for step 4 of the method. This
procedure is explained in Sect. 3.2.1.
P
(dðrÞ ¼ k2CN;p PkC;p ðr Þ Pp ðrÞ). Next, if dðrÞ 6¼ 0 a new constraint on accepted
contracts in next round of iterations (r þ 1) shall be added for the most problematic
P
unit. If dðrÞ [ 0: the constraint shall take form of k2CN;i PkC;p ðr þ 1Þ Pp ðr Þ dðrÞ,
P
and of k2CN;i PkC;p ðr þ 1Þ Pp ðrÞ þ dðrÞ when d\0. Since it is guaranteed that the
deviation for most problematic unit is minimal, it is obvious that constraining its
accepted contracts by d will drive the dispatch closer to feasibility. Once added, the
method returns to end of step 4 – negotiations between peers.
subject to:
X X
þ
PkC;i sG
P;i Pi k2CN;i
PkC;i þ sG
P;i 8i 2 N G ð3bÞ
k2CN;i
Pmin
i Pi Pmax
i 8i 2 N G ð3cÞ
G= þ
sP;i 0 8i 2 N G ð3dÞ
G þ =
where: cP;i – positive cost (price) of violation of upper/lower limits on generation of
unit i for active power.
Ensuring feasibility of power flow 125
In the paper I propose to add more freedom into the balancing energy market. As a
result, this should add more economical efficiency to the market, whose amount should
be maximal. Yet, the freedom is achieved through rounds of bilateral negotiations,
whose number should be minimal. Thus, given the above, I propose to assess the
performance of the method by considering two criteria - measure of market efficiency
and number of iterations it takes to come to consensus.
f E ¼ 2f S=L cG cA ; ð4Þ
where f S=L is the total amount of subjective benefits for making contracts between
suppliers and consumers (sometimes referred to as utility function), and cG is the total
cost of generation and cA is the total cost of adjustments of units operating points when
formulation (3) is used. Assuming that subjective benefits of consumer j for making a
bilateral contract with generating unit i is equal to subjective benefit of i making a
contract with j, f S=L is formulated as
X X X X
f S=L ¼ j2N L
a PD
i2N G ji ji
¼ j2N L
a PD ;
i2N G ij ji
ð5Þ
where aji – factor of benefit for delivering balancing power from i to j quantifying how
important the benefit is, PD ji – amount of delivered balancing power from i to j,
N L – set of indices of consumers.
For assessment, it is assumed that total generation cost of each unit is well known.
Normally it can be of any form, yet most likely quadratic formulation is used, as in
(6) where ai ; bi ; ci are cost coefficients and Pi is the active power output of unit i.
X
cG ¼ i2N
ðai P2i þ bi Pi þ ci Þ ð6Þ
G
In the tests we look at the measures described previously – i.e. market efficiency ðf E Þ
and number of negotiation rounds (max. Round). Resulting value of f E is benchmarked
with the result of standard OPF problem (1) – to check performance as opposed to
currently used dispatching methods. All simulations were coded in Matlab and solved
with MATPOWER MIPS solver [17].
Test system comprises of nine buses, three of whom being generator buses and
three load buses. All generating units have their technical minima equal to 10 MW and
technical maxima of unit 1, 2 and 3: 250 MW, 300 MW, 270 MW respectively, giving
the total generating capacity of 820 MW. The load is assumed constant and attached to
buses 5, 7 and 9. Value of load in each of these buses is as follows, given as pairs
(active demand, reactive demand): bus 5 – (90 MW, 30 MVAR); bus 7 – (100 MW,
35 MVAR); bus 9 – (125 MW, 50 MVAR), giving the total demand of (315 MW,
115 MVAR). The system consists also of nine transmission lines. Their ratings
(maximum flow capacities) can be found in [16].
For simulations it is assumed that system topology, with locations of load and
values of the load and with costs of generation are constant. It is also assumed that,
demanded load and technical limits of generating units are known and guarded in step 0
of the method. What is more, I assume that subjective benefit for unit i for making a
contract with consumer j is symmetrical and equal to the benefit for consumer j, i.e.
aij ¼ aji . Apart from this, I also assume that after estimation of transmission losses,
consumers are forced to contract their compensation evenly, i.e. in the discussed test
system each consumer contracts exactly one third of the losses.
Generation costs for each of the generating units are presented in Table 1. They
correspond to quadratic formulation given in (6). For simplicity it is assumed here that
generating units sell energy at its cost, meaning that sell price equals generation costs.
Benefit coefficients are assumed as given in Table 2. These reflect both benefit of the
consumers and of suppliers meaning that benefit should be summed only once into f E .
Ensuring feasibility of power flow 127
Previous statements yield that economic surplus of the market is produced only through
the benefits. It is assumed here that each consumer wants to contract as much demand
with its preferred unit, as possible. From generating unit’s perspective – the unit
accepts those contracts, which bring most benefit to it until its maximum capacity is
reached. What is more, it is also assumed that operator’s IT system guards that sum of
contracts made for each unit falls within this unit’s technical capacities – from mini-
mum to maximum.
For all of the scenarios, with the load as written previously, the objective of
standard OPF problem (1) is equal to 5,296.69 USD, which in terms of benefit function
equals f E ¼ 5; 296:69 USD. This is the cost value for optimal solution: P1 ¼
89:7986MW,P2 ¼ 134:3207MW,P3 ¼ 94:1874MW.
Scenario 2, however, is slightly different from the previous one. Here the preferred
unit for all of consumers was unit 3. Similarly to previously considered example,
benefit coefficients differed making it easy for the generating unit to decide on which
contracts to accept. However, this time most of the problem with network feasibility
resulted from transmission losses which were not considered by the participants at first
round of balancing. They were estimated by solving (2) and an additional constraint
was added on most problematic unit’s acceptable contracts. Market participants man-
aged to find consensus in two rounds of negotiations. Results are presented in Table 3.
Scenario 3 differs significantly from both previously presented cases. This time all
consumers wish to contract all energy supply from unit 1 without any will to com-
promise. After a considerable round of negotiations (here assumed 10), market par-
ticipants did not come to a conclusion on dispatch yielding a network- feasible power
flow. Therefore, the Operator used formulation (3) to arbitrarily adjust operating points
G þ =
of generating units. It was assumed that prices cP;i were all equal to 100 for all units.
They form the cost cA which is considered while calculating the value of f E .
The results show that under the assumptions made, adding more freedom to the
balancing energy market might add some significant benefit as opposed to the standard
dispatching method, e.g. solving the regular Optimal Power Flow. As shown in
Table 3, this is the case for all scenarios. Not surprisingly, it is more beneficial to the
market when participants manage to decide on contracts alone (as in Scenario 1 and in
Scenario 2) rather than as when these contracts are arbitrarily changed by the operator –
as in Scenario 3.
What is more, the results show that using the proposed method it is possible to find
a network-feasible units dispatch for every discussed scenario, which actually maxi-
mizes use of bilateral balancing energy contracts made between participants.
Power systems with high penetration of distributed energy resources are nowadays
becoming of interest to both academia and industry. Balancing of energy in such
a system ideally should performed via peer-to-peer trading between energy market
participants, as the participants should be given direct incentives to both develop DER
infrastructure and to actively control it to maximize their subjective profit.
Therefore, in this paper a multi-step method for assuring network-feasibility of
power flow resulting from generating unit dispatch obtained via bilateral, peer-to-peer,
balancing contracts is presented and analyzed. The proposed method allows market
participants to make bilateral contracts between consumers and suppliers for provision
of necessary amounts of balancing energy. The method assumes that market partici-
pants do not have sufficient knowledge on technical situation in the power system and
thus, do not know a priori what dispatch to agree on. Therefore, it is assumed that it is
system operator’s responsibility to guide participants in the search of network-feasible
dispatch. The method however, returns necessary information on direction of this
search, as a result of optimization problem (2) with heuristically added constraints on
accepted bilateral contracts. If despite having this information market participants are
unable to agree on technically feasible contracts, the method foresees the ability for the
operator to arbitrarily adjust those contracts using optimization problem (3). Yet, when
such a situation arises, suppliers are to be paid for adjusting their contractual positions.
The basis of the proposed method is formed by two optimization problems given in
formulations (2) and (3), which aim to maximize use of such contracts, while always
guaranteeing feasibility of resulting power flow.
Proposed approach was tested in simulations over three scenarios. All tests showed
that allowing free trade on the balancing energy market may bring more benefits to the
market than currently used centralized balancing options – namely when it is the
system’s operator who is always a party in any transaction made on the balancing
market. What is more, it has been shown in the paper, that it is possible to derive
methods which can satisfactorily help to achieve technical feasibility of power flow
resulting from so obtained dispatch of generating units.
130 M. Drabecki
Some perspective for further research can be identified. First of all, one may
consider designing architecture for an operator’s IT system for accepting contracts –
such as assumed in step 0 of described method. Another perspective may be devel-
opment of new market efficiency measures or new models for assuring technical fea-
sibility of power flow. They could be developed basing on different optimization,
modelling and simulation problems, possibly including development of agent systems.
When all of the above issues are deeply considered and followed-up by academic
communities worldwide, hopefully it will be possible to establish highly effective
bilateral balancing energy markets, that may be able to give incentives for transition
towards high penetration of green energy. Specific design of these markets shall be also
given thorough consideration, to make room for all their participants, regardless of their
size and power supply capabilities.
References
1. Wang, Q., et al.: Review of real-time electricity markets for integrating distributed energy
resources and demand response. Appl. Energy 138, 695–706 (2015)
2. Guerrero, J., et al.: Towards a transactive energy system for integration of distributed energy
resources: Home energy management, distributed optimal power flow, and peer-to-peer
energy trading. Renew. Sustain. Energy Rev. 132, 110000 (2020)
3. Lee, W., et al.: Optimal operation strategy for community-based prosumers through
cooperative P2P trading. In: 2019 IEEE Milan PowerTech. IEEE (2019)
4. Guerrero, J., Archie, C., Verbič, G.: Decentralized P2P energy trading under network
constraints in a low-voltage network. IEEE Trans. Smart Grid 10(5), 5163–5173 (2018)
5. Pasha, A.M., et al.: A utility maximized demand-side management for autonomous
microgrid. In: 2018 IEEE Electrical Power and Energy Conference (EPEC). IEEE (2018)
6. Okawa, Y., Toru, N.: Distributed optimal power management via negawatt trading in real-
time electricity market. IEEE Trans. Smart Grid 8(6), 3009–3019 (2017)
7. Zhang, Y., Chow, M.Y.: Distributed optimal generation dispatch considering transmission
losses. In: 2015 North American Power Symposium (NAPS), Charlotte (2015)
8. Kar, S., Hug, G.: Distributed robust economic dispatch: a consensus + innovations approach.
In: 2012 IEEE Power and Energy Society General Meeting, San Diego (2012)
9. Lin, C., Lin, S.: Distributed optimal power flow with discrete control variables of large
distributed power systems. IEEE Trans. Power Syst. 23(3) (2008)
10. Garrity T.F.: Innovation and trends for future electric power systems. In: 2009 Power
Systems Conference, Clemson (2009)
11. Machowski, J., Lubosny, Z., Bialek, J.W., Bumby, J.R.: Power System Dynamics: Stability
and Control. Wiley (2020)
12. Drabecki, M., Toczyłowski, E.: Comparison of three approaches to the security constrained
unit commitment problem. Zeszyty Naukowe Wydziału Elektrotechniki i Automatyki
Politechniki Gdańskiej (62) (2019). http://yadda.icm.edu.pl/baztech/element/bwmeta1.ele
ment.baztech-0345a0bb-f574-41c1-be5d-71f50b8e060c?q=5838bcf6-a5ee-44ec-93f8-ea79d
2cd037b$1&qt=IN_PAGE
13. Drabecki, M., Toczyłowski, E.: Obtaining feasibility of power flows in the deregulated
electricity market environment. Przegląd Elektrotechniczny 95 (2019)
Ensuring feasibility of power flow 131
14. Zhu, J.: Optimization of Power System Operation. Wiley, Hoboken (2015)
15. Drabecki, M.: A method for enhancing power system’s steady-state voltage stability level by
considering active power optimal dispatch with linear grid models. In: Zeszyty Naukowe
Wydziału Elektrotechniki i Automatyki Politechniki Gdańskiej (62) (2019). http://yadda.
icm.edu.pl/baztech/element/bwmeta1.element.baztech-2eb9eb53-193d-4970-827d-4dba236e
0936?q=5838bcf6-a5ee-44ec-93f8-ea79d2cd037b$2&qt=IN_PAGE
16. Chow, J.H. (ed.): Time-Scale Modelling of Dynamic Networks with Applications to Power
Systems. Lecture Notes in Control and Information Sciences vol. 26, pp. 59–93. Springer.
Berlin (1982)
17. Zimmerman R.D., Murillo-Sanchez C.E., Thomas R.J.: MATPOWER: steady-state oper-
ations, planning and analysis tools for power systems research and education. IEEE Trans.
Power Syst. 26(1) (2011)
Sustainable Optimization,
Metaheuristics and Computing
for Expert System
The Results of a Compromise Solution, Which
Were Obtained on the Basis of the Method
of Uncertain Lagrange Multipliers
to Determine the Influence of Design Factors
of the Elastic-Damping Mechanism
in the Tractor Transmission
1 Introduction
Optimizing the operation of certain elements and replacing outdated mechanisms with
new developments can significantly improve performance. A large number of studies
exist in the field of improving mobile technology.
2 Purpose of Research
Our previous works [14–17] were aimed at finding the optimal parameters of the
Elastic Damping Mechanism (EDM) in the 14 kN class tractor transmission. Con-
ducted research is described in detail in these papers [14–17]. The indicator «P» is
proposed to assess the protective qualities of the mechanism («transmission trans-
parency degree») as the ratio of the current amplitude of vibrations of the engine speed
to its maximum value [14, 15, 18]. If P = 1 is reduction gear is absolutely «trans-
parent», the engine is not protected from fluctuations in the traction load (this happens
in serial transmissions). If P = 0, reduction gear is absolutely «not transparent» and will
The Results of a Compromise Solution, Which Were Obtained 137
completely absorb vibrations transmitted to the engine. Research has been conducted
for next agricultural operations: plowing, cultivation and seeding.
Purpose of this research is finding the conditional extremum of the “transmission
transparency degree” function regarding restrictions as a result of a compromise
solution based on Lagrange multipliers.
The following tasks were solved to achieve this goal:
1. Acquisition a compromise model that expresses the EDM main parameters
influence on the «degree of transparency of the transmission»;
2. Finding the factors values included in the resulting Lagrange function.
Main limitations for conducting this research the restrictions were chosen based on our
previous research [14, 15, 17, 18]. The name and designation of the elastic damping
mechanism factors are given in Table 1.
Table 1. Name and designation of factors of the elastic damping mechanism installed in the
tractor transmission.
N Factor name Factor identification Code mark
1 The throttle cross-sectional area Sth, m2 x1
2 Volume of hydropneumatic accumulator (HPA) Vhpa, m3 x2
3 The air pressure in HPA Pa, Pa x3
4 Inertia moment of additional load Jth, kgm2 x4
5 The oscillation frequency of the traction load f, Hz x5
F ðx1 ; x2 ; . . .; xn ; k1 ; k2 ; . . .; km Þ ¼ f þ k1 u1 þ k2 u2 þ þ km um ð1Þ
4 Discussion Results
The model (4) and (5), which were taken as the conditions and restrictions.
All further calculations were made in the Maple and MATLAB software envi-
ronment for convenience.
The Lagrange function is an Eq. (6) in general form in this case:
that takes the form (7) when Eqs. (3), (4), and (5) are substituted into it:
Partial derivatives were taken from the variables x1, x2, x3, x4, x5 in Eq. (7) to find
the extremes of the Lagrange function, then they were equated to zero. The resulting
equations, the constraint Eqs. (4) and (5), those are equated to zero, constitute a system
of equations, during the solution of which all the variables x (x1, x2, x3, x4, x5) were
expressed in terms of k2 and k3 those are means L2 and L3 in the program text [21].
Next, the values x1…x5 were inserted into the main equation and the constraint
equations, then into the Lagrange function. This part of the calculations was performed
in the Maple 5 software environment (Fig. 1).
140 S. Senkevich et al.
Two small programs were written to find Lagrange coefficients in the MATLAB
software environment (Fig. 2).
The results of calculations of variables x1…x5 and coefficients k2 and k3 and the
resulting value of the Lagrange function are presented in Table 2.
Analysis of the Table 2 shows that the best version of the Lagrange function is
shown in the first row of the table. But these values are not applicable because the
factors x3 and x5 are outside the scope of research. So, the values of line 2 are assumed
to be optimal values.
The Lagrange compromise model (8) was obtained by substituting optimal values:
Graphical dependencies of the Lagrange function are shown in Fig. 3. They are
constructed using Eq. (9). Factors not shown on the graph are fixed at zero (equal to
zero) in the function (8).
Figure 3 shows the surface plot of the Lagrange function depending on the vari-
ables x1 and x5 which is obtained using the Eq. (9):
5 Conclusions
A compromise solution to this problem was found in this study. The extremum of the
“transmission transparency degree” function is found using Lagrange multipliers.
A compromise model is obtained that expresses the influence of the main EDM
parameters on the «transmission transparency degree». The values of factors included
in the resulting Lagrange function are found.
We have determined the value of the desired function is as a solution to the
compromise problem L = 0.68; the factor values are x1 = −0.1258, x2 = 0.1496,
x3 = 0.4461, x4 = 0.0435, x5 = 0.1184, they are included in the Lagrange function.
Finding natural values is determined by formulas those are described in detail in the
research [14].
The values are: Sth = 2.285 10–4 m2, Vhpa, = 4.030 10–3 m3, Pa, = 4.446
10 Pa, Jth, = 4.740 10–3 kgm2, f, = 0.940 Hz.
5
Acknowledgments. The team of authors expresses recognition for the organization of the
Conference ICO'2020, Thailand, and personally Dr. Pandian Vasant. The authors are grateful to
anonymous referees for their helpful comments.
References
1. Li, B., Sun, D., Hu, M., Zhou, X., Liu, J., Wang, D.: Coordinated control of gear shifting
process with multiple clutches for power-shift transmission. Mech. Mach. Theory 140, 274–
291 (2019). https://doi.org/10.1016/j.mechmachtheory.2019.06.009
2. Chen, X., Hu, Q., Xu, Z., Zhu, C.: Numerical modeling and dynamic characteristics study of
coupling vibration of multistage face gearsplanetary transmission. Mech. Sci. 10, 475–495
(2019). https://doi.org/10.5194/ms-10-475-2019
3. Kirchner, M., Eberhard, P.: Simulation model of a gear synchronisation unit for application
in a real-time HiL environment. Veh. Syst. Dyn. 55(5), 668–680 (2017). https://doi.org/10.
1080/00423114.2016.1277025
4. Guercioni, G.R., Vigliani, A.: Gearshift control strategies for hybrid electric vehicles: a
comparison of powertrains equipped with automated manual transmissions and dual-clutch
transmissions. Proc. Inst. Mech. Eng. Part D J. Autom. Eng. 233(11), 2761–2779 (2019).
https://doi.org/10.1177/0954407018804120
5. Zhu, S., Xu, G., Tkachev, A., Wang, L., Zhang, N.: Comparison of the road-holding abilities
of a roll-plane hydraulically interconnected suspension system and an anti-roll bar system.
Proc. Inst. Mech. Eng. Part D J. Autom. Eng. 231(11), 1540–1557 (2016). https://doi.org/10.
1177/0954407016675995
The Results of a Compromise Solution, Which Were Obtained 143
6. Sau, J., Monteil, J., Bouroche, M.: State-space linear stability analysis of platoons of
cooperative vehicles. Transportmetrica B Transp. Dyn. 1–26 (2017).https://doi.org/10.1080/
21680566.2017.1308846
7. Kim, J.: Design of a compact 18-speed epicyclic transmission for a personal mobility
vehicle. Int. J. Automot. Technol. 17(6), 977–982 (2016). https://doi.org/10.1007/s12239-
016-0095-9
8. Park, T., Lee, H.: Optimal supervisory control strategy for a transmission-mounted electric
drive hybrid electric vehicle. Int. J. Automot. Technol. 20(4), 663–677 (2019). https://doi.
org/10.1007/s12239-019-0063-2
9. Chen, S., Zhang, B., Li, B., Zhang, N.: Dynamic characteristics analysis of vehicle
incorporating hydraulically interconnected suspension system with dual accumulators.
Shock Vib. 2018, 1–5 (2018). https://doi.org/10.1155/2018/6901423
10. Ding, F., Zhang, N., Liu, J., Han, X.: Dynamics analysis and design methodology of roll-
resistant hydraulically interconnected suspensions for tri-axle straight trucks. J. Franklin Inst.
353(17), 4620–4651 (2016). https://doi.org/10.1016/j.jfranklin.2016.08.016
11. Kuznetsov, N.K., Iov, I.A., Iov, A.A.: Reducing of dynamic loads of excavator actuators. In:
Journal of Physics: Conference Series, vol. 1210, no. 1, p. 012075. IOP Publishing (2019).
https://doi.org/10.1088/1742-6596/1210/1/012075
12. Ziyadi, M., Ozer, H., Kang, S., Al-Qadi, I.L.: Vehicle energy consumption and an
environmental impact calculation model for the transportation infrastructure systems.
J. Clean. Prod. 174, 424–436 (2018). https://doi.org/10.1016/j.jclepro.2017.10.292
13. Melikov, I., Kravchenko, V., Senkevich, S., Hasanova, E., Kravchenko, L.: Traction and
energy efficiency tests of oligomeric tires for category 3 tractors. In: IOP Conference Series:
Earth and Environmental Science, vol. 403, p. 012126 (2019). https://doi.org/10.1088/1755-
1315/403/1/012126
14. Senkevich, S., Kravchenko, V., Duriagina, V., Senkevich, A., Vasilev, E.: Optimization of
the parameters of the elastic damping mechanism in class 1, 4 tractor transmission for work
in the main agricultural operations. In: Advances in Intelligent Systems and Computing,
pp. 168–177 (2018). https://doi.org/10.1007/978-3-030-00979-3_17
15. Senkevich, S.E., Sergeev, N.V., Vasilev, E.K., Godzhaev, Z.A., Babayev, V.: Use of an
elastic-damping mechanism in the tractor transmission of a small class of traction (14 kN):
Theoretical and Experimental Substantiation. In: Handbook of Advanced Agro-Engineering
Technologies for Rural Business Development, pp. 149–179. IGI Global, Hershey (2019).
https://doi.org/10.4018/978-1-5225-7573-3.ch006
16. Senkevich, S., Duriagina, V., Kravchenko, V., Gamolina, I., Pavkin, D.: Improvement of the
numerical simulation of the machine-tractor unit functioning with an elastic-damping
mechanism in the tractor transmission of a small class of traction (14 kN). In: Vasant, P.,
Zelinka, I., Weber, G.W. (eds.) Intelligent Computing and Optimization. ICO 2019.
Advances in Intelligent Systems and Computing, vol. 1072, pp. 204–213. Springer, Cham
(2020). https://doi.org/10.1007/978-3-030-33585-4_20
17. Senkevich, S.E., Lavrukhin, P.V., Senkevich, A.A., Ivanov, P.A., Sergeev, N.V.:
Improvement of traction and coupling properties of the small class tractor for grain crop
sowing by means of the hydropneumatic damping device. In: Kharchenko, V., Vasant,
P. (eds.) Handbook of Research on Energy-Saving Technologies for Environmentally-
Friendly Agricultural Development, pp. 1–27. IGI Global, Hershey (2020). https://doi.org/
10.4018/978-1-5225-9420-8.ch001
144 S. Senkevich et al.
18. Senkevich, S., Kravchenko, V., Lavrukhin, P., Ivanov, P., Senkevich, A.: Theoretical study
of the effect of an elastic-damping mechanism in the tractor transmission on a machine-
tractor unit performance while sowing. In: Handbook of Research on Smart Computing for
Renewable Energy and Agro-Engineering, pp. 423–463. IGI Global, Hershey (2020). https://
doi.org/10.4018/978-1-7998-1216-6.ch017
19. Nocedal, J., Wright, S.: Numerical Optimization, p. 664. Springer, Heidelberg (2006)
20. Härdle, W., Simar, L.: Applied Multivariate Statistical Analysis, p. 580. Springer, Berlin
(2015)
21. Malthe-Sorenssen, A.: Elementary Mechanics Using Matlab: A Modern Course Combining
Analytical and Numerical Techniques, p. 590. Springer, Heidelberg (2015)
Multiobjective Lévy-Flight Firefly Algorithm
for Multiobjective Optimization
Abstract. The firefly algorithm (FA) was firstly proposed during 2008–2009 as
one of the powerful population-based metaheuristic optimization techniques for
solving continuous and combinatorial optimization problems. The FA has been
proved and applied to various real-world problems in mostly single objective
optimization manner. However, many real-world problems are typically for-
mulated as the multiobjective optimization problems with complex constraints.
In this paper, the multiobjective Lévy-flight firefly algorithm (mLFFA) is
developed for multiobjective optimization. The proposed mLFFA is validated
against four standard multiobjective test functions to perform its effectiveness.
The simulation results show that the proposed mLFFA algorithm is more effi-
cient than the well-known algorithms from literature reviews including the
vector evaluated genetic algorithm (VEGA), non-dominated sorting genetic
algorithm II (NSGA-II), differential evolution for multiobjective optimization
(DEMO) and multiobjective multipath adaptive tabu search (mMATS).
1 Introduction
During 2008–2009, the firefly algorithm (FA) was firstly proposed by Yang [6]
based on the flashing behavior of fireflies and a random drawn from the uniform
distribution for randomly generating the feasible solutions. The FA has been widely
applied to solving real-world problems, such as industrial optimization, image pro-
cessing, antenna design, business optimization, civil engineering, robotics, semantic
web, chemistry, meteorology, wireless sensor networks and so on [7, 8]. Afterwards in
2010, the Lévy-flight firefly algorithm (LFFA) was proposed by Yang [9]. The LFFA
utilized a random drawn from the Lévy distribution to randomly generate new solu-
tions. The LFFA was tested and its results outperformed the genetic algorithm
(GA) and the particle swarm optimization (PSO). The state-of-the-art and its applica-
tions of the original LF and LFFA algorithms are reported [7–9].
In early 2020, the multiobjective Lévy-flight firefly algorithm (mLFFA) is proposed
for particularly solving multiobjective optimization problems [10]. In this paper, the
mLFFA is applied to solve well-known standard multiobjective optimization problems.
After an introduction given in Sect. 1, the original FA and LFFA algorithms are briefly
described, and the proposed mLFFA algorithm is illustrated in Sect. 2. Four standard
multiobjective test functions used in this paper are detailed in Sect. 3. Results are
discussions of the performance evaluation are provided in Sect. 4. Conclusions are
followed in Sect. 5.
In this section, the original FA and LFFA algorithms are briefly described. The pro-
posed mLFFA algorithm is then illustrated as follows.
Rule-1: All fireflies are assumed to be unisex. A firefly will be attracted to others
regardless of their sex.
Rule-2: Attractiveness is proportional to the brightness. Both attractiveness and
brightness decrease when distance increases. The less brighter firefly will
move towards the brighter one. If there is no brighter one, it will move
randomly.
Rule-3: Brightness of each firefly will be evaluated by the objective function of
interest.
The light intensity I of the FA will vary according to the inverse square law of the
distance r, I = Is/r2, where Is is the intensity at the source. For a fixed light absorption
coefficient, the light intensity I varies with the distance r as stated in (1), where I0 is the
original light intensity. The attractiveness b will vary according to the distance r as
stated in (2), where b0 is the attractiveness at r = 0. From (1) and (2), the scaling factor
Multiobjective Lévy-Flight Firefly Algorithm 147
c is defined as the light absorption coefficient. The distance rij between two fireflies
i and j at their locations xi and xj can be calculated by (3). In FA algorithm, a new
solution xt+1 can be obtained by an old solution xt as stated in (4), where ei is a vector
of random numbers drawn from a uniform distribution. The randomization parameter at
can be adjusted as stated in (5), where a0 is the initial randomness scaling factor. The
original FA algorithm can be described by the pseudo code as shown in Fig. 1.
I ¼ I0 ecr ð1Þ
b ¼ b0 ecr
2
ð2Þ
vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
u d
u X
rij ¼ xi xj ¼ t
ðxi;k xj;k Þ2 ð3Þ
k¼1
at ¼ a0 dt ; ð0\d\1Þ ð5Þ
Initialize:
The objective function f(x), x = (x1,…,xd)T
Generate initial population of fireflies xi = (i = 1, 2,…,n)
Light intensity Ii at xi is determined by f(x)
Define light absorption coefficient γ
while (Gen ≤ Max_Generation)
for i =1 : n all n fireflies
for j =1 : i all n fireflies
if (Ij > Ii)
- Move firefly i towards j in d-dimension via
Lévy-flight distributed random
end if
- Attractiveness varies with distance r via exp[-γr]
- Evaluate new solutions and update light intensity
end for j
end for i
Rank the fireflies and find the current best x*
end while
Report the best solution found x*
ð6Þ
ð7Þ
s ¼ u=jvj1=b ð8Þ
PF ¼ fs 2 Sj9s 2 S : s sg ð13Þ
Multiobjective Lévy-Flight Firefly Algorithm 149
Multiobjective function:
F(x) = {f1(x), f2(x),…,fn(x)}, x = (x1,…,xd)T
Initialize LFFA1,…,LFFAk, and Pareto optimal set P*
Generate initial population of fireflies xi = (i = 1, 2,…,n)
Light intensity Ii at xi is determined by F(xi)
Define light absorption coefficient γ
while (Gen ≤ Max_Generation)
for z =1: k all k LFFA
for i =1 : n all n fireflies
for j =1 : i all n fireflies
if (Ij > Ii)
- Move firefly i towards j in d-dimension via
Lévy-flight distributed random
end if
- Attractiveness varies with distance r via exp[-γr]
- Evaluate new solutions and update light intensity
end for j
end for i
end for z
- Rank the fireflies and find the current best x*
- Sort and find the current Pareto optimal solutions
end while
- Pareto optimal solutions are found and
- Pareto front PF* is performed.
In order to perform its effectiveness, the mLFFA is thus evaluated against several
multiobjective test functions. In this work, four widely used standard multiobjective
functions, ZDT1 – ZDT4, providing a wide range of diverse properties in terms of
Pareto front and Pareto optimal set are conducted [11, 12]. ZDT1 is with convex front
as stated in (14) where d is the number of dimensions. ZDT2 as stated in (15) is with
non-convex front, while ZDT3 with discontinuous front is expressed in (16), where
g and xi in functions ZDT2 and ZDT3 are the same as in function ZDT1. For ZDT4, it
is stated in (17) with convex front but more specific. In evaluation process, the error Ef
between the estimated Pareto front PFe and its corresponding true front PFt is defined
in (18), where N is the number of solution points.
9
ZDT1 : >
>
f1 ðxÞ ¼ x1 ; pffiffiffiffiffiffiffiffiffi >
=
f2 ðxÞ ¼ gð1 f1 =g Þ; ð14Þ
P . >
>
>
g ¼ 1 þ 9 di¼2 xi ðd 1Þ; xi 2 ½0; 1; i ¼ 1; . . .; 30 ;
150 S. Sumpunsri et al.
9
ZDT2 : =
f1 ðxÞ ¼ x1 ; ð15Þ
;
f2 ðxÞ ¼ gð1 f1 =gÞ2
9
ZDT3 : >
=
f1 ðxÞ ¼ x1; ð16Þ
pffiffiffiffiffiffiffiffiffi >
f2 ðxÞ ¼ g 1 f1 =g f1 sinð10pf1 Þ=g ;
9
ZDT4 : >
>
f1 ðxÞ ¼ x1 ; pffiffiffiffiffiffiffiffiffi =
f2 ðxÞ ¼ gð1 f1 =gÞ; ð17Þ
P >
>
;
g ¼ 1 þ 10ðd 1Þ þ di¼2 x2i 10 cosð4pf1 Þ ; xi 2 ½0; 1; i ¼ 1; . . .; 30:
X
N
Ef ¼ kPFe PFt k ¼ ðPFej PFt Þ2 ð18Þ
j¼1
The proposed mLFFA algorithms were coded by MATLAB version 2018b run on Intel
(R) Core(TM) i5-3470 CPU@3.60 GHz, 4.0 GB-RAM. Search parameters of each
LFFA in the mLFFA are set according to Yang’s recommendations [6, 9], i.e. the
numbers of fireflies n = 30, a0 = 0.25, b0 = 1, k = 1.50 and c = 1. These parameter
values are sufficient for most optimization problems because the LFFA algorithm is
very robust (not very sensitive to the parameter adjustment) [6, 9]. In this work, the
termination criteria (TC) either use a given tolerance or a fixed number of generations.
As simulation results, it was found that a fixed number of generations is not only easy
to implement, but also suitable to compare the closeness of Pareto front of test func-
tions. Therefore, for all test functions, Max_Gen = 2,000 is set as the TC.
For comparison, the results obtained by the proposed mLFFA over all four standard
multiobjective test functions are compared with those obtained by the well-known
algorithms from literature reviews, i.e. VEGA [3], NSGA-II [4], DEMO [5] and
multiobjective multipath adaptive tabu search (mMATS) [13]. The performance of all
algorithms is measured via the error Ef stated in (18) and for all algorithms, a fixed
number of generations/iterations of 2,000 (Max_Gen) is set as the TC. The results
obtained from all test functions are summarized in Table 1 – Table 2, and the estimated
Pareto fronts and the true front of functions ZDT1 – ZDT4 are depicted in Fig. 3,
Fig. 4, Fig. 5, Fig. 6, respectively. It was found from Fig. 3, Fig. 4, Fig. 5 and Fig. 6
that the mLFFA can satisfactory provide the Pareto front containing all Pareto optimal
solutions of all test functions very close to the true front of each multiobjective test
function. Referring to Table 1 – Table 2, the mLFFA shows superior results to other
algorithms with the less Ef and search time consumed.
Multiobjective Lévy-Flight Firefly Algorithm 151
1.0
0.9 True Pareto front
0.8 mLFFA
0.7
0.6
f2(x)
0.5
0.4
0.3
0.2
0.1
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7
f1(x)
1.0
0.9
0.8
0.7
0.6
f2(x)
0.5
0.4
0.3
0.2
True Pareto front
0.1 mLFFA
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
f1(x)
1.0
0.8 True Pareto front
mLFFA
0.6
0.4
0.2
f2(x)
0
-0.2
-0.4
-0.6
-0.8
-1.0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
f1(x)
1.0
0.9 True Pareto front
mLFFA
0.8
0.7
0.6
f2(x)
0.5
0.4
0.3
0.2
0.1
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
f1(x)
5 Conclusions
The multiobjective Lévy-flight firefly algorithm (mLFFA) has been presented in this
paper for global optimization. Based on the original FA and LFFA, the mLFFA has
been tested against four standard multiobjective test functions to perform its effec-
tiveness of multiobjective optimization. Details of the performance evaluation of the
mLFFA algorithm have been reported. As simulation results, it was found that the
mLFFA algorithm performs more efficient than the well-known algorithms including
the VEGA, NSGA-II, DEMO and mMATS. This can be noticed that the mLFFA
algorithm is one of the most effective metaheuristic optimization search techniques for
solving global optimization problems. For future research, the proposed mLFFA
algorithm will be applied to various real-world engineering optimization problems.
References
1. Zakian, V.: Control Systems Design: A New Framework. Springer, London (2005)
2. Talbi, E.G.: Metaheuristics form Design to Implementation. Wiley, Hoboken (2009)
3. Schaffer, J.D.: Multiple objective optimization with vector evaluated genetic algorithms. In:
The 1st International Conference on Genetic Algorithms, pp. 93–100 (1985)
4. Deb, K., Pratap, A., Agarwal, S., Mayarivan, T.: A fast and elitist multiobjective algorithm:
NSGA-II. IEEE Trans. Evol. Comput. 6, 182–197 (2002)
5. Robič, T., Filipič, B.: DEMO: Differential Evolution for Multiobjective Optimization.
Lecture Notes in Computer Sciences, vol. 3410, pp.520–533 (2005)
6. Yang, X.S.: Firefly algorithms for multimodal optimization. In: Stochastic Algorithms,
Foundations and Applications. Lecture Notes in Computer Sciences, vol. 5792, pp. 169–178
(2009)
7. Fister, I., Fister Jr., I., Yang, X.S., Brest, J.: A comprehensive review of firefly algorithms.
In: Swarm and Evolutionary Computation, vol. 13, pp. 34–46. Springer, Heidelberg (2013)
8. Fister, I., Yang, X.S., Fister, D., Fister Jr., I.: Firefly algorithm: a brief review of the
expanding literature. In: Yang, X.S. (ed.) Cuckoo Search and Firefly Algorithm, pp. 347–
360 (2014)
9. Yang, X.S.: Firefly algorithm, Lévy flights and global optimization. In: Research and
Development in Intelligent Systems, vol. XXVI, pp. 209–218. Springer, Heidelberg (2010)
10. Sumpunsri, S., Puangdownreong, D.: Multiobjective Lévy-flight firefly algorithm for
optimal PIDA controller design. Int. J. Innovative Comput. Inf. Control 16(1), 173–187
(2020)
11. Zitzler, E., Thiele, L.: Multiobjective evolutionary algorithms: a comparative case study and
the strength pareto approach. IEEE Trans. Evol. Comput. 3, 257–271 (1999)
12. Zitzler, E., Deb, K., Thiele, L.: Comparison of multiobjective evolutionary algorithms:
empirical results. Evol. Comput. 8, 173–195 (2000)
13. Puangdownreong, D.: Multiobjective multipath adaptive tabu search for optimal PID
controller design. Int. J. Intell. Syst. Appl. 7(8), 51–58 (2015)
Cooperative FPA-ATS Algorithm
for Global Optimization
1 Introduction
In this section, the original FPA and ATS algorithms are briefly described. Then, the
proposed cooperative FPA-ATS algorithm is elaborately illustrated as follows.
kCðkÞ sinðpk=2Þ 1
L ; ðs [[ s0 [ 0Þ ð2Þ
p s1 þ k
1=ðb aÞ; a q b
eðqÞ ¼ ð4Þ
0; q\a or q [ b
Start
No
rand > p No
f(x)<f(x*)
Yes
- Draw ε from a uniform Yes
- Draw a step vector L distribution in (4)
via Lévy flight in (2) - Randomly choose j and k - Update x* = x
- Activate cross- among all the solutions
pollination in (1) - Invoke self-pollination
for generating in (3) for generating - Store x in TL
new solutions new solutions
- Activate AR mechanism by R = αR
No
No TC ?
TC ? Yes
Yes
- Report the current best solution g* - Report the best solution x*
Stop Stop
Start
For FPA:
- Objective function f(x), x = (x1, x2,…, xd)
- Initialize a population of n flowers/pollen gametes with random solutions
- Find the best solution g* in the initial population
- Define a proximity probability p ∈ [0, 1]
For ATS:
- Objective function f(x), x = (x1, x2,…, xd)
- Initialize a search radius R, weighting search radius α, 0<α<1, and TL
For TC:
- Let Max_Gen is the maximum generation and Gen = 1
- Set SG for switching from FPA to ATS
Yes
- Update Gen = Gen + 1 Gen ≤ SG
No
- Set the initial solution x* = g*
- Activate ATS algorithm in Fig. 2
Yes Gen ≤
- Update Gen = Gen + 1
Max_Gen
No
Stop
3 Benchmark Functions
The proposed cooperative FPA-ATS algorithm will be tested against the standard
function optimization to perform its effectiveness for global minimum finding. In this
work, ten multimodal benchmark functions are selected [11, 12] as detailed in Table 1,
where x* = (x*, y*) is the optimal solution and f(x*) is the optimal function values. All
selected functions are nonlinear and multimodal functions which are very complex for
global minimum finding. They consist of (1) Griewank function (GF), (2) Salomon
function (SF), (3) Michaelwicz function (MF), (4) Rastrigin function (RF), (5) Peaks
function (PF), (6) Beale function (BF), (7) Giunta function (GiF), (8) Lévy function
(LF), (9) Yang function (YF) and (10) Drop-Wave function (DWF), respectively.
Names, Functions,
Functions Surface sketches
Search space and Coefficients
6
5
f1(x1, x2) 4
3
2
1 100
0 50
100 0 x1
50
0 -50
x2 -50
-100 -100
2.5
2
f2(x1, x2)
1.5
1
0.5
0 4
4 2
2 0 x1
0 -2
x2 -2
-4 -4
1
0.5
0
f3(x1, x2)
-0.5
-1
-1.5
4
-2 3
4 2x
3 1
2 1
x2 1
0 0
80
60
f4(x1, x2)
40
20
5
0
5 0 x1
0
x2 -5
-5
5
f5(x1, x2)
-5
4
-10 2
4 0 x1
2
0 -2
x2 -2
-4 -4
(Continued)
Cooperative FPA-ATS Algorithm for Global Optimization 159
Table 1. (Continued)
Names, Functions,
Functions Surface sketches
Search space and Coefficients
4
×10
15
f6(x1, x2)
10
5
4
0 2
4 0 x
2 1
0 -2
x2 -2 -4
-4
0.6
f7(x1, x2)
0.4
0.2
1
0
1 0 x1
0.5
0
x2 -0.5 -1
-1
400
300
f8(x1, x2)
200
100
10
0 5
10 0 x1
5
0 -5
x2 -5
-10 -10
0.5
f9(x1, x2)
-0.5
10
-1 5
10 0x
5 1
0 -5
x2 -5
-10 -10
-0.2
-0.4
f10(x1, x2)
-0.6
-0.8
5
-1
5 0x
1
0
x2 -5
-5
To perform its search performance, the proposed cooperative FPA-ATS will be tested
against 10 selected benchmark functions as summarized in Table 1. For the perfor-
mance evaluation and comparison, the ATS, FPA and cooperative FPA-ATS algo-
rithms were coded by MATLAB version 2017b (License No.#40637337). Searching
parameters of the ATS are set according to recommendations [6] and the preliminary
study as shown in Table 2. Those of the FPA are set according to recommendations [3]
and the preliminary study for all benchmark functions, i.e. numbers of flowers n = 20
160 T. Niyomsat et al.
Table 3. (continued)
Func-tions Search parameters of FPA-ATS
FPA ATS SG
n p N R(%) Jmax AR mechanism
Stage-I Stage-II Stage-III
GiF 20 0.5 5 20 5 f(x) < f(x*)+1e-2 f(x) < f(x*)+1e-3 f(x) < f(x*)+1e-4 350
R = 1e-2 R = 1e-3 R = 1e-4
LF 20 0.5 5 20 5 f(x) < f(x*)+1e-1 f(x) < f(x*)+1e-2 f(x) < f(x*)+1e-3 750
R = 1e-2 R = 1e-3 R = 1e-4
YF 20 0.5 5 20 5 f(x) < f(x*)+1e-1 f(x) < f(x*)+1e-2 f(x) < f(x*)+1e-3 700
R = 1e-2 R = 1e-3 R = 1e-4
DWF 20 0.5 5 20 5 f(x) < f(x*)+1e-1 f(x) < f(x*)+1e-2 f(x) < f(x*)+1e-3 750
R = 1e-2 R = 1e-3 R = 1e-4
Note:n is numbers of flowers, p is a switch probability and SG is a switching generation
Table 4. Results of performance evaluation of the ATS, FPA and cooperative FPA-ATS.
Func- Algorithms
tions ATS FPA Cooperative FPA-ATS
GF 2.608e+4 ± 8.786e+3 1.914e+4 ± 2.149e+3 1.809e+4 ± 4.386e+1
(19%) (25%) (100%)
SF 1.937e+4 ± 1.260e+03 8.484e+3 ± 5.945e+3 1.276e+4 ± 5.782e+2
(33%) (64%) (94%)
MF 8.106e+3 ± 8.432e+3 4.406e+3 ± 8.904e+2 2.454e+3 ± 7.009e+2
(69%) (100%) (100%)
RF 6.642e+4 ± 5.678e+3 1.153e+4 ± 1.591e+3 1.010e+4 ± 6.267e+1
(93%) (100%) (100%)
PF 3.583e+3 ± 7.092e+2 2.406e+3 ± 2.012e+2 2.096e+3 ± 2.840e+1
(100%) (100%) (100%)
BF 5.274e+3 ± 4.896e+3 4.232e+3 ± 7.790e+2 2.269e+3 ± 2.560e+2
(97%) (100%) (100%)
GiF 2.218e+3 ± 6.745e+2 2.013e+3 ± 6.116e+2 1.903e+3 ± 6.421e+1
(100%) (100%) (100%)
LF 1.239e+4 ± 7.379e+3 4.861e+3 ± 8.239e+2 4.152e+3 ± 2.280e+2
(64%) (100%) (100%)
YF 8.305e+3 ± 4.204e+3 4.887e+3 ± 4.351e+3 1.807e+4 ± 9.435e+1
(93%) (98%) (99%)
DWF 3.846e+4 ± 1.794+4 1.399e+4 ± 2.842e+3 8.085e+3 ± 4.898e+1
(37%) (96%) (100%)
162 T. Niyomsat et al.
50
40
30
20
10
0
x2
-10
-20
-30
-40
-50
-50 -40 -30 -20 -10 0 10 20 30 40 50
x1
and a switch probability p = 0.5 (50%). Searching parameters of the cooperative FPA-
ATS algorithm are set according to the preliminary study as summarized in Table 3.
100-trial runs are conducted for each algorithm. All algorithms will be terminated once
two termination criteria (TC) are satisfied, i.e. (1) the function values are less than a
given tolerance e 10–5 or (2) the search meets the maximum generation (Max_-
Gen = 1,000). The former criterion implies that the search is success, while the later
means that the search is not success.
The comparison results between ATS, FPA and cooperative FPA-ATS are sum-
marized in Table 4. The numeric data in Table 4 are presented in the following format,
i.e. AE ± SD(SR%), where the AE is the average number (mean) of function evalu-
ations, the SD is the standard deviation and the SR is the success rate. The AE value
implies the searching time consumed. The less the AE, the less the searching time
consumed. The SD value implies the robustness of the algorithm. The less the SD, the
more the robustness. From Table 4, the proposed cooperative FPA-ATS performs more
efficient in finding the global minima of all selected benchmark functions than the FPA
and ATS, respectively, with the highest SR and smallest AE and SD values.
In addition, movements and convergent rates of the cooperative FPA-ATS for
global minimum finding of GF are plotted in Fig. 4 as an example. It was found that the
FPA-ATS can reach the global minima with the 1st TC for all trial runs.
5 Conclusions
results, it was found that the proposed cooperative FPA-ATS is more efficient than the
FPA and ATS with the highest success rate of minimum finding and smallest number
of function evaluations (least time-consumed) and smallest standard deviation of
solution finding (highest robustness). This can be concluded that the proposed coop-
erative FPA-ATS algorithm is one of the most effective metaheuristic optimization
techniques alternatively used for solving the global optimization problems.
References
1. Glover, F., Kochenberger, G.A.: Handbook of Metaheuristics. Kluwer Academic Publishers,
Dordrecht (2003)
2. Talbi, E.G.: Metaheuristics Forn Design to Implementation. Wiley, Hoboken (2009)
3. Yang, X.S.: Flower Pollination Algorithm for Global Optimization. Unconventional
Computation and Natural Computation. LNCS, vol. 7445, pp. 240–249 (2012)
4. Chiroma, H., Shuib, N.L.M., Muaz, S.A., Abubakar, A.I., Ila, L.B., Maitama, J.Z.: A review
of the applications of bio-inspired flower pollination algorithm. Procedia Comput. Sci. 62,
435–441 (2015)
5. He, X., Yang, X.S., Karamanoglu, M., Zhao, Y.: Global convergence analysis of the flower
pollination algorithm: a discrete-time Markov chain approach. In: International Conference
on Computational Science (ICCS2017), pp. 1354–1363 (2017)
6. Sujitjorn, S., Kulworawanichpong, T., Puangdownreong, D., Areerak, K-N.: Adaptive tabu
search and applications in engineering design. In: Zha, X.F., Howlett, R.J. (eds.) Integrated
Intelligent Systems for Engineering Design, pp. 233–257. IOS Press, Amsterdam (2006)
7. Glover, F.: Tabu Search - part i. ORSA J. Comput. 1(3), 190–206 (1989)
8. Glover, F.: Tabu Search - part ii. ORSA J. Comput. 2(1), 4–32 (1990)
9. Puangdownreong, D., Sujitjorn, S., Kulworawanichpong, T.: Convergence analysis of
adaptive tabu search. ScienceAsia J. Sci. Soc. Thai. 30(2), 183–190 (2004)
10. Puangdownreong, D., Kulworawanichpong, T., Sujitjorn, S.: Finite Convergence and
Performance Evaluation of Adaptive Tabu Search. LNCS, vol. 3215, pp. 710–717. Springer,
Heidelberg (2004)
11. Ali, M.M., Khompatraporn, C., Zabinsky, Z.B.: A Numerical evaluation of several
stochastic algorithms on selected continuous global optimization test problems. J. Global
Optim. 31, 635–672 (2005)
12. Jamil, M., Yang, X.S., Zepernick, H-J.: Test functions for global optimization: a
aomprehensive survey. In: Swarm Intelligence and Bio-Inspired Computation Theory and
Applications, pp. 193–222. Elsevier Inc. (2013)
Bayesian Optimization for Reverse Stress
Testing
Peter Mitic1,2(&)
1
Department of Computer Science, UCL, Gower Street,
London WC1E 6BT, UK
p.mitic@ucl.ac.uk
2
Santander UK, 2 Triton Square, Regents Place, London NW1 3AN, UK
1 Introduction
The need for banks to be resilient to adverse change due to economic circumstances has
been a principle, and indeed a regulatory requirement, for many years. As a result,
stress testing has become a routine component in banking operations. The overall
purpose of stress testing is to ensure that national banking systems are resilient to
economic shocks. The testing principles involved originate from the Bank for Inter-
national Settlements [1] and are reflected in central bank publications such as from the
European Central Bank [2]. Most countries seek to implement these principles but
almost nothing is said about how that should be done.
A central bank would typically issue a set of stressed economic factors and expect
regional banks to calculate the reserves they need to cover anticipated losses. The
process is referred to simply as stress testing. The alternative is termed reverse stress
testing. Given a predetermined target, the level of stressed resources is calculated so as
to hit the target. The back calculations involved are often very onerous. In this paper we
present a method to do them efficiently, and illustrate it in the context of operational
risk (losses resulting from failed processes, fraud and operational failures).
The optimization problem defined by Eq. (2) should be regarded as a search for a
minimum relative error. Alternatively, it could be interpreted as a search for a value, ^x,
that satisfies an equality gð^xÞ ¼ 0, but with a very wide error bound specified by the
parameter L.
The objective function g used to calculate VaR in this context is an implementation
of the Loss Distribution Approach (hereinafter LDA), due to Frachot, Georges, and
Roncalli [3]. It uses a non-linear data fit, and simulates losses that represent, typically,
one year's activity. LDA valuations of g involve lengthy Monte Carlo calculations. In
this paper we say g is ‘expensive’. Therefore, we seek to minimize the number of
evaluations of g, and do it using a Gaussian Process (GP) embedded in Bayesian
Optimization (BO). Although BO is a well-established optimization technique,
empirical evidence presented here indicates a lack of success when using standard
forms of BO in the context of financial loss data. We suggest reasons and propose a
solution in Sect. 4.
2 Supporting Literature
In this section we summarize the main events in BO development, and stress its
relationship to a GP. BO with a GP is a method of using a Gaussian distribution of
functions as a proxy for optimizing a function that is difficult to optimize in any other
way. It is originally due to Mockus [4]. Two ideas were key to the original argument.
166 P. Mitic
The first was to calculate covariances using a kernel function. The original kernel took
ðxi xj Þ2
the form e 2 , in which the term xi xj is a distance measure between two feature
vectors xi and xj. The second was a proof that a sequence of evaluation points for the
‘expensive’ function proposed using the GP converges to a solution of the original
optimization problem. Mockus and co-workers proposed applications of the method
shortly after [5].
In [5], the concept of an acquisition function as a means of proposing the next
‘expensive’ evaluation point was formalised. A GP is characterised by two parameters:
a vector of means l and a covariance matrix K. From K, a vector, r, which can be
interpreted as a vector of standard deviations, can be derived. Specifically, r is the root
mean square error vector when a normal model parameterized by l and K is fitted to
data. The vectors l and r feature in the Expected Improvement (EI) acquisition
function, which was formalised by Jones [13], based on ideas in [5]. A Bayesian
predictor is tested at a sequence of possible evaluation points, and the point with the
highest expected value of providing an optimal next ‘expensive’ evaluation point is
selected. A similar acquisition function, Probability of Improvement - POI, was also
proposed by Mockus [6].
The innovation in this paper is to extend the Confidence Bound - CB acquisition
function, introduced by Cox and John [7]. Given elements of µ and r indexed by n and
with a real-valued constant j, the metric µn - jrn defines the Lower Confidence Bound
(LCB) acquisition function, which is used for minimisations. For maximizations, the
metric µn + jrn defines the Upper Confidence Bound (UCB) acquisition function. The
relationship between the objective function g(x) and the parameters of a GP, was
established by Srinivas et al. [8]. The absolute difference between g(x) and µn−1(x) at
stage n in a sequence of evaluations was shown to be bounded, with high probability,
pffiffiffiffiffi pffiffiffiffiffi
by bn rn1 ð xÞ, where bn is a real number that depends on logðn2 Þ. That result is
significant in the proof presented in Appendix A of this paper.
General discussion of the BO method, including development steps, may be found
in Rana et al. [9], Rasmussen and Williams [10], or Murphy [11]. Berk et al. [12]
highlight a general problem that GPs are prone to detecting a local rather than a global
optimum. That is the origin of the ‘sticking’ phenomenon, where the acquisition
function persistently proposes a non-optimal next evaluation point. We aim to make
progress in solving this problem in our analysis.
the Markov chain. Kwon and Mertikopoulos [17] used the idea of ‘regret’ (defined in the
Appendix of this paper) in a general method to replace discrete-time optimizations by
continuous-time optimizations. Ascher [18] approximates a discrete process by a con-
tinuous system of differential equations. In our context, such a replacement might take
the form of an implicit or explicit expression of g(x) (Eq. 2), from which a solution to
g(x) = V could be obtained numerically. Samakpong et al. [21] used Particle Swarm
optimization methods in conjunction with random sampling and Monte Carlo simula-
tions. Particle Swarms are unlikely to be useful in our context because the ‘particles’ are
competing candidate solutions. Any competition implies more ‘expensive’ evaluations
of function g (Eq. 2). Chance-constrained optimization was used by Yang and Sutanto
[19] in conjunction with scenarios to model parameter uncertainty. As with Particle
Swarms, too many ‘expensive’ evaluations of g would be needed in our context.
In this section we give a brief overview of BO, and stress the roles of the GP, and the
acquisition function part of the BO. The intention of the combined BO-GP process is to
minimize the number of slow or difficult evaluations of g. It proceeds as follows.
1. Define an initial set of evaluation points
2. Repeat until converged
2:1. Define a Bayes prior with the GP
2:2. Supply data and calculate a Bayes likelihood
2:3. Calculate a Bayes posterior
2:4. Define sample points to calculate a Bayes predictor
2:5. Propose a next evaluation point using the BO Acquisition function
Rasmussen and Williams [10] define a GP as a distribution over functions. There
are two key points in GP usage. First, ‘expensive’ function evaluation is only necessary
at a finite, but arbitrary, set of initial evaluation points x = {x1, x2, …, xn}. Thereafter,
the BO executes multiple non-‘expensive’ function evaluations of a multi-variate
normal distribution, and only then proposes a next ‘expensive’ evaluation point xn+1.
Second, general applicability of a GP arises because if a Gaussian prior is defined and
combined with Gaussian data, both posterior and predictive distributions are Gaussian,
and evaluation of them is very fast.
calculations the vectors l and r are generated using the BayesianOptimization function in
the R package rBayesianOptimization [22]. Both are inputs to the LCB acquisition
function, which is the actual predictor of the next evaluation point, xn+1. Using j, the
M components of l and r are combined in the form li jri , and a minimum can be
sought by searching a range of values of those components, as in Eq. (3). Full details can
be found in Murphy [11]. Note that both l and r are both implied functions of x 2 I.
We have found that LCB acquisition, as formulated in Eq. (3), does not produce
satisfactory results for any reasonable value of j for the noisy data and noise-inducing
optimizing function g. This result was unexpected, since selecting larger j values
‘ought’ to have diversified the search (see [13], for example). The same is true for
UCB, EI and POI acquisitions. Therefore, we have introduced a small change to the
LCB acquisition, described in Sect. 4.
It is likely that LCB acquisition fails because the optimization rule in Eqs. 1 and 2
contains the additional requirement that the minimum deviation from zero must be
within a pre-determined limit L. The following amendment to LCB acquisition makes a
dramatic difference. We call it the ZERO acquisition function since the optimal solution
should result in an error (relative to the target) of approximately zero.
The intuition behind the proposal in Eq. (4) is that a quantity has to be as close as
possible to a target, and a simple way to measure closeness is “deviation-squared”
(absolute deviation works just as well). Ultimately the justification for using Eq. (4) is
empirical, and results are given in Sect. 5.
5 Results
The data used for this study (Y in Eq. 1) are daily operational losses collected from
January 2010 to December 2018. Outlier losses have a significant effect on the BO
process, resulting in unsatisfactory proposals.
Table 1 shows a comparison of the EI, POI, UCB, LCB and ZERO acquisition
functions. In each case, all evaluations were repeated 25 times to obtain a mean and
standard deviation run length. Three ‘expensive’ evaluations of g were done to ini-
tialize the GP. Also shown is a random search for the optimal solution. In general, the
ZERO, LCB and UCB parameters j were not significant in determining the run length,
and were both set to 1. The number of Monte Carlo iterations used in evaluating g was
170 P. Mitic
more important, and Table 1 shows the results for the maximum (5 million) and
minimum (1 million) considered. Normally for this type of calculation we would opt
for the minimum number (and often fewer) to save time.
Table 1. Mean and SD of run lengths for Random search, and the EI, POI, UCB, LCB and
ZERO acquisition functions: 1 and 5 million Monte Carlo iterations of g.
Acquisition function 5m Mean 1m Mean 5m SD 1m SD
EI 11.45 18.08 10.43 12.15
LCB (j = 1) 9.85 17.56 7.72 15.70
UCB (j = 1) 12.16 13.44 9.88 10.14
POI 14.60 22.92 14.41 15.87
Random 10.84 12.42 10.88 12.56
ZERO (j = 1) 5.60 7.28 3.85 6.62
The most notable observation from Table 1 is the poor performance of the
‘established’ acquisition functions relative to a random search, and the superior per-
formance of ZERO acquisition compared to all others. The high standard deviations in
all cases except ZERO acquisition indicate instability (and hence unreliability) of the
methods used. The ZERO acquisition standard deviations are also high (compared to
the corresponding means), but are acceptable given alternatives in the table.
Figure 2 shows an alternative view of the ZERO acquisition results, in the form of
surfaces that show the variation of the number of ‘expensive’ evaluations of g with the
number of Monte Carlo cycles used in those evaluations and values of j. Using µR and rR
for the mean and standard deviation of the Run Number, R, the upper surface shows
µR + rR, the middle surface shows µR, and the lower surface shows µR − rR (with a floor
of 3 – the initial number of evaluations). Overall, the surfaces are “choppy”, particularly
for fewer than 2 million Monte Carlo cycles. That is due to sampling and method-induced
uncertainty. The surfaces show that the ‘tightest’ confidence limits are also the most
stable regions. They are for low values of j and a high number of Monte Carlo cycles.
5.1 Discussion
There are two surprising elements in the foregoing discussion. The first is the obser-
vation that “traditional” GP acquisition functions underperformed when used with
noisy data and with an optimization function (g in Eq. 2) which is necessarily
stochastic. Comparisons indicated that points (xn) proposed by the GP did not lead to
convergence as fast as with a random selection process. The second is that a minor
modification of the LCB acquisition function (ZERO) performs significantly better than
the “traditional” acquisition functions. We have attempted to explain this improvement
informally by comparing the proposals in the LCB and ZERO cases. The ZERO case
appears to present more opportunity to propose a potential solution because there are
two possibilities for every one that LCB is able to propose. ZERO acquisition is helped
further because the target is expressed as a range rather than a fixed value.
Bayesian Optimization for Reverse Stress Testing 171
More formally, the argument can be made more secure using the concept of “re-
gret”. Ultimately, the reason for using ZERO acquisition is strictly practical: it works.
That is an important consideration in commercial data analysis.
The simple transition from LCB to ZERO acquisition has resulted in a very significant
improvement in solving the optimization problem in Eq. 1. Since the objective function
concerned is stochastic, it is not possible to guarantee consistency of results (as
measured by Run Length) other than by using a very large number of Monte Carlo
cycles. Consequently, choosing an optimization method with a low Run Length stan-
dard deviation is an important consideration.
We are therefore considering two distinct approaches. The first is to improve the
performance of ZERO acquisition even further by providing feedback to each GP
evaluation from the previous GP evaluation (apart from the first). The intention is to
vary the j parameter in Eq. 4 so as to optimize the next evaluation, and to stabilize
oscillation around what emerges as the next proposal point. So far only a minimal
improvement has been noted. This feedback approach is similar to that used by Ozer
et al. [20] in the context of forecasting ATM cash withdrawals.
The second approach does not use BO. We have found that using a binary or linear
interpolation search to solve Eq. 1 does not necessarily result in a lower run length, but
does result in a much more consistent run length.
172 P. Mitic
Appendix A
This proof shows that the expected number of ‘expensive’ function evaluations of
g (Eq. 2) is greater for LCB acquisition than for ZERO acquisition.
We first define Local Regret, which is the difference between a proposed function
estimate and the actual function value. That is followed by Cumulative Regret, which a
sum of local regrets. All other notation is defined in the main body of this paper. The
following proof depends on a probability bound on the error estimate for g(xn) at an
evaluation point xn due to Srinivas et al. [8]. With that bound, the general strategy is to
calculate an upper bound for local regret, use that bound to determine the expected
value of local regret, and then to calculate the expected number of ‘expensive’ function
evaluations using the ZERO and LCB acquisition functions.
Definition: bn Srinivas [8], Appendix A.1, Lemma 5.1, for constants Cn > 0 that
P
satisfy n 1 Cn1 ¼ 1, and small d:
jI jCn
bn ¼ 2 log : ðA3Þ
d
Using the definition of local regret (Srinivas [8] Appendix A.1, Lemma 5.1),
Equation (A4) provides upper and lower confidence bound for the GP “mu” term.
pffiffiffiffiffi
P jgð xÞ ln1 ð xÞj bn rn1 ð xÞ 1 d; 8x 2 I; 1 n N ðA4Þ
Then since ^x is an optimal solution (so that gð^xÞ\gðxn Þ), these upper and lower
bounds apply respectively:
pffiffiffiffiffi pffiffiffiffiffi
ln1 ð^xÞ þ bn rn1 ð^xÞ ln1 ðxn Þ þ bn rn1 ðxn Þ
pffiffiffiffiffi pffiffiffiffiffi ðA5Þ
ln1 ðxn Þ bn rn1 ðxn Þ ln1 ð^xÞ þ bn rn1 ð^xÞ:
The general strategy in this proof is to estimate the maximum cumulative regret in
each of these cases for LCB and ZERO acquisitions, and then calculate the expected
difference between the two.
Proposition: ZERO acquisition converges faster than LCB acquisition in the case
PðÞ [ 1 d (Eq. A4).
Proof: First, from Eq. A4 with probability greater than 1 d, the Local Regret cal-
culation proceeds as in Eq. A6. The first line uses Eq. A4 with x ¼ ^x, the second line
Bayesian Optimization for Reverse Stress Testing 173
uses Eq. A5 and the third line uses Eq. A4 with x ¼ xn . The fourth line uses an upper
bound: bn ¼ maxðbn Þ.
n
pffiffiffiffiffi
rn ln1 ð^xÞ þ bn rn1 ð^xÞ gðxn Þ
pffiffiffiffiffi
ln1 ðxn Þ þ bn rn1 ðxn Þ gðxn Þ
pffiffiffiffiffi ðA6Þ
2 bn rn1 ðxn Þ
pffiffiffi
2 brn1 ðxn Þ
Now consider the Local Regret in the cases of LCB and ZERO acquisition. ZERO
acquisition is always non-negative but LCB acquisition can be negative. So we partition
the values of n into those which result in zero or positive LCB acquisition (set S) and
those which result in negative LCB acquisition (the complement, set S0 ). These sets are
shown in Eq. A7.
For S, the evaluation points proposed are identical for the two acquisition functions,
since they both correspond to the same minimum. Therefore, using superscripts to
denote the regrets for the two acquisition functions, the following equality applies.
This leads to the following expressions for the two regrets, using Eq. A6.
8 pffiffi
> pffiffiffi
< max rnðZEROÞ ¼ 2 brn1 ðxn Þ ¼ 2 j b ln1 ðxn Þ
n
pffiffi ðA10Þ
> pffiffiffi
: max rnðLCBÞ ¼ 2 brn1 ðxn Þ ¼ 2 b ðln1 ðxn Þ þ /n Þ
n j
Equation (A11) shows the partitioning the Cumulative Regret between sets S
and S0 .
X X
RN ¼ r
n2S n
þ r
n2S0 n
ðA11Þ
Then, Eq. A12 show the maximum Cumulative Regret for ZERO and LCB
acquisitions.
174 P. Mitic
8 pffiffi
> P P
< max RðNZEROÞ ¼ n2S rn þ 2 b n2S0 ln1 ðxn Þ
j
p ffiffi ðA12Þ
> P 2 b P
: max RðNLCBÞ ¼
n2S rn þ j n2S0 ln1 ðxn Þ þ /n
Equation (A13) is a strong indication that ZERO acquisition leads to faster con-
vergence than LCB acquisition, since it applies with high probability 1 d. This
completes the proof.
References
1. Basel Committee on Banking Supervision, Stress testing principles d450. Bank for
International Settlements (BIS) (2018). https://www.bis.org/bcbs/publ/d450.htm
2. European Banking Authority: 2020 EU-wide stress test methodological note (2019). https://
www.eba.europa.eu/sites/default/documents/files/documents/10180/2841396/ba66328f-476f
-4707-9a23-6df5957dc8c1/2020%20EU-wide%20stress%20test%20-%20Draft%20Method
ological%20Note.pdf
3. Frachot, A., Georges, P., Roncalli, T.: Loss Distribution Approach for operational risk,
Working paper, Groupe de Recherche Operationnelle, Credit Lyonnais, France (2001).
https://ssrn.com/abstract=1032523
4. Mockus, J.: On Bayesian methods for seeking the extremum. In: Proceedings of IFIP
Technical Conference, pp. 400–404 (1974). https://dl.acm.org/citation.cfm?id=646296.
687872
5. Mockus, J., Tiesis, V., Zilinskas, A.: The application of Bayesian methods for seeking the
extremum. Towards Global Optimisation (eds. Dixon,L. and Szego,G.P.) vol. 2 (1978)
6. Mockus, J.: The Bayesian approach to local optimization. In: Bayesian Approach to Global
Optimization. Mathematics and Its Applications, vol. 37. Springer, Heidelberg (1989).
https://doi.org/10.1007/978-94-009-0909-0_7
7. Cox, D.D., John, S.: SDO: a statistical method for global optimization. In: Multidisciplinary
Design Optimization, pp. 315–329. SIAM, Philadelphia (1997)
8. Srinivas, N., Krause, A., Kakade, S., Seeger, M.: Gaussian process optimization in the bandit
setting: no regret and experimental design. In: Proceedings of ICML 2010, pp. 1015–1022
(2010). https://dl.acm.org/citation.cfm?id=3104322.3104451
9. Rana, S., Li, C., Gupta, S.: High dimensional Bayesian optimization with elastic Gaussian
process. In: Proceedings of 34th International Conference on Machine Learning, Sydney,
PMLR 70 (2017)
10. Rasmussen, C.E., Williams, C.K.I.: Gaussian Processes for Machine Learning. MIT Press,
Cambridge (2016)
11. Murphy, K.P.: Machine Learning: A Probabilistic Perspective, Chapter 15. MIT Press,
Cambridge (2015)
12. Berk, J., Nguyen, V., Gupta, S., et al.: Exploration enhanced expected improvement for
Bayesian optimization. In: Joint European Conference on Machine Learning and Knowledge
Discovery in Databases. LNCS, vol. 11052, pp. 621–637 (2018)
Bayesian Optimization for Reverse Stress Testing 175
13. Jones, D.R.: A taxonomy of global optimization methods based on response surfaces.
J. Global Optim. 21(4), 345–383 (2001)
14. Kara, G., Özmen, A., Weber, G.: Stability advances in robust portfolio optimization under
parallelepiped uncertainty. Central Eur. J. Oper. Res. 27, 241–261 (2019). https://doi.org/10.
1007/s10100-017-0508-5
15. Özmen, A., Weber, G.W., Batmaz, I., Kropat, E.: RCMARS: robustification of CMARS
with different scenarios under polyhedral uncertainty set. Commun. Nonlinear Sci. Numer.
Simul. 16(12), 478–4787 (2011). https://doi.org/10.1016/j.cnsns.2011.04.001
16. Savku, E., Weber, G.: Stochastic differential games for optimal investment problems in a
Markov regime-switching jump-diffusion market. Ann. Oper. Res. (2020). https://doi.org/10.
1007/s10479-020-03768-5
17. Kwon, J., Mertikopoulos, P.: A continuous-time approach to online optimization. J. Dyn.
Games 4(2), 125–148 (2017). https://doi.org/10.3934/jdg.2017008
18. Ascher, U.M.: Discrete processes and their continuous limits. J. Dyn. Games 7(2), 123–140
(2020). https://doi.org/10.3934/jdg.2020008
19. Yang, Y., Sutanto, C.: Chance-constrained optimization for nonconvex programs using
scenario-based methods. ISA Trans. 90, 157–168 (2019). https://doi.org/10.1016/j.isatra.
2019.01.013
20. Ozer, F., Toroslu, I.H., Karagoz, P., Yucel, F.: Dynamic programming solution to ATM cash
replenishment optimization problem. In: Intelligent Computing & Optimization. ICO 2018,
vol. 866. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-00979-3_45
21. Samakpong, T., Ongsakul, W., Nimal Madhu, M.: Optimal power flow considering cost of
wind and solar power uncertainty using particle swarm optimization. In: Intelligent
Computing and Optimization. ICO 2019, vol. 1072. Springer, Cham (2020). https://doi.org/
10.1007/978-3-030-33585-4_19
22. Yan, Y.: (2016). https://cran.r-project.org/web/packages/rBayesianOptimization/index.html
Modified Flower Pollination Algorithm
for Function Optimization
Abstract. The flower pollination algorithm (FPA) was firstly proposed in 2012
as one of the population-based metaheuristic optimization search techniques. It
is conceptualized by the pollination behavior of flowering plants. In this paper,
the new enhanced version of the original FPA named the modified flower
pollination algorithm (MoFPA) is proposed to improve its search performance
for function optimization. The switching probability of the original FPA used for
selection between local and global pollinations is changed from the fixed
manner to the random manner according to the pollination behavior of flowering
plants in nature. To perform its effectiveness, the proposed MoFPA is tested
against ten standard benchmark functions compared with the original FPA. As
simulation results, it was found that the proposed MoFPA performs superior
search performance for function optimization to the original FPA with higher
success rates and faster search time consumed.
1 Introduction
Last few decades, the nature-inspired metaheuristic algorithms have become very
popular in a wide range of real-world applications. This is due to their success in
finding good solutions for complex problems, especially the NP-complete problems [1,
2]. By literatures, most nature-inspired metaheuristic algorithms are developed by
mimicking biological or physical phenomena. They can be grouped in four main
categories, i.e. (i) evolution-based metaheuristic algorithms, (ii) physics-based meta-
heuristic algorithms, (iii) bio-swarm-based metaheuristic algorithms and human-based
metaheuristic algorithms [1, 2].
The flower pollination algorithm (FPA) was firstly proposed by Yang in 2012 [3]
for solving both continuous and combinatorial, single-objective and multi-objective
optimization problems [3, 4]. As one of the most popular and efficient metaheuristic
algorithms, the FPA algorithm imitates the pollination behavior of flowering plants.
Optimal plant reproduction strategy involves the survival of the fittest as well as the
optimal reproduction of plants in terms of numbers. These factors represent the fun-
damentals of the FPA and are optimization-oriented. The FPA algorithm was proved
for the global convergent property [5]. Since 2012, the FPA has shown superiority to
other metaheuristic algorithms in solving various real-world problems, such as
economic/emission dispatch, reactive power dispatch, optimal power flow, solar PV
parameter estimation, load frequency control, wireless sensor networks, linear antenna
array optimization, frames and truss systems, structure engineering design, multilevel
image thresholding, travelling transportation problem and control system design. The
state- of-the-art developments and significant applications of the FPA have been
reviewed and reported [6].
Following the literature, many variants of FPA have been developed by modifi-
cation, hybridization, and parameter-tuning manners in order to cope with the complex
nature of optimization problem. One of the modified flower pollination algorithms
called the MFPA [7]. The MFPA hybridized the original FPA with the clonal selection
algorithm (CSA) in order for generating some elite solutions. The binary flower pol-
lination algorithm (BFPA) was developed for solving discrete and combinatorial
optimization problems [8]. Another significant modification was proposed as the
modified global FPA (mgFPA) [9]. The mgFPA was designed to better utilize features
of existing solutions through extracting its characteristics, and direct the exploration
process towards specific search areas [9].
In this paper, the novel enhanced version of the original FPA named the modified
flower pollination algorithm (MoFPA) is proposed for function optimization. The
switching probability of the original FPA used for selection between local and global
pollinations is changed from the fixed manner to the random manner. The proposed
MoFPA is tested against ten benchmark optimization problems compared with the
original FPA in order to perform its search performance. This paper consists of five
sections. After an introduction presented in Sect. 1, the rest of the paper is arranged as
follows. The original FPA and the proposed MoFPA algorithms are described in
Sect. 2. Ten selected standard benchmark functions used in this paper for function
optimization are detailed in Sect. 3. Results and discussions of the performance
evaluation of the FPA and MoFPA algorithms against ten selected benchmark func-
tions are illustrated in Sect. 4, while the conclusions are followed in Sect. 5.
In this section, the original FPA algorithm is briefly detailed. Then, the proposed
MoFPA algorithm is elaborately illustrated.
kCðkÞ sinðpk=2Þ 1
L ; ðs s0 [ 0Þ ð2Þ
p s1 þ k
motivation, the random value of pollination process is still varied from 0 to 1, i.e.
rand 2 [0, 1], but the fixed value of the switching probability p = 0.2 will be changed to
the randomly switching probability randp 2 [pmin, pmax], pmin, pmax 2 [0, 1] and
pmin < pmax, regarding to the natural flower pollination behavior.
Start
Initialize:
- Objective function f(x), x = (x1, x2,…, xd)
- Initialize a population of n flowers/pollen gametes with random solutions
- Find the best solution g* in the initial population
- Define a switch probability p ∈ [0, 1]
rand > p No
(random) (fixed)
Yes
global pollination
local pollination
- Draw a step vector L - Draw ε from a uniform
via Lévy flight in (2) distribution ∈ [0, 1] in (4)
- Activate cross - Randomly choose j and
pollination in (1) k among all the solutions
- Invoke self pollination in (3)
No
f(x)<f(g*)
Yes
- Update g* = x
No
TC ?
Yes
- Report the current best solution g*
Stop
pollination will be invoked. The algorithm of the proposed MoFPA can be represented
by the flow diagram of the original FPA as shown in Fig. 1, but the condition of
selection between local and global pollinations is changed to rand > randp, where
rand 2 [0, 1] and randp 2 [pmin, pmax] are random numbers drawn from a uniform
distribution. With a new selecting condition, if rand > randp, the global pollination
will be activated as shown in Fig. 2(b). Otherwise, the local pollination will be
invoked. Form this modification, the mathematical relations in (1)–(4) of the original
FPA will be conducted in the proposed MoFPA. The algorithm of the proposed
MoFPA can be represented by the flow diagram shown in Fig. 3.
random random
pollinations
1.0 1.0
global
pollinations
rand ∈[0, 1]
global
pollinations
0.5
local
pollinations
0.2
p = 0.2
(fixed)
local
0 0
(a) Original FPA (b) Proposed MoFPA
Hundreds benchmark test functions have been proposed for global minimum [13–15].
In this work, ten standard benchmark test functions are selected for testing the proposed
MoFPA. They are (1) Sinusoid function (SF), (2) Bohachevsky function (BF),
(3) Rastrigin function (RF), (4) Griewank function (GF), (5) Michaelwicz function
(MF), (6) Shubert function (ShF), (7) Rosenbrock function (RoF), (8) Schwefel
function (SchF), (9) Keane function (KF) and (10) Cosine-Mixture function (CMF),
respectively. Details of all selected benchmark test functions are summarized in
Table 1, where x* = (x*, y*) is the optimal solution and f(x*) is the optimal function
values. Ten selected benchmark test functions are different in that they are nonlinear,
multimodal and unsymmetrical functions which are very difficult for global minimum
finding [13–15].
Modified Flower Pollination Algorithm for Function Optimization 181
Start
Initialize:
- Objective function f(x), x = (x1, x2,…, xd)
- Initialize a population of n flowers/pollen gametes with random solutions
- Find the best solution g* in the initial population
- Define a switch probability randp ∈ [pmin, pmax]
Yes
global pollination
local pollination
- Draw a step vector L - Draw ε from a uniform
via Lévy flight in (2) distribution ∈ [0, 1] in (4)
- Activate cross - Randomly choose j and
pollination in (1) k among all the solutions
- Invoke self pollination in (3)
No
f(x)<f(g*)
Yes
- Update g* = x
No
TC ?
Yes
- Report the current best solution g*
Stop
The proposed MoFPA will be tasted against ten selected benchmark test functions to
perform its search performance for function optimization. Both original FPA and
MoFPA algorithms were coded by MATLAB version 2018b (License No.#40637337)
run on Intel(R) Core(TM) i5-3470 CPU@3.60 GHz, 4.0 GB-RAM for comparison.
Searching parameters of the original FPA are set according to recommendations of
Yang [3, 4], and the preliminary studies on the selected benchmark test functions as
detaied in Table 2, where n is number of flower and p is the switching probability. For
182 N. Pringsakul and D. Puangdownreong
(Continued)
Modified Flower Pollination Algorithm for Function Optimization 183
Table 1. (Continued)
184 N. Pringsakul and D. Puangdownreong
the proposed MoFPA, searching parameters are set for a fair comparison, i.e. n is the
same value of the original FPA for each function. Values of randp 2 [pmin, pmax] are
set as four intervals, i.e. the 1st interval: randp 2 [0, 0.25], the 2nd interval: randp
[0.25, 0.5], the 3rd interval: randp 2 [0.5, 0.75] and the 4th interval: randp 2 [0.75, 1],
respectively. 100-trial runs are conducted for each algorithm in order to carry out
meaningful statistical analysis. Both algorithms will be terminated once two termina-
tion criteria (TC) are satisfied, i.e. (1) the function values are less than a given tolerance
d 10–5 or (2) the search meets the maximum generation (Max_Gen = 1,000). The
former criterion implies that the search is success, while the later means that the search
is not success. The simulation results are summarized in the Table 3, where the global
optima are reached. The numeric data in Table 3 are expressed in the format of
ANE ± STD(PSR), where ANE is the average number of evaluations, STD is the
standard deviation and PSR is percent success rate.
Table 3. Results of performance evaluation of the original FPA and proposed MoFPA.
Modified Flower Pollination Algorithm for Function Optimization 185
From Table 3, it was found that the proposed MoFPA with randp 2 [0, 0.25] is not
better than the original FPA because it almost uses only global pollination. Also, the
MoFPA with randp 2 [0.75, 1] is not better than the original FPA because it almost
uses only local pollination. The MoFPA with randp 2 [0.25, 0.5] is not better than the
original FPA, but it is better than the MoFPA with randp 2 [0, 0.25] and that with
randp 2 [0.75, 1]. However, the proposed MoFPA with randp 2 [0.5, 0.75] is the bset
amont those intervals. It performs more efficient in finding the global optima with faster
(less average number of evaluations) and higher success rates superior to the original
FPA. This is because it is balanced between local and global pollinations.
Figure 4(a), 4(b) and 4(c) show the results of the global optimal finding by the
MoFPA against the SF function at the 1st, 50th and 100th generation, respectively. The
convergent rates over 100-trial runs of the SF’s solution finding proceeded by the
original FPA and the proposed MoFPA are shown in Fig. 5(a) and 5(b), respectively.
The convergent rates of other functions are omitted because they have a similar form to
those of SF function in Fig. 5. Form Fig. 5, it can be observed that the proposed
MoFPA is more robust than the original FPA in global optimum finding.
(a) At the 1st generation (b) At the 50th generation (c) At the 100th generation
Fig. 4. MoFPA movement for global optimal finding of Sinusoid function (SF).
Fig. 5. Convergent rates over 100-trial runs of the SF’s solution finding.
186 N. Pringsakul and D. Puangdownreong
5 Conclusions
The modified flower pollination algorithm (MoFPA) has been proposed in this paper.
Based on the original FPA formed from the pollination behavior of flowering plants,
the proposed MoFPA is more flexible by using the random switching probability for
selection between local and global pollinations. The MoFPA has been tested against ten
selected standard benchmark functions. Results of function optimization obtained by
the MoFPA have been compared with those obtained by the original FPA. As simu-
lation results, it was found that the proposed MoFPA has performed the best perfor-
mance once the random switching probability is varied in the interval of 0.5 – 0.75.
Moreover, the MoFPA yields superior search performance for global optimization to
the original FPA with higher success rates and faster search time consumed.
References
1. Glover, F., Kochenberger, G.A.: Handbook of Metaheuristics. Kluwer Academic Publishers,
Dordrecht (2003)
2. Talbi, E.G.: Metaheuristics Forn Design to Implementation. Wiley, Hoboken (2009)
3. Yang, X.S.: Flower pollination algorithm for global optimization. In: Unconventional
Computation and Natural Computation. LNCS, vol. 7445, pp. 240–249 (2012)
4. Yang, X.S., Karamanoglu, M., He, X.S.: Flower pollination algorithm: a novel approach for
multiobjective optimization. Eng. Optim. 46(9), 1222–1237 (2014)
5. He, X., Yang, X.S., Karamanoglu, M., Zhao, Y.: Global convergence analysis of the flower
pollination algorithm: a discrete-time Markov chain approach. In: International Conference
on Computational Science (ICCS 2017), pp. 1354–1363 (2017)
6. Chiroma, H., Shuib, N.L.M., Muaz, S.A., Abubakar, A.I., Ila, L.B., Maitama, J.Z.: A review
of the applications of bio-inspired flower pollination algorithm. Procedia Comput. Sci. 62,
435–441 (2015)
7. Nabil, E.: A modified flower pollination algorithm for global optimization. Expert Syst.
Appl. 57, 192–203 (2016)
8. Rodrigues, D., Yang, X.S., De Souza, A.N., Papa, J.P.: Binary flower pollination algorithm
and its application to feature selection. In: Recent Advances in Swarm Intelligence and
Evolutionary Computation, pp. 85–100. Springer, Cham (2015)
9. Shambour, M.Y., Abusnaina, A.A., Alsalibi, A.I.: Modified global flower pollination
algorithm and its application for optimization problems. Interdisc. Sci. Comput. Life Sci. 11,
1–12 (2018)
10. Willmer, P.: Pollination and Floral Ecology. Princeton University Press, Princeton (2011)
11. Balasubramani, K., Marcus, K.: A study on flower pollination algorithm and its applications.
Int. J. Appl. Innov. Eng. Manag. 3, 320–325 (2014)
12. Pavlyukevich, I.: Lévy flights, non-local search and simulated annealing. J. Comput. Phys.
226, 1830–1844 (2007)
13. Ali, M.M., Khompatraporn, C., Zabinsky, Z.B.: A numerical evaluation of several stochastic
algorithms on selected continuous global optimization test problems. J. Global Optim. 31,
635–672 (2005)
14. Jamil, M., Yang, X.S.: A literature survey of benchmark functions for global optimization
problems. Int. J. Math. Model. Numer. Optim. 4(2), 150–194 (2013)
15. Kashif, H., Mohd, N.M.S., Shi, C., Rashid, N.: Common benchmark functions for
metaheuristic evaluation: a review. Int. J. Inf. Visual. 1(4–2), 218–223 (2017)
Improved Nature-Inspired Algorithms
for Numeric Association Rule Mining
1 Introduction
Association Rule Mining (ARM) is used for discovering the dependence rules
between features in a transaction database. On the other hand, Numeric Associ-
ation Rule Mining (NARM) extends the idea of ARM, and is intended for mining
association rules where attributes in a transaction database are represented by
numerical values [4]. Usually, traditional algorithms, e.g. Apriori, requires numer-
ical attributes to be discretized before use. Discretization is sometimes trivial,
and sometimes does not have a positive influence on the results of mining. On
the other hand, many methods exist for ARM that do not require the discretiza-
tion step before applying the process of mining. Most of the these methods are
based on population-based nature-inspired metaheuristics, such as, for example,
Differential Evolution or Particle Swarm Optimization. NARM has recently also
been featured in some review papers [3,7] which emphasize its importance in the
data revolution era.
The objective of this short paper is to extend the paper of Fister et al. [5],
where the new algorithm for NARM was proposed, based on Differential Evolu-
tion. Indeed, the practical experiments revealed some problems/bottlenecks that
can be summarized into two issues:
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 187–195, 2021.
https://doi.org/10.1007/978-3-030-68154-8_19
188 I. Fister Jr. et al.
2 Background Information
2.1 Association Rule Mining
This section presents briefly a formal definition of ARM. Let us suppose, a set
of objects O = {o1 , . . . , om }, where m is the number of objects, and transaction
set D is given, where each transaction T is a subset of objects T ⊆ O. Then, an
association rule can be defined as the implication:
X ⇒ Y, (1)
where X ⊂ O, Y ⊂ O, in X ∩ Y = ∅. The following two measures are defined
for evaluating the quality of the association rule [2]:
n(X ∪ Y )
conf (X ⇒ Y ) = , (2)
n(X)
n(X ∪ Y )
supp(X ⇒ Y ) = , (3)
N
where conf (X ⇒ Y ) ≥ Cmin denotes confidence and supp(X ⇒ Y ) ≥ Smin
support of association rule X ⇒ Y . Thus, N in Eq. (3) represents the number of
transactions in transaction database D, and n(.) is the number of repetitions of
the particular rule X ⇒ Y within D. Here, Cmin denotes minimum confidence
and Smin minimum support determining that only those association rules with
confidence and support higher than Cmin and Smin are taken into consideration,
respectively.
Improved Nature-Inspired Algorithms for Numeric Association Rule Mining 189
The basis of our study presents Differential Evolution for ARM using mixed
attributes (ARM-DE) proposed by Fister et al. in [5]. Development of this algo-
rithm is divided into three steps:
– domain analysis,
– representation of solution,
– evaluation function definition.
In that study, domain analysis of the observed sport database identified three
numerical and eleven categorical attributes. Thus, the former are determined
with intervals of feasible values limited by their minimum lower and maximum
upper bound values.
The individuals in the population are represented as real-valued vectors,
where every numerical attribute consists of corresponding lower and upper
bounds. On the other hand, each categorical attribute is identified by the real-
value drawn from the interval [0, 1] that is mapped to the corresponding discrete
attribute according to dividing the interval into equally-sized non-overlapping
sub-intervals. Additionally, the last element in the representation xi,D denotes
the so-called cut point, determining which part of the vector belongs to the
antecedent and which to the consequent of the mined association rule. The fol-
lowing fitness function was defined for evaluation of solutions:
(t) (t) (t)
(t) α ∗ conf (xi ) + γ ∗ supp(xi )/α + γ, if feasible(xi ) = true,
f (xi ) = (4)
−1, otherwise,
where conf (.) is confidence, supp(.) support, α and γ are weights, function
feasible(xi ), denotes if the solution is feasible. The task of the optimization
is to find the maximum value of the evaluation function.
(t)
where elements xi,4πj +k , for i = 1, . . . , Np and j = 0, . . . , D − 1 and k = 0, . . . , 3,
denote the attributes of features in association rules, t is an iteration counter,
and D the number of attributes. Indeed, each numerical attribute is expressed
as a quadruple:
(t)
At i,j = xi,4πj , xi,3πj +1 , xi,4πj +2 , xi,4πj +3 , (6)
where the first term denotes the lower bound, the second the upper bound, the
third the threshold value, and the fourth determines the ordering of the attribute
in the permutation.
The threshold value determines the presence or absence of the corresponding
numerical feature in the association rule, in other words:
⎧
⎪ (t)
⎨ [N U LL, N U LL], if rand(0, 1) < xi,4πj +2 ,
⎪
At(t)
πj =
(t) (t) (t) (t)
[xi,4πj , xi,4πj +1 ], if xi,4πj > xi,4πj +1 , (7)
⎪
⎪
⎩ [x(t) ,x
(t)
], otherwise,
i,4πj +1 i,4πj
and Lb πj and Ub πj denote the corresponding lower and upper bounds. The
motivation behind the proposed equation is to shrink the whole interval of the
feasible values.
(t)
A permutation Π = (π1 , . . . , πD ) is assigned to each solution xi , which
(t)
orders the attributes At j according to the following equation:
(t) (t) (t)
xi,4π0 +3 ≥ xi,4πj +3 ≥ xi,4πD−1 +3 , for j = 0, . . . , D − 1. (9)
(t)
Thus, the attributes with the higher value of the fourth element xi,4πj +3 are
ordered at the start of the permutation, while the attributes with the lower
Improved Nature-Inspired Algorithms for Numeric Association Rule Mining 191
values at the end of the permutation. In this case, each numerical attribute has
an equal chance to be selected as an antecedent or consequent of the mined
association rule.
(t)
The last element in the vector determines the cut point Cp i , expressed as:
(t) (t)
Cp i = xi,D · (D − 2) + 1, (10)
(t)
where Cp i ∈ [1, D −1]. In summary, the length of solution vector is the 4·D +1.
The mapping of the solution representation to the corresponding association
rule is expressed as follows:
(t)
Ante(X ⇒ Y ) = {oπj |πj < Cp i ∧ At π(t)j = [NULL, NULL]},
(t)
(11)
Cons(X ⇒ Y ) = {oπj |πj ≥ Cp i ∧ At π(t)j = [NULL, NULL]},
1
The first number denotes the number of antecedents, while second denotes the num-
ber of consequent.
Improved Nature-Inspired Algorithms for Numeric Association Rule Mining 193
4.1 Discussion
The results of the mentioned nature-inspired algorithms for NARM showed that
selection of the algorithm has a big influence on the quality of the results. Thus,
an advantage of the DE algorithm is in the total discovered rules, average fitness
and inclusion, while the PSO was better regarding the average support, confi-
dence, and shrink. On the other hand, working with the numerical attributes
revealed a lot of issues that need to be considered for the future work. Let us
mention only the more important ones:
– How to consider shrinking as the statistical measure? In our results, we con-
sidered the shrinking intervals of all attributes, including those that did not
arise in the mined rules.
– How to balance the weights of four terms in the proposed evaluation function?
In our case, all weights were set to the value of 1.0, which means that all the
contributions were weighted equally.
– Is the best mined association rule according to fitness value also the most
interesting?
– How to find the balance between shrink and inclusion?
The mentioned issues confirm that the development of the proposed algo-
rithm for pure NARM is far from completion. A lot of researches would be
necessary in order to find the proper answers to the mentioned issues.
194 I. Fister Jr. et al.
5 Conclusion
Development of a nature-inspired algorithm for pure NARM demands answers
to new issues, such as, for instance: How to shrink the lower and upper bounds of
numerical attributes? or How to find the proper evaluation function? The former
issue is confronted with the exploration of the search space, while the latter with
evaluating the quality of the mined association rules.
This paper proposes usage of the shrinking coefficient, that is determined as
a ratio between the difference of the generated upper and lower bounds, and
difference of the maximum upper and minimum lower bounds. As an evalua-
tion function, a weighted sum of support, confidence, inclusion, and shrinking
coefficient are taken into consideration. However, the weights were set to the
same value of 1.0 in our preliminary study. The nature-inspired algorithms for
pure NARM were employed to a sample sport dataset generated by the random
generator located on the web. Even four nature-inspired algorithms were tested
in our comparative study, as follows: DE, PSO, CS, and FPA.
The results of the comparative analysis revealed that, although there are dif-
ferences between the specific nature-inspired algorithms, these could be applied
for solving the problem in practice. On the other hand, a lot of work is necessary
in order to find the proper weights for determining the particular contributions
of terms in the evaluation function. However, all this work could be a potential
direction for the future work.
References
1. Sportydatagen: An online generator of endurance sports activity collections. In:
Proceedings of the Central European Conference on Information and Intelligent
Systems, Vara ždin, Croatia, 19, 21 September 2018, pp. 171–178 (2018)
2. Agrawal, R., Srikant, R., et al.: Fast algorithms for mining association rules. In:
Proceedings 20th international conference very large data bases, VLDB, vol. 1215,
pp. 487–499 (1994)
3. Altay, E.V., Alatas, B.: Performance analysis of multi-objective artificial intelligence
optimization algorithms in numerical association rule mining. J. Ambient Intell.
Human. Comput. 1–21 (2019)
4. Fister, Jr.I., Fister, I.: A brief overview of swarm intelligence-based algorithms for
numerical association rule mining. arXiv preprint arXiv:2010.15524 (2020)
5. Fister Jr., I., Iglesias, A., Galvez, A., Del Ser, J., Osaba, E., Fister, I.: Differential
evolution for association rule mining using categorical and numerical attributes. In:
Yin, H., Camacho, D., Novais, P., Tallón-Ballesteros, A.J. (eds.) Intelligent Data
Engineering and Automated Learning - IDEAL 2018, pp. 79–88. Springer Interna-
tional Publishing, Cham (2018)
6. Storn, R., Price, K.: Differential Evolution – A Simple and Efficient Heuristic for
Global Optimization over Continuous Spaces. J. of Global Optimization 11(4), 341–
359 (dec 1997). https://doi.org/10.1023/A:1008202821328
7. Telikani, A., Gandomi, A.H., Shahbahrami, A.: A survey of evolutionary computa-
tion for association rule mining. Information Sciences (2020)
Improved Nature-Inspired Algorithms for Numeric Association Rule Mining 195
8. Yang, X.S.: Flower pollination algorithm for global optimization. In: Durand-Lose,
J., Jonoska, N. (eds.) Unconventional Computation and Natural Computation, pp.
240–249. Springer, Heidelberg (2012)
9. Yang, X.S.: Bat algorithm and cuckoo search: a tutorial, pp. 421–434. Springer,
Heidelberg (2013). https://doi.org/10.1007/978-3-642-29694-9 17
Verification of the Adequacy of the Topological
Optimization Method of the Connecting Rod
Shaping by the BESO Method in ANSYS
APDL System
1 Introduction
The main parts of piston of internal combustion engines are exposed to several
variables in magnitude and direction of forces; therefore, their stress-strain state turns
out to be difficult. Because of this, without calculation or experiment, it is difficult to
determine which parts of the detail are loaded more, and which parts are loaded less. In
addition, when acting on a variable of magnitude forces, it is necessary to take into
account fatigue phenomena in the metal. The experience of design, testing and oper-
ation shows that in most cases the destruction of the main parts of internal combustion
engines is precisely fatigue. Often the material of the part is loaded unevenly: in some
areas, the stress values approach the maximum permissible and in some – the material
is practically not loaded. This means that the material is not optimally used. The task of
finding the optimal shape of the part is difficult, but relevant, since such optimization
will reduce the material consumption. In addition, if the part is movable, then reducing
its mass will lead to a decrease in inertial forces acting on it.
The performance of the part can be checked by experimental or calculated methods.
Experimental methods are accurate, but their preparation and implementation take a
very long time even with the use of rapid prototyping technologies. Their cost is also
quite large. For this reason, it is preferable to use computational methods for assessing
the performance of parts, especially in the initial stages of finding the optimal design.
There is a large number of ready-made software products for solving optimization
problems. However, most often they have some disadvantages, such as cost, closed
source code or lack of flexibility. For example, in the “Topology Optimization” module
included in the ANSYS software package it is impossible to adequately describe a
cylindrical sliding bearing, since the types of contacts that allow significant relative
displacement of the contacting parts are not applicable. For this reason, it was decided
to develop its own optimization method. In order to formulate the requirements for the
developed optimization method, a review of existing methods was performed.
2 Analysis of Publications
The optimization problem is described by a goal function and restrictions. The function
of the goal in problems of optimizing the shape of parts is the mass of the
part. Restrictions may be different, but most often they are somehow related to the
performance of the part. Currently, the evaluation of the reliability and performance of
parts is almost always carried out by the finite element method. In most cases, com-
mercial software products are used, for example, ANSYS [1–6], ABAQUS [7–9],
COMSOL [10], MSC Patran [11], LS-DYNA [12].
To search the optimal shape, some method of creating various variants of the design
of the investigated part is necessary, from which the optimal variant is selected.
A set of design options can be created in advance [3–5, 7, 13]. In this case, the best
design option is selected from among the available options. Additional changes in the
design are not made. However, in this case only a limited number of options can be
investigated.
Another approach is to gradually change the design of the part, depending on the
results of testing its performance. This approach can be called iterative [1, 6, 8–12, 14–
16]. Design changes can be made either manually [1, 17] or automatically by creating
198 S. Chirskiy and V. Panchenko
various feedbacks between the results of the performance check and the design. To
automate the change in the shape of the part, use either its parametric description [3–5, 8,
10, 14], either topological [9, 11, 12, 15, 16]. Topological description is more compli-
cated than parametric, but it allows getting much more diverse forms of parts.
Different optimization methods use different ways to find the optimal solution. In
the event that the shape of the part changes during the optimization process, the
algorithm for finding the optimal variant describes precisely the law of shape change.
The approximation method can be used in tasks of finding the optimal shape of parts [5]
or method based on sensitivity analysis [4]. Often used evolutionary [6] or genetic
algorithms [6, 8–10, 12, 14, 15]. An example of a more exotic method is the modified
cuckoo algorithm [10].
The selected or developed algorithm for finding the best variant can be imple-
mented using the finished software product, for example, TOSCA [15], mode-
FRONTIER [8], OPTIMUS [12]. In addition, the program for finding the optimal
variant can be written in any programming system, for example, Python [9].
Fig. 1. The general block diagram of the developed algorithm for finding the optimal shape of
the detail
As a result, the simulation of the stress-strain state of the model being optimized is
performed for two sets of boundary conditions corresponding to the states of maximum
and minimum loading. The simulation results allow calculating the safety factors by the
criterion of cyclic strength.
The result of the review was the following requirements for the optimization method
being developed: the optimization method should be iterative and based on the topo-
logical method of describing the shape of the detail. Since parts of piston engines
operate under conditions of variable periodic forces, it is necessary to take into account
the phenomena of metal fatigue. To model the stress-strain state, the ANSYS finite-
element analysis package was chosen and the program for finding the optimal variant is
implemented in the built-in programming language ANSYS APDL.
To assess the adequacy of the developed algorithm, test models were optimized
with its help and with the help of the topology optimization module “Topology
Optimization”, included in the ANSYS software package. The piston engine con-
necting rod is selected as the model to be optimized. The sizes and acting forces are
determined according to the tables of average statistical values for forced diesel
engines. Three cases of loading are considered:
1) Asymmetrical compression-tension, the compressive force is 22122 N, tensile force
is 5482,8 N (Fig. 2).
200 S. Chirskiy and V. Panchenko
Fig. 3. Calculation model: boundary conditions for symmetric bending by forces equal modulo
3) Asymmetrical complex loading, approximately describing the real load cycle acting
on the engine connecting rod (Fig. 4).
In all cases, the tightening of the connecting rod bolts is modeled by a pair of
oncoming forces equal to the modulus of 13316 N, compressing the middle part of the
connecting rod bolt. For the inner cylindrical surface of the connecting rod, a boundary
condition of the type “Fixed Support” is set, describing the termination. Such boundary
conditions do not fully describe the operating conditions of the connecting rod, but they
can also be set as boundary conditions in the ANSYS Mechanical APDL system and in
the “Topology Optimization” module.
Some surfaces of the model are used to set the boundary conditions – the upper
head bushing, the lower head liners, the connecting rod bolts and the surfaces with
which the bolts come into contact, the surface of the connecting rod and the connecting
rod cap. In order for these surfaces to remain unchanged during optimization, they are
excluded from the region being optimized. In both cases, the optimization criterion is
the mass of the part. Restrictions are set so that the maximum equivalent stress does not
exceed 600 MPa.
Optimization results for all three loading variants are presented below. In all cases
the left side of the figures shows the result of optimization using the developed method
and the right side of the figures shows the result of optimization in the “Topology
Optimization” module, included in the ANSYS software package.
The result in the Fig. 5 is obtained for the case of asymmetric tension-compression
(Fig. 2).
Figure 6 shows the optimization result for the case of symmetric bending (Fig. 3).
202 S. Chirskiy and V. Panchenko
For the case of a complex asymmetric loading (Fig. 4), the results are obtained, as
shown in the Fig. 7.
Fig. 7. Comparison of optimization results for the case of a complex asymmetric loading
The forms obtained in the middle part of the connecting rod are almost identical.
The difference in shape in the areas of the upper and lower heads is due to the
difference in the algorithms. The “Topology Optimization” module, included in the
ANSYS software package, while optimizing, provides sufficient rigidity for all areas,
and since the cylindrical surface of the lower connecting rod head is fixed in all degrees
of freedom, then a minimum volume of material is sufficient to ensure its rigidity.
However, in the developed method, the criterion for the removal of material is
Verification of the Adequacy of the Topological Optimization Method 203
associated with the value of the equivalent stress, which in the sealing zone will be
significantly greater than zero. All this leads to the conclusion about the adequacy of
the developed method. The duration of solving the considered optimization tasks by the
developed method and using the “Topology Optimization” module is almost the same.
The main advantages of the developed method are the possibility of using movable
types of contacts and setting thermal loads on parts. Movable types of contacts are
necessary for an adequate description of cylindrical sliding bearings, such as a con-
necting rod bearing or a piston pin bearing. Heat loads are convenient, for example, to
simulate the pre-tightening of bolts or studs.
5 Conclusion
The proposed method allows searching for a new detail shape within a certain initial
model, as well as optimizing the shape of existing products. The program is written in
the integrated programming language APDL. The optimization goal function is the
minimum part mass. The limitations are related to the performance of the detail,
described by the strength or stiffness of the detail. The ability to install a new detail into
an existing structure is provided both at the stage of preparing the initial model and
when converting an optimized finite element model to a solid model. Thus, the pro-
posed method is quite universal and can be used for various tasks related to finding the
optimal form of details.
The optimized form of the detail often turns out quite complicated. Its production
by traditional methods is likely to be too complicated and expensive. Such details can
be produced using additive technologies, although in this case their cost remains high.
The duration of the optimization process is almost completely determined by the
time spent on modeling the stress-strain state of the detail. The processing of results
and switching on/off elements takes several minutes, even if the finite element model
contains a large number of elements.
The developed algorithm in the three test tasks showed the same results as the
“Topology Optimization” module of the ANSYS Workbench package. The observed
small difference is due to the fundamental difference between optimization algorithms.
At the same time, the developed algorithm has several advantages that are important in
optimizing the main details of piston internal combustion engines.
References
1. Shenoy, P., Fatemi, A.: Connecting Rod Optimization for Weight and Cost Reduction. SAE
Technical Paper 2005-01-0987 (2005). https://doi.org/10.4271/2005-01-0987
2. Charkha, P.G., Jaju, S.B.: Analysis & optimization of connecting rod. In: Second
International Conference on Emerging Trends in Engineering and Technology, ICETET
2009, pp. 86–91 (2009). https://doi.org/10.1109/ICETET.2009.30
3. Bin, Z., Lixia, J., Yongqi, L.: Finite element analysis and structural improvement of diesel
engine connecting rod. In: Second International Conference on Computer Modelling and
Simulation, pp. 175–178 (2010). https://doi.org/10.1109/ICCMS.2010.238
204 S. Chirskiy and V. Panchenko
4. Hou, X., Tian, C., Fang, D., Peng, F., Yan, F.: Sensitivity analysis and optimization for
connecting rod of LJ276M electronic gasoline engine. In: Computational Intelligence and
Software Engineering (2009). https://doi.org/10.1109/CISE.2009.5363219
5. Limarenko, A.M., Romanov, A.A., Aleksejenko, M.A.: Optimizaciya shatuna avtomo-
bil'nogo dvigatelya metodom konechnyh ehlementov (Optimization of the connecting rod of
a car motor by the method of final elements). In: Trudy Odesskogo Politekhnicheskogo
Universiteta (Proceedings of the Odessa Polytechnic University), no. 2(39), pp. 98–100
(2012). (in Russian)
6. Roos, D., Nelz, J., Grosche, A., Stoll, P.: Workflow-Konzepte zum benutzerfreundlichen,
robusten und sicheren Einsatz automatischer Optimierungsmethoden. In: 21th CAD-FEM
Users’ Meeting International Congress on FEM Technology (2003)
7. Bhandwale, R.B., Nath, N.K., Pimpale, S.S.: Design and analysis of connecting rod with
abaqus. Int. J. Recent Innovation Trends Comput. Commun. 4(4), 906–912 (2016)
8. Clarich, A., Carriglio, M., Bertulin, G., Pessl, G.: Connecting rod optimization integrating
modefrontier with FEMFAT. In: 6-th BETA CAE International Conference (2016). https://
www.beta-cae.com/events/c6pdf/12A_3_ESTECO.pdf
9. Zuo, Z.H., Xie, Y.M.: A simple and compact Python code for complex 3D topology
optimization. Adv. Eng. Softw. 85, 1 (2015). https://doi.org/10.1016/j.advengsoft.2015.02.
006
10. Moezi, S.A., Zakeri, E., Bazargan-Lari, Y., Zare, A.: 2&3-dimensional optimization of
connecting rod with Genetic and modified Cuckoo optimization algorithms. IJST Trans.
Mech. Eng. 39(M1), 39–49 (2015)
11. Shaari, M.S., Rahman, M.M., Noor, M.M., Kadirgama, K., Amirruddin, A.K.: Design of
connecting rod of internal combustion engine: a topology optimization approach. In:
National Conference in Mechanical Engineering Research and Postgraduate Studies,
pp. 155–166 (2010)
12. Fonseka, S.: Development of new structural optimization methodology for vehicle
crashworthiness. Honda R&D Tech. Rev. 22(2), 59–65 (2010)
13. Jia, D., Wu, K., Wu, S., Jia, Y., Liang, C.: The structural analysis and optimization of diesel
engine connecting rod. In: International Conference on Electronic & Mechanical Engineer-
ing and Information Technology, pp. 3289–3292 (2011). https://doi.org/10.1109/EMEIT.
2011.6023712
14. García, M.J., Boulanger, P., Henao, M.: Structural optimization of as-built parts using
reverse engineering and evolution strategies. Struct. Multidisc. Optim. 35, 541 (2008).
https://doi.org/10.1007/s00158-007-0122-6
15. Boehm, P., Pinkernell, D.: Topology optimization of main medium-speed diesel engine
parts. In: CIMAC Congress, Bergen (2010)
16. Ogata, Y., Suzuki, S., Iijima, Y.: Optimization method for reduction of transmission housing
weight. Honda R&D Tech. Rev. 16(2), 103–108 (2004)
17. Shedge, V.A., Munde, K.H.: Optimization of connecting rod on the basis of static & fatigue
analysis. IPASJ Int. J. Mech. Eng. (IIJME) 3(5), 7–13 (2015)
18. Myagkov, L.L., Chirskiy, S.P.: The implementation of the BESO method for topology
optimization in ANSYS APDL and its application for optimization of the connecting rod shape
of a locomotive diesel engine. In: Proceedings of Higher Educational Institutions. Machine
Building, no. 11, pp. 38–48 (2018). https://doi.org/10.18698/0536-1044-2018-11-38-48
19. Myagkov, L., Chirskiy, S., Panchenko, V., Kharchenko, V., Vasant, P.: Application of the
topological optimization method of a connecting rod forming by the BESO technique in
ANSYS APDL. In: Intelligent Computing & Optimization. Advances in Intelligent Systems
and Computing, vol. 1072, pp. 239–248 (2020). https://doi.org/10.1007/978-3-030-33585-
4_24
Method for Optimizing the Maintenance
Process of Complex Technical Systems
of the Railway Transport
Abstract. The article discusses one of the ways to improve the reliability of
technical systems by optimizing the maintenance process. The structure of the
engineering method for finding the optimal frequency of maintenance is pro-
posed. Practical recommendations have been developed to optimize the main-
tenance process of the “PALM” automatic identification system based on the
proposed method.
1 Introduction
Figure 1 shows the structural diagram of a method for optimizing the parameters of a
maintenance system. It is an algorithm for performing work in solving the problem of
optimizing the maintenance process of complex systems.
In blocks 1, 2, 5 the stages of work on compiling a meaningful description of the
operation of the technical system are displayed. Here, individual operational parameters
are highlighted and the corresponding functional units and elements are set.
The maintenance of any technical system or facility begins with the control of the
technical condition. The task of determining the technical condition of a technical
system is extremely important for the following reasons:
1. The determination of the technical condition of a technical system is the stage after
which it is concluded that it is advisable to carry out maintenance and select specific
maintenance operations;
2. If the technical system fails during its intended use, then the possible damage
associated with this failure greatly exceeds the costs associated with the
maintenance;
3. By itself, control of a technical system is always associated with time and material
costs, which are always desirable to minimize, but not to the detriment of the
sufficiency and reliability of information about the technical condition of the
facility.
These factors make the task of choosing the initial operations difficult, multifaceted
and responsible. The optimization task is to find such a number of initial maintenance
operations that would take a minimum of time, and the information received would
completely characterize the technical condition of the technical system under study.
The selection of the required number of initial maintenance operations can be
carried out in various ways. The simplest task is solved on the basis of graph theory
[2, 3]. This theory allows, based on the analysis of the functional diagram of a par-
ticular device, to select a set of determining parameters for monitoring the technical
condition and determine the sequence of their control. The choice of parameters for
monitoring the technical condition based on the analysis of the functional diagram of
the technical system as a whole presents considerable difficulty. Therefore, the selec-
tion of parameters for monitoring the technical condition is advisable to carry out for
individual functional devices.
Denoting the set of outputs of functional devices by X = {xi}, i = 1… n, and the set
of external inputs by U = {Uv}, v = 1… m, which are necessarily controlled, we find
some set of outputs k Є x, which is necessary monitor to assess the technical condition
of the functional device. We represent each output xi Є X as the vertex of some oriented
graph G (x, v), where x is the set of vertices, and v is the set of arcs.
The principle of constructing the graph is as follows: if the output xi is the input of a
block having the input xj then these two points forming the vertices of the graph xi and
xj are connected by an arc (xi, xj). According to the constructed oriented graph of
outputs, the smallest set of controlled parameters is determined, with the help of which
the technical state of the functional device is evaluated. Finding the minimum number
of monitored parameters reduces to finding the smallest external stable set in the output
graph. An external stable set is a certain set of vertices of a graph into which arcs from
all other vertices that do not belong to K. The parameters of outputs belonging to an
external stable set fully characterize the technical state of the functional device, and
their number will be minimal [4].
208 V. Apatsev et al.
The specified method allows selecting the minimum number of initial operations to
control the maintenance of a functional device, the implementation of which will
reliably judge the technical condition of the device.
In addition to the minimum number of initial maintenance operations, it is neces-
sary to determine the sequence of their implementation. The choice of the optimal
sequence is a rather complicated mathematical task, for the solution of which dynamic
programming methods are usually used [5, 6]. For practical purposes, it is convenient
to use the method when each previous maintenance operation would provide infor-
mation on the technical condition of the largest number of units or functional devices.
To determine the sequence of maintenance operations, some preference function and
preference rules are used. As a preference function, it can be used the ratio value:
li
FPREF ðli Þ ¼ ; ð1Þ
n
where li(i = 1… n) is the index of precedence of the output xi; n is the number of
blocks in the functional device. Obviously, the closer FPREF(li) is to unity, the greater
the number of blocks is controlled by the operation, and, therefore, the greater the
specific gravity of the control this operation has. Selected maintenance operations
should be carried out in accordance with the priority series, arranged in descending
order of the values of FPREF(li).
The developed method was applied to optimize the maintenance periods of the auto-
matic identification system “PALM”, which is the information basis of the automated
control systems for railway transport and freight traffic.
The automatic identification system is intended for automatic fixation of the rolling
stock following via pre-selected reference points along the route [5].
For this, the entire rolling stock is equipped with on-board encoders that carry
information about each moving object. Coded on-board sensors operate on the prin-
ciple of reflection of the irradiating signal of a reader with amplitude modulation by an
identification code. Wheel sensors record the moments when the wheelset passes a
given point on a path segment and sets the pace for the entire data reading process.
At reference points, the primary elements of automatic information collection are
established – reading points with irradiating and reading equipment, which automati-
cally reads information from the code-based sensors of the rolling stock. The com-
position of the reading point includes (Fig. 2):
– irradiating and reading equipment, consisting of a reader 1 and antenna 2;
– controller for counting axles of wheelsets 3;
– cold-resistant TGSA 4 modem with power supply 6;
– two devices for fixing wheelsets 7, 8;
– two-channel power supply system 9;
– reading point heater 5.
Method for Optimizing the Maintenance Process of Complex Technical Systems 209
The reading point equipment (Fig. 2) is a two-channel polling system for param-
eters of rolling stock passing by: a high-frequency polling channel of on-board enco-
ders installed on locomotives and rolling stock cars, consisting of antenna 2, a high-
frequency, low-frequency channel, reader 1 and its final formation unit information
from the sensor, the low-frequency channel for fixing the moments of passage of the
wheels of the rolling stock over the devices for fixing the wheelsets, which includes
two devices for fixing the wheelsets 7, 8, the axle counting controller 3 and the reader
assembly to form the final message about the moments of passage of the wheels.
Fig. 2. Functional diagram of the reading point of the automatic identification system “PALM”
The reading point emits radio signals only in those periods when the rail circuit of
the block section to which the emitter is “tied” is occupied by rolling stock. In the
absence of rolling stock, the reading point is in standby mode.
The received information is decrypted and transmitted with fixed time stamps via a
local communication line as primary information to a line level concentrator, which is
built on the basis of a personal electronic computer.
The line concentrator collects information from all reading points of the railway
junction, processes the received information and transmits the final message about the
rolling stock via the network of the “Russian Railways” Company to the road level
concentrator. It is at this level that the operational management of traffic flows is
ensured.
In a personal electronic computer of the hub of the automatic identification system,
a message is generated with a list of past rolling stock numbers, location, time,
direction of movement. By the signals from the electronic pedals, serial numbers of
each mobile unit are also recorded, which makes it possible to take into account rolling
stocks with a missing code on-board sensor. The message is transmitted to the road
computer center and to the main computer center.
In the automatic identification system “PALM”, an algorithm using temporary
redundancy is implemented to control the operability of the reading point equipment.
For this, a passive test identifier is introduced into the reader, which is located opposite
210 V. Apatsev et al.
the reader's antenna on the opposite side of the path. The reader periodically turns on
the radiation. In the absence of rolling stock in front of the reader, information is read
from the test identifier, after which a decision is made on the health of the reader. Since
the stationary identifier is part of the reader, in the absence of information during its
repeated reading, the maintenance personnel determines which elements are faulty.
After a decision is made on the health of the reader, the radiation is turned off and on in
the standard mode when the rolling stock approaches the reader. Thus, the operability
of the radiating and reading equipment is determined.
The automatic identification system also implements an algorithm for verifying the
operability of the equipment of a reading point by software polling of the state of
functional blocks as the rolling stock approaches. Such a check is started when the
rolling stock of the rail circuit of the block section is occupied, to which the emitter is
“tied”. In this case, a time reserve is also used, the value of which is determined from
the moment the train is occupied by the rolling stock to the next point through the
reading point.
Thus, the “PALM” automatic identification system meets the needs of the “Russian
Railways” Company for the necessary electronic control devices for rolling stock and
transported goods.
In accordance with the Fig. 1, to solve the problem of optimizing the maintenance
parameters, it is necessary to have information about the existing system and main-
tenance parameters, as well as information about the failures of the technical system.
Information on the maintenance schedule and inspection procedures for the
equipment of the “PALM” automatic identification system is contained in [7]. In [8],
time norms are provided for planning, organizing, and regulating the labor of electrical
engineers and electricians for servicing and repairing alarm, centralization and locking
devices engaged in maintenance of the equipment of a reading point of an automatic
identification system.
Statistical information about the failures of the automatic identification system
“PALM” was obtained from specialists of the “Industry Center for the Implementation
of New Equipment and Technologies” Company of the “Russian Railways” Company,
who were involved in the development of the automatic identification system and
monitoring the operation of the automatic identification system on the railway network.
According to [9], the initial data for calculating the necessary reliability indicators
should be generated using automated systems, where the main attention should be paid
not only to failure statistics, but generally to control the completeness and correctness
of the technological processes of operation and repair of technical equipment, as well as
the process of organizing transportation. Currently, in railway transport, such a system
is the “Integrated automated system for recording, monitoring the elimination of
technical equipment failures and analyzing their reliability”.
Within the framework of this system, the entire technological chain has been
implemented, starting from fixing the fact of failure to eliminating the cause and
assigning responsibility, forming materials for the investigation of the failure.
At present, the system under consideration contains data on technical equipment
failures from the automated system for maintaining the train schedule “Ural - All-
Russian Scientific Research Institute of Railway Transport”, a number of industrial
automated systems. The main source of primary data on technical equipment failures is
the above-mentioned system of the schedule of executed train traffic, which accounts
for about 70% of the data recorded in the “Integrated automated system for recording,
monitoring the elimination of technical equipment failures and analyzing their
reliability”.
The above-mentioned train execution schedule system is designed to control the
progress of the transportation process from automated workstations of the dispatching
and managing apparatus at all levels of operational management. In addition, the
information capabilities of the system are used by employees of other services and
departments. It includes the functions of forecasting, planning, control, regulation,
accounting and analysis.
The most important advantage of the “Integrated automated system for recording,
monitoring the elimination of technical equipment failures and analyzing their relia-
bility” is that the system allows receiving dynamically updated information about
technical equipment and equipment infrastructure failures directly.
212 V. Apatsev et al.
Guided by the above approach, on the basis of the analysis of functional diagrams of
the automatic identification system “PALM” and [10, 11], a set of maintenance
operations was selected. At the same time, the results of an expert survey of specialists
of the “Industry Center for the Implementation of New Equipment and Technologies”
Company on the operation of an automatic identification system were taken into
account. The method for the maintenance of the automatic identification system is
shown in the Table 1.
Table 1. (continued)
№ Work in progress Duration,
min
16 Cable rack 9
Checking the internal state of the cable rack
17 Cable glands 20
Checking the status of cable glands with opening
18 Ground loop 4
Checking the ground loop and ground circuits
19 Checking the correct connection of the grounding devices of the floor 2
readers cabinet
20 Checking and adjusting lightning protection devices 7
21 Checking the status of the visible elements of the grounding devices of 4,5
the floor readers cabinet
22 Selective digging and inspection of the elements of grounding devices 37
located in the ground
23 Checking the condition and serviceability of spark gaps 4
24 Checking the condition and serviceability of diode earthing switches of 14
floor reading devices
Based on the developed method, a study of the maintenance process of the auto-
matic identification system was conducted, the optimal values of the frequency of
maintenance of the main functional devices of the automatic identification system were
determined, which are summarized in the Table 2.
Table 2. Maintenance periods for the main functional devices of the automatic identification
system, used in practice before optimization and optimal
№ Functional device Maintenance period before Optimum maintenance
optimization, days period, days
1 Reading point 30 43
2 Floor readers cabinet 120 120
3 Reader 120 120
4 Sensors 7 10
5 Wheel axle counting 120 178
controller
6 Rail chain 30 22
7 Track transformer 30 24
box
8 Cable rack 180 192
9 Cable glands 180 278
10 Ground loop 180 200
11 System as a whole 360 450
214 V. Apatsev et al.
7 Conclusion
The proposed method for optimizing the parameters of a maintenance system may be
applicable to various systems in terms of structure, size and characteristics. The
modularity of the method and the convenience of its implementation on an electronic
computer allow considering the method as a possible element of an automated control
system for the operational reliability of train support systems and railway transport
infrastructure facilities.
Based on the proposed method, a way for automating the calculation of reliability
indicators and the frequency of technical maintenance of train support systems and
railway transport infrastructure facilities using statistical information obtained from the
“Integrated automated system for recording, monitoring the elimination of technical
equipment failures and analyzing their reliability” and economic automated systems of
the “Russian Railways” Company is proposed. An example of the application of the
method for optimizing the maintenance process of the automatic identification system
“PALM” is considered. Measures are proposed to adjust the existing maintenance
system in order to achieve a maximum availability factor.
References
1. Novikov, E.V.: Nauchno-metodicheskij podhod k razrabotke inzhenernoj metodiki rascheta
pokazatelej nadezhnosti [Scientific and methodological approach to the development of an
engineering methodology for calculating reliability indicators]. Sbornik nauchnyh trudov
kafedry “Transportnye ustanovki” [The collection of scientific works of the department
“Transport units”], T. 1, 111–116 (2008). (in Russian)
2. Rainshke, K., Ushakov, I.A.: Ocenka nadezhnosti sistem s ispol'zovaniem grafov
[Reliability assessment of systems using graphs]. Moskva: Radio i svyaz' [Moscow: Radio
and communications], 208 p. (1988). (in Russian)
Method for Optimizing the Maintenance Process of Complex Technical Systems 215
3. Herzbach, I.B., Kordonsky, H.B.: Modeli otkazov [Failure Models]. Moskva, Soviet Radio
[Moscow, Sovetskoe radio], 168 p. (1966). (in Russian)
4. Latinskij, S.M., SHarapov, V.I.: Teoriya i praktika ekspluatacii radiolokacionnyh sistem
[Theory and practice of the operation of radar systems]. Moskva, Sovetskoe radio [Moscow,
Soviet Radio], 432 p. (1970). (in Russian)
5. Beichelt, F., Franken, N.: Nadezhnost' i tekhnicheskoe obsluzhivanie. Matematicheskij
podhod [Reliability and maintenance. The mathematical approach]. Moskva, Radio i svyaz'
[Moscow, Radio and communications], 392 p. (1988). (in Russian)
6. Bezrodniy, B.F., Gorelik, A.V., Nevarov, P.A., Shaliagin, A.V.: Principy upravleniya
nadezhnost'yu sistem zheleznodorozhnoj avtomatiki i telemekhaniki [The principles of
reliability management of railway automation and telemechanics systems]. Avtomatika,
svyaz', informatika [Automation, communications, informatics], № 7, 13–14 (2008). (in
Russian)
7. Legkij, N.M., Kozlov, V.I.: Postroenie punktov schityvaniya SAI “Pal'ma” [Construction of
reading points of the AIS “Palm”]. Opyt proektirovaniya, vnedreniya i ekspluatacii
[Experience in design, implementation and operation], 55–67 (2007). (in Russian)
8. Punkt schityvaniya sistemy avtomaticheskoj identifikacii “Pal'ma” [Reading point of the
automatic identification system “Palm”]. Rukovodstvo po ekspluatacii. Moskva: “Otraslevoj
centr vnedreniya novoj tekhniki i tekhnologij” [Operation manual. Moscow, “Industry
Center for the Implementation of New Equipment and Technologies”] (2002). https://pandia.
ru/text/77/357/77520.php. (in Russian).
9. Funkcional'naya strategiya obespecheniya garantirovannoj nadyozhnosti i bezopasnosti
perevozochnogo processa [Functional strategy to ensure guaranteed reliability and safety of
the transportation process]. Moskva, “Rossijskie zheleznye dorogi” [Moscow, “Russian
Railways”] (2007). https://annrep.rzd.ru/reports/public/ru?STRUCTURE_ID=4422. (in
Russian)
10. Vremennye otraslevye normy vremeni i normativy chislennosti na tekhnicheskoe
obsluzhivanie punktov schityvaniya sistemy avtomaticheskoj identifikacii podvizhnogo
sostava “Pal'ma” [Temporary industry time standards and personnel standards for
maintenance of reading points of the automatic identification system of rolling stock
“PALM”]. Moskva, “Centr organizacii truda i proektirovaniya ekonomicheskih normativov”
“Rossijskih zheleznyh dorog” [Moscow: “Center for the organization of labor and designing
economic standards” of “Russian Railways”] (2004). https://www.railway.kanaries.ru/index.
php?act=attach&type=post&id=8065. (in Russian)
11. Instrukciya po tekhnicheskomu obsluzhivaniyu i remontu ustrojstv elektrosnabzheniya
signalizacii, centralizacii, blokirovki i svyazi na federal'nom zheleznodorozhnom transporte
[Instructions for the maintenance and repair of power supply devices for signaling,
centralization, blocking and communication on federal railway transport]. Departament
elektrifikacii i elektrosnabzheniya Ministerstva putej soobshcheniya Rossijskoj Federacii,
Moskva, Transizdat [Department of Electrification and Power Supply of the Ministry of
Railways of the Russian Federation, Moscow, Transizdat] (2002). https://files.stroyinf.ru/
Data2/1/4293764/4293764042.htm (in Russian)
Optimization of Power Supply System
of Agricultural Enterprise with Solar
Distributed Generation
Abstract. Today, the activities of most enterprises are partially, and in some
organizations, fully automated, that is why power outages lead to the inter-
ruption of all technological processes, financial damage to enterprises and
various problems in the communal sector during power outages. One of the
options for ensuring uninterrupted energy supply, improving the quality and
reliability of power supply to consumers in rural and remote areas in cases of
accidents in distribution networks can be considered the use of renewable
energy sources. The purpose of the research is to optimize the configuration of
the power supply system of agricultural enterprise by using solar power plants.
The object of the research is the power supply system of the production facility
of the agro-industrial complex. There was conducted calculation of the number
of photovoltaic modules that can be placed on the roofs of agricultural enterprise
to achieve maximum electrical energy generation during the day. On the basis of
experimentally obtained data of currents and voltages, the generated power of
photovoltaic modules was determined. The ratio of the maximum generated
power of the south-oriented module located at tilt angle of 15° to the maximum
generated power of the south-oriented module located at the optimal tilt angle of
39° was also determined. The average value of the ratio of the maximum
generated power of the south-oriented PV module located at tilt angle of 15° to
the maximum generated power of the south-oriented panel located at optimal tilt
angle of 39° was calculated. The total daily production of electrical energy by all
solar power plants is 885.22 kWh per day for a given date (Nov. 1). The ratio of
daily electricity generated by solar power plant to that consumed by production
facility in the specially selected city of Russia (Zernograd) is 71.8%.
1 Introduction
The main distribution networks in rural areas are overhead power transmission lines
with the voltage of 0.38 kV, which today have high degree of technical and economic
wear and tear. The annual growth in electricity consumption is observed in certain
sectors of industry, in agriculture, in the household sector, where the overwhelming
majorities are consumers of electricity with nonlinear and non-sinusoidal nature of
loads [1, 2]. In addition, the quality of electricity and the reliability of electricity supply
to rural consumers, and especially to remote ones, remain at low level. Until now, there
are territories on the planet and in Russia where there are no traditional power supply
systems even with low efficiency [3]. The reason for this state of affairs can be a
number of economic and technical difficulties faced by both consumers and utilities.
Today, the activities of most enterprises are partially, and in some organizations,
fully automated, therefore power outages lead to the fact that all technological pro-
cesses are interrupted, financial damage to enterprises is inflicted, and various problems
arise in the utilities sector during interruptions in the supply of electricity.
Besides traditional energy sector has very strong impact on the ecological situation
on Earth and is increasingly accused of global warming and climate change.
One of the options for uninterrupted energy supply, improving the quality and
reliability of power supply to consumers in rural and remote areas is considered to be
the use of renewable energy sources [4], united in small microgrid [5]. The use of solar
energy for the operation of solar power plants (SPP) and photovoltaic installations
(PI) is profitable, environmentally friendly and not so expensive source of primary
energy. Today solar power plants are used not only abroad, but also in Russia. Pho-
tovoltaic modules (PM), assembled on the basis of solar cells of various designs with
high characteristics [6], are now actively used in industry and in number of economy
sectors [7].
Besides a lot of modern approaches to computing and optimization applied by
growing number of scientists, pedagogues, advisors, deciders and practitioners now
exists [8]. Actually, very strong and smart (“intelligent”) computational techniques
arose for handling a very large number of real-world challenges and crises, in partic-
ular, in relation with the main areas of optimization in theory and practical use [9].
Purpose of Research
The purpose of the research is to optimize the configuration of the power supply system
of agricultural enterprise by means of solar power plants to provide additional tech-
nological capacities.
To carry out the analysis and substantiate the made decisions, there were use traditional
general methods of assessing the value of the solar potential in particular territory
[10, 11]; the methods of design, calculation and selection of the configuration of the
internal power supply network of production facilities adopted in the electric power
218 Yu. V. Daus et al.
industry; algorithm for calculating power consumption and methods for assessing the
power generated by photovoltaic installations [12].
To solve the set research problem, it is necessary to choose location and position
and determine the parameters of a photovoltaic installation, at which the consumption
of electrical energy from the network will be minimal. The theoretical basis for such a
search is the mathematical formulation of the optimization problem using the criterion-
the maximum generation of electrical energy by photovoltaic power plants:
X
m X
m
Wi ¼ Rbci gi Si Ni ! max
i¼1 i¼1
where Wi – electric power generation by the i-th generation source; m is the number
of distributed generation sources (photovoltaic power plants), pcs.; Rbci – the intensity
of solar radiation on the receiving surface, oriented at tilt angle bi relatively to the
horizon and ci relatively to the cardinal points, kWh/m2; Si, ηi – area and efficiency of
photovoltaic modules used for distributed generation source, m2 and relative units
respectively; Ni – number of photovoltaic modules at the power plant, pcs.
Since the purpose of the research is to increase the efficiency of the power supply
system of agricultural facility, this imposes number of restrictions on the choice of
parameters and composition of photovoltaic power plants - sources of distributed
generation. Agricultural object is area with dense buildings or extensive areas set aside
for agricultural activities. Therefore, the construction of photovoltaic installations on
the ground can cause number of difficulties, namely, shading of modules from nearby
structures and buildings, the impossibility of removing the land from processing,
blocking up territories for the passage of equipment, etc. It is advisable to place
generating units on the structures of existing buildings and constructions, although it
imposes certain restrictions on their parameters, it makes their construction and
maintenance more convenient [13].
Therefore, when modeling, the following restrictions must be taken into account:
1) dimensions of the roof where the photovoltaic installation is supposed to be placed:
a b, m;
2) spatial orientation of the photovoltaic installation location: slope angle bi, ° and
orientation to the cardinal points ci, °;
3) coordinates of the location of the photovoltaic power plant, u, °N, k, °E.
The proposed approach is implemented in the “Search for solutions” add-on of the
Microsoft EXCEL program. The result of solving the optimization problem is the
location, parameters of the sources of solar distributed generation, at which the con-
sumption of electrical energy from the network will be minimal.
Optimization of Power Supply System of Agricultural Enterprise 219
The analyzed production facility consists of oil mill, weighing tower, drying and
purification tower, intake pit, machine building, hydraulic fracturing unit, laboratory, fuel
and lubricants station, garage, warehouses and utility rooms. The facility receives power
from a 10/0.4 kV transformation substation and has its own 0.38 kV electrical network.
Using the generally accepted methods [12], it is calculated for the power consumption of
the electrical equipment of this agricultural enterprise, which in the daytime is 136.1 kW,
and the total consumed daily electricity is 1232.5 kWh/daily.¼ 1232; 5kWh.
Based on the data obtained from the energy survey, daily graph of the power
consumption of the studied production facility in the summer-autumn period was
conducted (Fig. 1).
Fig. 1. The daily schedule of the power consumption of the production facility in the summer-
autumn period: 1 - lighting, 2 - street lighting, 3 - power equipment, 4 - household equipment, 5 -
office equipment
Since the production facility has seasonal load, which has maximum value in July
and August, as well as from September 15 to October 1 and from October 15 to
November 1, during these periods of the year the total consumed daily electricity is
1232.5 kWh, and during the rest of the year it decreases to 180–230 kWh. To achieve
maximum generation of electrical energy during the day, it is necessary to calculate the
number of photovoltaic modules that can be located on the roofs of pre-selected
constructions (Table 1).
220 Yu. V. Daus et al.
The number of PMs located on the roofs of industrial facilities is limited by their
geometric dimensions. The configuration of a rooftop solar power plant was designed
using photovoltaic monocrystalline solar modules of the JAM72-365/PR brand from
HEVEL, with rated power of 365 W, rated voltage of 39.2 V, efficiency of 18.8%, and
dimensions of 1960 991 40 mm [13].
The optimal inclination angle of the receiving surface relatively to the horizon, at
which the module’s generation of electrical energy is maximum per year, is 39° for the
selected city of Zernograd [14, 15].
The amount of total solar radiation arriving at the receiving area, inclined at the
optimal angle (39°), on November 1 will be determined using the program [16], while
the insolation is 2; 82kWh=m2 .
However, not all roofs are optimally oriented in space. Some of them have tilt
angles of 15° and 30° with south and east orientation. As a result, to determine the ratio
of the solar insolation amount at south-oriented receiving surface located at tilt angle of
15° to the receiving surface located at optimal angle, series of experiments was carried
out, including the recording of current-voltage characteristics (I–V characteristics) of
three PMs during the day which are oriented at south at inclination angles of 15°, 30°
and 39°. Based on the experimentally obtained data of currents and voltages, the
generated power of the PV module was calculated. By calculation, the ratio of the
maximum generated power of south-oriented PV array located at tilt angle of 15° to the
maximum generated power of south-oriented PV array located at tilt angle of 39° was
also determined. Also, the average value of the ratio of the maximum generated power
of the south-oriented PM located at tilt angle of 15° to the maximum generated power
of the south-oriented PM located at tilt angle of 39° was calculated, and equals 77%.
To determine the percentage of solar insolation for the eastern orientation of the
receiving surface located at tilt angle of 39° to the southern orientation of the receiving
surface located at tilt angle of 39°, series of additional experiments was performed,
including taking the I–V characteristics of three PMs oriented to the east, south and
west at optimal inclination angle of 39°. Based on the obtained data, the calculation of
the maximum generated power of the PMs, oriented to the east, south and west at
optimal tilt angle of 39°, was made. The average value of this ratio was 89%.
The working day at this production facility starts at 8:00 (Fig. 1). Due to the
number of features of annual technological processes, the maximum generated power
of the projected solar power plant is required by consumers until November 1 of the
calendar year. Thus, it is necessary to determine the minimum distance between the
lower points of the PM in adjacent rows for the time t = 8 h on November 1 at the
optimal inclination angle of the receiving area for the geographical point 46.8 °N And
40.3 °E (Zernograd)
Thus, one PM located at tilt angle of 39° per day on November 1 generates
0.772 kWh. There were determined the daily amount of electrical energy generated by
the south-oriented photovoltaic module on November 1, at tilt angle of 15°, supple-
menting with factor of 0.77, which takes into account the angular orientation of the PM
15°/39°. One south-facing PM located at tilt angle of 15° generates 0.594 kWh. The
calculation results are summarized in Table 1.
Optimization of Power Supply System of Agricultural Enterprise 221
There was determined the daily amount of electrical energy generated by the east-
oriented PM on November 1, at tilt angle of 15°, supplementing with a factor of 0.77,
which takes into account the angular orientation of the PM 15°/39°, and factor of 0.89,
which takes into account the spatial orientation of the PM East/South. One east-facing
PM located at tilt angle of 15° generates 0.529 kWh. The calculation results are
summarized in Table 1.
Table 1. Daily generation of PM located on the roofs of the agricultural enterprise’s facilities
Building Tilt angle of Orientation Size, a b, m Number Daily
the roof bi, ° ci, ° of PM, generation by
pcs. SPP, kWh/day
Warehouses 15 0 11 85 425 252.45
No. 3–5
Administrative 15 −90 12 22 132 69.82
building
Warehouses 0 0 84 24 + 57 13 393 303.39
No. 1–2
Elevator 15 −90 11.8 33 160 84.64
Warehouses 15 −90 10.5 54 270 142.83
No. 6–7
Oil mill 15 0 7.9 18 54 32.08
The total daily production of electricity by all solar power plants is 885.22 kWh/day
for November 1. The ratio of the daily electricity generated by the solar power plant to
the consumed one by production facility in Zernograd is 71.8%.
4 Conclusion
The use of renewable energy sources can be considered one of the options for unin-
terrupted energy supply, improving the quality and reliability of power supply to
consumers in rural and remote areas. The use of solar energy for the operation of solar
power plants (SPP) and photovoltaic installations (PI) is profitable, environmentally
friendly and not so expensive source of primary energy. Today solar power plants are
used not only abroad, but also in Russia. Photovoltaic modules (PMs) are now actively
used in industry and in number of sectors of the economy [17].
Conducted studies have revealed that on day with minimal solar radiation, solar
power plants located at differently oriented roof slopes with different tilt angles of roof
slopes of seasonal industrial agricultural facilities is able to generate 71.8% of the
required daily electricity consumption and significantly reduce the consumption of
electricity from the centralized power supply system.
222 Yu. V. Daus et al.
References
1. Hirsch, A., Parag, Y., Guerrero, J.: Microgrids: a review of technologies, key drivers, and
outstanding issues. Renew. Sustain. Energy Rev. 90, 402–411 (2018)
2. Molyneaux, L., Wagner, L., Foster, J.: Photovoltaic/battery system sizing for rural
electrification in Bolivia: considering the suppressed demand effect. Appl. Energy 235,
519–528 (2019)
3. Pode, R., Pode, G., Diouf, B.: Solution to sustainable rural electrification in Myanmar.
Renew. Sustain. Energy Rev. 59, 107–118 (2016)
4. Yudaev, I., Daus, Yu., Zharkov, A., Zharkov, V.: Private solar power plants of ukraine of
small capacity: features of exploitation and operating experience. Appl. Solar Energy 1(56),
54–62 (2020)
5. Adomavicius, V., Kharchenko, V., Valickas, J., Gusarov, V.: RES-based microgrids for
environmentally friendly energy supply in agriculture. In: Proceedings of 5th International
Conference TAE 2013, Trends in Agricultural Engineering, 2–3 September 2013, Prague,
Czech Republic, pp. 51–55 (2013)
6. Kharchenko, V.V., Nikitin, B.A., Tikhonov, P.V.: Estimation and forecasting of PV cells
and modules parameters on the basis of the analysis of interaction of a sunlight with a solar
cell material. In: Proceedings of 4th International Conference TAE 2010, Trends in
Agricultural Engineering, Prague, Czech Republic, 7–10 September 2010, pp. 307–310
(2010)
7. Daus, Yu., Yudaev, I., Taranov, M., Voronin, S., Gazalov, V.: Reducing the costs for
consumed electricity through the solar energy utilization. Int. J. Energy Econ. Policy 9(2),
19–23 (2019)
8. Intelligent computing & optimization. In: Conference Proceedings, ICO 2018. Springer,
Cham. ISBN 978-3-030-00978-6
9. Intelligent computing and optimization. In: Proceedings of the 2nd International Conference
on Intelligent Computing and Optimization 2019 (ICO 2019). Springer. ISBN 978-3-030-
33585-4
10. Daus, Yu., Yudaev, I.: Estimation of solar energy potential under conditions of urban
development. In: Actual Issues of Mechanical Engineering (AIME 2017). Advances in
Engineering Research, vol. 133, pp. 156–161. Atlantis-Press, Paris (2017)
11. Liu, B., Jordan, R.: Daily insolation on surfaces tilted towards the equator. ASHRAE J. 3
(10), 53–60 (1961)
12. Komenda, T., Komenda, N.: Morphometrical analysis of daily load graphs. Int. J. Electr.
Power Energy Syst. 1(42), 721–727 (2012)
13. Belenov, A., Daus, Y., Rakitov, S., Yudaev, I., Kharchenko, V.: The experience of operation
of the solar power plant on the roof of the administrative building in the town of Kamyshin,
Volgograd oblast. Appl. Solar Energy (English translation of Geliotekhnika) 52(2), 105–108
(2016). https://doi.org/10.3103/S0003701X16020092
14. Hevel Homepage. http://www.hevelsolar.com. Accessed 15 01 2020
Optimization of Power Supply System of Agricultural Enterprise 223
15. Daus, Yu., Pavlov, K., Yudaev, I., Dyachenko, V.: Increasing solar radiation flux on the
surface of flat-oplate solar power plants in Kamchatka krai conditions. Appl. Solar Energy 2
(55), 101–105 (2019)
16. Daus, Yu., Kharchenko, V., Yudaev, I.: Managing spatial orientation of photovoltaic module
to obtain the maximum of electric power generation at preset point of time. Appl. Solar
Energy 6(54), 400–405 (2018)
17. Daus, Yu., Yudaev, I.: Designing of software for determining optimal tilt angle of
photovoltaic panels. In: 2016 International Conference on Education, Management and
Applied Social Science (EMASS 2016), pp. 306–309. DEStech Publications Inc., Lancaster
(2016)
18. Harder, E., Gibson, J.: The costs and benefits of large–scale solar photovoltaic power
production in Abu Dhabi, United Arab Emirates. Renew. Energy 2(36), 789–796 (2011)
19. Kharchenko, V., Panchenko, V., Tikhonov, P.V., Vasant, P.: Cogenerative PV thermal
modules of different design for autonomous heat and electricity supply. In: Handbook of
Research on Renewable Energy and Electric Resources for Sustainable Rural Development,
pp. 86–120 (2018). https://doi.org/10.4018/978-1-5225-3867-7.ch004
Crack Detection of Iron and Steel Bar Using
Natural Frequencies: A CFD Approach
Abstract. Solid composite materials such as steel, fiber plastics, etc. are now
particularly widely used in industry and civil construction, where a high
strength-to-weight ratio is required. Its dynamic behavior is influenced by
cracks, body deflection, or other defects in a structural element and changes its
stiffness. Therefore, crack identification and observation of load distribution and
deflection in the solid components of health monitoring systems is important.
The constant study provides a vibration-based method for detecting the location
of open or hidden cracks as well as a tool for comparing the best material
between iron and steel. In this analysis, a numerical simulation using the Finite
Element Method (FEM) has been performed using the COMSOL Multiphysics
software to optimize the frequency and strength of iron and steel bar along with
two semi-circular surface cracks. It is found that frequencies are proportional to
the increase in load and maximum frequency (around 2184.2 Hz for iron and
6197.7 Hz for steel) reserved at the cracked stage. Finally, we get that deflection
of steel bodies is higher than that of iron and steel is capable of tolerating so
much load as iron is deformed without actually breaking through cracked or un-
cracked bodies. From this study, very small sizes crack (minimum 0.05 mm)
can be defined in any structural body and the most acceptable solid materials can
be chosen for the particular structure.
1 Introduction
The monitoring of structural health and the identification of structural damage are
common areas of study. Infrastructural developments in civil structures and consid-
erable building materials have rapidly evolved around the world. Damage can be
incurred at various stages of the service life of the structure. The natural frequencies
and mode shapes of the structure provide details on the location and the size of the
damage [1]. The unexpected occurrence of damage can lead to catastrophic failure and
therefore poses a potential threat to human life [2]. Cracks are an important measure of
the safety status of infrastructure. Cracking is one of the most common damages that
occurs during the service life of engineering structures and, in some cases, causes a
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 224–236, 2021.
https://doi.org/10.1007/978-3-030-68154-8_23
Crack Detection of Iron and Steel Bar Using Natural Frequencies 225
major reduction in their load-bearing capacity and can even lead to failure. Crack
detection and study of cracked structures are therefore critical and fascinating topics for
researchers. The presence of cracks in the material structures can be detected by using
variations in local stiffness, which have a major impact on the natural frequency and
shape of the mode.
To avoid the failure caused by cracks, several researchers have performed broad
investigations over the few decades to develop structural behavior observance tech-
niques. An excellent methodology was proposed by D. P. Patil and S. K. Maiti [3] to
find multiple cracks within the beam victimization frequency measuring. Their results
offer a linear relationship between injury parameters and the natural frequency of
vibration of beam. A. K. Darpe et al. [4] studied dynamics of a bowed rotor with a
transversal surface crack. They finished that amplitude and directional nature of upper
harmonic elements of bowed rotor remain unmoved, but rotating frequency element
changes in magnitude. In another analysis Athanasios C. Chasalevris and Chris A.
Papadopoulos [5] studied the identification of multiple cracks in beams below bending.
They formulate a compliance matrix of two degrees of freedom as an operator of each
crack depth and angle of rotation of the shaft. Their declared methodology provides not
solely depth and size of the crack however conjointly spatial relation of the crack.
Ashish K. Darpe [6] proposes a completely unique process to find the transverse
surface crack in a very rod. He studied the behavior of the merely supported shaft with
transverse crack subjected to each bending and torsional vibration. Prabhakar [7]
applied vibration analysis of a cantilever beam with two open transversal cracks, to
check the response characteristics. In an initial part native compliance matrices of
various degrees of freedom are accustomed model transversal cracks in the beam on the
offered expression of stress intensity factors. Lee et al. [8] studied the Leg Mating Unit
(LMU), which is an element that supports the elastomeric bearing and steel plate
topside of the offshore plant. The hybrid metamodel-based optimization technique and
the satisfying trade-off approach were applied in their research. S.K. Georgantzinos, N.
K. Anifantis [9] bestowed the study of the respiration mechanism of a crack in a solid
rod. He studied the behavior of the transversal crack in cantilever shaft beam with two
completely different cases of straight and semicircular front of the shaft. He concludes
that the respiration behavior depends on depth and form of the crack front. Regenhardt
et al. [10] uses systematic reliability tools to obtain details about how the failure of
multiple constituent members of offsh platforms contributes to overall system failure.
In this way, systemic and time-dependent information can be segregated, allowing
reliability to be computed flexibly and computationally efficient.
The main objective of our study is to build a finite element analysis based on a non-
cracked and cracked solid cylindrical shaped iron and steel bar consisting of two half-
circular, open, transverse cracks and to optimize the relationship between modal natural
frequencies with different crack locations subject to free vibration. The aim of this
paper is to show a way to predict crack parameters (crack location) in the association in
nursing iron and steel rod from changes in natural frequencies using simulation and to
find out the most important material for construction.
226 R. Karmaker and U. K. Deb
2 Mathematical Modeling
Over the last few decades, most techniques are focused on vibration measurement and
analysis because vibration-based approaches can provide an efficient and easy way to
detect fatigue cracks in structures [11]. In this constant analysis, we used a vibration-
based approach to identify cracks in our domain. This approach is mainly focused on
shifts in dynamic characteristics, such as natural frequency and crack location
parameters [12]. The purpose of this analysis is to compare the behavior and frequency
of shape curvatures with the existence and absence of cracks in iron and steel structures
subject to free vibration. We will also observe the effects of cracks on the natural
frequency and deflection behavior of the bar. In the current work, the natural frequency
of the domain with and without crack was investigated by FEM.
d2 uðxÞ
EA þ qðxÞ ¼ 0 ð1Þ
dx2
where,
q(x) = cx = Distributed load [c = constant]
A = Cross sectional area
E = Young’s Modulus of elasticity
u(x) = Displacement vector
ðxÞ
Subject to the condition: uð0Þ ¼ 0 and EA dudx
x¼L
The strain energy release rate at the cracked section Irwin et al. [13] is,
1
Es ¼ ðKI1 þ KI2 Þ2 : ð2Þ
E_
where, the stress intensity factors are KI1, KI2 of mode I (opening of the crack) under
load P1 and P2 respectively.
Defining the flexibility influence co-efficient Cij per unit depth,
where,
Uc = strain energy
JC ¼ dU
dh = strain energy release rate.
c
Crack Detection of Iron and Steel Bar Using Natural Frequencies 227
Using the value of the strain energy release rate (JC) we get,
B d2 hZ1
Cij ¼ ðKI1 þ KI2 Þ2 dh ð4Þ
E_ dPi d Pj 0
The local stiffness matrix can be obtained by taking the inverse of the compliance matrix,
1
K11 K12 C11 C12
½K ¼ ¼ ð5Þ
K21 K22 C21 C22
All types of changes at the corresponding cracked points can be measured by Eqs. (6)
and (7).
Where [Q] is the coefficient matrix defined in terms of the cracked beam parameters
and A is the constants to be determined by the boundary conditions.
nu¼0 ð9Þ
A solid cylindrical shaped iron and steel bar is considered as computational domain with
Length = 0.36 m and Radius = 0.015 m are shown in Fig. 1. We construct the suitable
mesh design of the computational domain for with and without cracked iron and steel
bar whose are generated by COMSOL Multiphysics software is shown in Fig. 2. The
mesh properties of the computational domain are shown in Table 1 and Table 2.
In this study, we have investigated the frequency and load distribution of Iron and Steel
bar containing half-circular double crack using the finite element method. For our
simulation, we construct a solid cylindrical Iron and Steel bar and have used different
parameter values according to the Table 3 and the Table 4.
In Fig. 3 it is seen that a different amount of load has been applied vertically to the
crown edge of the bar. The deflection of the body after applying the load and the
simulation provided in the Fig. 4.
230 R. Karmaker and U. K. Deb
(a) Position of the first crack = 0.01m and (b) Position of the first crack = 0.01m and
the second crack = 0.10 m of the Iron bar the second crack = 0.10 m of the Steel bar
(c) Position of the first crack = 0.10m and (d) Position of the first crack = 0.10m and
the second crack = 0.20 m of the Iron bar the second crack = 0.20 m of the Steel bar
(e) Position of the first crack = 0.20m and (f) Position of the first crack = 0.20m and
the second crack = 0.30 m of the Iron bar the second crack = 0.30 m of the Steel bar
Fig. 5. The phase of the computational domain for various crack positions
232 R. Karmaker and U. K. Deb
In Fig. 5, we noticed that the applied load distributed across the domain of com-
putation. The full load for both metals is absorbed in the cracked position as shown in
Fig. 5 [(a)–(f)]. By comparing to our legend, we see the highest value found in the
cracked stage, however in un-cracked iron and steel bars, shown in Fig. 5 [(g)–(h)] that
the load ingested flows continuously to the end of the body and the steel is deviate
more than iron.
Figure 6 shows the absorption of loads at any cracked domain level. We observe
that the steel body transfers the load to the bottom of the cracked point which is
maximum and is continuously spread to the domain boundary, which is the main
deflection factor. The iron body still performs the same but iron does not spread the
load as much as steel to the entire body.
Based on Fig. 7 [(g)–(h)] we found that, for the un-cracked steel and iron bar, the
frequency remains almost the same. But from Fig. 7 [(a)–(f)] for both bar we found a
surprising shift for the different crack positions. The frequency graph reaches the
highest point of the Iron body at the cracked point, and we get two peak points at the
corresponding cracked spot. We see a remarkable contrast between the iron line graph
of the highest point and its nearest point. The graph even reaches the steel bar at the
respective cracked point at the center, but the line graph is almost regular. This also
implies the highest frequency spread to its nearest location. Moreover, there is a large
difference between different points in the iron line graph for irregular load distribution.
In previous figures, we find that the load distributed across the whole body in the steel
structure, where the iron body is not. In this figure, this concept fits completely with our
previous description.
Crack Detection of Iron and Steel Bar Using Natural Frequencies 233
(a) Position of the first crack at 0.01m and the (b) Position of the first crack at 0.01m and the
second crack at 0.10 m of Iron bar second crack at 0.10 m of Steel bar
(c) Position of the first crack at 0.10m and the (d) Position of the first crack at 0.10m and the
second crack at 0.20 m of Iron bar second crack at 0.20 m of Steel bar
(e) Position of the first crack at 0.20m and the (f) Position of the first crack at 0.20m and the
second crack at 0.30 m of Iron bar second crack at 0.30 m of Steel bar
3.00E+05
2.50E+05
Deflection (mm)
2.00E+05
1.50E+05
1.00E+05
5.00E+04
Iron Steel
0.00E+00
0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16
Poisition (m)
8.0E+07
6.0E+07
4.0E+07
2.0E+07
0.0E+00
0 0.1 0.2 0.3
Poisition (m)
According to Fig. 8, we found that two deflection graphs of both studies have
almost the same pattern for un-cracked steel and iron bars. But the form of Fig. 9
reveals the steel deflection graph is higher than the body of iron at the cracked points.
Finally, significant differences in deflection are found in the steel and iron body due to
the presence of the crack.
Crack Detection of Iron and Steel Bar Using Natural Frequencies 235
5 Conclusion
This paper deals with the optimization of frequency sensitivity of the solid cylindrical
iron and steel bar consists of half-circular double crack which varies with the change in
position of the crack for all vibration modes. In this analysis, in COMSOL Multi-
Physics software the Finite Element Analysis of the Iron and Steel bar with two
transverse cracks was performed for simulation. In our study, we also found that
applied load spread evenly across the steel bar and deformed the structure without
breaking. It is seen that the structure of steel is capable of absorbing various amounts of
load, causing vibration, and eventually deflected. Steel retains its elasticity by
remaining un-cracked in the body. On the other hand, Iron stores the load at any
defined section, such that the force applied cannot move through the entire body. That’s
why an amount of load creates too much vibration and the frequency at that portion or
affected region of the body, which creates a chance of damaging the body. It can be
said that due to the durability, good elasticity, and deflection facilities steel structure
can be considered as an acceptable and lot of property metal for each cracked and un-
cracked state of affairs than others. Finally, we can reach the conclusion that in the
construction sector steel has become a very popular and reliable structural material day
by day.
6 Future Study
In the present study, the domain under consideration have a uniform cross-section but
this method can be extended to components with varying cross section, different
geometry, and any boundary condition. The proposed method can be extended for
defect diagnosis in beams, shafts, or rotating machine elements.
Acknowledgement. The authors gratefully acknowledge the technical supports to The Centre of
excellence in Mathematics, Department of Mathematics, Mahidol University, Bangkok,
Thailand.
References
1. Khan, I.A., Yadao, A., Parhi, D.R.: Fault diagnosis of cracked cantilever composite beam by
vibration measurement and RBFNN. J. Mech. Des. Vib. 1, 1–4 (2013)
2. Ramanamurthy E.V.V., Chandrasekaran K.: Vibration analysis on a composite beam to
identify damage and damage severity using finite element method. Int. J. Eng. Sci. Technol.
(IJEST) 3, 0975–5462 (2011)
3. Patil, D.P., Maiti, S.K.: Detection of multiple cracks using frequency measurements. Eng.
Fract. Mech. 70, 1553–1572 (2002)
4. Darpe, A.K., Gupta, K., Chawla, A.: Dynamics of a bowed rotor with a transverse surface
crack. J. Sound Vib. 296, 888–907 (2006)
5. Chasalevris, A.C., Papadopoulos, C.A.: Identification of multiple cracks in beams under
bending. Mech. Syst. Signal Process. 20, 1631–1673 (2006)
236 R. Karmaker and U. K. Deb
6. Darpe, A.K.: A novel way to detect transverse surface crack in a rotating shaft. J. Sound Vib.
305, 151–171 (2007)
7. Prabhakar M.S.: Vibration analysis of cracked beam. Thesis of M school, National Institute
of Technology (2009)
8. Lee, S.H., Jeong, K-II, Lee, K.H.: Structural design of an LMU using approximate model
and satisficing trade-off method. In: 2nd International Conference on Intelligent Computing
and Optimization (ICO 2019), pp. 118–127. Springer (2019)
9. Georgantzinos, S.K., Anifantis, N.K.: An insight into the breathing mechanism of a crack in
a rotating shaft. J. Sound Vib. 318, 279–295 (2008)
10. Regenhardt, T.E., Azad, M.S., Punurai, W., Beer, M.: A novel application of system survival
signature in reliability assessment of offshore structure. In: International Conference on
Intelligent Computing and Optimization (ICO 2018), pp. 11–20. Springer (2018)
11. Kute, D., Hredeya Mishra, H.: Two crack detection in tapered cantilever beam using natural
frequency as basic criterion. Global J. Eng. Sci. Res. (2018). ISSN 2348–8034
12. Kocharla, R.P.B., Bandlam, R.K., Kuchibotla, M.R.: Finite element modelling of a turbine
blade to study the effect of multiple cracks using modal parameters. J. Eng. Sci. Technol. 11,
1758–1770 (2016)
13. Irwin, G.R.: Analysis of stresses and strains near the end of a crack transverse in a plate.
J. Appl. Mech. 24, 361–364 (1956)
Belief Rule-Based Expert System to Identify
the Crime Zones
Abstract. Crime is the most significant and dominant problem in our society, and
its prevention is an important task. A large number of crimes are being committed
daily. This requires tracking all crimes and maintaining a database that can be used
for future reference. Analyzing this data is the current problem we face to maintain
a proper dataset of crime and to assess and solve future crimes. With the increasing
advent of computerized systems, crime data analysts are helping law enforcement
officers speed up the process of solving crimes. 10% of offenders commit 50% of
crimes. Although we cannot assume that everyone is a victim of crime, we can
estimate where it will occur. This paper focuses on crime zone identification. Then,
it clarifies how we conducted the Belief Rule Base algorithm to produce interesting
frequent patterns for crime hotspots. The paper also shows how we used an expert
system to forecast potential types of crime. To further analyze the crime datasets,
the paper introduces an analysis study by combining our findings of the Chat-
togram crime dataset with demographic information to capture factors that could
affect neighbourhood safety. The results of this solution could be used to raise
awareness of the dangerous locations and to help agencies predict future crimes at a
specific location in a given time.
1 Introduction
Crimes are a common social issue that affects a society’s quality of life and economic
growth. It is considered an essential factor in deciding how people migrate to a new city
and what areas should be avoided while travelling. When crime increases, law
enforcement agencies continue to demand innovative expert solutions to enhance crime
detection and protect their communities better. While crimes may occur anywhere, it is
normal for offenders to focus on the crime opportunities they face in the most familiar
places. Through offering a Belief Rule Expert System to identify the crime zones and to
evaluate the type of the location and time of the crimes committed, we hope to raise
awareness of the dangerous locations in specific periods [1].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 237–249, 2021.
https://doi.org/10.1007/978-3-030-68154-8_24
238 A. Pathak et al.
Therefore, our proposed solution and saving lives can potentially help people stay
away from the locations at a particular time. Therefore, gaining this kind of awareness
help people make better choices about their places of living. Police forces, on the other
hand, can use this approach to improve crime detection and prevention. Also, this
would be useful for allocating police resources. It can help to ensure efficient use of
police resources in the distribution of police at most likely places of crime for any given
time. By providing all of this information, we hope to make our community safer for
the people who live there and others who will travel there.
The expected results cannot be assured of 100% accuracy. Still, the results show
that our program helps to minimize the crime rate by providing security in sensitive
areas of crime. So we have to collect and review crime reports to create such a useful
crime analytics tool [2] (Fig. 1).
The expert system is an evolution of the standard law-based system and can rep-
resent more complex casual relationships using various types of information with
ambiguity. A degree of belief is associated with every possible consequence of a rule in
an expert system based on belief system rules. In the creed rule-based expert system,
the expert system based knowledge is developed using the evidential reasoning
approach to replace the knowledge base inferential engine. The creed rule-based expert
system can capture complicated and continuous casual relationships between different
factors that traditional IF-THEN rules cannot apply [3]. Research would include
evaluating, developing, implementing, and reviewing the best system for defining the
Crime environment. It will concentrate on the processes performed to define crime
zones for many reasons, and how to determine how to detect crime hotspots before
decisions are made on which and when hotspots can be dangerous. It will also focus on
how ordinary people and law enforcement can know the crime hotspots properly. The
study of the crime zone and the detection of the effects of the crime zone is based on
paper. Our study is aimed at finding spatial and temporal criminal hotspots using a set
of real-world crime datasets. We must try to locate the most likely places for the crime
Belief Rule-Based Expert System to Identify the Crime Zones 239
and their frequent occurrence. We will also predict what type of crime could occur next
in a specific location within a given time frame [4]. Develop a computer-based system
that best identifies the crime zone and analyzes the results of crime hotspot tests and the
underlying conditions to recommend the best safe zone for the people.
• Characterize and significance of the hotspot selection process.
• Develop a user-friendly expert system with specific knowledge to advise a person
on selecting crime zone identifiers.
• Assess the performance of the Expert System developed.
To achieve the expected goals, the following objectives are addressed:
• Defining system requirements to assess the information needs of different users so
that the best crime hotspot identification system can be delivered effectively and
effectively.
• Assess the processes involved in the monitoring of the crime zone, and the
parameters used to assess the crime areas can be unsafe for ordinary people.
• Develop a new concept for the crime zone identification system that analyzes the
hotspot and provides not only specific feedback on crime hotspots but also mea-
sures possible crime zone.
2 Related Works
Countless work relating to crime has been done. Large datasets were reviewed, and
information such as location and type of crime was extracted to help people follow the
law enforcement. Existing methods used those databases to identify location-based
crime hotspots. Several map applications show the exact location of the crime along
with the type of crime for any given city. Although crime locations have been iden-
tified, there is no information available that includes the date and time of the crime and
techniques that predict precisely what crimes will occur in the future [5].
On the other hand, the previous work and its existing methods mainly identify crime
hotspots based on the location of high crime density without considering either the crime
type or the date and time of the crime. Related research work, for example, containing a
dataset for the city of Philadelphia with crime information from 1991–1999. It focused
on the existence of complex multi-scale relationships between time and space [6].
Another study titled “The effectiveness of hotspot analysis to forecast spatial crime
trends” looks at the various types of crime to see if they vary in their predictive capa-
bilities. Many studies investigate the relationship between criminal activity and socio-
economic variables such as age, race, income, and unemployment [7].
None of them considers the three elements (location, time, type of crime) together,
despite all the current work. In our study, we provide a Belief Rule-Based Expert
System to Identify the Crime Zones.
240 A. Pathak et al.
3 Methodology
In the proposed system, various components interact with each other. Their proper
interaction creates a perfects BRB and ER based system. During the implementation of
BRB and ER methodology first, important things are to choose a programming lan-
guage to implement ER methodology and user interface, knowledge-based need to
implement using the database. After designing the perfect Graphical User Interface
(GUI), it will be easy for the general user to interact with the developed system and
perform meditation assessment [8]. Thus, the implementation of the proposed system
of experts is one of the main tasks for this work. It is understood that BRB and ER
methods for the system proposed can be applied using an expert knowledge-based
system. An expert knowledge-based system works with data from input and output.
Interaction between the different components of the knowledge-based expert system for
our proposed system is shown in the following (Fig. 2):
It’s known that BRB and ER methodology for the proposed system is possible to
implement using a knowledge-based expert system. A knowledge-based expert system
works with input and output data. Interaction among various components of the
knowledge-based expert system for our proposed system is shown below:
• The crime zone identifier system must have a database that can support the storage
and retrieval of user details, crime details, and details that users who log in to the
system will access later.
• The crime zone identification system must have a client interface that allows
privileged users, such as administrators, to perform tasks such as adding/criminal
editing information, adding/editing hotspot details, and creating/editing user details.
• The crime zone identifier system must have a viewing functionality to allow regular
and privileged users to view the details of a particular entity from the system
database.
• The crime zone identification program must also have a software interface for
regular users to sign up for user accounts, enter data for review and measurement.
• The identifying crime zone system must-have functionality that produces summary
reports from analyses and calculations.
Belief Rule-Based Expert System to Identify the Crime Zones 241
more. (like women population) • Traffic in the area; • Income level in the area
(Unemployment, population density); • Education rate in the area; • Social cause
(Poverty, inequality, etc.);
It has been shown how to calculate input transformation in Sect. 3. The input
transformation process enables the system to transform human readable input data into
system readable data for best crop selection. It’s been possible due to RIMER
methodology in the proposed system, which is a contribution of this project in the best
crop selection process. For instance, let’s take the consequent X (behavioural impact),
for which one of the antecedent Land Types, Water Removal Situation, Drainage
System, Soil Texture, and pH value. Evolution grade for each factor are:
Let take the following portion of BRB as an example to show initial rule base
construction (Fig. 4):
In the proposed system, various components interact with each other. Their proper
interaction creates a perfects BRB and ER based system. During the implementation of
BRB and ER methodology first, important things are to choose a programming lan-
guage to implement ER methodology and user interface, knowledge-based need to
implement using the database. After designing the perfect Graphical User Interface
(GUI), it will be easy for the general user to interact with the developed system and
perform meditation assessment. So the implementation of the proposed expert system is
one of the primary tasks to deal with. System implementation was achieved using
MySQL, JAVA scripting language. The database was designed using MySQL because
it is highly efficient and an effective solution for data storage and handling even in a
networked environment [12].
244 A. Pathak et al.
Table 2. (continued)
Rule Rule Types of Water romoval Drainage Soil pH High Mid Low
ID Wt. land situation system texture
R-26 1 1 1 0 0 0.5 0.4 0.2 0.4
R-27 1 1 1 0 0 0 0.4 0 0.6
R-28 1 1 0.5 1 1 1 0.8 0.2 0
R-29 1 1 0.5 1 1 0.5 0.6 0.4 0
R-30 1 1 0.5 1 1 0 0.6 0.2 0.2
R-31 1 1 0.5 1 0.5 1 0.6 0.4 0
R-32 1 1 0.5 1 0.5 0.5 0.4 0.6 0
R-33 1 1 0.5 1 0.5 0 0.4 0.4 0.2
R-34 1 1 0.5 1 0 1 0.6 0.2 0.2
R-35 1 1 0.5 1 0 0.5 0.4 0.4 0.2
R-36 1 1 0.5 1 0 0 0.4 0.2 0.4
R-37 1 1 0.5 0.5 1 1 0.6 0.4 0
R-38 1 1 0.5 0.5 1 0.5 0.4 0.6 0
R-39 1 1 0.5 0.5 1 0 0.4 0.4 0.2
R-40 1 1 0.5 0.5 0.5 1 0.4 0.6 0
R-41 1 1 0.5 0.5 0.5 0.5 0.2 0.8 0
R-42 1 1 0.5 0.5 0.5 0 0.2 0.6 0.2
R-43 1 1 0.5 0.5 0 1 0.4 0.4 0.2
R-44 1 1 0.5 0.5 0 0.5 0.2 0.6 0.2
R-45 1 1 0.5 0.5 0 0 0.2 0.4 0.4
R-46 1 1 0.5 0 1 1 0.6 0.2 0.2
R-47 1 1 0.5 0 1 0.5 0.4 0.4 0.2
R-48 1 1 0.5 0 1 0 0.4 0.2 0.4
R-49 1 1 0.5 0 0.5 1 0.4 0.4 0.2
R-50 1 1 0.5 0 0.5 0.5 0.2 0.6 0.2
……. ……. ……. ……. ……. ……. ……. ……. ……. …….
Each node (button) works as a link to a particular form for data input, without
calculating nodes in subsequent order, the user can’t move further, as it’s been seen that
some nodes are greyed out. The syntax grid is showing the syntax for corresponding
nodes. After calculating all nodes, aggregated values for the corresponding input value
will be shown in nodes. Finally, aggregated fuzzy values will be converted into
numerical value to get the desired result, which will indicate a value in the scale of
meditation assessment [14].
Data for identifying crime zone of the BRB framework in (previous) have been col-
lected from the CMP. Fifty expert data have been collected, and for simplicity, some of
them have been shown in Table. The expert has selected the major five significant
factors of nine factors and the associated risk factors data from the expert. The collected
data of the expert has been used as input data in this expert system to assess the crop
selection. The BRBES has been compared with a fuzzy rule-based expert system
(FRBES), developed in the MATLAB environment [17]. Illustrates the results of crime
zone prediction carried out by the BRBES (Column 2), the expert (Column 3), and by
the FRBES (Column 4). If the value of an expert is more excellent or equal, 50 positive
than the benchmark is considered as “1” and otherwise, it is considered as “0”. This
data has been considered as the benchmark data, as shown in column 5 of the table
(Table 3 and Fig. 6).
Belief Rule-Based Expert System to Identify the Crime Zones 247
Fig. 6. Comparison of reliability among BRBES, FRBES, and Expert opinion by using ROC
curves
The Receiver Operating Characteristics (ROC) curve is widely used to analyze the
effectiveness of assessment, having ordinal or continuous results. Therefore, it has been
considered in this research to test the accuracy of the BRBES output against expert
opinion and FRBES. The accuracy or performance of the BRBES in crop selection can
be measured by calculating the Area under Curve (AUC) (Table 4).
Having less AUC against expert opinion in Crime Zone Prediction is that expert is
not aware of the uncertainty associated with the factor and risk factors of Crime Zone
selection. Moreover, during our research, I have observed while asking an expert about
the procedures of measuring factor, we have noticed their approach is Boolean. The
reason for less performance of the Fuzzy rule-based expert system than from the belief
rule-based expert system is that the belief rule base considers uncertainty and incom-
pleteness. Also, the inference procedures of BRBES consists of input transformation,
rule activation weight calculation, belief update, and rule aggregation using the evi-
dential reasoning approach [15]. Evidential reasoning can process various uncertainties,
which is not the case with the fuzzy-based inference engine.
In this paper, we have demonstrated the design, development, and application of
BRBES to identify the crime zone by taking five significant factors of crime zone
prediction. This BRBES employed a novel methodology known as RIMER, which
allows the handling of various types of uncertainties. Hence, the results generated from
the BRBES are reliable than from expert opinion or fuzzy rule-based expert systems
[16].
6 Conclusion
The project was successful in implementing the objectives stipulated in earlier chapters.
This system offers several benefits to the users; average persons and researchers can
automatically register, update, and view records. The system administrator can manage
and configure the various parameters of system functionality. The system can also
authenticate the users, display crop yields and total costs, and generate reports. To
improve and increase the use of information technology solutions in day-to-day life,
there is a need for further research to:
• Come up with more advanced software that can remotely connect all the other
remote major towns in districts round and globally.
• Enable an average person or researcher to use the internet, log in to this system and
remotely analyze test results and recommend crime hotspots and get all the infor-
mation they need, say, real danger, and so on.
• Other demands may arise with the ongoing evolution and use of the system.
The study demonstrated how a short text specification of a system could be
modelled in analysis, expanded and detailed into a design model, and finally imple-
mented and programmed in PHP, HTML, and JavaScript.
References
1. RIMER Conference on Belief Rule-Base Inference Methodology Using the Evidential
Reasoning Approach, pp. 41, 44 (2006)
2. Bogomolov, Lepri, B., Staiano, J., Oliver, N., Pianesi, F., Pentland, A.: Once Upon a Crime
3. Buczak, Gifford, C.: Fuzzy association rule mining for community crime pattern Auckland,
New Zealand, pp. 31–38 (2014)
Belief Rule-Based Expert System to Identify the Crime Zones 249
1 Introduction
The work was supported by the Optronic Sensor Systems department, at the Council
for Industrial and Scientific Research (CSIR).
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 250–262, 2021.
https://doi.org/10.1007/978-3-030-68154-8_25
Meta-Heuristic Algorithms in PID Tuning 251
optical systems require stabilizing the sensor’s line-of-sight (LOS) towards a tar-
get of interest [3]. How well the system stabilizes the LOS is highly dependent
on how well the controller is tuned to the system needs. The PID controller is
popular and extensively used in industry because it is simple to tune and has
been extensively implemented and tested [4]. Classical PID tuning methods (e.g.
Ziegler-Nichols) at times do not produce desired results, compared to intelligent
tuning methods [5–7]. This dilemma makes the optimization of controller param-
eters a prominent research area. There are various control systems where the PID
controllers have been tuned using meta-heuristic algorithms [8–13]. Rajesh and
Kavitha [8] are the few researchers who have used meta-heuristic algorithms to
tune the PID controller particularly for gimbal stabilization, which feeds into
the motivation of this work. The meta-heuristic algorithms used in the current
study are the Teaching-Learning-Based Optimization (TLBO) and the Flower
Pollination Algorithm (FPA), of which two variations of the FPA were evalu-
ated. The TLBO algorithm was chosen firstly due to it’s unique ‘parameterless’
quality unlike other algorithms which require careful parameter adjustment for
optimal performance, and secondly because it performed well in past studies [13].
The FPA was chosen because it also performed well in the study conducted in
[10]. This current study made use of two implementations of the FPA: the static
FPA with a switch probability of 80%, and a dynamic FPA with a deterministic
switch probability adapted from [14]. Therefore, the contribution of this paper
lies in investigating different heuristic algorithms, and the parameter settings for
these algorithms, that can best determine the PID controller parameters for an
inertial stabilization platform (ISP).
The model of the gimbal system (the plant) used in this study is adapted from
[15], with some differences. The first difference is that this current study imple-
mented a rate control loop as the focus was on stabilizing the gimbal and LOS
rate, rather than tracking the target and focusing on the LOS angle. Secondly,
only gyro III from the original study is considered here, as the purpose of the
study in [15] was to compare the performances of different gyros which differs
from the purpose of this current study. Another difference is the controller in
this current study is a PID controller with a low pass filter in the derivative
path whereas the study in [15] did not use a low pass filter for the derivative
controller implemented. The model of this study was particularly chosen to be
used because it represents a simple generic system which can be applied in differ-
ent applications. This model was also chosen because the results were able to be
repeated thus model validation was achieved. Other elements and details of the
control system, including the DC motor specifications, the gimbal specifications,
and the gyro specifications, are shown in study [15] and will not be repeated
here.
252 S. Baartman and L. Cheng
The fitness function quantifies the system performance for a set of PID parame-
ters. The fitness function used in this study included weighted averages on both
the ISE and the ITAE because the ISE will help decrease the large initial system
error and aid in the fast response of the system, and the ITAE will reduce oscil-
lations caused by the ISE whilst ensuring a small steady-state error [16]. The
fitness function used in this study is:
where μ = 0.6, α = 0.4. Note that the smaller the value of the fitness function,
the better the solution.
3 Nature-Inspired Algorithms
3.1 Flower Pollination Algorithm
The FPA algorithm is inspired by flower pollination in nature [17]. Global polli-
nation occurs when pollen grains are transferred from one plant to the next using
a pollinator, like an insect. Whereas local pollination occurs when the plant does
not use a pollinator and sheds its pollen from its anther to its stigma [18]. The
switch probability (p) controls whether local or global pollination occurs at any
iteration, which has the range [0,1].
There are many different ways to determine step size when global pollination
occurs. One of which is using the Lévy flight distribution, which is shown below:
λΓ (λ) sin(πλ/2) 1
L(s) = , (2)
π s1+λ
where Γ (λ) represents the standard gamma function, with this study using λ
representing the scaling factor of value 1.5 adapted from [17]. This distribution
is valid for steps s > 0.
Static and Dynamic Switch Probability. The static switch probability (p)
is 0.8 because this works best in most applications [17]. The dynamic switch
probability (pn+1 ) in iteration n + 1 is adapted from [14]:
Niter − n
pn+1 = pn − λ , (3)
Niter
where pn is the value of the switch probability in the previous iteration, Niter is
the maximum number of iterations, n is the current iteration, and the scaling
factor λ = 1.5. The value of the dynamic switch probability begins from 0.9 and
decreases using Eq. (3). Algorithm 1 shows the flow of the FPA.
Meta-Heuristic Algorithms in PID Tuning 253
This study used a basic design of experiments (DOE) [20] methodology which is
the 2k factorial design technique to compare the outcome effect of two values for
each parameter. Table 1 shows these two parameters, along with the two values
of each parameter.
254 S. Baartman and L. Cheng
These parameter values were chosen considering various factors. A low pop-
ulation size may result in the algorithm not having enough population members
to find the optimal solution, and a high population size may result in a waste
of computational space. A low search space bound may result in the optimal
solution not included in the search space, and a high search space bound may
reduce the probability of the algorithm obtaining the optimal solution. The detail
of each algorithm design is shown in Table 2.
The algorithm design and operation are evaluated for two problem instances:
a step input which represents a constant target velocity of 1 rad s−1 and a ramp
input which represents a constant acceleration of 1 rad s−2 . The study ran each
algorithm ten times (with random seed) in order to observe their reactive nature,
as also done in [11]. This research used the best PIDN solutions for performance
comparison because in a real-life scenario, the best solution from the number
of runs would be chosen [21]. The maximum iteration counts for the FPA and
TLBO algorithm were kept constant throughout the experiments at 50 and 25
respectively, as the TLBO performs two fitness function evaluations at each
iteration.
Meta-Heuristic Algorithms in PID Tuning 255
Table 3 shows that the TLBO algorithm, implemented with algorithm design 4
(AD = 4), obtained the lowest fitness value. Furthermore, the lowest mean fitness
value was obtained by AD = 4 for all the algorithms, showing that AD = 4 is the
best performing for this problem instance when observing the fitness value.
The results on Table 3 also show that the dynamic FPA generally obtained
lower fitness values compared to the static FPA, i.e., the problem benefits from
fine-tuning and exploitation in later iterations rather than a constant higher
probability of global pollination throughout all iterations. This could also be
the reason why the TLBO algorithm obtained the lowest fitness value, as this
algorithm works by fine-tuning the solutions at each iteration.
Both the FPA and TLBO resulted in a PI controller solution, as the FPA
resulted in a low derivative gain, and the TLBO resulted in a low filter coefficient
256 S. Baartman and L. Cheng
1.2
1
A D1 A D1
A D2 A D2
1 0.95
A D3 A D3
A D4 A D4
0.9
journal journal
0.8
input input
Step Response
Step Response
0.85
0.6
0.8
0.4 0.75
0.7
0.2
0.65
0
0 0.5 1 1.5 0.05 0.1 0.15 0.2 0.25 0.3 0.35
Time (s) Time (s)
Fig. 1. Step response comparing the algorithm design performance for the TLBO
as seen on Table 4. The PI controller solution means that the system is suffi-
ciently damped as the purpose of the derivative controller is to reduce overshoot
and oscillations, i.e. increasing damping.
The best performing algorithm design is AD = 2 as seen on Fig. 1 for the
TLBO algorithm, and Fig. 2 for the FPA, with the static FPA performing
the best. In order to find the overall best performing algorithm and algorithm
design, the best performing FPA algorithm (implemented with the best per-
forming algorithm design) was compared with the TLBO algorithm, to observe
whether this problem instance prefers algorithms with algorithm-specific param-
eters. The algorithm design chosen as best performing for the all the algorithms
is AD = 2 as it showed a good compromise with the most favourable response, yet
with a reasonable fitness value. Since the static FPA (implemented using AD = 2)
resulted in the lowest rise and settling time, and followed the commanded input
the closest, this algorithm was chosen to be compared with the TLBO algorithm.
Figure 3 shows that the TLBO algorithm reaches the commanded input faster.
Meta-Heuristic Algorithms in PID Tuning 257
1.2
D-AD1 1.02 D-AD1
D-AD2 D-AD2
1 D-AD3 D-AD3
1
D-AD4 D-AD4
S-A D1 S-A D1
0.8 0.98
S-A D2 S-A D2
Step Response
Step Response
S-A D3 S-A D3
0.96
0.6 S-A D4 S-A D4
journal journal
input 0.94 input
0.4
0.92
0.2
0.9
0
0 0.5 1 1.5 0.045 0.05 0.055 0.06 0.065 0.07 0.075
Time (s) Time (s)
Fig. 2. Step response comparing the algorithm design performance for the FPA
1.2
1
TLBO TLBO
Static FPA 0.99 Static FPA
1 input input
0.98
0.97
0.8
Step Response
Step Response
0.96
0.95
0.6
0.94
0.4 0.93
0.92
0.2 0.91
0.9
0
0 0.5 1 1.5 0.03 0.035 0.04 0.045 0.05 0.055 0.06 0.065 0.07 0.075
Time (s) Time (s)
Thus, the TLBO algorithm, using AD = 2 was chosen to be the best performing
for this problem instance.
AD KP KI Kd N
Adaptive 1 68.4011 34.2160 13.2017 100
FPA 2 97.7425 0 13.5612 94.1247
3 22.6570 500 0 36.0330
4 25.2946 500 0 165.944
Static 1 59.0788 55.6679 14.3738 89.4286
FPA 2 37.3178 79.6967 13.6054 99.8358
3 25.2989 500 0 84.1660
4 24.0282 500 0 415.586
TLBO 1 42.9984 66.3752 13.491 99.8793
2 43.2900 83.3955 13.4509 100
3 24.806 500 0 500
4 25.0514 499.932 0 358.493
This problem favoured the dynamic FPA instead of the static FPA, as the
dynamic FPA gave the lowest fitness value from the two. This means this problem
instance is not multi-modal and requires fine-tuning of a specific region in the
search space. This experiment resulted in AD = 4 performing best with the
lowest mean fitness value and best response behaviour for all the algorithms.
Because the dynamic FPA gave the lowest fitness function value when com-
pared to the static FPA, the dynamic FPA was chosen for comparison, along
with the TLBO using the algorithms ramp response shown in Fig. 6. This Figure
illustrates that though the TLBO algorithm resulted in the lowest fitness value,
the best performing algorithm is the dynamic FPA with it following the ramp
input closest. The low fitness value from the TLBO algorithm, when compared
to the dynamic FPA, may be a result of the dynamic FPA taking slightly longer
to settle. Thus, there is no best performing algorithm in this instance as the
Meta-Heuristic Algorithms in PID Tuning 259
1.6
0.068
A D1 A D1
1.4 A D2 A D2
0.066
A D3 A D3
1.2 A D4 A D4
0.064
input input
Step Response
Step Response
1
0.062
0.8 0.06
0.6 0.058
0.4 0.056
0.2 0.054
0
0 0.5 1 1.5 0.052 0.054 0.056 0.058 0.06 0.062 0.064 0.066 0.068 0.07 0.072
Time (s) Time (s)
Fig. 4. Ramp response comparing the algorithm design performance for the TLBO
1.6
0.075
D-AD1 D-AD1
1.4 D-AD2 D-AD2
0.07
D-AD3 D-AD3
1.2 D-AD4 D-AD4
0.065
S-A D1 S-A D1
S-A D2 S-A D2
Step Response
Step Response
1
S-A D3 0.06 S-A D3
0.8 S-A D4 S-A D4
input 0.055 input
0.6
0.05
0.4
0.045
0.2
0.04
0
0 0.5 1 1.5 0.055 0.06 0.065 0.07 0.075 0.08 0.085
Time (s) Time (s)
Fig. 5. Ramp response comparing the algorithm design performance for the FPA
1.6
0.887
input input
TLBO TLBO
1.4
Dynamic FPA Dynamic FPA
0.886
1.2
0.885
Step Response
1
Step Response
0.8 0.884
0.6
0.883
0.4
0.882
0.2
0 0.881
0 0.5 1 1.5 0.8825 0.883 0.8835 0.884 0.8845 0.885 0.8855 0.886
Time (s) Time (s)
TLBO algorithm gave the lowest fitness value, but the dynamic FPA follows the
input the fastest, though it results in a longer time to settle.
6 Conclusion
The tuning of PID controllers is imperative to ensuring a good performance
as classical tuning methods have proven to only produce sub-optimal results
and the need for better tuning method remains. Nature-inspired meta-heuristic
algorithms have shown to be a good alternative solution to PID controller tuning.
However, these algorithms have their own (meta) parameters which need to be
tuned in order to ensure good algorithm performance. This study made use of the
2k factorial method for tuning non-algorithm specific parameters, and compared
self-tuning methods with static parameter values for tuning algorithm-specific
parameters. The performance of the algorithms were then compared on two
different simulation conditions which represent two different target motions. It
was found that the TLBO algorithm resulted in the lowest (best) fitness value
and reached the near optimal solution with the lowest iteration number.
Most gimbal stabilization systems have two, three or even more axes. Future
work should consider how these algorithms perform when optimising PID sys-
tems for multi-axis gimbal systems that include Coriolis-force cross-coupling
disturbance torque between the axes.
References
1. Jan, J.A., Šulc, B.: Evolutionary computing methods for optimising virtual reality
process models. In: International Carpathian Control Conference, May 2002
Meta-Heuristic Algorithms in PID Tuning 261
2. Memari, A., Ahmad, R., Rahim, A.R.A.: Metaheuristic algorithms: guidelines for
implementation. J. Soft Comput. Decis. Support Syst. 4(7), 1–6 (2017)
3. Hilkert, J.M.: Inertially stabilized platform technology concepts and principles.
IEEE Control Syst. Mag. 28, 26–46 (2008)
4. Bujela, B.W.: Investigation into the robust modelling, control and simulation of a
two-dof gimbal platform for airborne applications. Msc. Thesis, University of the
Witwatersrand, Johannesburg, SA (2013)
5. Ribeiro, J.M.S., Santos, M.F., Carmo, M.J., Silva, M.F.: Comparison of PID
controller tuning methods: analytical/classical techniques versus optimization
algorithms. In: 2017 18th International Carpathian Control Conference (ICCC),
pp. 533–538, May 2017
6. Jalilvand, A., Kimiyaghalam, A., Ashouri, A., Kord, H.: Optimal tuning of pid con-
troller parameters on a DC motor based on advanced particle swarm optimization
algorithm, vol. 3, no. 4, p. 9 (2011)
7. Salem, A., Hassan, M.A.M., Ammar, M.E.: Tuning PID controllers using artifi-
cial intelligence techniques applied to DC-motor and AVR system. Asian J. Eng.
Technol. 02(02), 11 (2014)
8. Rajesh, R.J., Kavitha, P.: Camera gimbal stabilization using conventional PID
controller and evolutionary algorithms. In: 2015 International Conference on
Computer Communication and Control (IC4), (Indore), IEEE, September 2015.
https://doi.org/10.1109/IC4.2015.7375580
9. Dash, P., Saikia, L.C., Sinha, N.: Flower pollination algorithm optimized PI-PD
cascade controller in automatic generation control of a multi-area power system.
Int. J. Electr. Power Energy Syst. 82, 19–28 (2016)
10. Jagatheesan, K., Anand, B., Samanta, S., Dey, N., Santhi, V., Ashour, A.S., Balas,
V.E.: Application of flower pollination algorithm in load frequency control of multi-
area interconnected power system with nonlinearity. Neural Comput. Appl. 28,
475–488 (2017)
11. Fister, D., Fister Jr., I., Fister, I., Šafarič, R.: Parameter tuning of PID controller
with reactive nature-inspired algorithms. Robot. Autonom. Syst. 84, 64–75 (2016)
12. Sabir, M.M., Khan, J.A.: Optimal design of PID controller for the speed control
of DC motor by using metaheuristic techniques. Adv. Artif. Neural Sys. 2014, 1–8
(2014). https://doi.org/10.1155/2014/126317
13. Rajinikanth, V., Satapathy, S.C.: Design of controller for automatic voltage regu-
lator using teaching learning based optimization. Procedia Technol. 21, 295–302
(2015)
14. Li, W., He, Z., Zheng, J., Hu, Z.: Improved flower pollination algorithm and its
application in user identification across social networks. IEEE Access 7, 44359–
44371 (2019)
15. Jia, R., Nandikolla, V., Haggart, G., Volk, C., Tazartes, D.: System performance
of an inertially stabilized gimbal platform with friction, resonance, and vibra-
tion effects. J. Nonlinear Dyn. 2017, 1–20 (2017). https://doi.org/10.1155/2017/
6594861
16. Pan, F., Liu, L.: Research on different integral performance indices applied
on fractional-order systems. In: 2016 Chinese Control and Decision Conference
(CCDC), pp. 324–328, May 2016
17. Yang, X.-S.: Flower pollination algorithm for global optimization. In: Durand-
Lose, J., Jonoska, N. (eds.) Unconventional Computation and Natural Computa-
tion vol. 7445, pp. 240–249. Springer, Heidelberg (2012)
262 S. Baartman and L. Cheng
18. Alyasseri, Z.A.A., Khader, A.T., Al-Betar, M.A., Awadallah, M.A., Yang, X.-S.:
Variants of the flower pollination algorithm: a review. In: Yang, X.-S. (ed.) Nature-
Inspired Algorithms and Applied Optimization, vol. 744, pp. 91–118. Springer,
Cham (2018)
19. Rao, R.V.: Teaching-Learning-Based Optimization and Its Engineering Applica-
tions. Springer, Cham (2016)
20. Adenso-Dı́az, B., Laguna, M.: Fine-tuning of algorithms using fractional experi-
mental designs and local search. Oper. Res. 54, 99–114 (2005)
21. Eiben, A.E., Jelasity, M.: A critical note on experimental research methodology
in EC. In: Proceedings of the 2002 Congress on Evolutionary Computation, CEC
2002, vol. 1, pp. 582–587. IEEE, Honolulu, 12 May 2002
Solving an Integer Program by Using
the Nonfeasible Basis Method Combined
with the Cutting Plane Method
1 Introduction
In many real-world problems, there are a lot of problems which require integer solu-
tions such as travelling salesman problems [1], packing problems [2], knapsack
problems [3], assignment problems [4], hub location problems [5] etc. To solve these
problems, integer linear programming problems are formulated which there are many
methods such as branch and bound method [6], cutting plane method [7, 8] and branch
and cut method [9] can be used to find the optimality. The steps of these methods are
solving an LP relaxation problem which the integer condition is dropped, then linear
programming subproblems are solved for finding the integer solution.
One of the most widely used to solve the integer programming problem is branch
and bound method [6] which is a divide-and-conquer strategy. This method starts by
considering the fractional optimal solution to the LP relaxation problem. Then, it is
branched on generating sub-problems and bounded by the integer solution, infeasible
solution and a fractional solution that gives the worse objective value than the current
best solution. However, for solving a large problem, the branch and bound method has
a lot of sub-problems for solving which this leads to the expensive computational time.
In 1958, Gomory [7] introduced the technique named the cutting plane method to
solve an integer linear programming problem. This technique starts by solving the LP
relaxation problem. If an optimal solution to the LP relaxation problem is not an
integer, then the cut is added to find the integer solution. The construction of an
efficient added cut is one of the interesting issues to improve the cutting plane method.
The well-known cut is the Gomory’s cut which is obtained from the optimal
simplex tableau after the LP relaxation problem is solved. This cut is constructed by the
coefficients of nonbasic variables in the row that has a fractional solution. Later in
2005, Glover and Sherali [10] proposed the improvement of the Gomory’s cut, named
Chvatal-Gomory-tier cut. In 2011, Wesselmann et al. [11] developed the Gomory’s cut
enhance the performance of the Gomory mixed integer cut. Additionally, the Gomory’s
cut was improved by many researchers [8, 10].
From the above researches, each method which is used to solve the optimal integer
solution requires the algorithm to solve subproblems which are linear programming
problems. Therefore, the efficient algorithm for solving linear programming problems
has been investigated.
The well-known method for solving a linear programming problem is the simplex
method which is the traditional iterative method introduced by Dantzig in 1947 [12].
However, the simplex method is not the best method since Klee and Minty [13]
demonstrated the simplex method has poor worst-case performance.
Many issues that researchers attempt to improve the simplex method are proposed,
for example, proposed the technique for removing the redundant constraints [14],
introduced the new method for solving the specific linear programming [15, 16] and
developed the artificial-free techniques [17–24].
Before the simplex method starts, the linear programming problem in the standard
form is required. So, the slack or surplus variables are added. Then, a coefficient matrix
of constraints is separated into two sub matrices called the basic matrix (the basis) and
the nonbasic matrix. The variables corresponding to the basis are the basic variables
while the remaining variables are the nonbasic variables. The simplex method starts by
choosing an initial basis to construct a basic feasible solution, then the basis is updated
until the optimal solution is found. However, it is hard to construct an initial feasible
basis by choosing basic variables among decision variables. In the case that surplus
variables are added, the basic variables should consist of slack variables and artificial
variables. It leads to the enlarged problem size. To solve this issue, some researchers
develop a new method for solving linear programming problems which start by
infeasible initial basis.
In 1969, Zionts [17] introduced the artificial-free technique named criss-cross
technique. This technique focuses on both of primal and dual problems. It alternates
between the (primal) simplex method and the dual simplex method until a feasible
solution of a primal problem or a dual problem is found.
Later in 2000, Pan [18] proposed two novel perturbation simplex variants combined
with the dual pivot rules for solving a linear programming problem without using the
artificial variable.
Solving an Integer Program by Using the Nonfeasible Basis Method 265
In 2003, Paparrizos et al. [19] introduced a method for solving a linear program-
ming by using the infeasible basis. The chosen basis is generated by two artificial
variables associated with the additional constraint with a big-M number. Furthermore,
this infeasible basis returns the dual feasible solution.
In 2009, Nabli [20] described the artificial-free technique name the nonfeasible
basis method. The nonfeasible basis method starts when the initial basis gives the
primal infeasible solution. It performs on the condensed tableau which is the briefly
version of the simplex tableau. In the condensed tableau, the columns of nonbasic
variables are only considered because the column of basic variables for each iteration is
not changed. By the operation of the nonfeasible basis method, its performance is better
than two-phase method. Later in 2015, NFB was improved by Nabli and Chahdoura
[21] by using the new pivot rule for choosing an entering variable. In 2020, Sangngern
and Boonperm [22] proposed the method for constructing an initial basis which closes
to the optimal solution by considering the angles between the objective function and
constraints. If the current basic is infeasible, then the nonfeasible method will be
performed.
In 2014, Boonperm and Sinapiromsaran [23, 24] introduced the artificial-free
techniques by solving the relaxation problem which is constructed by considering the
angle between the objective function and constraints.
Due to the above researches, we found that one of techniques for solving an integer
program and one of techniques for solving a linear programming problem are com-
patible. That is the nonfeasible basis method and the cutting plane method with the
Gomory’s cut. Since the added Gomory’s cut is generated by only the coefficients of
nonbasic variables like the nonfeasible basis method, which is performed on the
condensed tableau, we will use this compatibility to develop the algorithm for solving
an integer linear programming problem by the combination of the nonfeasible basis
method and the cutting plane method. Then, it is named the nonfeasible-basis cutting
plane method (NBC).
In this paper, it is organized into five sections. First section is the introduction.
Second, all tools that are used for the proposed method are described. Third, the
proposed method is detailed. The computational result is shown in the fourth section.
Finally, the last section is the conclusion.
2 Preliminaries
In this section, the cutting plane method with the Gomory’s cut for solving an integer
linear programming problem and the artificial-free technique named the Nonfeasible
Basis method are described. We begin with the cutting plane method.
266 K. Sangngern and A. Boonperm
max cT x
ðILPÞ s:t: Ax b
x 0 and integer;
max cT x
ðLPRÞ s:t: Ax b
x 0:
To solve LPR, if the origin point is a feasible point, then the simplex method can
start. Otherwise, the two-phase method is performed.
T
Let A ¼ ½ B N ; x ¼ xTB xTN and cT ¼ cTB cTN . The standard form of LPR
associated to the basis B, the nonbasic matrix N, the basic variable xB and the nonbasic
variable xN is written as follows:
max 0T xB þ cTN cTB B1 N xN ¼ cTB B1 b
s:t: I m xB þ B1 NxN ¼ B1 b
xB ; xN 0
For the cutting plane method, after LPR is solved and its optimal solution is not an
integer solution, then the cuts are added to the LP relaxation problem for cutting of a
non-integer solution until an optimal integer solution is found. In this paper, we are
interested in the Gomory’s cut which can be defined below.
Let B be the optimal basis of LPR. The associated tableau is as follows:
Im B 1N B 1b xB
0 cTN cTB B 1 N cTB B 1b
xB xN
Suppose that there is at least one row of the optimal tableau, called the ith row, with
ðB1 bÞi is fraction. This row is corresponding to the equation
X
xBi þ aij xj ¼ bi ; ð1Þ
j2IN
where aij ¼ ðB1 N Þij ; bj ¼ ðB1 bÞj and IN is the set of indices of nonbasic variables.
Solving an Integer Program by Using the Nonfeasible Basis Method 267
Im B 1N B 1b xB
0 cT T
c B N 1 T 1
c B b
N B B
xB xN
For all iterations, the column matrix of basic variables ðxB Þ is the identity matrix.
Thus, the condensed tableau is introduced instead of the original simplex tableau by
removing the column of basic variables. It can be written as follows:
268 K. Sangngern and A. Boonperm
B 1N B 1b xB
T T 1 T 1
cN c B N
B c B b
B
xN
For using the simplex method with the condensed tableau, the updated condensed
tableau for each iteration is defined by the theorem in [9].
NFB was introduced to construct the feasible basis without using artificial variables
and deals with the condensed tableau. So, the size of a matrix of each problem handled
by this method is smaller than the original simplex method. The NFB method starts
when the chosen basis B is an infeasible basis (i.e., there exists the ith row that
ðB1 bÞi \0). Finally, this method returns a basic feasible solution. The detail of the
algorithm is shown as follows:
r arg max
basic nonbasic
End
Apply pivoting
End
The current basis is feasible.
Apply the simplex method by the desired pivot rule.
Solving an Integer Program by Using the Nonfeasible Basis Method 269
max cT x
ðILPÞ s:t: Ax b ð3Þ
x 0 and integer;
max cT x
s:t: Ax þ Im s ¼ b ð4Þ
x; s 0:
If b 0, then the simplex method can be performed for solving the problem (4).
Otherwise, the origin point is not a feasible point and artificial variables are required.
From the previous section, the nonfeasible basis method is better than the two-
phase method because it performs on the smaller tableau and it is not involved the
artificial variable. Consequently, NFB is used for solving the problem (4). After the
problem (4) is solved, the optimal integer solution to the problem (3) can be found by
the following cases:
1. If the optimal solution to the LP relaxation problem is found and it is integer, then it
is the optimal solution to the problem (3).
2. If the optimal solution to the LP relaxation problem is not integer, then the cutting
plane method is performed in the condensed tableau.
3. If the LP relaxation problem is unbounded, then the problem (3) is unbounded.
4. If the LP relaxation problem is infeasible, then the problem (3) is infeasible.
From the problem (4), the optimal condensed tableau is written as seen below:
B 1N B 1b xB
cTN cTB B 1 N cTB B 1b
xN
From the above tableau, the reduced cost cTN cTB B1 N 0 is obtained, that is the
dual solution is feasible. Suppose that there exists the fractional solution. Then, the
cutting plane method is preferred for finding the integer solution.
270 K. Sangngern and A. Boonperm
If the Gomory’s cut is used, then one of rows in the optimal tableau, say the kth
row, which has a fractional solution is selected to generate an added cut. Then, the
added cut is
ðbak c ak ÞxN þ xm þ n þ 1 ¼ bk
bk ð5Þ
where ak is the k th row of matrix B1 N, bak c ¼ ½ b a1k c bamk c and bk ¼
ðB1 bÞk .
From the added cut (5), if xm þ n þ 1 is chosen to be a basic variable, then this cut can
add all coefficients of nonbasic variables to the optimal condensed tableau which is
suitable for the cutting plane method. Hence, the condensed tableau with the added cut
can be written as follows:
B 1N B 1b xB
ak ak bk bk xm n 1
Since bk bk cannot be positive, the dual simplex method will be performed for
solving the condensed tableau with the added cut. After the LP relaxation problem is
solved, if the obtained solution is not integer, then repeat steps by adding a new cut one
by one until the optimal integer solution is found. The algorithm of this method can be
summarized as follows:
4 Computational Results
In Sect. 3, the combination method for solving an integer linear programming problem
without using artificial variables was designed. To show the performance of NBC, the
computational time for solving random generated integer programs between the NBC
method and the traditional method (the simplex method or the two-phase method
combined with the cutting plane method) is compared. Both methods are implemented
through Python. Finally, the computational results will be shown and discussed.
max cT x
s:t: Ax b ð6Þ
x 0 and integer;
All methods are implemented by using Python coding on the Google Colaboratory,
and these tests were run on Intel® Core™ i7-5500U 8 GB 2.4 GHz.
For the traditional method, the two-phase method is used for solving the LP
relaxation problem, and then the original cutting plane method is used for finding the
optimal integer solution and they are performed on the full simplex tableau.
272 K. Sangngern and A. Boonperm
Table 2 exhibits the average computational time before the cutting plane method is
performed and the average total computational time. In Table 2, the main column
named TRAD and NBC are represented the computational time for solving problem by
using the traditional method and the NBC method, respectively, and the sub column
named BC and TC are represented the average computational time before the cutting
plane method is performed and the average total computational time, respectively. Note
that the bold face numbers indicate the smallest average computational time.
From Table 2, the average computational time before the cutting plane method is
performed and the average total computational time of NBC are less than TRAD.
Since NBC is not involved the artificial variable and is performed on the condensed
tableau while TRAD requires artificial variables and the original cutting plane method,
NBC is faster than TRAD.
From the computational result, we found that the average total computational time
5436:39 0:4438 or approximate 44.38% of the average total computational
of NBC is 2412:86
time of TRAD. It means that the proposed method can reduce the computational time
approximately 55.62% from the computational time of the traditional time (Fig. 1).
Solving an Integer Program by Using the Nonfeasible Basis Method 273
4.3 Discussion
From the computational results, we found that the average of the total computational
time of the traditional method is greater than the NBC method. There are two main
reasons why the NBC method has more effective than the traditional method. First, the
size of the performed tableau for each method is different since the traditional method
performs on the full simplex tableau which is bigger than the condensed tableau that
used in the NBC method.
To compare the size of the tableaux between the original simplex tableau and the
condensed tableau, if one cut is added, then the original simplex tableau will be
expanded with one row and one column
Im B 1N 0 B 1b xB
0 as as 1 bs bs xm n 1
T T 1 T 1
0 cN c B N
B
0 c B b
B
xB xN xm n 1
while the condensed tableau will be expanded with only one row.
B 1N B 1b xB
as as bs bs xm n 1
T T 1 T 1
cN c B N
B c B b
B
xN
274 K. Sangngern and A. Boonperm
Moreover, if k cuts are added, then the original simplex must be expanded with k
rows and k columns while the condensed tableau is expanded with only k rows.
Therefore, the size of tableaux performed by NBC is smaller than the origin simplex
method causing that the NBC can reduce the computational time.
The remaining reason is about the use of artificial variable. The traditional method
is involved in the artificial variables. This leads to the number of variables in the system
of the traditional method more than the NBC method which is the artificial-free
technique.
5 Conclusion
There are two types of methods for solving an integer linear programming problem, the
exact method and the heuristic method. In this paper, we focus on the exact method.
The traditional exact method consists of the method for solving the LP relaxation
problem and the technique for finding the integer solution. For solving the LP relax-
ation problem, the simplex method or the two-phase method is preferred. If the two-
phase method is chosen, then artificial variables are added, and it implies that the
problem must be expanded. To avoid the artificial variable, the artificial-free technique
is required. The artificial-free technique that we choose for dealing with the artificial
variable is the nonfeasible basis method (NFB). NFB is not only the artificial-free
technique but also it performs on the small simplex tableau named condensed tableau.
The nonfeasible-basis cutting plane method (NBC) is a combination of the non-
feasible basis method and the cutting plane method performed on the condensed
tableau. The step of NBC starts with NFB for solving the LP relaxation problem, then
the cutting plane method is used to find the integer solution. Since NBC is not use the
artificial variable and perform on the condensed tableau, the computational time of
NBC must less than the traditional method. From the computational results, we found
that the computational time of NBC is approximate 44% of the computational time of
the traditional method.
References
1. Fatthi, W., Haris, M., Kahtan H.: Application of travelling salesman problem for minimizing
travel distance of a two-day trip in Kuala Lumpur via Go KL city bus. In: Intelligent
Computing & Optimization, pp. 227–284 (2018)
2. Torres-Escobar, R., Marmolejo-Saucedo, J., Litvinchev, I., Vasant, P.: Monkey algorithm for
packing circles with binary variables. In: Intelligent Computing & Optimization, pp. 547–
559 (2018)
3. Yaskov, G., Romanova, T., Litvinchev, I., Shekhovtsov, S.: Optimal packing problems:
from knapsack problem to open dimension problem. In: Advances in Intelligent Systems and
Computing, pp. 671–678 (2019)
4. Marmolejo-Saucedo, J., Rodriguez-Aguilar, R.: A timetabling application for the assignment
of school classrooms. In: Advances in Intelligent Systems and Computing, pp. 1–10 (2019)
Solving an Integer Program by Using the Nonfeasible Basis Method 275
5. Chanta, S., Sangsawang, O.: A single allocation P-Hub maximal covering model for
optimizing railway station location. In: Intelligent Computing & Optimization, pp. 522–530
(2018)
6. Land, A.H., Doig, A.: An automatic method of solving discrete programming problems.
Econometrica 28, 497–520 (1960)
7. Gomory, R.E.: Outline of an algorithm for integer solutions to linear programs. Bull. Am.
Math. Soc. 64, 275–278 (1958)
8. Gomory, R.E.: Solving linear programming problems in integers. Proc. Symposia Appl.
Math. 10, 211–215 (1960)
9. Padberg, M., Rinaldi, G.: A branch-and-cut algorithm for the resolution of large-scale
symmetric traveling salesman problems. SIAM Rev. 33, 60–100 (1991)
10. Glover, F., Sherali, H.D.: Chvatal-gomory-tier cuts for general integer programs. Discrete
Optim. 2, 51–69 (2005)
11. Wesselmann, F., Koberstein, A., Suhl, U.: Pivot-and-reduce cuts: an approach for improving
gomory mixed-integer cuts. Eur. J. Oper. Res. 214, 15–26 (2011)
12. Dantzig, G.B.: Activity Analysis of Production and Allocation. Wiley, New York (1951)
13. Klee, V., Minty, G.: How Good is the Simplex Algorithm. In Equalities. Academic Press,
New York (1972)
14. Paulraj, S., Sumathi, P.: A comparative study of redundant constraints identification methods
in linear programming problems. Math. Problems Eng. 2010, 1–16 (2010)
15. Gao, C., Yan, C., Zhang, Z., Hu, Y., Mahadevan, S., Deng, Y.: An amoeboid algorithm for
solving linear transportation problem. Physica A 398, 179–186 (2014)
16. Zhang, X., Zhang, Y., Hu, Deng, Y., Mahadevan, S.: An adaptive amoeba algorithm for
constrained shortest paths. Expert Syst. Appl. 40, 7607–7616 (2013)
17. Zionts, S.: The criss-cross method for solving linear programming problems. Manage. Sci.
15, 426–445 (1969)
18. Pan, P.Q.: Primal perturbation simplex algorithms for linear programming. J. Comput. Math.
18, 587–596 (2000)
19. Paparrizos, K., Samaras, N., Stephanides, G.: A new efficient primal dual simplex algorithm.
Comput. Oper. Res. 30, 1383–1399 (2003)
20. Nabli, H.: An overview on the simplex algorithm. Appl. Math. Comput. 210, 479–489
(2009)
21. Nabli, H., Chahdoura, S.: Algebraic simplex initialization combined with the nonfeasible
basis method. Eur. J. Oper. Res. 245, 384–391 (2015)
22. Sangngern, K., Boonperm, A.: A new initial basis for the simplex method combined with the
nonfeasible basis method. J. Phys. Conf. Ser. 1593, 012002 (2020)
23. Boonperm, A., Sinapiromsaran, K.: The artificial-free technique along the objective direction
for the simplex algorithm. J. Phys. Conf. Ser. 490, 012193 (2014)
24. Boonperm, A., Sinapiromsaran, K.: Artificial-free simplex algorithm based on the non-acute
constraint relaxation. Appl. Math. Comput. 234, 385–401 (2014)
A New Technique for Solving a 2-Dimensional
Linear Program by Considering the Coefficient
of Constraints
1 Introduction
In 1976, Shamos and Hoey [22] proposed the algorithm which is used to solve a 2-
dimensional linear program. This algorithm forms the intersection of geometric objects
in the plane. Then, it starts by considering n line segments in the plane and finding all
intersecting pairs. After that, they use the Oðn log nÞ algorithm to determine whether
any 2 intersection points and use it to find out whether two simple plane polygons
intersecting. They proved that a linear program with 2 variables and n constraints can
be solved in Oðn log nÞ time while the simplex method requires Oðn2 Þ time.
In 1983, Megiddo [23] introduced the linear-time algorithm to solve a linear pro-
gram in R2 . This method searches for the smallest circle enclosing n given points in the
plane. Moreover, it disproves a conjecture by Shamos and Hoey that their algorithm
requires Oðn log nÞ time. In addition, a 2-dimensional linear program which has a
closely related problem called a separability problem in R2 is constructed and solved in
OðnÞ time. That is, this algorithm corrects the error in Shamos and Hoey’s paper that it
can solve in the Oðn log nÞ time.
In 1984, Dyer [24] also proposed the algorithm which requires OðnÞ time to solve a
2-dimensional linear program. The concept of this algorithm is utilization of convexity
which is well-known ideas for creating the linear time algorithm. This algorithm starts
at any feasible point and coefficient values of the objective function, c1 and c2 are
considered. If c1 ¼ c2 ¼ 0, then this feasible point is optimal. Otherwise, the problem
is transformed to provide a gradient direction of the objective function. After that, the
optimal solution is gained by performing the pairing algorithm. All processes of Dyer’s
algorithm require OðnÞ time for each process.
Recently, Vitor and Easton [25] proposed a new method to solve a 2-dimensional
linear program which is called the slope algorithm such that the problem must satisfy
the following conditions: the coefficient values of the objective function are positive,
the right-hand side values of all constraints are also positive and both variables are
nonnegative. Not only this algorithm can solve the 2-dimensional problem, but also can
be applied to identify the optimal basis for the simplex method. They apply the slope
algorithm into the simplex framework for solving a general linear program. The sim-
plex method is improved by exchanging two variables into the basis which are iden-
tified by using the slope algorithm and this method is called the double pivot simplex
method. For each iteration of this method, it requires the optimal basis from a 2-
dimensional linear program to indicate leaving variables for the original problem. Then,
the slope algorithm is used to identify such variables. After the slope algorithm is
performed, the simplex tableau is updated corresponding to the new basis. It repeats
steps until the optimal solution to the original problem is obtained. From the compu-
tational result, this algorithm can reduce the iterations of the simplex method.
Although the slope algorithm can identify the basis and can reduce the iterations of
the simplex method, it takes more computations for each iteration. So, in this paper, we
propose a new method for solving a 2-dimensional linear program. It can improve the
double pivot simplex method by avoiding the computation of the slope algorithm in
such case that is there exists only one pivot. We specifically determine the 2-
dimensional linear program such that the problem satisfies the following conditions: the
cost coefficient values are positive, the right-hand side values are nonnegative and both
variables are also nonnegative. This approach can identify the optimal solution by
278 P. Jamrunroj and A. Boonperm
Maximize z = c1 x1 + c2 x2
Subject to a11 x1 + a12 x2 ≤ b1
≤ ð1Þ
am1 x1 + am 2 x2 ≤ bm
x1 , x2 ≥ 0,
where c1 ; c2 [ 0 and bi 0 for all i ¼ 1; . . .; m. Since the origin point is feasible, the
problem (1) is either optimal or unbounded solution (see in Fig. 1).
Theorem 1. Consider the 2-dimensional linear program (1) with ai2 [ 0 for all
i ¼ 1; . . .; m. Let
n o n o
k ¼ arg min abi1i jai1 [ 0; i ¼ 1; . . .; m and r ¼ arg min abi2i jai2 [ 0; i ¼ 1; . . .; m .
i) If ak1 \ak2 ðc1 =c2 Þ, then x ¼ ðx1 ; 0Þ is the optimal solution to problem (1) where
x1 ¼ bk =ak1 .
ii) If ar1 [ ar2 ðc1 =c2 Þ, then x ¼ ð0; x2 Þ is the optimal solution to problem (1) where
x2 ¼ br =ar2 .
Proof.
i) Suppose that ak1 \ak2 ðc1 =c2 Þ. Let x1 ¼ bk =ak1 . We need to show that x ¼ ðx1 ; 0Þ
is the optimal solution by using the optimality conditions.
First, we show that ðx1 ; 0Þ is a feasible solution of the primal problem. Consider the
following system:
Since bi 0 for all i ¼ 1; . . .; m, if ai1 0, then ai1 x1 bi and if ai1 [ 0, then
ai1 x1 ¼ ai1 abk1k ai1 abi1i ¼ bi . Therefore, ðx1 ; 0Þ is a feasible solution.
Next, we will show that there exists a dual feasible solution. Consider the dual
problem:
Minimize b1 w1 + b2 w2 + + bm wm = bT w
Subject to a11 w1 + a21 w2 + + ak1 wk + + am1 wm ≥ c1
a12 w1 + a22 w2 + + ak 2 wk + + am 2 wm ≥ c2
w1 ,..., wm ≥ 0.
Since ak1 \ak2 ðc1 =c2 Þ, c2 \ak2 ðc1 =ak1 Þ. Hence, w ¼ ½ 0 c1 =ak1 0 T
is a dual feasible solution.
Finally, consider the objective values of the primal and the dual problems. We get
cT x ¼ c1 ðbk =ak1 Þ and bT w ¼ bk ðc1 =ak1 Þ.
Since x and w are a primal feasible solution and a dual feasible solution,
respectively, and cT x ¼ bT w , x is the optimal solution to problem (1).
ii) Suppose that ar1 [ ar2 ðc1 =c2 Þ. Let x2 ¼ br =ar2 . We need to show that x ¼ ð0; x2 Þ
is the optimal solution by using the optimality conditions. First, we will show that
ð0; x2 Þ is a primal feasible solution. Consider the following system:
Since bi 0 for all i ¼ 1; . . .; m, if ai2 0, then ai2 x2 bi and if ai2 [ 0, then
ai2 x2 ¼ ai2 abr2r ai2 abi2i ¼ bi . Therefore, the primal problem is feasible.
Next, we will show that there exists a dual feasible solution. Consider the dual
problem:
Minimize b1 w1 + b2 w2 + + bm wm = bT w
Subject to a11 w1 + a21 w2 + + ak1 wk + + am1 wm ≥ c1
a12 w1 + a22 w2 + + ak 2 wk + + am 2 wm ≥ c2
w1 , , wm ≥ 0.
Since ar1 [ ar2 ðc1 =c2 Þ, c1 \ar1 ðc2 =ar2 Þ. Hence, w ¼ ½ 0 c2 =ar2 0 T
is a dual feasible solution.
Finally, we will prove that the objective values of primal and dual problems are
equal.
Consider cT x ¼ c2 ðbr =ar2 Þ and bT w ¼ br ðc2 =ar2 Þ.
A New Technique for Solving a 2-Dimensional Linear Program 281
Since x and w are a primal feasible solution and a dual feasible solution,
respectively, and cT x ¼ bT w , x is the optimal solution to problem (1).
Theorem 2. Consider the 2-dimensional linear program (1) with ai1 [ 0 for all
i ¼ 1; . . .; m. Let
n o n o
k ¼ arg min abi1i jai1 [ 0; i ¼ 1; . . .; m and r ¼ arg min abi2i jai2 [ 0; i ¼ 1; . . .; m .
i) If ar2 \ar1 ðc2 =c1 Þ, then x ¼ ð0; x2 Þ is the optimal solution to problem (1) where
x2 ¼ br =ar2 .
ii) If ak2 [ ak1 ðc2 =c1 Þ, then x ¼ ðx1 ; 0Þ is the optimal solution to problem (1) where
x1 ¼ bk =ak1 :
For this section, we give two examples for solving 2-dimensional linear problems and
one example for applying with the simplex method. The first and second examples
present that the 2-dimensional linear program gives the optimal solution. For the third
example, we apply the proposed approach with the general linear program for selecting
a leaving variable.
Example 1. Consider the following 2-dimensional linear program:
Maximize z = 3x1 + 4 x2
Subject to −3 x1 + 5 x2 ≤ 15
−2 x1 + x2 ≤ 22
− x1 + 4 x2 ≤ 16 ð2Þ
x1 + 7 x2 ≤ 30
x1 + 2 x2 ≤ 11
x1 , x2 ≥ 0 .
Since ai2 [ 0 for i ¼ 1; . . .; m and ak1 ¼ 1\ak2 ðc1 =c2 Þ ¼ 2ð3=4Þ ¼ 3=2, the
optimal solution is x1 ; x2 ¼ ðbk =ak1 ; 0Þ ¼ ð11; 0Þ with z ¼ 44.
282 P. Jamrunroj and A. Boonperm
For solving this problem by the simplex method, it requires 4 iterations while our
method can identify the optimal solution by considering the coefficients.
Example 2. Consider the following 2-dimensional linear program:
Maximize z = 3x1 + 2 x2
Subject to 5 x1 + x2 ≤ 5
− x1 + x2 ≤ 5
ð3Þ
2 x1 + x2 ≤ 4
5 x1 + 4 x2 ≤ 25
x1 , x2 ≥ 0.
ai2 [
Since 0 for i ¼ 1; . . .; m and ar1 ¼ 2 [ ar2 ðc1 =c2 Þ ¼ 3=2, the optimal solu-
tion is x1 ; x2 ¼ ð0; br =ar2 Þ ¼ ð0; 4Þ with z ¼ 8.
The next example, we will apply our algorithm to solve a general linear program for
identifying an entering variable.
Example 3. Consider the following linear program:
Maximize z = 5 x1 + 3x2 + x3 + 2 x4
Subject to 2 x1 + x2 + 2 x3 + x4 ≤ 10
7 x1 + 4 x2 + 2 x4 ≤ 12
6 x1 + 3x2 + 3x3 ≤ 15
ð4Þ
3x1 + x2 + x3 + 3x4 ≤ 8
9 x1 + 2 x2 + 4 x3 + x4 ≤ 9
3x1 + x2 + 2 x4 ≤ 5
x1 , x2 , x3 , x4 ≥ 0.
First, construct the initial tableau by adding the slack variables and they are set to
form a basis. The initial tableau can be obtained below.
x1 x2 x3 x4 x5 x6 x7 x8 x9 x10 RHS
z −5 −3 −1 −2 0 0 0 0 0 0 0
x5 2 1 2 1 1 0 0 0 0 0 10
x6 7 4 0 2 0 1 0 0 0 0 12
x7 6 3 3 0 0 0 1 0 0 0 15
x8 3 1 1 3 0 0 0 1 0 0 8
x9 9 2 4 1 0 0 0 0 1 0 9
x10 3 1 0 2 0 0 0 0 0 1 5
A New Technique for Solving a 2-Dimensional Linear Program 283
After we use the proposed method, we get x ¼ ð0; 3Þ is the optimal solution to (5).
So, x2 is selected to be an entering variable. The variable x6 is chosen to be a leaving
variable because its corresponding constraint gives the minimum value of bi =ai2 for all
i ¼ 1; . . .; m. The simplex tableau is updated as the following tableau.
x1 x2 x3 x4 x5 x6 x7 x8 x9 x10 RHS
z 0.25 0 −1 −0.5 0 0.75 0 0 0 0 9
x5 0.25 0 2 0.5 1 −0.25 0 0 0 0 7
x2 1.75 1 0 0.5 0 0.25 0 0 0 0 3
x7 0.75 0 3 −1.5 0 −0.75 1 0 0 0 6
x8 2.25 0 1 2.5 0 −0.25 0 1 0 0 5
x9 5.5 0 4 0 0 −0.5 0 0 1 0 3
x10 1.25 0 0 1.5 0 −0.25 0 0 0 1 2
From the above tableau, x3 and x4 are considered. Since no positive coefficient
vector of both variables, the proposed method cannot determine entering and leaving
variables. Then, we back to use the slope algorithm to solve it and get that x3 ; x4 are
selected to be entering variables and x9 ; x10 are chosen to be leaving variables. The
updated simplex tableau is as follows:
x1 x2 x3 x4 x5 x6 x7 x8 x9 x10 RHS
z 2.042 0 0 0 0 0.542 0 0 0.25 0.333 10.417
x5 −2.917 0 0 0 1 0.083 0 0 −0.5 −0.333 4.833
x2 1.333 1 0 0 0 0.333 0 0 0 −0.333 2.333
x7 −2.125 0 0 0 0 −0.625 1 0 −0.75 1 5.75
x8 −1.208 0 0 0 0 0.292 0 1 −0.25 −1.667 0.917
x3 1.375 0 1 0 0 −0.125 0 0 0.25 0 0.75
x4 0.833 0 0 1 0 −0.167 0 0 0 0.667 1.333
From Example 3, at the first iteration, the proposed method selects the second most
negative reduced cost to be an entering variable while the simplex method chooses the
most negative reduced cost to be an entering variable. At the second iteration, the
proposed method performed double pivots while the simplex method performed only
one a pivot. In summary, we can reduce the number of iterations for this case because
the proposed method performed only 2 iterations to solve the problem in Example 3
while the simplex method performed 4 iterations. In addition, at the first iteration, we
can reduce a computation because an entering variable was indicated without using the
slope algorithm.
In this section, we report the computational results by comparing the number of iter-
ations between the simplex method with the proposed technique and the original
simplex method. By solving the following canonical linear program:
Maximize z = cT x
Subject to Ax ≤ b ð6Þ
x ≤ 0.
The sizes of tested problems have six different sizes of the coefficient matrix which
are m n ¼ 6 4; 6 8; 6 10; 8 6; 8 10 and 8 12. In addition, most of the
tested problems are generated by strictly at the first iteration that requires the second
most negative reduced cost to be an entering variable. Then, the computational results
are shown in Table 1. We can see that the average number of iterations of the simplex
method with the proposed technique are less than the average number of iterations of
the original simplex method. So, we can reduce the number of iterations when a
problem is solved by the simplex method with the proposed technique.
5 Conclusions
In this paper, a new method for solving a 2-dimensional linear program by considering
only the coefficient of constraints is proposed. If there exists a positive coefficient
vector of one variable, then the optimal solution can be identified immediately.
Moreover, this technique can fix some cases of the double pivot simplex method that is
the case of having only one entering variable. In this case, the double pivot simplex
method performs the slope algorithm to solve the relaxed 2-dimensional linear program
for identifying an entering variable that has more computation while the proposed
method considers only the coefficient constraints for specifying an entering variable.
By this reason, we can reduce computation for this case. But if the relaxed 2-
dimensional linear program does not satisfy the conditions, it is solved inevitably by
the slope algorithm. Now, we can fix only the case of one entering variable for the
double pivot simplex method. In the future work, we need to find a method that can fix
the case of two entering variables.
References
1. Dantzig, G.B.: Linear Programming and Extensions. RAND Corporation, Santa Monica
(1963)
2. Klee, V., Minty, G.: How Good is the Simplex Algorithm. In Equalities. Academic Press,
New York (1972)
3. Terlaky, T.: A finite crisscross method for oriented matroids. J. Comb. Theory B 42(3), 319–
327 (1987)
4. Pan, P.: A new perturbation simplex algorithm for linear programming. J. Comput. Math. 17
(3), 233–242 (1999)
5. Pan, P.: Primal perturbation simplex algorithms for linear programming. J. Comput. Math.
18(6), 587–596 (2000)
6. Elhallaoui, I., Villeneuve, D., Soumis, F., Desaulniers, G.: Dynamic aggregation of set-
partitioning constraints in column generation. Oper. Res. 53(4), 632–645 (2005)
7. Elhallaoui, I., Desaulniers, G., Metrane, A., Soumis, F.: Bi-dynamic constraint aggregation
and subproblem reduction. Comput. Oper. Res. 35(5), 1713–1724 (2008)
8. Elhallaoui, I., Metrane, A., Soumis, F., Desaulniers, G.: Multi-phase dynamic constraint
aggregation for set partitioning type problems. Math. Program. 123(2), 345–370 (2010)
9. Elhallaoui, I., Metrane, A., Desaulniers, G., Soumis, F.: An improved primal simplex
algorithm for degenerate linear programs. INFORMS J. Comput. 23(4), 569–577 (2010)
10. Raymond, V., Soumis, F., Orban, D.: A new version of the improved primal simplex for
degenerate linear programs. Comput. Oper. Res. 37(1), 91–98 (2010)
11. Jeroslow, R.G.: The simplex algorithm with the pivot rule of maximizing criterion
improvement. Discr. Math. 4, 367–377 (1973)
12. Bland, R.G.: New finite pivoting rules for the simplex method. Math. Oper. Res. 2(2), 103–
107 (1977)
13. Pan, P.: Practical finite pivoting rules for the simplex method. OR Spectrum 12, 219–225
(1990)
14. Bixby, R.E.: Solving real-world linear programs: a decade and more of progress. Oper. Res.
50(1), 3–15 (2002)
286 P. Jamrunroj and A. Boonperm
15. Pan, P.: A largest-distance pivot rule for the simplex algorithm. Eur. J. Oper. Res. 187(2),
393–402 (2008)
16. Pan, P.: A fast simplex algorithm for linear programming. J. Comput. Math. 28(6), 837–847
(2010)
17. Csizmadia, Z., Illés, T., Nagy, A.: The s-monotone index selection rules for pivot algorithms
of linear programming. Eur. J. Oper. Res. 221(3), 491–500 (2012)
18. Liao, Y.: The improvement on R. G. Blands method. In: Qi, E., Shen, J., Dou, R. (eds.) The
19th International Conference on Industrial Engineering and Engineering Management,
Changsha, pp. 799–803 (2013)
19. Pan, P.: Linear Programming Computation, 1st edn. Springer, Heidelberg (2014)
20. Etoa, J.: New optimal pivot rule for the simplex algorithm. Adv. Pure Math. 6, 647–658
(2016)
21. Ploskas, N., Samaras, N.: Linear Programming Using MATLAB®, volume 127 of Springer
Optimization and Its Applications, 1st edn. Springer, Cham (2017)
22. Shamos, M., Dan Hoey: Geometric intersection problems. In: 17th Annual Symposium on
Foundations of Computer Science (SFCS 1976), pp. 208—215 (1976)
23. Megiddo, N.: Linear-time algorithms for linear programming in R3 and related problems.
SIAM J. Comput. 12(4), 759–776 (1983)
24. Dyer, M.: Linear time algorithms for two- and three-variable linear programs. SIAM J. Com-
put. 13(1), 31–45 (1984)
25. Vitor, F., Easton, T.: The double pivot simplex method. Math. Meth. Oper. Res. 87, 109–137
(2018)
A New Integer Programming Model
for Solving a School Bus Routing Problem
with the Student Assignment
1 Introduction
One of the four elements of the marketing mix is distribution, which is the process
involving selling and delivering products and services from a manufacturer to cus-
tomer. Distribution is also an essential part of manufacturers’ operations since a poor
distribution leads to a loss of trust of customers, retailers and suppliers. Therefore, it
becomes important to improve product distribution to ensure that customers are
pleased. There are many ways to improve the efficiency of the distribution. One of
these improvements is finding a route for distributing products to maximize the sat-
isfaction of a manufacturer and customer. For instance, in 2019, Mpeta et al. [1]
introduced the binary programming model for solving a municipal solid waste col-
lection problem. The obtained solution is an optimal route to reduce waste collection
time and costs.
The vehicle routing problem (VRP) is a problem for finding the route to distribute
products from center depot to customers in different locations and return to center
depot. The VRP is commonly investigated in companies for product distribution. There
are many factors to seek the route in the VRP such as the capacity of vehicles, the
number of vehicles, the driver cost, the travel distance, etc. In 1959, the VRP was first
studied by Dantzig et al. [2]. The purpose in their work is to find the shortest route for
delivery gasoline between a bulk terminal and many stations corresponding to the
demand in each station.
The VRP is adjusted to the wider problem named the open vehicle routing problem
(OVRP). The OVRP relaxes the condition that the vehicles must return to the center
depot from the VRP. It aims that the OVRP is a much more precise model for con-
forming business compare to the VRP with close routes returning to the center depot.
There are several algorithms for solving the OVRP. For instance, Tarantilisl et al. [3]
presented the threshold-accepting algorithm called the list based threshold accepting
(LBTA) to solve the OVRP on artificial case with 50–199 depots.
The school bus routing problem (SBRP) is one type of the OVRP which is designed
for planning the route to pick up students from bus stops to a school or from a school to
bus stops. The SBRP was first considered by Newton and Thomas [4]. They solve this
problem by generating routes using a computer. A case study of this research is in
Indiana with 1500 students. Afterward, many researchers improved a model for solving
the SBRP. For instance, In 1997, Braca et al. [5] attempted to minimize the number of
buses with many constraints involving the capacity of buses, riding time, school time
window, walking distance of student from home to bus stop and the number of stu-
dents. They also studied the case study in New York with 838 bus stops and 73
schools. More importantly, they solved all the problems of every school in one state. In
2012, Jalel and Rafaa [6] developed a hybrid evolutionary computation based on an
artificial ant colony with a variable neighborhood local search algorithm for the Tunisia
school bus routing problem. In 2013, Taehyeong Kim et al. [7] formulated a mixed-
integer programming problem for the SBRP, and developed a heuristic algorithm based
on the harmony search to find the solution of this problem. The obtained solution from
the developed heuristic algorithm are compared with the exact solution by CPLEX. The
solution of both methods is the same, but the developed heuristic method is calculated
faster. In 2016, Hashi et al. [8] used Clarke & Wright’s algorithms for solving the
school bus routing and scheduling problem with time window. They also tested this
algorithm with the Scholastica School in Dhaka, Bangladesh. Different methods for
solving the SBRP can be seen from results of Park and Kim [9], Ellegood et al. [10],
Bodin and Berman [11], Schittekat et al. [12], Eldrandaly and Abdallah [13] and
references therein.
We now discuss the research about the SBRP due to Baktas et al. [14], which is the
main motivation for writing this paper. In their research, they presented a model to
minimize the number of buses with capacity and distance constraints was presented.
Both constraints prevent a sub-tour. They used this model for solving the SBRP with a
real case in Turkey. In this model, they solved the SBRP under the situation that the
number of students in each bus stop is known. However, the satisfaction of all students
before arrival at each bus stop is not considered in their results. This leads to the
inspiration for investigation and improving the process for solving the SBRP in this
view.
In this paper, we are interested for solving the SBRP with capacity and distance
constraints in view of the above mentioned inspiration. Distance and capacity con-
straints in our model are revised from the model in [14] to handle the satisfaction of all
students together with the total distance to all buses. In addition, we provide a com-
parison of results providing from our models between two situations based on two
distinct objectives including to maximize the total satisfaction of students and to
minimize the total travel distance of all buses.
A New Integer Programming Model 289
In this paper, the SBRP with the student assignment depending on the satisfaction score
of students at each bus stop is investigated. For this problem, all students go to bus
stops under two situations including to maximize the total score of all students and to
minimize the total travel distance of all bus stops. Then, each bus starts from some bus
stops and picks-up students for going to a school before the school starts, and it returns
to the same route for dropping-off students. The number of students in each bus must
not exceed the capacity of the bus. Each bus is involved in a single path, and the total
distance is limited by the specified distance. An example of a school, bus stops and
home of each student are illustrated as Fig. 1.
Fig. 1. An example of the node of school, bus stops and home of students
For the proposed model, sets, parameters and decision variables are as follows:
Sets:
I is the set of all bus stops.
V is the set of all bus stops and a school, that is, V ¼ I [ f0g; where 0 is a school
node.
V 0 ¼ V [ fdg; where d is a dummy node which is the first node of each path.
L is the set of all students.
Parameters:
dij is the distance between nodes i and j for all i; j 2 I:
0; if i ¼ d; j 2 I
dij ¼ where M is a large number.
M; if i ¼ d; j ¼ 0;
sil is the satisfaction score of a student l waiting a school bus at bus stop i:
290 A. Lekburapa et al.
Before developing the capacity constraint of Bektas et al. [14], we give its formula
as follows:
where ui is the total number of students after pick-up from bus stop i; and qi is the
number of students waiting at bus stop i. In our focused problem, the number of
students in each bus stop is unknown because the number ofP students is obtained from
the selection of the model. So, the variable qi is replaced by yil : We also change the
l2L
index to make it easier to understand. The formula is as follows:
P P P
ui uj þ Qxij þ Q yil yjl xji Q yjl 8i; j 2 I with i 6¼ j
P l2L l2L l2L
ui yil 8i 2 I
P
l2L
ui yil xdi þ Qxdi Q 8i 2 I:
l2L
We can see that the first and third constraints are nonlinear. Then, the modification is
obtained by linearization to make it easier to find theanswer. To make alinearization,
P P
we will introduce new decision variables as zji ¼ Q yil yjl xji and pi ¼
P l2L l2L
yil xdi : Then, the capacity constraints can be rewritten as follows:
l2L
A New Integer Programming Model 291
P
ui uj þ Qxij þ zji Q yjl 8i; j 2 I with i 6¼ j
l2L
zji Qxji P P 8i; j 2 I with i 6¼ j
zji Q yil yjl 8i; j 2 I with i 6¼ j
P l2L P l2L
Q yil yjl ð1 xji ÞQ zji 8i; j 2 I with i 6¼ j
P
l2L l2L
ui yil 8i 2 I
l2L
ui pi þ Qxdi Q 8i 2 I
pi Qx
Pdi 8i 2 I
pi yil 8i 2 I
P l2L
yil ð1 xdi ÞQ pi 8i 2 I:
l2L
Next, we present a model for maximizing the total student satisfaction with the
improved constraint as follows:
PP
Maximize sil yil ð1Þ
i2I l2L
P
s:t: xi0 k ð2Þ
i2I
X
xdi k ð3Þ
i2I
P
yil ¼ 1 8l 2 L ð4Þ
i2I
P
xij ¼ 1 8i 2 I ð5Þ
j2I [ f0g
P
xij ¼ 1 8j 2 I ð6Þ
i2I [ fdg
P
ui uj þ Qxij þ zji Q yjl 8i; j 2 I with i 6¼ j ð7Þ
l2L
ui pi þ Qxdi Q 8i 2 I ð12Þ
292 A. Lekburapa et al.
pi Qxdi 8i 2 I ð13Þ
P
pi yil 8i 2 I ð14Þ
l2L
P
yil ð1 xdi ÞQ pi 8i 2 I ð15Þ
l2L
pi 0 8i 2 I ð22Þ
For the objective function (1), the total student satisfaction is maximized. In con-
straints (2) and (3), the number of vehicles used does not exceed the number of vehicles
specified. Constraint (4) ensures that each student can only board at one bus
stop. Constraints (5) and (6) ensure that vehicles visiting bus stop i must exit bus stop i:
Constraints (7)–(15) ensure that each path does not occur sub tour by the number of
students and the capacity of buses. Constraints (16)–(18) ensure that the cumulative
distance at a bus stop i; denoted by vi ; does not exceed the limit T: Constraints (19)–
(20) guarantee that the decision variables are binary. Constraints (21)–(22) claim that
decision variables are nonnegative.
In addition, we consider the other model by changing the objective function from
(1) to a different view of the problem, that is, to minimize the total travel distance of all
buses. In this considering this problem, we solve the following objective function:
XX
Minimize dij xij ð23Þ
i2V 0 j2V 0
3 Experimental Results
In this section, an experimental result is given to show the usage of our model. This
experiment solves the SBRP in the case of a kindergarten with 70 students using the
school bus service. Each school bus has 13 seats and a kindergarten has 8 school buses.
A New Integer Programming Model 293
There are 12 bus stops in our experiment. We solve this experiment in two situations by
using CPLEX.
Results of the first and the second objective functions are presented in Table 1 and
Table 2, respectively. In these tables, the first column is the number of buses. The
second column is the path of each bus for picking up students. The number of all
students who are picked up by each bus shows in the third column and the satisfaction
of students (%) for waiting the bus at each bus stop shown in the fourth column. The
final column presents the total travel distance of buses for picking up students at each
bus stop in the path to school.
Table 1. Results from the proposed model with the objective function (1).
Bus Bus route Number of students Total satisfaction score (%) Total distance (km.)
1 d-3-5-0 11 100 20.5
2 d-4-2-0 12 100 37.1
3 d-6-9-0 11 95.46 14
4 d-7-8-0 11 97.23 14.3
5 d-11-1-0 12 99.17 34.3
6 d-12-10-0 13 100 13.7
98.71 133.9
Table 2. Results from the proposed model with the objective function (23).
Bus Bus route Number of students Total satisfaction score (%) Total distance (km.)
1 1-2-0 12 73.33 22.3
2 3-8-0 12 84.17 15
3 5-10-0 10 90 13.35
4 7-4-0 12 76.67 8.9
5 9-6-0 11 83.64 11
6 12-11-0 13 78.46 12.5
80.71 83.05
In Table 1, The total distance of buses is 133.9 km. and the optimal solution is 691
scores, which is calculated as 98.71%. It implies that all students are extremely satisfied
using the school bus service. Table 2 shows the solution of the second objective
function. The total distance of buses is 83.05 km. The total satisfaction score is 565.
A comparison of the satisfaction score and the total distance of the first and the second
objective functions is given in Table 3. The solutions for these two objective functions
give different values. The choice to use to solve the problem depends on the purpose of
the user of this algorithm.
294 A. Lekburapa et al.
Figure 2 shows an example route in the first model. It can be seen that each student
travels to the designated bus stop. The distance for each student will vary depending on
the satisfaction score the student provides. The path of the school bus is quite
complicated because we do not take into account the shortest route. But we want the
distance of each route to not exceed the specified distance.
Figure 3 shows an example route in the second model. The satisfaction of the
students with the bus stop was 80.71%. The shortest distance is found caused a
reducing travel time and expenses. It follows that the school can be saved money and
so the school bus price for students is reduced. As traveling time is reduced, students
have more time for homework or family activities.
4 Conclusion
In this paper, we presented two integer linear programming models formulated to solve
the SBRP with capacity and distance constraints. In these two models, students can be
assigned to each bus stop based on the choice of an objective function. The first model
is to maximize the total satisfaction score of students, and the second model is to
minimize the total travel distance of buses. The selection of the model will depend on
the needs of the user. That is, if the user takes the student’s needs in mind, the first
objective function should be chosen. If the user takes into account the distance, time or
cost of the travel, then the second objective function should be chosen.
Acknowledgement. This work was supported by Thammasat University Research Unit in Fixed
Points and Optimization.
References
1. Mpeta, K.N., Munapo, E., Ngwato, M.: Route optimization for residential solid waste
collection: Mmabatho case study. In: Intelligent Computing and Optimization 2019.
Advances in Intelligent Systems and Computing, vol. 1072, pp. 506–520 (2020)
2. Dantzig, G.B., Ramser, J.H.: The Truck dispatching problem. INFORMS. Manage. Sci. 6(1),
80–91 (1959)
3. Tarantilis, C.D., Ioannou, G., Kiranoudis, C.T., Prastacos, G.P.: Solving the open vehicle
routing problem via a single parameter metaheuristic algorithm. J. Oper. Res. Soc. 56(5),
588–596 (2005)
4. Newton, R.M., Thomas, W.H.: Design of school bus routes by computer. Socio-Econ. Plan.
Sci. 3(1), 75–85 (1969)
5. Braca, J., Bramel, J., Posner, B., Simchi-Levi, D.: A computerized approach to the New
York City school bus routing problem. IIE Trans. 29, 693–702 (1997)
6. Jalel, E., Rafaa, M.: The urban bus routing problem in the Tunisian case by the hybrid
artificial ant colony algorithm. Swarm Evol. Comput. 2, 15–24 (2012)
7. Kim, T., Park, B.J.: Model and algorithm for solving school bus problem. J. Emerg. Trends
Comput. Inf. Sci. 4(8), 596–600 (2013)
8. Hashi. E.K., Hasan, M.R., Zaman, M.S.: GIS based heuristic solution of the vehicle routing
problem to optimize the school bus routing and scheduling. In: Computer and Information
Technology (ICCIT), 2016 19th International Conference, pp. 56–60. IEEE (2016)
9. Park, J., Kim, B.I.: The school bus routing problem: a review. Eur. J. Oper. Res. 202, 311–
319 (2010)
296 A. Lekburapa et al.
10. Ellegood, W.A., Solomon, S., North, J., Campbell, J.F.: School bus routing problem:
contemporary trends and research directions. Omega 95, 102056 (2020)
11. Bodin, L.D., Berman, L.: Routing and scheduling of school buses by computer. Transp. Sci.
13(2), 113–129 (1979)
12. Schittekat, P., Sevaux, M., Sorensen, K.: A mathematical formulation for a school bus
routing problem. In: Proceedings of the IEEE 2006 International Conference on Service
Systems and Service Management, Troyes, France (2006)
13. Eldrandaly, K.A., Abdallah, A.F.: A novel GIS-based decision-making framework for the
school bus routing problem. Geo-spat. Inform. Sci. 15(1), 51–59 (2012)
14. Bektas, T., Elmastas, S.: Solving school bus routing problems through integer programming.
J. Oper. Res. Soc. 58(12), 1599–1604 (2007)
Distributed Optimisation of Perfect Preventive
Maintenance and Component Replacement
Schedules Using SPEA2
1 Introduction
2 Related Work
Muhammad Ali Ismail [15] has presented his work by using a master-slave para-
digm on a Beowulf Linux cluster using MPI programming library. With the help of this
he has written a pseudo code that first initializes the basic population and also the
number of nodes present in the cluster. Then it assigns the fitness function to the slave
or other nodes in the cluster. The slave ones compute the fitness objective and do the
mutation. After this, they send back the result to the master where it makes the final
counterpart. This paper presented a view of implementation and realization of algo-
rithms on a parallel architecture.
Lim, Y. Ong, Y. Jin, B. Sendhoff, and B. Lee, proposed a Hierarchical Parallel
Genetic Algorithm framework (GE-HPGA) [16] based on standard grid technologies.
The framework offers a comprehensive solution for efficient parallel evolutionary
design of problems with computationally expensive fitness functions, by providing
novel features that conceals the complexity of a Grid environment through an extended
GridRPC API and a metascheduler for automatic resource discovery (Fig. 1).
4 MoHPBPSPEA2 Implementation
In the proposed approach, the problem is first decomposed based on the number of
objectives, i.e., tasks and population P, on individual node Ni. Then the second stage
deals with population balancing by parallel decomposition of the given population on
each node. Each node is offered P/N number of individuals in the population, where P
is the population size and N is the number of nodes considered for solving the problem.
The smaller population or subpopulation; which forms part of the jobs or tasks is then
assigned to each node. Each node further distributes its workload or job among its
working threads and the jobs are performed in parallel in the cluster. This dual-phase
distribution approach also contributes to the efficiency and effectiveness of the
developed MoHBPSPEA2.
The dataset is first processed and stored in the datastore and almost all the time-
consuming computations and evaluations- population generation, fitness assignment,
selection and objective functions are moved to the mapper for processing while the
output is sent to the reducer as input for crossover and mutation, finally, the reducer
output the results. These strategies help to save time of dataset processing (input-output
process) that would have been done at different stages of the entire processing thereby
saving both computation time and communication time amongst the nodes. Then the
results from all the participating nodes in the clusters are combined and the results sent
to the server for execution of local EA for best results which are also displayed by the
server.
Distributed Optimisation of Perfect Preventive Maintenance 303
5 Results
A set of optimal PPMR schedules for the fuel oil system was obtained for (i) serial
processing and (ii) parallel and distributed processing. Also, a set of trade-off PM
schedules shown in the Pareto frontier of Table 1 and Fig. 6; Table 2 and Fig. 5 is
produced. Each of the solutions or PM schedule represents an optimal trade-off
between unavailability and cost.
5.1 Discussion
A total number of 53 optimal solutions are obtained under FOSS PPMR scheduling
optimisation via serial and 64 optimal solutions via parallel and distributed process.
The first and last 5 of the optimal PPM schedules for serial and parallel and distributed
processes are as presented in the Table 1 and Table 2 above respectively and their
graphic representation as shown in Figs. 5 and 6 respectively. Their statistical tool
result comparison is shown in Table 3.
Distributed Optimisation of Perfect Preventive Maintenance 305
Table 1. Optimal PPMR schedules – parallel and distributed optimisation via NUST Cluster of
Servers
Fig. 5. Pareto frontier of PPMR schedules via serial optimisation on a NUST Server
Fig. 6. Pareto frontier of PPMR schedules via parallel and distributed optimisation on NUST
Cluster of Servers
Distributed Optimisation of Perfect Preventive Maintenance 307
Table 3. Statistical tool comparison for serial on server versus parallel and distributed
optimisation of FOSS PPMR scheduling
From Table 3 above, the unavailability minimum and cost minimum values for
both serial and, parallel and distributed processes are the same.
However, this is not the same for cost, via parallel and distributed process has cost
maximum value of 0.012276 greater than for serial and, the difference in range is also
0.012276 greater than for serial. In the same vein, parallel and distributed process has
cost maximum value of 166000 with range variance of 166000 greater than serial. This
implies that parallel and distributed process explored wider space into the feasible
region than serial, and also obtained more optimal PPMR schedules than serial.
The unavailability standard deviation for parallel and distributed process is 0.00151
smaller than for serial. But the reverse is the case for cost standard deviation as parallel
and distributed process is 12026 greater than for serial. Similarly, the unavailability
mean for parallel and distributed process is 0.014682 smaller than for serial. But the
cost mean for parallel and distributed process is 16347 greater than for serial. However,
to have a better and clearer perception of what happened in the optimisation of
FOSS PPMR schedules, the distances in objective function values for both via serial
and, parallel and distributed processes are required to be established further.
From Table 2 (via serial), the least optimal schedule has unavailability as
0.094808367 and cost as 4161191 in generation 500 while the highest optimal schedule
was also found in generation 500 with 0.37103334 as unavailability and cost as
2741967. While the least optimal schedule in Table 1 (via parallel and distributed
process) has unavailability as 0.094808367 and cost as 4373191 at generation 497 and
the highest optimal schedule obtained at generation 499 has unavailability as
0.383308865 and 2704967 as cost. The unavailability and cost variance for serial are
0.276225 and 1419224 respectively and the unavailability and cost variance for parallel
and distributed process are 0.2885 and 1668224 respectively.
Comparing the two objective functions variance values above, it is seen that dis-
tances (spread) between the optimal solutions obtained by parallel and distributed
process shrunk more than that for serial. This means that the optimal PPMR schedules
308 A. O. Ikechukwu et al.
obtained via parallel and distributed process are closer in distance to themselves than
that of serial and this satisfies one of the goals of multiobjective optimisation that says
the smaller the distances of the objective function values between one solution and the
other obtained non-dominated solutions, the better diverse the non-dominated solutions
are.
Interestingly, from Fig. 5 (serial), the first and last optimal solutions of the Pareto
frontier for serial have objective functions as approximately 0.38 unavailability and
2.810 cost and; 0.08 unavailability and 4.310 cost respectively, giving objective
function space value difference as 0.30 and 1.500 for unavailability and cost respec-
tively. In the same vein, from Fig. 6 (parallel and distributed process), the first and last
optimal solutions of the Pareto frontier for parallel and distributed process have
objective functions as approximately 0.39 unavailability and 2.810 cost and; 0.07
unavailability and 4.410 cost respectively, giving objective function space value dif-
ference as 0.32 and 1.600. From the analysis, the difference of the objective functions
for parallel and distributed process is seen to strike a balance more than that for serial.
That is, the optimal schedules (solutions) obtained via parallel and distributed process
has the objective functions (unavailability and cost) better trade-off. It can be seen from
Fig. 5, Fig. 6, Table 1 and Table 2 that the PPMR schedules obtained via parallel and
distributed process are superior to PPMR schedules obtained via serial.
The computational time for optimisation via parallel and distributed process is
3792 s lower than the time for serial as shown in Table 3. In other words, parallel and
distributed process improved the optimisation run time by 79%.
6 Conclusion
(4) Investigate and demonstrate the effect of PPMR schedules execution on more than
4 nodes or more than one cluster. This will significantly further reduce the com-
putational time.
References
1. Nggada, S.H.: Reducing component time to dispose through gained life. Int. J. Adv. Sci.
Technol. 35(4), 103–118 (2011)
2. Ikechukwu, O.A, Nggada, S.H, Quenum, J.G, Samara, K.: Optimisation of perfect
preventive maintenance and component replacement schedules (PPMCRS) Using SPEA2.
In: Proceedings of ITT 2019 6th International Conference on Emerging Technologies on
Blockchain and Internet of Things, HCT, Women Campus, Ras Al Khaimah, UAE 20–21
November 2019. IEEE inspec access no: 19569157 (2019). https://doi.org/10.1109/itt48889.
2019.9075124. ISBN: 978-1-7281-5061-1
3. Pawar, R.S.: Parallel and distributed evolutionary algorithms. seminar computational
intelligence in multi-agent systems, 10 July 2018
4. Cantú-Paz, E.: A survey of parallel genetic algorithms. Calculateurs Paralleles, Reseaux et
Systems Repartis 10(2), 141–171 (1998)
5. Cantú-Paz, E., Goldberg, D.E.: Are multiple runs of genetic algorithms better than one? In:
Proceedings of the 2003 International Conference on Genetic and Evolutionary Computation
(GECCO 2003) (2003)
6. Hiroyasu, T., Kaneko, M., Hatanaka, K.: A parallel genetic algorithm with distributed
environment scheme.In: Proceedings of the 1999 IEEE International Conference on
Systems, Man, and Cybernetics (SMC 1999) (1999)
7. Abramson, D., Abela, J.: A parallel genetic algorithm for solving the school timetabling
problem. In: Proceedings of the 15th Australian Computer Science Conference, Hobart,
pp 1–11, February 1992
8. De Toro, F., Ortega, J., Fernandez, J., Diaz, A.: PSFGA: a parallel genetic algorithm for
multiobjective optimization. In: Proceedings of the 10th Euromicro Workshop on Parallel,
Distributed and Network-based Processing (2002)
9. Schaffer, J.D.: Multiple objective optimization with vector evaluated genetic algorithms. In:
Proceedings of the International Conference on Genetic Algorithm and their applications
(1985)
10. Fonseca, C.M., Fleming, P.J.: Multiobjective genetic algorithms. In: IEEE Colloquium on
Genetic Algorithms for Control Systems Engineering (Digest No. 1993/130) (1993)
11. Horn, J., Nafpliotis, N., Goldberg, D.E.: A Niched Pareto genetic algorithm for
multiobjective optimization. In: Proceedings of the 1st IEEE Conference on Evolutionary
Computation. IEEE World Congress on Computational Intelligence (1994)
12. Murata, T., Ishibuchi, H.: MOGA: multi-objective genetic algorithms. In: Proceedings of the
1995 IEEE International Conference on Evolutionary Computation, Perth (1995)
13. Srinivas, N., Deb, K.: Multiobjective optimization using nondominated sorting in genetic
algorithms. J. Evol. Comput. 2(3), 221–248 (1994)
14. Zitzler, E., Thiele, L.: Multiobjective evolutionary algorithms: a comparative case study and
the strength Pareto approach. IEEE Trans. Evol. Comput. 3(4), 257–271 (1999)
15. Ismail, M.A.: Parallel Genetic Algorithms (PGAS): master slave paradigm approach using
MPI, ETech (2004)
310 A. O. Ikechukwu et al.
16. Lim, D., Ong, Y., Jin, Y., Sendhoff, B., Lee, B.: Efficient hierarchical parallel genetic
algorithms using grid computing. Proc. Future Gener. Comput. Syst. 23(4), 658–670 (2007)
17. Sefrioui, M., Srinivas, K., Periaux, J.: Aerodyanmic shape optimization using a hierarchical
genetic algorithm. In: European Conference on Computational Methods in Applied Science
and Engineering (ECCOMAS 2000) (2000)
18. Verma, A., Llora, X., Goldberg, D.E., Campbell, R.H.: Scaling genetic algorithms using
MapReduce. In: Proceedings of the 9th International Conference on Intelligent Systems
Design and Applications, ISDA 2009 (2009)
19. Nggada, S.H.: Multi-objective system optimisation with respect to availability, maintain-
ability and cost. Ph.D. dissertation, The University of Hull, vol. 3 2012 (2012)
20. Zitzler, E.K., Deb, L.T.: Comparison of multiobjective evolutionary algorithms: Empirical
results (revised version). Technical Report 70, Computer Engineering and Networks
Laboratory (TIK), Swiss Federal Institute of Technology (ETH) Zurich, Gloriastrasse 35,
CH 8092 Zurich, Switzerland (1999)
A Framework for Traffic Sign Detection
Based on Fuzzy Image Processing
and Hu Features
1 Introduction
Traffic sign is an indispensable component to assist traffic police and to pro-
vide safety mechanism by controlling flow of traffic and guiding walkers and
drivers in every aspect of road environment. Automatic traffic sign recognition
system using computer vision attracts special interest in real time applications
such as intelligent vehicles and traffic maintenance. In case of automatic TSR
system, on board mounted module is used to detect and return the information
of traffic signs for proper navigation of the vehicle. Regarding transportation
maintenance, well positioning and accurate and exact information of traffic signs
definitely enhance road security and safety. In a word, TSR is an important part
of Intelligent Transportation System (ITS) [1,2,4].
Traffic sign is designed using distinct colours and shapes, which are really
salient and visible, easily discriminable by human from the environment. Most of
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 311–325, 2021.
https://doi.org/10.1007/978-3-030-68154-8_30
312 Z. Abedin and K. Deb
the signs are made using red, blue and yellow colors where shapes are triangular,
circular, square and rectangle. A sign is made distinctive by three things :border
color, pictogram and background. The signs are installed in a location along side
with the road, so that driver can easily detect the signs [11]. Also, the shape and
colors determine the categories of signs. The properties of Bangladesh road signs
are outlined in Table 1 and some examples of traffic signs are illustrated in the
Fig. 1.
countries. Another important research gap is that most of the research did not
consider the scenario of night mode sign detection [10].
In the light of above situation, the aim of this research is to propose a frame-
work for the detection of Bangladesh traffic signs, where along with the com-
mon challenges of the detection algorithm, night mode scenarios are emphasized.
The key contributions of the framework are intensity transformation using fuzzy
interference that can handle uncertainties in image processing and a computing
tool provided for incorporating human knowledge, segmentation by a statisti-
cal threshold and shape verification using geometric invariant Hu moments and
QDA model.
The rest of the paper is structured as such that Sect. 2 presents related
research, Sect. 3 presents the proposed framework while result and analysis is
demonstrated in Sect. 4. Finally, conclusion is drawn in Sect. 5.
2 Related Works
To detect and localize the RoIs from an image is called Traffic Sign Detection
(TSD) which is the first step of TSR. According to the comprehensive reviews
regarding TSD accomplished by the contributors [9,10], the major categorizes of
TSD algorithms are color based, shape based and the machine learning models.
In color based method, the images are segmented using a threshold where the
threshold is chosen on the basis of color information while in shape base method,
shapes of the signs are detected using some popular edge detection algorithms
such as Hough transformation. Recently, machine learning based sign detec-
tion has become very popular due to better accuracy. In this approach, pixels
are classified from the background by using different features and classification
models.
In [11], a color based detection algorithm was proposed for Spanish traffic
signs where the threshold was determined by observing the Hue and Satura-
tion histogram of HSI images of the manually cropped traffic signs. Also, this
algorithm used Distances to Borders (DtBs) vectors and SVM model to ver-
ify the shapes of the signs. Similar segmentation method was applied by the
authors of [4] where the shape verification was implemented by Hu features and
the cross correlation values of the detected shapes and the ground truth shapes
were calculated and the shapes within an empirical threshold were considered as
314 Z. Abedin and K. Deb
positive shapes. Similarly, in [8], the authors proposed a detection algorithm for
Bangladeshi road signs using YCbCr color space and used statistical threshold
values for segmentation and DtBs vectors to verify shapes.
In [12], a shape and color based detection algorithm was proposed with f-
measure of 0.93. The potential region of interests were searched in the scene
area by using Maximally Stable Extremal Regions (MSERs) and then, used
HSV threshold to detect the sign regions based on text. The detected regions
were further filtered by using temporal information through consecutive images.
Yang et al. proposed a computational efficient detection algorithm which can
detect signs at a rate of six frames per second on image of size 1360 × 800 [13].
He developed probability model for red and blue channel to increase the traffic
colors visibility using Gaussian distribution and Ohta space; then, the MSERs
were applied to locate potential RoIs. Next, the RoIs are classified either signs
or not using HOG features and SVM classifier.
The authors of [2] presented a machine learning based approach to segment
the image where positive colors are converted to foreground and others are
mapped to black. Mean shift clustering algorithm based on color information
was used to cluster the pixels into regions and then, centroid of the cluster was
used to classify the cluster whether it is sign or not by random forest. In addi-
tion, a shape verification was included in the framework to discard the false alarm
and this verification was based on log-polar transformation of the detected blob
and correlations values with the ground truth. The authors of [15] proposed an
efficient detection algorithm by using AdaBoost and Support Vector Regression
(SVR) for discriminating sign area from the background where a salient model
was designed by using specific sign features (shape, color and spatial informa-
tion). By employing the salient information, a feature pyramid was developed
to train an AdaBoost model for the detection of sign candidates in the image
and a SVR model was trained using Bag of Word (BoW) histograms to classify
the true traffic signs from the candidates. In [1], the chromatic pixels were clus-
tered using k-NN algorithm taking the values of a and b in the L*a*b space and
then, by classifying these clusters using Support Vector Machine. Alongside, the
achromatic segmentation was implemented by threshold in HSI space and shape
classification was done by Fourier descriptors of the regions and SVM model.
Most of the machine learning algorithms methods are based on handcrafted
features to identify signs. Recently, deep learning has been very popular in the
computer vision research to detect objects due to their the capabilities of fea-
tures representation from the raw pixels in the image. The detection algorithm
proposed in [19], used dark area sensitive tone mapping to enhance the illumi-
nation of dark regions of the image and used optimized version of YOLOV3 to
detect signs. The authors of [20] used to detect signs using Convolution Neural
Network (CNN) where modified version of YOLOv2 was applied.
An image segmentation by fuzzy rule based methods in YCbCr color space
was presented where triangular membership function and weighted average
defuzzification approach were selected [18].
A Framework for Traffic Sign Detection 315
3 Detection Framework
In general, the TSR is comprised of two steps: sign detection from an image or
video frame and sign classification of the detection. This research concentrates
on the detection of the signs from an image as a mandatory step before sign
recognition. The proposed framework for traffic signs detection is illustrated in
Fig. 3. According to the framework, the image is captured by a camera and
then some pre-processing such as reshaping of the image into 400 × 300 pixels
and color space mapping of RGB to HSV are applied. After pre-processing, the
image is transformed into gray level by using fuzzy inference where the inputs
are the crisp value of hue and saturation of a pixel. Then, the gray image is seg-
mented by applying a statistical threshold. The segmented image is subjected
to some processing such as morphological closing and filtering using area and
aspect ratio to discard some unwanted regions. Next, the shape of remaining
regions are verified by extracting Hu features and a trained Quadratic Discrim-
inate Analysis (QDA) model. The regions which are not classified as circle or
triangle or rectangle are discarded in this stage. After shape verification, the con-
nected components remaining in the binary image are considered as the Region
of Interest (RoI). Finally, the RoIs are detected as the potential traffic signs in
the input image by bounding box parameters.
3.1 Preprocessing
Nonlinear color space mapping of RGB to Hue Saturation Values (HSV) is
applied in the input image to reduce the illumination sensitivity of the RGB
model. Hue and saturation levels of HSV image are isolated for the crisp inputs
in the fuzzy inference system.
In this work, we use Mamdani Fuzzy Inference System (FIS) to map the hue and
saturation values of a pixel to gray value. This inference consists of mainly three
steps: fuzzification, inference using fuzzy knowledge base and defuzzification [3,7]
316 Z. Abedin and K. Deb
with two inputs and one output. The inputs are the hue and saturation value of a
pixel and the output is the gray value of a pixel.
Fuzzification: it converts crisp sets into fuzzy sets with the assignment of degree
of membership values. There are many techniques for fuzzification process. Here,
the fuzzification is based on intuition of an expert. A fuzzy set is defined by lin-
guistic variable where degree of the membership values is assigned in the range
(0, 1) by a characteristic function. Intuition is used to make the membership func-
tions and the knowledge of an expert in the domain is utilized to develop the intu-
ition. For the universe of ‘hue’, we consider five linguistic labels such as reddish-1,
reddish-2, bluish, noise-1, noise-2, for ‘saturation’, they are reddish and bluish and
for the output they are reddish, bluish and black. To assign the membership val-
ues to each linguistic variable, triangular and trapezoidal functions are used. The
Table 2 presents the parameters values of each membership function.
Inference: it is the main step of FIS where the fuzzy knowledge base is used
to infer a set of rules on the input fuzzy set. In fuzzy logic systems, the fuzzy
knowledge base is developed by the rules and linguistic variables based on the
fuzzy set theory to do inference. This knowledge base is constructed by using a
set of If-Then rules. The structure of the rules in the standard Mamdani FIS is
as follows: the kth rule is Rk : IF x1 is F1k and . . . and xp is Fpk , THEN y is Gk ,
where p is the number of inputs. Example of a rule is R1 = ‘If hue is reddish-1
AND saturation is reddish, THEN output is reddish’. In our knowledge base,
five rules are incorporated.
The inference is done by four steps. Step1: the inputs are fuzzified. Here
the crisp value of hue and saturation value of a pixel are converted to fuzzy
variables. Step2: all the antecedents of a rule are combined to yield a single
fuzzy value using the fuzzy logic max operation, depending on whether the parts
are connected by ORs or by ANDs. Step3: the output of each rule is inferred
by an implication method. We use AND operator for the implication which do
the min operation. Step4: This step aggregates the output of each rule to yield
a single fuzzy output. Here fuzzy max operator is used.
A Framework for Traffic Sign Detection 317
3.3 Segmentation
The gray image, obtained by fuzzy inference, is converted to binary image using
the statistical threshold presented by the Eq. 1 which is derived on the statistical
parameters of the image.
1, if I(x, y) > μ + σ
Binaryimage (x, y) = (1)
0, otherwise
where, μ is mean and σ is standard deviation of the gray image, I(x, y).
In this step, regions after the post processing, are verified to filter false alarm.
To verify the regions, geometric invariant Hu moments are extracted for every
regions and Quadratic Linear discriminant (QDA) classifier is used to verify the
shapes. The steps of shape classification are depicted in Fig. 5. After the shape
verification, the regions which are not triangular or circular or rectangle are
filtered.
According to [3,17], the 2D moment of order(p + q) of a digital image f(x,y)
of size M X N is defined as
M −1 N
−1
mpq = xp y q f (x, y) (2)
x=0 y=0
M −1 N
−1
μpq = (x − x̄)p (y − ȳ)q f (x, y) (3)
x=0 y=0
h6 = (η20 − η02 )((η30 + η12 )2 − (η21 + η03 )2 ) + 4η11 (η30 + η12 )(η21 + η03 ) (10)
h7 = (3η21 − η03 )(η30 + η12 )((η30 + η12 )2 − 3(η21 + η03 )2 )
(11)
+(3η12 − η30 )(η21 + η03 )(3(η30 + η12 )2 − (η21 + η03 )2 )
The classification model, discriminate analysis is defined by the Eq. 12.
where δk (x) is the discriminate function and definition of this is given by the
Eq. 13.
1 1
deltak (x) = − log |Σk | − (x − μk )T Σk−1 (x − μk ) (13)
2 2
where Σk is co-variance matrix and πk is prior probability.
3.6 Detection
After the shape verification, the potential candidate of sign region is finalized.
The bounding box parameters of the remaining connected component in the
binary image after shape filtering, are determined. Then, using these parameters,
the traffic sign area is detected in the input frame and can be easily cropped for
recognition of the sign.
Fig. 6. Steps of detection framework (image from Bangladeshi road scene): a) input
image b) HSV image c) fuzzified image d) binary image e) after morphological process-
ing f) after filtering g) after shape verification h) detection of sign.
Fig. 7. Steps of detection framework (image from GTSD): a) input image b) HSV
image c) fuzzified image d) binary image e) after morphological processing f) after
filtering g) after shape verification h) detection of sign.
A Framework for Traffic Sign Detection 321
Fig. 8. Steps of detection framework (night mode): a) input image b) HSV image c)
fuzzified image d) binary image e) after morphological processing f) after filtering g)
after shape verification h) detection of sign.
Fig. 11. Illustration of IoU (red BB is ground truth and blue BB is detected with TP
is 1, FN = 0 and =0).
A data sets of 900 shapes for circle, triangle and rectangle are formed where the
shapes are taken from the segmented road images. Next, the seven Hu features
are extracted for every shape and a feature vector of size 900 × 8 is created.
Then, the models are trained with 10-fold cross validation. The Fig. 12 illustrates
the accuracy of different models. From Fig. 12, it is evident that the accuracy
of the QDA model is better than others, with 96% accuracy. Hence, this model
is deployed in our detection framework for shape verification.
A Framework for Traffic Sign Detection 323
5 Conclusion
In this paper, a detection framework is presented by utilizing fuzzy image pro-
cessing where crisp sets of hue and saturation are converted into fuzzy sets and
Mamdani Fuzzy Inference scheme is applied to transform those into gray val-
ues. A statistical threshold is applied to segment the fuzzified image. After some
filtering, to reduce the false alarm, shape classification is done by QDA model
trained on Hu features. In addition, an analysis is drawn to compare accuracy of
some popular models where QDA outperforms. The framework is evaluated by
calculating the IoU scores where the average recall and precision rate are 94.19%
and 93.69% respectively. The future work aims recognizing the traffic sign based
on deep learning model.
References
1. Lillo, C., Mora, J., Figuera, P., Rojo, Á.: Traffic sign segmentation and classification
using statistical learning methods. Neurocomputing 15(3), 286–299 (2015)
2. Ellahyani, A., Ansari, M.: Mean shift and log-polar transform for road sign detec-
tion. Multimed. Tools Appl. 76, 24495–24513 (2017)
3. Rafael, C.G., Richard, E.W.: Digital Image Processing, 3rd edn. Pearson Educa-
tion, Chennai (2009)
4. Ayoub, E., Mohamed, E.A., Ilyas, E.J.: Traffic sign detection and recognition based
on random forests. Appl. Soft Comput. 46, 805–815 (2015)
5. Road Sign Manual Volume-1.pdf. www.rhd.gov.bd/documents/ConvDocs/.
Accessed 4 Nov 2020
6. Hung, P.D., Kien, N.N, : SSD-Mobilenet implementation for classifying fish species.
In: Vasant, P., Zelinka, I., Weber, G.W. (eds.) Intelligent Computing and Optimiza-
tion. ICO 2019, Advances in Intelligent Systems and Computing, vol. 1072, pp.
399–408. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-33585-440
7. Borse, K., Agnihotri, P.G.: Prediction of crop yields based on fuzzy rule-based sys-
tem (FRBS) using the Takagi Sugeno-Kang approach. In: Vasant, P., Zelinka, I.,
Weber, G.W. (eds.) Intelligent Computing and Optimization. ICO 2018, Advances
in Intelligent Systems and Computing, vol. 866, pp. 438–447. Springer, Cham
(2018). https://doi.org/10.1007/978-3-030-00979-3 46
8. Soumen, C., Kaushik, D.: Bangladeshi road sign detection based on YCbCr color
model and DtBs vector. In: 2015 International Conference on Computer and Infor-
mation Engineering (ICCIE) on Proceedings, Rajshahi, Bangladesh, pp. 158–161.
IEEE (2015)
9. Safat, B.W., Majid, A.A., Mahammad, A.H., Aini, H., Salina, A.S., Pin, J.K.,
Muhamad, B.M.: Vision Based traffic sign detection and recognition systems: cur-
rent trends and challenges. Sensors 19(9), 2093 (2019)
10. Chunsheng, L., Shuang, L., Faliang, C., Yinhai, W.: Machine vision based traffic
sign detection methods: review, analyses and perspectives. IEEE Access 7, 86578–
86596 (2019)
11. Maldonado, S., Lafuente, S., Gil, P., Gomez, H., Lopez, F.: Road sign detection
and recognition based on support vector machines. IEEE Trans. Intell. Transp.
Syst. 8(2), 264–278 (2007)
12. Greenhalgh, J., Mirmehdi, M.: Recognizing text-based traffic signs. IEEE Trans.
Intell. Transp. Syst. 16(3), 1360–1369 (2015)
A Framework for Traffic Sign Detection 325
13. Yang, Y., Luo, H., Xu, H., Wu, F.: Towards real time traffic sign detection and
classification. IEEE Trans. Intell. Transp. Syst. 17(7), 2022–2031 (2016)
14. Cao, J., Song, C., Peng, S., Xiao, F., Song, S.: Improved Traffic sign detection and
recognition algorithm for intelligent vehicles. Sensors 19(18), 4021 (2019)
15. Tao, C., Shijian, L.: Accurate and efficient traffic sign detection using discriminative
AdaBoost and support vector regression. IEEE Trans. Veh. Technol. 65(6), 4006–
4015 (2016)
16. Everingham, M., Van, G.L., Williams, C.K.I.: The pascal visual object classes
(VOC) challenge. Int. J. Comput. Vis. 88, 303–338 (2010)
17. Hu, M.K.: Visual pattern recognition by moment in-variants. IRE Trans. Inf. The-
ory 8(2), 179–187 (1962)
18. Alvaro, A.R., Jose, A.M., Felipe, G.C., Sergio, G.G.: Image segmentation using
fuzzy inference system on YCbCr color model. Adv. Sci. Technol. Eng. Syst. J.
2(3), 460–468 (2017)
19. Jameel, A.K., Donghoon, Y., Hyunchul, S.: New dark area sensitive tone mapping
for deep learning based traffic sign recognition. Sensors 18(11), 3776 (2018)
20. Zhang, J., Huang, M., Jin, X., Li, X.: A real time Chinese traffic sign detection
algorithm based on modified YOLOv2. Algorithms 10(4), 127 (2017)
21. Xue, Y., Jiaqi, G., Xiaoli, H., Houjin, C.: Traffic sign detection via graph-based
ranking and segmentation algorithms. IEEE Trans. Syst. Man Cybern. Syst.
45(12), 1509–1521 (2015)
Developing a Framework for Vehicle
Detection, Tracking and Classification
in Traffic Video Surveillance
1 Introduction
As vehicle detection and classification system has attracted many researchers from both
institutions and industry in the field of Intelligent Transportation System (ITS). We
proposed a new method in this paper to effectively and robustly track and classify
vehicles. We also compared our own improved system with various existing algorithms
of detection and tracking them. The suggested approach is evaluated using self-
collected video sequences in condition of motorway driving. Experimental results show
how effective and feasible our approach is.
In recent years the rise of vehicles has caused significant transport problems. As a
result of the increase in automobile traffic, some issues emerged. Due to the rise in
vehicles, the jamming of traffic in cities is a major problem and people’s health is at
risk. Thus, road traffic video monitoring is becoming crucial work on Intelligent
Transportation System (ITS) is attracting considerable attention. The area of automated
video surveillance systems is actually of immense concern because of its safety
impacts.
The objective of this work is vehicle detection, tracking and classifying from the
high quality video sequence. The paper will be carried out to achieve following goals:
detection and tracking of multiple moving vehicles in traffic video sequence. Also
classifying vehicles into their types based on the dimension of the objects. It is an
important and difficult job to monitor frame by frame of identifying objects in video. It
is a vital part of the smart surveillance system because it would not be possible for the
system to obtain coherent temporal information about objects without monitoring
objects, and measures for higher level behavioral analysis. The first step in video
processing is identification of moving objects. Several regions may be used, such as
video monitoring, traffic controls, etc.
According to the relevant works mentioned in Sect. 2, the remainder of this paper is
structured. In Sect. 3, the system architecture and design is suggested. Section 4 and 5
display experimental results and conclusions.
2 Related Work
For vehicle detection, tracking and classification several different methods have been
suggested in the literature review. Detection and tracking of vehicles are major computer
vision issues that serve as the basis for further study of target motion. Vehicle detection
takes advantage of technology such as digital image processing to distinguish the
vehicle and its context [15], multimedia automation [16] statistics etc. In [11] an efficient
algorithm to track multiple people is presented. For feature extraction and neural net-
works for identification, PCA was used in Matthews et al. [12]. A method of measuring
traffic parameters in real time is defined here. It uses a feature-based approach to track
vehicles in congested traffic scenes, along with occlusion logic. Instead of monitoring
whole vehicles, sub-features of the vehicle are monitored for handling occlusions.
However this approach is very expensive in terms of computation [13].
For vehicle identification and spatio-temporal surveillance, Chengcu et al. [1]
suggested a system for adaptive context learning where in order to analyze image/video
segmentation process using SPCPE, this technique uses unsupervised vehicle detection
and spatio-temporal tracking. This work’s difficulty that lacks user interaction and
doesn’t know a priori can pixel belongs to the class. Gupte et al. [2] suggested a
method, where processing is carried out at three levels: raw images, the level of the area
and the level of the vehicle, and classification of two items of the type. The limitation
of this work is cannot classify a large number of category object and error occur for
occlusion. The experiment result shows the highway scene which can fail in case of
narrow road video sequence.
328 R. Saha et al.
Peng et al. [3] implemented a system using data mining techniques for vehicle type
classification where the front of the vehicle was precisely determined using the position
of the license plate and the technique of background-subtraction. Also vehicle type
probabilities were extracted from eigenvectors rather than explicitly deciding the type
of vehicle and using SVM for classification. The limitation of this method is that all
vehicle license plate can’t found so there is an error occur in detection object and also
handling complex.
Sun et al. [4] proposed a structure for a real-time precrash vehicle detection method
utilizing two forms of detection algorithms, such as the generation of multicycle-driven
hypotheses and the verification of appearance-based hypotheses, and the classification
of SVM algorithm used. The shortcoming of this work is that parameter failed in the
case of environment and also cause error in the detection moving object.
Jayasudha et al. [5] set out a summary of road traffic and accident data mining in the
report. They provide an overview of the techniques and applications of data mining in
accident analysis working with different road safety database and also need data dis-
criminator which compares the target class one or more set.
Chouhan et al. [6] suggested a technique for retrieving images using data mining
and image processing techniques to introduce an image’s size, texture and dominant
color factors. If the shape and texture same, they use weighted Euclidean distance.
Chen et al. [7] developed a framework on PCA based vehicle classification
framework where they used for feature extraction segment individual vehicles using
SPCPE algorithm and for classification use two algorithms – Eigenvehicle and PCA-
SVM. The weakness of this method detects vehicle have to use an unsupervised
algorithm SPCP. On-road vehicle detection technique with training linear and non-
linear SVM classifiers using modified HOG based features in Y. Zakaria et al. [9]. For
training dataset, they used the KITTTI dataset.
Vehicle detection, tracking and classification is complicated as any traffic
surveillance system due to the interference of illumination, blurriness and congestion of
narrow road in rural area but exposed to real world problems. This demand makes the
task very challenging for requirements of increasing accuracy.
This paper, we developed a robust vehicle detection, tracking and classification
method by extraction HOG features and Haar like features, with Support Vector
Machine (SVM) and cascade based classifier for classifying the types of vehicles. Then
Kalman filter technique is used. The system increases the tracking and classification
accuracy by using a proper algorithm with less computational complexity and reduces
the processing time.
As we discuss previous related works step about the related works on vehicle detection,
tracking and classification procedure. We notice that there are some limitations of all
the procedure for which reason those methods can’t very efficient framework in the
Intelligent Transportation System. For solving all of the limitations of those methods
we proposed an efficient framework of this field, according to Fig. 6 which can reduce
the limitations of the background work. The system architecture of vehicle detection,
Developing a Framework for Vehicle Detection 329
Tracking and classification System comprises some basic modules: Dataset initializa-
tion, feature extraction, Haar like feature, classifier, tracking and length calculation.
The function of dataset initialization is to set up the database with positive and negative
samples. The feature extraction module is for generating important feature from the
positive samples. Classifier module is for training samples for separate vehicles and
non- vehicles. Tracking module is for detecting all moving vehicles and also reduce the
number of missing objects. Finally, the most important part of the system is classifi-
cation module for classifying vehicle types for better recognition based on the
dimension of the object. Algorithm 3 shown the whole evaluating procedure of Vehicle
detection and Tracking through an Algorithm.
3.1 Dataset
Samples are collected to evaluate the algorithm presented, in various scenes, city
highway, wide roads, narrow roads, etc. In the first stage,7325 samples were collected
for training and testing, including 3425 samples of vehicles (positive samples) and
3900 samples of non-vehicles (negative samples). The vehicle samples include various
vehicle types such as car, truck and buses with various directions such as front, rear,
left and right. In addition, the vehicle samples include both vehicles near the camera-
mounted vehicle, as well as those far away. At both stages, traffics sign etc. Illustration
Fig. 1 and Fig. 2 shows example of training on vehicles and non-vehicles images. The
training samples were resized to 64 by 64 RGB images.
The key element of HOG features extraction in the object detection are as follows:
The images are converted to grayscale input color. The Gamma correction proce-
dure adopts the standardization of the color space (normalized) of the input image; the
aim is to adjust the contrast of the image, decrease the shadow of the local image and
the effect of changes in the illumination. Though it can restrain interference with noise.
Developing a Framework for Vehicle Detection 331
Measure the gradient; primarily with a view to collecting contour details and reducing
light interference. Project gradient to a cell steering gradient. All cells within the block
are normalized; normalization has compressed the light, shadow, and edges, and the
block descriptor is called the HOG descriptor after normalization. Collect HOG fea-
tures from all blocks in the detection space; this step is intended to collect and merge
HOG features into the final classification feature vectors to overlap blocks in the
detection window.
In this Fig. 4 and 5 HOG feature extraction shown from vehicle and non- vehicle
image.
The variants of HOG features use the same technique for extraction of the function
but variations in the parameters below:
1. The number of angles to orientation.
2. The number of bins with histograms.
3. The method by which the pixel gradient is determined.
We may also optionally apply a color transformation to your HOG function vector
and attach binned color characteristics as well as color histograms.
332 R. Saha et al.
1: If (time==0) {
2: Consider a new track for any vehicle detected
3: }
4: Else {
5: For (All current tracks) {
6: New position and track size predict by Kalman-filter
7: Overlap new place with all blobs found
8: If (Overlap (track (a), blob (b))! =0)
9: Mark match_matrix [a] [b];
10: Analyze (match_matrix); }
1: Input: The video sequence that is being monitored by the target specified
2: Output: The outcome of tracking
3: for t 1 to N do
4: if It has been detected for 5 seconds, compared with
the last detection then
5: HOG and SVM to detect the vehicle;
6: If objects are found to be of concern (ROI)
Then
7: Tracking the entity in the bounding box using the Kalman filter
tracking algorithm
8: if It has been detected for 10 seconds
9: Comparing with the last detection then
Normal SVM and HOG to judge whether
3.10 Classification
The classification of vehicle types has recently gained a great deal of research interest.
It follows two main directions on the basis of the vehicle’s shape or appearance.
Classification is achieved by categorizing the vehicles, according to the size of the
vehicles into three groups, namely, large, medium and small. Since the length of the
vectors is easy to find, the length was taken as the parameter for classifying vehicles
according to the given size. When any vehicle object comes to our region of interest,
the length of the vector has been calculated by using length calculation equation,
Classification was carried out for that vehicle then pass the length requirements
(Fig. 9).
336 R. Saha et al.
In this portion, providing our development framework with the implementation pro-
cedure and performance review.
4.2 Implementation
Figure 10(a) shown the results only where there is some false position for detection.
But the false positive rate can be reduced after applying the tracking algorithm which is
shown in Fig. 10 (b), (c) and (d) and also track multiple vehicles. The detection boxes
are in this bounding box red Color, the tracking boxes are the blue boxes.
(a) (b)
(c) (d)
Fig. 10. Vehicle detection and tracking example in different scenes such as (b) urban road with
high light, (c) rural congested road, (d) low light.
Developing a Framework for Vehicle Detection 337
Vehicle tracking in town area is much more difficult because objects are nearby to
each other, while tress or other background might cast shadws both on the road and the
vehicles. Figure 10(b)-(d) shown some tracking results under different scenario where
our proposed system work very satisfactory. By using different time and scene traffic
sequence to check kalam-Filter based tracking system. We can see the system has a
good tracker both videos and images. It can also track fast moving objects. When a
vehicle enters the scene, a new tracking object is detected, a new number is transmitted
and a tracking window is configured for the new vehicle.
We tested the system on highway image sequences. Most vehicles can be suc-
cessfully tracked and classified by the system. Different videos and images were chosen
for testing of different types of vehicles, including 3 types, including buses, cars and
motorbikes. By plugging in haar cascade-based classifier algorithms based on vehicle
extract features or dimensions, we loop through them and draw a rectangle around each
frame of the videos. After those preprocessing, we have integrated the framework that
can automatically extract and classify vehicles into different traffic video scenes.
(e) (f)
(g) (h)
Fig. 11. Vehicle classification example in different types such as (e)-(f) motorbike types of
vehicle, (g) car type of vehicle, (h) Bus type of vehicle
Correct correspondence is the number of points which the proposed method detects
and tracks correctly. And the term for correspondence is the total number of match
points.
Table 3. Comparison with some existing method with our proposed method.
Methods Accuracy False rate
Gabor + SVM 94.5% 5.4%
PCA + NN 85% 7.5%
Wavelet + Gabor +NN 89% 6.2%
LDA 88.8% 5.5%
Hog + SVM + Mean-shift 94% 3.7%
HOG + PCA +SVM +EKF 95% 3.5%
Our proposed method 95.25% 3.7%
Robust and accurate vehicle detection, tracking and classification of frames collected
by a moving vehicle is a significant issue for autonomous self-guided vehicle appli-
cations. This paper provides a summary study of the techniques proposed that were
used in traffic video and images with a very high accuracy of tracking and classification
rate. We have given a comprehensive analysis of the state-of-the-art literature in this
paper dealing with computer vision techniques used in video-based traffic surveillance
and monitoring systems. A HOG and Haar-like feature-based, integrated vehicle
classification architecture is proposed in this paper that integrates several main com-
ponents relevant to an Intelligent Transportation Surveillance System.
The idea of using Histogram of gradients to extract features of different scales and
orientations is key to our approach. The Kalman filter is used to track multiple objects
in video. Classification has done based on the extracting Haar-like feature with cascade
based classifier which is categorize vehicle based on the threshold value. The goal of
the approach presented is to reduce the calculation time. Experimental findings show
that the system is accurate and efficient and that a high performance rate for vehicles
can be achieved by the system.
References
1. Zhang, C., Chen, S., Shy, M., Peeta, S.: Adaptive background learning for vehicle detection
and spatio- temporal tracking. In: Information, Communication and Signal Processing 2003
and the Fourth Pacific Rim Conference on Multimedia.0-7803-8185-8/03/$17.00 0, pp. 797-
801. IEEE (2003). https://doi.org/10.1109/icics.2003.1292566
2. Gupte, S., Masud, O., Martin, R.F.K.: Detection and classification of vehicles. Trans. Intell.
Transp. Syst. 3(1) 37–47 (2002)
3. Peng, Y., Jin, J.S., Luo, S., Xu, M., Au, S., Zhang, Z., Cui, Y.: Vehicle type classification
using data mining techniques. In: Jin, J.S. et al. (ed.) The Era of Interactive Media, LLC,
pp. 325–335. Springer, Cham (2013). https://doi.org/10.1007/978-1-4614-3501-3_27
4. Sun, Z., Miller, R., Bebis, G., DiMeo, D.: A real-time precrash vehicle detection system. In:
Proceedings on the sixth Workshop on Application Computer Vision (WACV 2002),
pp. 253–259. IEEE (2002)
5. Jayasudha, K., Chandrasekar, C.: An overview of data mining in road traffic and accident
analysis. J. Comput. Appl. 4(4), 32–37 (2009)
6. Chouhan, P., Tiwari, M.: Image retrieval using data mining and image processing
techniques, pp. 53–58 (2015). https://doi.org/10.17148/IJIREEICE.31212
Developing a Framework for Vehicle Detection 341
7. Zhang, C., Chen, X., Chen, W.: A PCA-based vehicle classification framework. In:
Proceedings of the 22nd International Conference on Data Engineering Workshops (ICDEW
2006) 0–7695-2571-7/06 $20.00 IEEE, pp. 451–458 (2006)
8. Sun, Z., Bebis, G., Mille, R.: On-road vehicle detection using gabor filters and support vector
machines. &7803-7S03- 3/02/$17.00 02002, pp. 1019–1022. IEEE (2002)
9. Zakaria, Y., Abd El Munim, H. E., Ghoneima, M., Hammad, S.: Modified HOG based on
road vehicle detection method. In: International Journal of Pure and Applied Mathematics,
vol. 118, no. 18, pp. 3277–3285 (2018)
10. Tian, Q., Zhang, L., Wei, Y., Zhao, W., Fei, W.: Vehicle detection and tracking at night. In:
Video Surveillance. iJOE – vol. 9, no. 6 “AIAIP2012”, pp. 60–64 (2013)
11. Tao, H., Sawhney, H., Kumar, R.: A sampling algorithm for tracking multiple objects. In:
Conference Proceedings of the International Workshop on Vision Algorithms (1999). https://
doi.org/10.1007/3-540-44480-7_4
12. Kul, S., Eken, S., Sayar, A.: A concise review on vehicle detection and classification. In:
2017 International Conference on Engineering and Technology (ICET), Antalya, Turkey,
IEEE (2017). https://doi.org/10.1109/IcengTechnol.2017.8308199
13. Cho, H., Rybski, P. E., Zhang, W.: Vision-based bicycle detection and tracking using a
deformable part Model and an EKF Algorithm. In: 13th International Conferenceon
Intelligent Transportation Systems, Funchal, Portugal. IEEE (2010). https://doi.org/10.1109/
ITSC,2010.5624993
14. Kong, J., Qiu, M., Zhang, K.: Authoring multimedia documents through grammatical
specification. In: IEEE ICME 2003, Baltimore, USA, pp. 629–632 (2003)
15. Qiu, M.K., Song, G.L, Kong, J., Zhang, K.: Spatial graph grammars for web information
transformation. In: IEEE Symposium on Visual/Multimedia Languages (VL), Auckland,
New Zealand, pp. 84–91. (2003)
16. Xu, W., Qiu, M., Chen, Z., Su, H.: Intelligent vehicle detection and tracking for highway
driving. In: IEEE International Conference on Multimedia and ExpoWorkshops, pp. 67–72
(2012). https://doi.org/10.1109/ICMEW.2012.19
17. Chitaliya, N.G., Trivedi, A.I.: Feature-extraction using Wavelet- PCA and neural network
for application of object classification and face recognition. In: Second International
Conference on Computer Engineering and Applications, pp. 510–514, IEEE (2010). https://
doi.org/10.1109/ICCEA.2010.104
Advances in Algorithms, Modeling
and Simulation for Intelligent Systems
Modeling and Simulation of Rectangular Sheet
Membrane Using Computational Fluid
Dynamics (CFD)
Abstract. The study demonstrates the modeling and simulation of the flow
phenomena inside the rectangular sheet-shaped membrane module using
Computational fluid dynamics (CFD) based solver. The module was imple-
mented to enhance the quality of effluent generated from the Rubber Industry.
Commercially available CFD software (ANSYS) implemented to mimic the
flow inside the porous membrane. The meshing of the developed model was
done using Gambit software. The grid independency study reports a grid size of
375000 was the best grid for the simulation procedure. To mimic the flow
pattern inside the membrane, the second-order laminar model was considered.
The accuracy of the simulation process is evaluated using error analysis. In the
error analysis, methods like percent bias, Nash-Sutcliffe, and the ratio of RMSE-
observation standard deviation are selected for error analysis. The assessed
estimations of PBIAS, NSE, and RSR are close to ideal value, justifying the
adequacy of the simulation. From model validation, it demonstrates that CFD
predicted values follow the experimental values with high precision.
1 Introduction
Over the last decades, membrane separation processes played a crucial role in the
industrial separation process. Numerous studies focus on the optimal way of using
membrane separation and filtration processes. Computational fluid dynamics
(CFD) techniques provides lots of information regarding the development of the
membrane process. Numerous advancements in membrane technology allowed the
selection procedure of the suitable membrane for different procedure easy and quick.
Membrane filtration has been used in broad range of application like wastewater
treatment [1, 2], protein separation [3], Food Processing industry [4] etc. In membrane
separation and filtration technique, the hydrodynamics of the fluid flow pattern inside
the membrane bed is very crucial. A membrane separation and filtration process is a
combination of free flow inside the membrane module and flow through the porous
membrane. Fluid dynamics description of free flow inside the membrane module is
easy to simulate by using Navier-Stokes equation for this part but most complex one to
reproduce is the flow of the fluid through the porous membrane bed which can be
modeled by coupling Darcy’s law with the Navier Stokes equation. And the validity of
Darcy law for simulating the incompressible flow through the porous zone with small
porosity has been found to be acceptable. The important part is to ensure the continuity
of flow field variables through the interface of laminar flow region and porous flow
region is adequately maintained [5]. Initial simulation of flow inside the membrane was
simulated using laminar condition along with porous wall [6]. Many authors have used
computational fluid dynamics (CFD) for optimizing the membrane processes [7].
Rahimi et al. [8] used hydrophilic polyvinylidene fluoride (PVDF) membrane for
preliminary work and studied using cross-flow velocities of 0.5, 1, and 1.3 m/s.
Fouling of the membrane are analyzed with the help of photographs and micrographs
of Scanning electron microscope. 3D CFD modelling of the developed porous mem-
brane bed was carried out using Fluent 6.2 which was used to pronounce the shear
stress distribution upon the membrane for explaining the fouling process. Particle
deposition pattern on the membrane surface have also been pronounced using discrete
phase model. Ghadiri et al. [9] developed model upon mass and momentum transfer
equation for solute in all phases including feed, solvent and membrane. The flow
behaviour inside the model was simulated using Navier Stokes equation by finite
element method under steady state condition. CFD predicted results are compared with
the experimental data to evaluate the accuracy of the model. A 2D mathematical model
was developed for illustrating seawater purification using direct contact membrane
desalination (DCMD) system [10]. The model was developed by coupling the equa-
tions of conservation for water molecule in three domains of the module and the
governing equations are solved by using finite element method. From the study of Reza
Kazemi [10], found that enhancing the channel length, the temperature of concentration
stream and e⁄(T.d) ratio or reducing the inlet velocity of the concentrated solution,
rejection of the salts can be enhanced.
Such detail CFD based study considering rectangular sheet shaped membrane to
enhance the quality of the rubber industrial effluent has not been reported earlier.
Modeling and Simulation of Rectangular Sheet Membrane 347
In the present paper, numerical finite volume method (FVM) is used to solve the
three-dimensional convective diffusion equations for solute transport in laminar flow
condition over a porous zone in stationary rectangular sheet membrane. CFD simu-
lation is used to predict the distribution pressure drop, velocity profile, wall shear stress
over the static rectangular sheet membrane. The CFD simulated data validated with the
experimental data.
2 Experiment Descriptions
feed tank. Table 1 demonstrate the characterization of the raw feed stream collected
from Rubber Industry of Tripura.
3.1 Assumptions
Following assumptions and ideas are considered for building up the CFD model
describing effluent flow through membrane bed [11]:
I. Effluents generated from rubber industry was considered to be Newtonian fluid
because of its dilution in water.
II. Effluent from rubber industry was considered to be Isothermal and
incompressible.
III. Mathematical model was assumed as laminar single phase pressure based
simple pressure-velocity coupling model for the simulation.
IV. Membrane was considered to be isotropic and homogeneous porous zone
throughout the geometry.
V. According to the necessity of the simulation work, under-relaxation factor
reduced to 0.7–0.3 or lower.
Modeling and Simulation of Rectangular Sheet Membrane 349
VI. Hexahedral grids considered for meshing the model for computational
simplicity.
VII. Large gradient of pressure and swirl velocities resolved by using refined grids.
VIII. The mathematical model developed for simulation work was confined to flow
model only.
Continuity equation
Equation of Continuity for effluent flow through the membrane bed can be defined by
Eq. 1,
! b d d d
V i þ bj þb
k ¼0 ð1Þ
dx dy dz
Darcy’s law
Darcy’s law for flow of effluent through the membrane bed is defined by Eq. 2,
l!
rP ¼ V ð2Þ
a
Momentum equation
Axial momentum equation for effluent flow through the membrane bed is given by
Eq. 3,
! dP
r qu V ¼ þ lr2 u ð3Þ
dx
Equation of radial momentum for rubber industrial effluent flow through membrane
is given by Eq. 4,
! dP
r qu V ¼ þ lr2 u ð4Þ
dr
2P
n 3
Yi Yi 100
6i¼1 7
PBIAS ¼ 6
4 P
m
7
5 ð6Þ
ðYi Þ
i¼1
2 P n 3
2
Y Yi
6 i¼1 i 7
NSE ¼ 1 6
4Pn
7
5 ð7Þ
2
ðYi Ymean Þ
i¼1
2 sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi 3
Pn
6 ðYi Yi Þ2 7
6 7
6 i¼1 7
RSR ¼ 6sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi7 ð8Þ
6 P n 7
4 ðY Y Þ2 5
i mean
i¼1
In Eqs. 6–8, Yi, Y*i , and Ymean represents actual dataset, predicted dataset, and mean
of the actual dataset, respectively. Where, n illustrate total number of observation in
actual dataset. The value of PBIAS, NSE, and RSR are considered to be best, if the
values are close to 0, 1, and 0, respectively. Table 3 shows the error analysis of the
developed CFD model and the result of PBIAS (0.317), RSR (0.022), and NSE (0.99)
are close to the ideal value. Error analysis demonstrate the accuracy of simulation
process and justify the use of CFD to demonstrate the flow through the membrane.
5 Conclusions
The model of the rectangular membrane was developed by using Gambit and effluent
flow pattern inside the membrane was simulated using Fluent. To pronounce outcomes
with least error percentage, hexahedral mesh with least PC memory was utilized for
meshing. From grid independence test, the mesh size of 375000 was selected for
carrying out the simulation work as any further refinement of the grid size does not
show any change in pressure profile of rectangular sheet membrane. The mass
imbalance (kg/sec) study proves that the membrane bed poses high separation effi-
ciency. CFD predicted results are validated against experimental data, where CFD
results follows the results of experiments with high precision and the error percentage
typically varies in the range of 1–5%. CFD simulation provides an insight to hydro-
dynamic properties such as wall shear stress and concentration profile over the
membrane. The error analysis was implemented to evaluate the accuracy of the sim-
ulation process. The methods such as PBIAS, NSE, and RSR are used for error analys.
Determined estimation of NSE, PBIAS, and RSR were 0.99, 0.317, and 0.022.
respectively, which were close to the ideal value. Pronounced results from error
analysis justify the accuracy of the simulation process. Results obtained from the study
can be used to develop a cost-effective membrane separation technique to treat the
rubber industrial effluent.
Modeling and Simulation of Rectangular Sheet Membrane 355
References
1. Chen, Z., Luo, J., Hang, X., Wan, Y.: Physicochemical characterization of tight
nanofiltration membranes for dairy wastewater treatment. J. Memb. Sci. 547, 51–63
(2018). https://doi.org/10.1016/j.memsci.2017.10.037
2. Noor, S.F.M., Ahmad, N., Khattak, M.A., et al.: Application of Sayong Ball Clay Membrane
Filtration for Ni (II) Removal from Industrial Wastewater. J. Taibah Univ. Sci. 11, 949–954
(2017). https://doi.org/10.1016/j.jtusci.2016.11.005
3. Emin, C., Kurnia, E., Katalia, I., Ulbricht, M.: Polyarylsulfone-based blend ultrafiltration
membranes with combined size and charge selectivity for protein separation. Sep. Purif.
Technol. 193, 127–138 (2018). https://doi.org/10.1016/j.seppur.2017.11.008
4. Nath, K., Dave, H.K., Patel, T.M.: Revisiting the recent applications of nanofiltration in food
processing industries: Progress and prognosis. Trends Food Sci. Technol. 73, 12–24 (2018).
https://doi.org/10.1016/j.tifs.2018.01.001
5. Pak, A., Mohammadi, T., Hosseinalipour, S.M., Allahdini, V.: CFD modeling of porous
membranes. Desalination 222, 482–488 (2008). https://doi.org/10.1016/j.desal.2007.01.152
6. Berman, A.S.: Laminar flow in channels with porous walls. J. Appl. Phys. 24, 1232–1235
(1953). https://doi.org/10.1063/1.1721476
7. Karode, S.K.: Laminar flow in channels with porous walls, revisited. J. Memb. Sci. 191,
237–241 (2001). https://doi.org/10.1016/S0376-7388(01)00546-4
8. Rahimi, M., Madaeni, S.S., Abolhasani, M., Alsairafi, A.A.: CFD and experimental studies
of fouling of a microfiltration membrane. Chem. Eng. Process. Process Intensif. 48, 1405–
1413 (2009). https://doi.org/10.1016/j.cep.2009.07.008
9. Ghadiri, M., Asadollahzadeh, M., Hemmati, A.: CFD simulation for separation of ion from
wastewater in a membrane contactor. J. Water Process. Eng. 6, 144–150 (2015). https://doi.
org/10.1016/j.jwpe.2015.04.002
10. Rezakazemi, M.: CFD simulation of seawater purification using direct contact membrane
desalination (DCMD) system. Desalination 443, 323–332 (2018). https://doi.org/10.1016/j.
desal.2017.12.048
11. Banik, A., Bandyopadhyay, T.K., Biswal, S.K.: Computational fluid dynamics simulation of
disc membrane used for improving the quality of effluent produced by the rubber industry.
Int. J. Fluid Mech. Res. 44, 499–512 (2017). https://doi.org/10.1615/InterJFluidMechRes.
2017018630
12. Banik, A., Biswal, S.K., Bandyopadhyay, T.K.: Predicting the optimum operating
parameters and hydrodynamic behavior of rectangular sheet membrane using response
surface methodology coupled with computational fluid dynamics. Chem. Papers 74(9),
2977–2990 (2020). https://doi.org/10.1007/s11696-020-01136-y
13. Myagkov, L., Chirskiy, S., Panchenko, V., et al.: Application of the topological optimization
method of a connecting rod forming by the BESO technique in ANSYS APDL. In: Vasant,
P., Zelinka, I., Weber, G. (eds) Advances in Intelligent Systems and Computing. Springer,
Cham (2020)
14. Vasant, P., Zelinka, I., Weber, G-W.: Intelligent computing and optimization. In:
Proceedings of the 2nd International Conference on Intelligent Computing and Optimization
2019 (ICO 2019), 1st edn. Springer, Cham (2020)
15. Banik, A., Biswal, S.K., Majumder, M., Bandyopadhyay, T.K.: Development of an adaptive
non-parametric model for estimating maximum efficiency of disc membrane. Int. J. Converg.
Comput. 3, 3–19 (2018)
356 A. Banik et al.
16. Panchenko, V.I., Kharchenko, A., Valeriy Lobachevskiy, Y.: Photovoltaic solar modules of
different types and designs for energy supply. Int. J. Energy Optim. Eng. 9, 74–94 (2020).
https://doi.org/10.4018/IJEOE.2020040106
17. Panchenko, V.A.: Solar roof panels for electric and thermal generation. Appl. Sol. Energy
(English Transl. Geliotekhnika) 54, 350–353 (2018). https://doi.org/10.3103/
S0003701X18050146
18. Banik, A., Dutta, S., Bandyopadhyay, T.K., Biswal, S.K.: Prediction of maximum permeate
flux (%) of disc membrane using response surface methodology (rsm). Can. J. Civ. Eng. 46,
299–307 (2019). https://doi.org/10.1139/cjce-2018-0007
19. Kalaycı, B., Özmen, A., Weber, G.-W.: Mutual relevance of investor sentiment and finance
by modeling coupled stochastic systems with MARS. Ann. Oper. Res. 295(1), 183–206
(2020). https://doi.org/10.1007/s10479-020-03757-8
20. Kuter, S., Akyurek, Z., Weber, G.W.: Retrieval of fractional snow covered area from
MODIS data by multivariate adaptive regression splines. Remote Sens. Environ. 205, 236–
252 (2018). https://doi.org/10.1016/j.rse.2017.11.021
21. Vasant, P., Zelinka, I., Weber, G-W.: Intelligent computing & optimization. In: Conference
proceedings ICO 2018. Springer, Cham (2019)
End-to-End Supply Chain Costs Optimization
Based on Material Touches Reduction
1 Introduction
1.1 Motivation and Objective
The global manufacturing industry requires the highest standards in productivity,
quality, delivery and flexibility. This is especially true for the truck manufacturing
industry, which is a relevant example due to the technical complexity of its assembly
process, as well as its tendency toward mass personalization, creating a very com-
plexsupply chain (Fogliatto 2003). There has been vast research on supply chain
optimization; nevertheless, most of it relates to lean manufacturing concepts (Monden
2011, Krajewski 2010) or to the minimum cost flow problem or MCFP (Hammamia
and Frein 2013). When related to lean concepts, it is normally applied to limited
procedures or steps in the supply chain (or SC); and when related to the MCFP, SC is
conceptualized as a net of indivisible levels whose internal operations are neglected.
This prevents us from realizing that inside these levels there are redundant activities
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 357–368, 2021.
https://doi.org/10.1007/978-3-030-68154-8_33
358 C. Pedrero-Izquierdo et al.
that do not add any value for further process (Closs 2010). In this particular research,
those internal activities are referred to as “material touches” or just “touches”, meaning
every action related to handling, moving or interacting in general with the product
flowing on the SC. Based on the background above, it follows that the general
objective of this research is to develop a different cost optimization methodology for a
SC of i 1 stages and i levels serving to a truck assembly company from suppliers to
final assembly process. This methodology must be capable of generating cost
improvements within the internal processes of the SC i levels by fragmenting them in j
basic touches and neglecting the traditional division between them.
1.2 Methodology
This research introduces a four-stage methodology. The first stage analyzes and seg-
ments the original i levels of the SC into individual j touches that are subsequently
cataloged according to specific attributes. The result of this stage plus the selection of
touches proposed to be externally outsourced, give rise to the next step that consists in
generating new sets of touches according to individual characteristics. These subsets
won´t satisfy the original SC touches individually; however, a combination of them
will satisfy them collectively. The third stage is calculating the costs of the brand-new
touch subsets. The last stage is dedicated to formulate and solve the mathematical
model as a “set covering problem”. The case studies used to validate this methodology,
come from a collaborative truck assembling company, which is the largest in the world
and the one with the highest offer for customization in Mexico and Latin America.
1.3 Contribution
The main objective and contribution to the state of the art of this research, was to
develop a new cost optimization strategy dedicated to the SC of the truck assembly
industry that could visualize improvement opportunities within the SC internal steps.
Out of the vast research dedicated to this topic, were lean concepts and MCFP stand
out, this research found the following outcome that hadn´t been considered within these
two branches:
• The concept of material touch is introduced, as well as its mathematical description
as the basis of a new methodology dedicated to SC costs optimization.
• An iterative process was developed to segment the entire supply chain in touches
based on individual attributes.
• A practical application of the “end to end” concept related to the SC was presented
as part of the cost optimization methodology.
• A mathematically strict methodology was provided in order to merge both strengths
and weaknesses of different logistics operators to optimize total costs.
• A four-stage methodology inspired by the aircrew-assigning problem “crew pair-
ing”, was developed as an IP in a “set covering” format.
End-to-End Supply Chain Costs Optimization 359
2 Literature Review
Kumar and Nambirajan (2014) suggested crossing borders between client and provider
to expand capabilities. Arif-Uz-Zaman and Ahsan (2014) suggested that the SC should
be treated as companies linked to each other. Stavrulaki and Davis (2010) suggested
that a SC improves the overall effectiveness when every stage is aligned. Krajewski
et al. (2010) integrated the design of the SC under “lean manufacturing system”.
Bezuidenhout (2015) added the agile concept to the existing lean concepts to describe
what he described as a more realistic “lean-agile” process. As an additional tool to
optimize a SC, Walters and Lancaster (2000) proposed the third party logistics (or 3PL)
figure as part of this process. Arias-Aranda et al. (2011) verified the correlation of 3PL
services and flexibility levels within a company. Wu et al. (2012) suggest flexibility
and specialization improvements trough 3PL services. Dolguia and Proth (2013)
considered that 3PL services can reduce non-essential activities and reduce costs.
Agrawal (2014) mention that using 3PL services enhances flexibility and innovation.
Lee (2010) used the concept “end to end” to suggest the elimination of redundant SC
processes. Closs et al. (2010) suggested also to migrate to a collaborative culture to
eliminate waste by eliminating silos along the SC. Trang (2016), Kram et al. (2015) and
Ageron (2013), propose the elimination of redundant activities through shared SC
design. González-Benito (2013) adopted the term “co-makership” as shared responsi-
bility for SC design. Ashley (2014) affirmed that SC co-responsibility must go from the
birth of the product to the end of its cycle. As of optimization methodologies, Castillo-
Villar (2014) added quality costs as optimization criterion. Hammamia and Frein
(2013) added delivery-time as optimization criterion. Sakalli (2017) included stochastic
parameters under uncertainty. Cheng and Ye (2011) adapted classic MCFP model to
handle parallel suppliers. Fahimniaa et al. (2013) introduced production planning and
SC parameters together. Ding (2009) proposed an optimization method using simul-
taneous production and distribution parameters. Paksoy and Ozceylan (2011) included
operation-balancing parameters to MCFP. Paksoy and Ozceylan (2012), particularized
their previous work by including U-shaped production balancing. Hamtaa (2015)
proposed an SC optimization problem by adding production with uncertain demand
parameters. Minoux (2006) solved the crew assignment problem by using heuristic
methods. Souai and Teghem (2009) proposed reassigning routes by using a genetic
algorithm. Deng and Lin (2010) proposed an ant colony algorithm to solve the crew
scheduling problem. The same problem but considering also legal requirements is
solved by Deveci and Demirel (2015) using a genetic algorithm. Tekiner et al. (2008)
proposed a column generation method incorporating disruptive events. Later on Muter
et al. (2010) also proposed a two-steps columns generation methodology solving firs
the crew pairing and then using these data to solve the crew rostering as well as
incorporating possible disruptions since the planning of the problem, what they called
“robust planning”. Özener et al. (2016) solved the aircraft allocation and flight
sequence assignation by using exact and meta heuristics methods.
360 C. Pedrero-Izquierdo et al.
Fig. 1. The concept of material touches is represented in this diagram, where level i ¼ 2 is
fragmented in j ¼ 8 material touches for component k.
End-to-End Supply Chain Costs Optimization 361
d Uijk
TtTMNijk ¼ TMNijk ð2Þ
Qijk
8i ¼ 1; 2; . . .; I; 8j ¼ 1; 2; . . .; J; 8k ¼ 1; 2; . . .; K
Where:
Tijk ¼ 1 if specific touch type j exists in level i for k:
ð3Þ
Tijk ¼ 0 in any other case:
Note that in (3) Tijk represents each of the nine type of touches TMNijk , TQNijk ,
TMENijk , TQENijk , TMXijk , TQXijk , TMEXijk , TQEXijk and TTRijk . Once Te and Tt are
calculated, the cost for each type of touch can be calculated. These costs are listed
below where the dimensional units for all of them are [monetary unit/day].
• CTMNi Internal manual touch costs.
• CTQNi Internal machinery touch costs.
• CTMENi Internal manual touch that implies packaging costs.
• CTQENi Internal machinery touch that implies packaging costs.
End-to-End Supply Chain Costs Optimization 363
Once the unitary costs per touch-type are obtained for level i, the total cost for each
individual touch is calculated by multiplying the unitary cost by the total touches Tt.
This calculation is done for each touch belonging to each subset in column n. Its sum
represents the total cost of every touch existing in that subset. The internal subset cost
CIntn ½monetary unit=day is the sum of every element belonging to column n:
0 1
CuTMNi TtTMNijk þ CuTQNi TtTQNijk þ CuTMENi
B TtTMEN þ CuTQEN TtTQEN þ CuTMX TtTMX C
B ijk i ijk i ijk C
CIntn ¼ Rijk B C
@ þ CuTQXi TtTQXijk þ CuTMEXi TtTMEXijk þ CuTQEXi A
TtTQEXijk þ CuTTRi TtTTRijk
ð5Þ
Calculation of costs for the subsets to be outsourced follows a different logic from
the previous one. These costs depend on two factors; the first one, CextTr , refers to the
commercial proposal from third parties. The second one, defined as approaching cost
CTextTrijk , is the relocation cost of the components to the point where they will con-
tinue to be processed. This cost makes external proposals cost-comparable with the
internal ones since it compensates the fact that the outsourced processes will take place
in another facility and later on, the components will return to the starting point to
continue the original flow. The total cost CExtr of subset Pr is the sum of both costs,
CextTr and CTextTrijk ½monetary unit=day; this cost is formulated in Eq. (6).
X
CExtr ¼ CextTr þ ijk
CTextTrijk 8r ð6Þ
The resulting cost matrix is shown in Table 2, were touches have already been
illustratively particularized.
364 C. Pedrero-Izquierdo et al.
Internal and external touches subsets are recorded in the incidence matrix A, where:
1 if element exists in the subset registered in n:
amn ¼
0 otherwise:
XN
Subject to : n¼1
amn PAn 18m ¼ 1; 2; . . .; M ð8Þ
PAn ¼ 0; 1 where n ¼ 1; 2; . . .; N
End-to-End Supply Chain Costs Optimization 365
4 Results Analysis
The methodology was evaluated in two different ways. The first validation consisted in
applying the proposed methodology to four designed instances under extreme condi-
tions. The purpose was to verify the correct functioning of the methodology even under
unusual circumstances. The second validation consisted in empirical tests, where three
real documented cases previously solved for the company in collaboration would be
solved again with the new methodology. Table 3 shows the seven instances.
Validation IC1 verified the right segmentation of the SC based on the interrelation
of touches. Validations of instances IC2 and IC3 consisted of proposing atypical
parameters in combination with a random interrelation assignment. For IC2, very high
value coefficients were selected for the external subsets. For IC3, equal parameters
were proposed for all decision variables. As expected, for IC2 the result was the
rejection of external subsets, and for IC3 the solution yielded to multiple solutions. IC4
validation consisted in using the longest and most complex internal supply chain found
in the collaborative company.
The empirical validation consisted in solving three real cases found in the col-
laborating industry and then comparing the results with the real solutions observed in
industrial practice. For the empirical instances IR1 and IR2, the results obtained were
consistent with the real observed solutions. However, IR3 presented a discrepancy
when compared to the real solution observed. The new methodology recommended
rejecting the outsourcing offers (in results table improvements show as zero), whereas
in practice, it was observed that external services were engaged. When comparing the
data of the business case in detail, it was confirmed that IR3 would be costlier if
performed externally. However, in the business case IR3 was part of a package of
services that, on the whole, was financially convenient. The methodology results are
shown in Table 4.
366 C. Pedrero-Izquierdo et al.
The proposed methodology offers improvements over the common cost reduction
methodologies used in the industrial practice, since it conveniently allows merging,
substituting or eliminating sections of the SC in the most cost effective combination.
This new methodology is capable of finding cost reduction opportunities in the internal
processes of the SC levels that some other methodologies cannot reach. In addition, it
allows combining touches belonging to one or more levels since the entire SC separates
in its basic elements and follows the “E2E” lean principle. This methodology allows
companies to search for cost opportunities that are hidden for other methodologies.
References
Dolgui, A., Proth, J.M.: Outsourcing: definitions and analysis. Int. J. Prod. Res. 51(23–24),
6769–6777 (2013). Enero 2020. De PROQUEST Base de datos
Agrawal, A., De Meyer, A., Van Wassenhove, L.N.: Managing value in supply chains: case
studies on the sourcing hub concept. California Management Review, University of
California, Berkeley, vol. 56, pp. 22–54 (2014)
David, A.: Differentiating through Supply chain innovation. Bus. Econ. Market. Purchas. 1, 1–4
(2014)
Fahimniaa, B., Farahani, R.Z., Sarkis, J.: Integrated aggregate supply chain planning using
memetic algorithm – a performance analysis case study. Int. J. Prod. Res. 51(18), 5354–5373
(2013). http://dx.doi.org/10.1080/00207543.2013.774492
Bratianu, C., Vătămănescu, E.-M., Anagnoste, S.: The influence of knowledge dynamics on the
managerial decision- making process. In: Proceedings of the European Conference on
Knowledge Management, vol. 1, pp. 104–111 (2018). Accessed from http://search.ebscohost.
com/login.aspx?direct=true&db=lih&AN=132145882&lang=es&site=ehost-live&custid=
s9884035
Kumar, C.G., Nambirajan, T.: Direct And Indirect Effects: SCM Componnets. SCMS J. Ind.
Manage. 1, 51–65 (2014)
Bezuidenhout, C.N.: Quantifying the degree of leanness and agility at any point within a supply
chain. School of Engineering, University of KwaZulu-Natal, Scittsville, South Africa and
SCION, Rotorua, New Zealand, vol. 118, no. 1, pp. 60–69, 16 September 2015
Arias-Aranda, D., Bustinza, O.F. and Barrales-Molina, V.: Operations flexibility and outsourcing
benefits: an empirical study in service firms. Routledge Taylor & Francis Group, vol. 31, no.
11, pp. 1849–1870. Enero 2020 (2011). De EBSCO Base de datos
Closs, D.J., Speier, C., Meacham, N.: Sustainability to support end-to-end value chains: the role
of supply chain management. Acad. Market. Sci. 39, 101–116 (2010)
Walters, D., Lancaster, G.: Implementing value strategy through the value chain. Manag. Decis.
38(3), 160–178 (2000)
Zeghal, F.M., Minoux, M.: Modeling and solving a crew assignment problem. Eur. J. Oper. Res.
175, 187–209 (2006). De ProQuest Base de datos
Cheng, F., Ye, F.: A two objective optimisation model for order splitting among parallel
suppliers. Int. J. Prod. Res. 49, 2759–2769 (2011). De EBSCO Base de datos
Fogliatto, F.S., Da Silveira, G.J.C., Royer, R.: Int. J. Prod. Res. 41(8), 1811–1829 (2003)
Deng, G.F., Lin, W.T.: Ant colony optimization-based algorithm for airline crew scheduling
problem. Expert Syst. Appl. 38, 5787–5793. (2011). De EBSCO Base de datos
Tekiner, H., Birbil, S.I., Bubul, K.: Robust crew pairing for managing extra flights. Manuf. Syst.
Ind. Eng. Sabanci Univ. 1, 1–30 (2008). De EBSCO Base de datos
Lee, H.L.: Don’t tweak your supply chain-rethink it end to end. Harvard Bus. Rev. I, 62–69
(2010)
Ding, H., Benyoucef, L., Xie, X.: Stochastic multi-objective production-distribution network
design using simulation-based optimization. Int. J. Prod. Res. 47(2), 479–505. (2009).
De PROQUEST Base de datos
Muter, I., Birbil, S.I., Bulbul, K., Sahin, G., Yenigun, H.: Solving a robust airline crew pairing
problem with column generation. algopt Alg. Optim. 40, 1–26 (2013). De EBSCO Base de
datos
González-Benito, J., Lannelonguea, G., Alfaro-Tanco, J.A.: Study of supply-chain management
in the automotive industry: a bibliometric analysis. Int. J. Prod. Res. 51(13), 3849–3863
(2013)
368 C. Pedrero-Izquierdo et al.
Wu, J.-Z., Chien, C.-F., Gen, M.: Coordinating strategic outsourcing decisions for semiconductor
assembly using a bi-objective genetic algorithm. Int. J. Prod. Res. 50(1), 235–260 (2012)
Arif-Uz-Zaman, K., Ahsan, A.M.M.N.: Lean supply chain performance measurement. Int.
J. Product. Perform. Manage. 63(5), 588–612 (2014)
Krafcik, J.F.: Triumph of the lean production system. Sloan Manag. Rev. 30(1), 41–52 (1988)
Krajewski, L.J., Ritzman, L.P., Malhotra, M.K.: Operations Management – Processes and Supply
Chains, vol. 07458, 9th edn. Pearson Education Inc, Upper Saddle River (2010)
Castillo-Villar, K.K., Smith, N.R., Herbert-Acero, J.F.: Design and optimization of capacitated
supply chain networks including quality measures. Hindawi Publishing Corporation,
Mathematical Problems in Engineering, pp. 1–17 (2014). Article ID 218913.
De PROQUEST Base de datos
Kram, M., Tošanović, N., Hegedić, M.: Kaizen approach to supply chain management: first step
for transforming supply chain into lean supply chain. Ann. Faculty Eng. Hunedoara – Int.
J. Eng. Tome XIII – Fascicule 13(1), 161–164 (2015)
Deveci, M., Demirel, N.C.: Airline crew pairing problem: a literature review. In: 11th
International Scientific Conference on Economic and Social Development, vol. 1, 17
December 2015–July 2019. De EBSCO Base de datos
Souai, N., Teghem, J.: Genetic algorithm based approach for the integrated airline crew-pairing
and rostering problem. Eur. J. Oper. Res. 199, 674–683 (2009). De ProQuest Base de datos
Trang, N.T.X.: Design an ideal supply chain strategy. Advances In Management 9(4), 20–27,
April 2016
Hamta, N., Shirazi, M.A., Fatemi Ghomi, S.M.T., Behdad, S.: Supply chain network
optimization considering assembly line balancing and demand uncertainty. Int. J. Prod.
Res. 53(10), 2970–2994 (2015). De PROQUEST Base de datos
Özener, O.Ö., Matoğlu, M.Ö., Erdoğan, G.: Solving a large-scale integrated fleet assignment and
crew pairing problem. Ann. Oper. Res. 253, 477–500 (2016). De ProQuest Base de datos
Paksoy, T., Ozceylan, E., Gokcen, H.: Supply chain optimization with assembly line balancing.
Int. J. Prod. Res. 50, 3115 (2011). https://doi.org/10.1080/00207543.2011.593052
Hammami, R., Frein, Y.: An optimisation model for the design of global multi-echelon supply
chains under lead time constraints. Int. J. Prod. Res. 51(9), 2760–2775 (2013). De EBSCO
Base de datos
Rubin, J.: A Technique for the solution of massive set covering problems, with application airline
crew scheduling. Transp. Sci. 7(1), 15–34 (1973)
Rushton, A., Croucher, P., Baker, P.: Logistics and distribution management. 4th edn. Kogan
Page, London, 636 p (2010). ISBN 978 0 7494 5714 3
Stavrulaki, E., Davis, M.: Aligning products with supply chain processes and strategy. Int.
J. Logist. Manage. Ponte Vedra Beach, 21(1), 127–151 (2010)
Paksoy, T., Özceylan, E.: Supply chain optimisation with U-type assembly line balancing. Int.
J. Prod. Res. 50(18), 5085–5105 (2012). De EBSCO Base de datos
Sakalli, U.S.: Optimization of production-distribution problem in supply chain management
under stochastic and fuzzy uncertainties. Hindawi, Mathematical Problems in Engineering,
pp. 1–29 (2017). Article ID 4389064. De EBSCO data base
Versión 2018 • Agenda Automotriz DIÁLOGO CON LA INDUSTRIA AUTOMOTRIZ 2018 •
2024 (2018). http://www.amia.com.mx/boletin/dlg20182024.pdf. Accessed 20 Jan 2019
Monden, Y.: Toyota Production System, An Integrated Approach to Just-In-Time, 4edn., CRC
Press, New York (2011)
Ageron, B., Lavastre, O., Spalanzani, A.: Innovative supply chain practices: the state of French
companies. Supply Chain Management Int. J. 18(3), 256–276 (2013)
Computer Modeling Selection of Optimal
Width of Rod Grip Header to the Combine
Harvester
1 Introduction
The results of conducted research [1–5] on the rationale the combine harvesters
parameters are that the main factors determining the effectiveness of combine har-
vesters fleet choosing with the rationale the optimal header width are qualitative and
quantitative composition of combine harvesters and the standard header size and the
optimal harvesting period duration. The optimal these parameters ratio can be estab-
lished only during the study of the combine harvester fleet operation by mathematical
modeling methods, because natural experiments require a lot of time and material
resources.
The operating experience and results of comparative combine harvesters tests in
different areas of the country show that the main factor determining their capacity
underuse for direct combine harvesting is the insufficient headers width.
Optimizing the harvester header width will allow to use effective it at different crop
yield.
Research and testing domestic and foreign combine harvesters have allowed to
develop a refined economic and mathematical model for optimizing the size of the
header for a combine harvester in recent years.
The crop yield in the Russian regions varies widely from 10 to 80 c/Ha (centners
per hectare), so there is a question of header purchasing by farms, it allows them to
realize the nominal combine harvester productivity and ensure the minimum cash
expenditures for bread harvesting.
Theoretical (system analysis, methods for optimizing the combine harvester fleet
composition, optimal header width calculation), logical (comparison, generalization
and analyzing scientific literature on the research problem), statistical (processing
experiment data). Mathematical modeling with the development deterministic eco-
nomic and mathematical model and calculation the optimal header width for the
combine harvester was used during the choosing header width.
Research objectives ensuring effective combine harvester operational and tech-
nological indicators due to the scientifically proven header width.
Research Tasks:
– to develop economical and mathematical model of choosing header width for the
combine harvester and a software for calculating by PC in relation to two levels of
winter wheat productivity;
– to conduct field research of a combine harvester with 12 kg/s throughput, that can
work with headers of different capture widths, with the definition of conditions and
operational and technological study indicators
– to calculate by PC for the choosing the optimal width of the header for combine
harvester with 12 kg/s throughput.
3 Discussion Results
The optimal header width is selected according to the thresher bandwidth and the
combine harvester engine power in specific zonal conditions. In this regard, a refined
economic and mathematical optimization model for the header size is proposed.
Computer Modeling Selection of Optimal Width of Rod Grip Header 371
Structural scheme of a refined economic and mathematical model for optimizing header
for the combine harvester size is shown in Fig. 1.
Fig. 1. Structural diagram of the algorithm of calculation and selection of the type size of the
header to the combine harvester.
The model consists interconnected blocks that determine the sequence of calcula-
tions for choosing the optimal combine harvester header size according to the formulas
(1−16) those are given in the text below.
The general version of the winter wheat harvesting organization using the direct
combining technology with grinding the non-grain crop part was adopted by us in order
to make the calculations as close as possible to the real economic task.
372 M. Chaplygin et al.
The following parameters change depending on the using different header width in
the combine unit:
– working movement speed V, km/h;
– operating combine harvester weight Gop ch , kg;
– combine harvester working moves coefficient K2 ;
– power for self-movement Nmov , h.p.;
– shift combine unit productivity Wsh , ton/h.
Combine harvester working movement speed with different header width, km/h, it
will be determined by the formula
W0n
V¼ ð2Þ
0; 1 Bh Ug
Gop
ch ¼ Gch þ Gh þ Gg ð3Þ
Gop
ch fst V
Nmov ¼ ð5Þ
3:6 75
Gg ¼ Vgt cg ð7Þ
Hh ¼ a V ð8Þ
Ch ¼ S þ F þ R þ D þ Lh ð10Þ
Ts
S¼ ð11Þ
Wsh
Nmov K2
qmov ¼ ð12Þ
W0W
where qthr is specific fuel consumption for threshing bread mass, kg/ton;
qmov is specific fuel consumption for self-movement, kg/ton;
qmov Nmov
qthr ¼ ð14Þ
Wsh
ðPch þ Ph Þsr
R¼ ð15Þ
Wsh Tzon
ðPch þ Ph Þd
D¼ ð16Þ
Wsh Tzon
Ug Hh Pg
Lh ¼ ð17Þ
100
where Hh is grain losses behind the header, %;
Pg is commercial grain selling price, RUB/ton.
Let us assume, that some farm has combine harvester, that works in wet work mode
with specified normative indicators. The manufacturer produces different width headers
for this combine harvester model. We will accept next condition: winter wheat pro-
ductivity is two-level: achieved 5,0 ton/ha and predictable 7,0 ton/ha.
Indicator are being calculated taking into account certain harvesting characteristics
and combine harvesting productivity, after that shift productivity should be determined
and harvesting cost will be calculated with different header width variants. It will be the
first task solution. However we know that the answer to the question – how many
combine harvesters are required for this farm with certain winter wheat productivity
area? – it need calculations sequence for last block.
Economical and mathematical model and PC software, that is named « ECO-
Adapter » were developed for solution this task [13, 14].
Software is used for calculating combine harvester work with different header width
economical efficiency at grain harvesting and comparative analysis with using different
width headers. Until 4 variants calculating can be in comparative analysis.
Software includes two guide: header and combine guides. The guides contains
information about technical tools, cost and another combine harvesters and headers
characteristics. Command « Task – New » from main window menu is for making
new task. Next, window « Add new task » appears. There must be additional source
task data:
– fuel cost RUB/kg;
– crop productivity, ton/ha;
– grain density, kg/m3;
– typical rut length, m;
– turn time, min;
– rolling over stubble coefficient;
– using shift time coefficient;
– hourly machine operators wage, RUB/h;
– repair cost, %;
– depreciation cost, %;
– commercial grain selling price, RUB/ton.
Task report is generating and displayed in Excel. It can be formed and printed by
menu command « Report – Task » in main form.
Tasks with equal crop productivity values and different headers (different width),
the same combine harvester is aggregated by those, are selected for comparative
analysis. Selected tasks are named « VARIANT 1 » , « VARIANT 2 » , « VAR-
IANT 3 » or « VARIANT 4 » by pushing right mouse button on the needful task line
and by choosing variant number from context menu. Selected variants are displayed at
the lower part of main window in yellow rectangle. Comparative analysis report is
376 M. Chaplygin et al.
Calculating results can be presented as a graph taking into account the entered
optimal working speed range limit 3 to 7 km/h (Fig. 2). Choosing optimal working
speed range limit is based on the research results, that were reviewed in study [15].
Fig. 2. Selection of the optimal width of the header, taking into account operating speed limits.
Graph show that optimal combine thresher workload with minimal grain losses and
crushing values is provided at choosing optimal header width in the crop productivity
range 2.5−9.5 ton/ha subject to entered working speed range limit 3−7 km/h. Thresher
workload with working speed limit will be optimal for header width of 6 m at crop
Computer Modeling Selection of Optimal Width of Rod Grip Header 377
productivity 4.6−9.8 ton/Ha, for 7 m is 4.2−9.0 ton/Ha and for 9 m is 3.0−7.6 ton/Ha.
All this indicates that header with 7 and 9 m width have bigger crop productivity range.
Made for the future calculating 12 m header shows that its using in conjunction
with combine harvester with 12 kg/sec capacity will be the most optimal at working
speed ranged workload and crop productivity 2.5−5.6 ton/Ha. Combine harvester
realizes its throughput potential fully under given conditions. To work with a larger
productivity range, engine unit power must be increased.
4 Conclusions
References
1. Zhalnin, E.V.: Matematicheskoye modelirovaniye protsessov zemledel’cheskoy mekhaniki
[Mathematical modeling of agricultural mechanics processes]. Tractors Agric. Mach. 1, 20–
23 (2000). (in Russian)
2. Lipkovich, E.I.: Osnovy matematicheskogo modelirovaniya sistemy mashin [Fundamentals
of mathematical modeling of a machine system]. Povysheniye effektivnosti uborochnykh
rabot. VNIPTIMESKH. Zernograd, pp. 13–22 (1984) (in Russian)
3. Tabashnikov, A.T.: Optimizatsiya uborki zernovykh i kormovykh kul’tur [Optimization of
harvesting of grain and forage crops]. Agropromizdat. Moscow, p 159 (1985) (in Russian)
4. Kavka, M., Mimra, M., Kumhala, F.: Sensitivity analysis of key operating parameters of
combine harvesters. Res. Agric. Eng. 62(3), 113–121 (2016). https://doi.org/10.17221/48/
2015-rae
5. Badretdinov, I., Mudarisov, S., Lukmanov, R., Permyakov, V.: Mathematical modeling and
research of the work of the grain combine harvester cleaning system. Comput. Electron.
Agric. 165, 104966 (2019). https://doi.org/10.1016/j.compag.2019.104966
6. Šotnar, M., Pospíšil, J., Mareček, J., Dokukilová, T., Novotný, V.: Influence of the combine
harvester parameter settings on harvest losses. Acta Technol. Agric. N21(3), 105–108
(2018). https://doi.org/10.2478/ata-2018-0019
7. Chen, J., Wang, S., Lian, Y.: Design and test of header parameter keys electric control
adjusting device for rice and wheat combined harvester. Trans. Chin. Soc. Agric. Eng. 34
(16), 19–26 (2018). https://doi.org/10.11975/j.issn.1002-6819.2018.16.003
8. Liu, H., Reibman, A.R., Ault, A.C., Krogmeier, J.V.: Video-based prediction for header-
height control of a combine harvester. In: IEEE Conference on Multimedia Information
Processing and Retrieval (MIPR), San Jose, CA, USA, pp. 310–315 (2019). https://doi.org/
10.1109/mipr.2019.00062
9. Zhang, K., Shen, H., Wang, H., et al.: Automatic monitoring system for threshing and
cleaning of combine harvester. IOP Conference Series: Materials Science and Engineering.
452(4), p. 042124 (2018). https://doi.org/10.1088/1757-899X/452/4/042124
10. Shepelev, S.D., Shepelev, V.D., Almetova, Z.V., et al.: Modeling the technological process
for harvesting of agricultural produce. In: IOP Conference Series: Earth and Environmental
Science. 115(1), p. 012053 (2018). https://doi.org/10.1088/1755-1315/115/1/012053
11. Almosawi, A.A.: Combine harvester header losses as affected by reel and cutting indices.
Plant Archives. 19, 203–207 (2019). http://www.plantarchives.org/PDF%20SUPPLEMENT
%202019/33.pdf
12. Zhang, Y., Chen, D., Yin, Y., Wang, X., Wang, S.: Experimental study of feed rate related
factors of combine harvester based on grey correlation. IFAC-PapersOnLine 51(17), 402–
407 (2018). https://doi.org/10.1016/j.ifacol.2018.08.188
13. Reshettseva, I.A., Tabashnikov, A.T., Chaplygin, M.E.: Certificate of state registration of the
computer program “ECO-Adapter” No. 2015613469. Registered in the Program Registry
03/17/2015. (in Russian)
14. Chaplygin, M.Y.: Ekonomiko-matematicheskaya model’ optimizatsii tiporazmera khedera k
zernouborochnomu kombaynu [Economic-mathematical model for optimizing the size of a
header to a combine harvester]. Machinery and Equipment for Rural Area. (2), pp. 23–24
(2012) (in Russian)
15. Chaplygin, M.Y.: Povishenie effektivnosti ispolzovaniya zernouborochnogo kombaina
putem obosnovaniya optimalnoi shirini zahvata jatki dlya uslovii yuga Rossii [Improving the
efficiency of the combine harvester by justification the optimal header width for the
conditions in southern Russia]. (Dissertation Candidate of Technical Sciences). Volgograd
State Agrarian University, Moscow (2015) (in Russian)
An Integrated CNN-LSTM Model for Micro
Hand Gesture Recognition
Abstract. Vision based micro gesture recognition systems enable the devel-
opment of HCI (Human Computer Interaction) interfaces to mirror real-world
experiences. It is unlikely that a gesture recognition method will be suitable for
every application, as each gesture recognition system rely on the user cultural
background and application domain. This research is an attempt to develop a
micro gesture recognition system suitable for the asian culture. However, hands
vary in shapes and sizes while gesture varies in orientation and motion. For
accurate feature extraction, deep learning approaches are considered. Here, an
integrated CNN-LSTM (Convolutional Neural Network- Long Short-Term
Memory) model is proposed for building micro gesture recognition system. To
demonstrate the applicability of the system two micro hand gesture-based
datasets namely, standard and local dataset consisting of ten significant classes
are used. Besides, the model is tested against both augmented and unaugmented
datasets. The accuracy achieved for standard data with augmentation is 99.0%,
while the accuracy achieved for local data with augmentation is 96.1% by
applying CNN-LSTM model. In case of both the datasets, the proposed CNN-
LSTM model appears to perform better than the other pre-trained CNN models
including ResNet, MobileNet, VGG16 and VGG9 as well as CNN excluding
LSTM.
1 Introduction
Gesture symbolizes the posturized instance of a body through which some information
is conveyed [3]. Gesture usually categorize as macro and micro. Where macro gesture
demonstrates the perpetuating motion of the hand in coordination to the body, while
micro gesture pictorializes the relative position of the fingers of the hand. This research
makes use of micro gestures with static images. Hand gesture recognition systems are
generally used to narrow the bridge between human and computer interactions (HCI).
Human interactions through hand gesture sometimes involve one hand or both hands.
An interactive machine which successfully mimics the natural way of human hand
interaction can be developed [16]. The research presented in this paper is an attempt to
develop a classification tool, enabling the accurate recognition of both single and
double hand gestures. Thus, this paper focuses on building an effective tool for the
recognition of hand gesture, by introducing an integrated [17] CNN-LSTM (Convo-
lutional Neural Network - Long Short Term Memory) model. The CNN model is
considered [13] because it can handle large amount of raw data with comparatively less
pre-processing effort [2], while LSTM can better optimize the back propagation.
Therefore, the integration of CNN with LSTM [4] would play an important role to
achieve better model accuracy [1]. This research is further enhanced by investigating
the impacts of data augmentation on the proposed CNN-LSTM model. To perform
these two different datasets, namely standard and local with augmentation as well as
without augmentation have been used to train the CNN-LSTM model. It is important to
note that the local dataset contains classes of gesture of both hands such as ‘dua’,
‘namaste’, ‘prayer’ and ‘handshake’ which are unique in the context of Indian sub-
continent. The accuracy of the proposed CNN-LSTM model has been compared with
LSTM excluded CNN model as well as with state-of-the-art pre-trained CNN models,
including ResNet, Mobile Net, VGG9 and VGG16 by taking account of both aug-
mented and unaugmented datasets. The real time validation against local dataset has
also been carried out, this will be discussed in result section. To check the overfitting
and underfitting aspects of the CNN-LSTM model four folds cross-validation has also
been carried out. The remaining of the paper is structured as follows. Section 2
analyses the different works related to hand gesture recognition. Section 3 outlines the
methodology undertaken in order to develop the model. Section 4 presents the results,
while Sect. 5 concludes the paper with a discussion on future research.
2 Related Work
There exist a plethora of research works in the area of hand gesture recognition sys-
tems. In [14] the different classes of hand gestures were classified using CNN on
augmented and unaugmented datasets. It was seen that augmented dataset produced an
accuracy of 97.18%, while unaugmented dataset produced an accuracy of 92.91%.
However, the recognition of gesture by both hands was not considered in this research.
In [19] Artificial Neural Networks (ANN) were used to classify ten different categories
of hand gestures with an average accuracy of 98%. Due to the low computational cost
to identify the different categories of hand gestures, this method became a good can-
didate to execute in real-time. However, the use of ANN requires pre-processing of raw
data and hence, becomes computationally expensive. Another effective method [20] to
recognize American Sign Language Hand Gesture for 26 different classes was pro-
posed. Principal Component Analysis (PCA) was used to extract features from the
dataset. These extracted features Micro Hand Gesture Recognition 3 were later fed into
the ANN for training the dataset. This method produced an accuracy of 94.3%. In spite
of this accuracy, the feature extraction technique, applied in this paper, was unable to
extract features based on depth and hand direction. In [21] twenty-six different classes
of American Sign language Hand Gesture dataset were trained using three different
CNN models. The hidden layers of each model increased from the other by one. It was
observed that, by increasing the number of hidden layers results in decreasing
An Integrated CNN-LSTM Model for Micro Hand Gesture Recognition 381
recognition rate of CNN model but increasing its run time. Consequently, testing
accuracy appears to steeply fall from 91% to 83%. In another approach [24], a self-
organizing as well as self-growing neural gas network along with an YCbCr (Green,
Blue, Red) color spacing method for hand region detection was proposed. The data
used in this paper consists of six different classes of gestures and after applying
classification produces an accuracy of 90.45%. Although it is fast to compute, the hand
region detection method sometimes produces wrong feature detection. In [27] two
different datasets were utilized, namely self-constructed and Cambridge Hand Gesture
Dataset. Each dataset consists of 1500 images, which are categorized into five different
classes. The self-constructed dataset was acquired from images of both left and right
hands separately. The dataset was preprocessed using Canny Edge method which
removed the effect of illuminations. In this study, a CNN-Architecture of five convo-
lution layers were used to train each dataset. Although the method produced an
accuracy of 94%, it was found suitable for recognizing complex gestures. It can be
observed that none of the study as discussed above considered the classification of both
hand gestures as well as considered context-based gestures such as ‘dua’, ‘namaste’,
‘handshake’ and ‘prayer’, resulting to inefficient nonverbal machine human interaction.
In addition, the use of CNN models produces better accuracy than other machine
learning models such as ANN [19]. However, layering of CNN’s introduce the
problems with saturated neurons and vanishing gradient, which can correspond to a
poor testing accuracy [21]. Such a drawback of CNN models can be overcome by
integrating the model with such a model like LSTM because it reduces saturation of
neurons as well as overcomes the vanishing gradient problem by optimizing back-
propagation. Therefore, in this research an integrated model of CNN-LSTM has been
proposed, which will be described in the following section.
3 Methodology
Figure 1 illustrates the main components of the methodology used in this research to
identify the different classes of hand gestures as mentioned previously. A brief dis-
cussion of each of the components is presented below.
an image, the LSTM layer supports by interpreting those features in each distributed
time stamps. These interpretations gives the model exposure to some complex temporal
dynamics of imaging [6]. As a result, the perceptual representation of images in the
convolutions become well defined. All the layers are connected to the output layer
which consists of ten nodes that indicate ten different classes of hand gestures used to
train the model. Afterwards, an activation function SoftMax is included in this layer to
obtain a probabilistic value for each of the classes. SoftMax is used in the very last
layer of the network to provide a nonlinear variant of multinomial logistic regression.
Stochastic Gradient Descent (SGD) of a learning rate 0.01 is used as an optimizer to
compile the model. SGD improves the optimization techniques by working with large
datasets as it is computationally faster [7] and hence, reduces the model loss. Table 1
summarizes the various components of CNNLSTM model.
PIL, Pandas and Numpy. Here, Tensor flow acts as the backend of the system while
Keras is used to build the CNNLSTM classifier since it provides built-in functions for
layering, optimization and activation. Further, Keras contains advanced tools for fea-
ture extraction. OpenCV is required for image processing [18]. On the other hand,
Sklearn gives access to many supervised and unsupervised algorithms. Matplotlib is
used to generate the confusion matrix. Tkinter supports easier configuration tools for
the user interface, required for real-time data collection and validation. PIL is an
integration tool for image processing, while Numpy is used to carry out the operations
of arrays [8]. Callbacks are used to train the model. Callbacks not only prevents
overfitting that occurs as a result of too many epochs but also avoids underfit models
[11]. The callbacks, which are implemented in this model include checkpoints, early
stopping and reducing learning rate on the plateau. Checkpoints allow saving the best
models by monitoring the loss invalidation. Once the model stops performing well on
the validation dataset, early stopping is applied to stop the training epochs. Reducing
the learning rate on the plateau is used when the validation loss abstains from any
further improvement. The data augmentation is directly applied to the CNN-LSTM
model through Keras built-in Data Augmentation API, which enables the generation of
the dataset while training is carried out on the model [5, 29]. Parameters namely
shifting, rotation, normalization and zoom are applied to induce robustness of dataset.
Therefore, the augmented dataset will expand the number of data points, enabling the
reduction of the distance between training and testing datasets. Thus, this will decrease
overfitting in the training dataset.
The section investigates the application of the developed CNN-LSTM model by taking
account of the both standard and local datasets as introduced in Sect. 3.
Fig. 9. Confusion Matrix Fig. 10. Model Accuracy Fig. 11. Model Loss
Loss of the augmented local dataset by applying CNN-LSTM model. It can be seen
from both the curves that an early stop at 54 epochs has been achieved. The training
accuracy as well as testing accuracy achieved are 95.0% and 96.1% respectively.
Hence, it can be concluded that the model is well fit.
To evaluate each model used in this research, performance matrices such as precision,
recall, F1-score and accuracy are used. These matrices calculate the performance by
using parameters such as TP (true positive) and FP (False Positive). Precision tries to
reduce the false positive values, in order to induce a precise model. Recall mainly
focuses on reducing the false negative value. F1- score reduces the false positive and
false negative values, in order to increase F1- score for the model. Accuracy focuses
solely on true positive and true negative values, in order to correctly identify the
classes. Table 2 provides the classification results of CNN-LSTM model for augmented
data of both standard and local datasets. Table 3 provides the classification results off
CNN-LSTM model for the unaugmented data of both the standard and local datasets.
From both the tables it can be observed that the values of evaluation metrics of
augmented data are much higher than that of unaugmented data.
Table 4 illustrates a comparison between the proposed CNN-LSTM model and four
pre-trained CNN models, namely RasNet, VGG9, VGG16 and MobileNet by taking
account of augmented local dataset. In addition, it also illustrates the comparison
between CNN-LSTM model and CNN model excluding LSTM by taking account of
augmented local dataset. It can be observed from the values of the evaluation metrics
against ResNet that the model performs poorly. It can also be observed that none of the
pre-trained CNN models are well fit because testing accuracy for VGG9 and VGG16
are much higher than training accuracy, meaning they are underfitted. On the other
hand, the testing accuracy of MobileNet is much lower than training accuracy meaning
it is overfitted. The CNN model excluding LSTM also appears overfitted since its
388 N. Basnin et al.
testing accuracy is much higher than training accuracy. Thus, it can be concluded that
none of the models are well fit. On the other hand, the proposed CNN-LSTM model is
well fit since its testing and training accuracy is much closer. Table 5 illustrates a
comparison between the proposed CNN-LSTM model and four pre-trained CNN
models, namely RasNet, VGG9, VGG16 and MobileNet by taking account of aug-
mented standard dataset. In addition, it also illustrates the comparison between CNN-
LSTM model and CNN model excluding LSTM by taking account of augmented
standard dataset. ResNet demonstrates a poor performance in comparison to other CNN
models because of its poor training accuracy. Both Mobile Net and CNN suffer from
overfitting because their training accuracies are much lower than their testing accura-
cies. In contrast, VGG9 and VGG16 are well fit models. However, the training
accuracy of VGG 9 is only 64%. Although VGG16 is a well fit model, its accuracy is
lower than the proposed CNN-LSTM model. Therefore, it can be concluded that the
proposed CNN-LSTM model is well fit as well as its performance in terms of accuracy
is better than other models.
Table 4. Comparision between CNN-LSTM model with other CNNmodels on augmented local
dataset
CNN model Precision Recall F1-score Train Acc. Test Acc.
ResNet 0.03 0.12 0.04 0.10 0.12
VGG9 0.11 0.10 0.10 0.27 0.72
MobileNet 0.44 0.28 0.24 0.40 0.28
CNN 0.80 0.80 0.81 0.89 0.75
VGG16 0.87 0.86 0.86 0.87 0.97
CNN-LSTM 0.83 0.81 0.82 0.96 0.95
accuracy. This means our proposed CNN-LSTM model is well fit. The average testing
loss is 19.0% while the average training loss is around 26.0%. This means the testing
loss is lower than the training loss (Table 6).
Table 6. Cross validation result of testing verses training accuracy and loss
Dataset Test accuracy Train accuracy Test loss Train loss
k-1 0.96 0.97 0.30 0.15
k-2 0.99 0.98 0.10 0.20
k-3 0.98 0.96 0.20 0.40
k-4 0.97 0.92 0.15 0.30
Average 0.98 0.96 0.19 0.26
In this research, two datasets were used namely, standard and local dataset for vision-
based micro hand gesture recognition. The local dataset is a heterogeneous dataset
which includes five double hand gesture classes out of the ten classes. These five
different classes of gestures include gestures used in the Indian sub-continent. Thus, a
socio-physical approach is taken into account. The CNNLSTM model proposed in this
study can accurately classify the micro gestures used in both the datasets. Besides, the
CNN-LSTM model for micro gesture datasets has been compared with four pre-trained
models namely ResNet, Mobile Net, VGG9 and VGG16 and with LSTM excluded
CNN model. It has been demonstrated in the result section the CNN-LSTM model
outperforms other pretrained models in terms of accuracy in case of both standard and
local dataset. Therefore, the system not only recognizes the gestures of one hand but
also two hands. In future, this research aims to build a dynamic dataset by involving a
greater number of micro gesture classes in the context of Asia as well as collect images
from a greater number of subjects. Few environmental factors such as the intensity of
light, a variety of image backgrounds and skin tone can be considered while building
the dataset. Furthermore, the performance of this robust dataset will present interesting
phenomena. This complexity can be overcome if an ensemble learning methodology
proposed in [22] is considered. This will not only result in better accuracy but also help
in selecting the optimal learning algorithm.
References
1. Ahmed, T.U., Hossain, M.S., Alam, M.J., Andersson, K.: An integrated cnn-rnn framework
to assess road crack. In: 22nd International Conference on Computer and Information
Technology (ICCIT), pp. 1–6. IEEE (2019)
2. Ahmed, T.U., Hossain, S., Hossain, M.S., ul Islam, R., Andersson, K.: Facial expression
recognition using convolutional neural network with data augmentation. In: Joint 8th
International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd
International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 336–341.
IEEE (2019)
3. Akoum, A., Al Mawla, N., et al.: Hand gesture recognition approach for asl language using
hand extraction algorithm. J. Softw. Eng. Appl. 8(08), 419 (2015)
4. Basnin, N., Hossain, M.S., Nahar, L.: An integrated cnn-lstm model for bangla lexical sign
language recognition. In: Proceedings of 2nd International Conference on Trends in
Computational and Cognitive Engineering (TCCE-2020) Springer Joint 8th International
Conference on Informatics (2020)
5. Chowdhury, R.R., Hossain, M.S., ul Islam, R., Andersson, K., Hossain, S.: Bangla
handwritten character recognition using convolutional neural network with data augmen-
tation. In: Joint 8th International Conference on Informatics, Electronics & Vision (ICIEV)
and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR),
pp. 318–323. IEEE (2019)
6. Donahue, J., Anne Hendricks, L., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko,
K., Darrell, T.: Long-term recurrent convolutional networks for visual recognition and
description. In: Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 2625–2634 (2015)
An Integrated CNN-LSTM Model for Micro Hand Gesture Recognition 391
7. Goyal, P., Dollar, P., Girshick, R., Noordhuis, P., Wesolowski, L., Kyrola, A., Tulloch, A.,
Jia, Y., He, K.: Accurate, large minibatch sgd: training imagenet in 1 hour. arXiv preprint
arXiv:1706.02677 (2017)
8. Greenfield, P., Miller, J.T., Hsu, J., White, R.L.: Numarray: a new scientific array package
for python. PyCon DC (2003)
9. Grundland, M., Dodgson, N.A.: Decolorize: fast, contrast enhancing, color to grayscale
conversion. Pattern Recogn. 40(11), 2891–2896 (2007)
10. Gti: Hand gesture recognition database (2018), https://www.kaggle.com/gtiupm/
leapgestrecog
11. Gulli, A., Pal, S.: Deep learning with Keras. Packt Publishing Ltd (2017)
12. Haralick, R.M., Sternberg, S.R., Zhuang, X.: Image analysis using mathematical morphol-
ogy. IEEE Trans. Pattern Anal. Mach. Intell. 4, 532–550 (1987)
13. Hossain, M.S., Amin, S.U., Alsulaiman, M., Muhammad, G.: Applying deep learning for
epilepsy seizure detection and brain mapping visualization. ACM Trans. Multimedia
Comput. Commun. Appl. (TOMM) 15(1s), 1–17 (2019)
14. Islam, M.Z., Hossain, M.S., ul Islam, R., Andersson, K.: Static hand gesture recognition
using convolutional neural network with data augmentation. In: Joint 8th International
Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International
Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 324–329. IEEE (2019)
15. Islam, R.U., Hossain, M.S., Andersson, K.: A deep learning inspired belief rulebased expert
system. IEEE Access 8, 190637–190651 (2020)
16. Jalab, H.A.: Static hand gesture recognition for human computer interaction. Inf. Technol.
J. 11(9), 1265 (2012)
17. Kabir, S., Islam, R.U., Hossain, M.S., Andersson, K.: An integrated approach of belief rule
base and deep learning to predict air pollution. Sensors 20(7), 1956 (2020)
18. Nandagopalan, S., Kumar, P.K.: Deep convolutional network based saliency prediction for
retrieval of natural images. In: International Conference on Intelligent Computing &
Optimization. pp. 487–496. Springer (2018)
19. Nguyen, T.N., Huynh, H.H., Meunier, J.: Static hand gesture recognition using artificial
neural network. J. Image Graph. 1(1), 34–38 (2013)
20. Nguyen, T.N., Huynh, H.H., Meunier, J.: Static hand gesture recognition using principal
component analysis combined with artificial neural network. J. Autom. Control Eng. 3(1),
40–45 (2015)
21. Oyedotun, O.K., Khashman, A.: Deep learning in vision-based static hand gesture
recognition. Neural Comput. Appl. 28(12), 3941–3951 (2017)
22. Oz¨o˘g¨ur-Aky¨uz, S., Otar, B.C., Atas, P.K.: Ensemble cluster pruning via convex- concave
programming. Comput. Intell. 36(1), 297–319 (2020)
23. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a
simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–
1958 (2014)
24. Stergiopoulou, E., Papamarkos, N.: Hand gesture recognition using a neural network shape
fitting technique. Eng. Appl. Artif. Intell. 22(8), 1141–1158 (2009)
25. Uddin Ahmed, T., Jamil, M.N., Hossain, M.S., Andersson, K., Hossain, M.S.: An integrated
real-time deep learning and belief rule base intelligent system to assess facial expression
under uncertainty. In: 9th International Conference on Informatics, Electronics & Vision
(ICIEV). IEEE Computer Society (2020)
26. Wang, W., Yang, J., Xiao, J., Li, S., Zhou, D.: Face recognition based on deep learning. In:
International Conference on Human Centered Computing. pp. 812– 820. Springer (2014)
392 N. Basnin et al.
27. Yingxin, X., Jinghua, L., Lichun, W., Dehui, K.: A robust hand gesture recognition method
via convolutional neural network. In: 6th International Conference on Digital Home (ICDH),
pp. 64–67. IEEE (2016)
28. Zhu, Y., Huang, C.: An improved median filtering algorithm for image noise reduction.
Phys. Procedia 25, 609–616 (2012)
29. Zisad, S.N., Hossain, M.S., Andersson, K.: Speech emotion recognition in neurological
disorders using convolutional neural network. In: International Conference on Brain
Informatics. pp. 287–296. Springer (2020)
30. Zivkovic, Z., Van Der Heijden, F.: Efficient adaptive density estimation per image pixel for
the task of background subtraction. Pattern Recogn. Lett. 27(7), 773–780 (2006)
Analysis of the Cost of Varying Levels
of User Perceived Quality for Internet
Access
Ali Adib Arnab1(B) , Sheikh Md. Razibul Hasan Raj1 , John Schormans2 ,
Sultana Jahan Mukta3 , and Nafi Ahmad2
1
University of Global Village, Barisal, Bangladesh
adib9877@yahoo.com
2
Queen Mary University of London, London, UK
3
Islamic University, Kushtia, Bangladesh
1 Introduction
Quality of Experience (QoE) has a significant but complex relationship with
Quality of Service (QOS) and it’s underlying factors [1]. Mean Opinion Score
(MOS) ranges from 1 to 5 (Bad to Excellent) and represents QoE [2]. Although
considerable work has been carried out in papers like [3–5], in this paper, we
have considered QOE within a budget with numerous metrics such as PLP,
bandwidth, and round trip time, TCP sending rate factor, packet buffer lengths
and packet bottleneck capacity. From the curve fitting, we obtained an analytical
expression for bandwidth and bandwidth cost. The goodness of fit is obtained
from SSE, R Square, Adjacent R Square and RMSE. Consequently, we found
one equation with variable MOS and bandwidth and another one with variable
bandwidth and bandwidth cost. The analysis has been performed multiple times
for varying the number of TCP sources. The major objective of this research is
to identify the mathematical relationship between MoS and bandwidth cost.
2 Related Work
Well-grounded bandwidth anticipated by the Internet Protocol is a pivotal
requirement for the advanced services, which also includes reliability amplifi-
cation arranged by TCP [6]. The paper signifies QoS (Quality of service) as a
contemporaneous considered issue worldwide that offers further reliable band-
width network traffic is based on TCP which keeps increasing the sending rate.
Packet loss could be resultant from above issues which is analyzed during
the QoS model improvement [7]. The author implies it is also possible for the
bottleneck links to have high utilization without any packet loss. According to
Roshan et al. [8] paper evaluates the relationships between QoS metrics packet
loss probability (PLP) and the user-perceived Quality of Experience (QoE) for
video on demand (VoD).
QoS has major implications from a policy standpoint [9]. The authors indi-
cates about a present situation where users acquire internet service according
to a fixed price. A simple monthly subscription fee or price according to the
connection time is common [10]. Dial-up modem, Ethernet, cable Ethernet, or
DSL are utilized to provide bandwidth. It is very clear from the analysis of the
authors that there may be no bandwidth guarantee or only statistical band-
width guarantees either one can be offered. Conventional Internet applications
mainly utilize TCP as their transport layer protocol [11]. We can understand
from the paper that Packet loss is used as a sign of congestion. As soon as the
sender experiences network congestion, TCP responds quickly by reducing its
transmission rate multiplicative [12].
The UK is ranked third in terms of the services and metrics included in
this analysis followed by France and Germany [13]. In this paper, The UK has
occupied third place in a table of the weighted average and lowest available
basket prices across the services which are covered in analyzing bandwidth cost
at the varying level. USA is the most expensive regarding both average and
lowest available prices across all the services while France is the least expensive.
We can also learn that UK’s highest ranking is acquired by the mobile phone
service price, triple-play bundle price, in which category the UK is ranked second
followed by France.
There are numerous analysis of QoE, QoS and relationship with band-
width where our paper actually provides relationship between QoE and Band-
width cost. This can cover up business aspects of many organizations. It also
implements proper idea about within allocating budget how much QoE can be
achieved which has not yet been discussed in mentioned papers.
building, enterprise voice quality patrons generally expect constant quality while
employing their telephone [14].
3 Method
Our first requirement is to determine PLP and MOS for different numbers of
TCP sources, round trip time, bottleneck capacity in packets and buffer length.
We take buffer length values from 10 to 1000 [8]. Then we obtained a MOS vs
bandwidth graph for different settings of buffer lengths. Different countries pays
different amounts of money for internet access. It largely depends on the internet
providers, facilities, internet access availability of networking telecommunication
product, infrastructure, etc. For example, someone in North America may not
be paying the same as someone in Africa. We considered values for UK which
will give us range of bandwidth and estimated average cost for that bandwidth.
From our analysis we can obtain a graph for bandwidth and bandwidth cost.
Our goal is to relate MOS score and bandwidth cost analytically.
The next step was to use existing formulas for PLP and MOS again. Initially
we used bottleneck capacity as one of the parameters to determine PLP. We
obtained a formula for MOS and bandwidth cost and plotted various bandwidth
cost value and evaluated MOS scores against those values. Also, we obtained a
similar curve which we previously obtained from MOS vs bandwidth. If we want
to differentiate for big organizations and small family houses we need to modify
our formula for MOS by changing number of TCP sources. In big organizations,
the number of TCP sources will be large, and for small household, it can be
much smaller. We modify the number of TCP sources in our initial formula for
MOS. We obtain one graph of MOS vs bandwidth cost for a larger number of
TCP sources and one graph of MOS vs bandwidth cost for smaller number of
TCP sources.
Q PLP MOS
10 7.465263e−01 4.750997e−01
100 6.503074e−01 6.280123e−01
200 5.637028e−01 8.073142e−01
400 4.353297e−01 1.171447e+00
600 3.462922e−01 1.516566e+00
800 2.820191e−01 1.827308e+00
1000 2.341107e−01 2.099708e+00
Method NonlinearleastSquares
Robust LAR
Algorithm Trust Region
DiffMinChange 1.0e-8
DiffMaxChange 0.1
MaxFunEvals 600
Maxlter 400
TolFun 1.0e-6
TolX 1.0e-6
We will take bandwidth as ‘x’ which lies X axis and bandwidth cost as ‘f(x)’
which lies in Y axis. We can acquire values for ‘a’ and ‘b’ from curve fitting tool
which is visible in Fig. 1. After that by implementing the Eq. (3) we obtain Eq.
(4) below:
Cost = 27.13 ∗ bandwidth0.0986 (4)
Fig. 1. Setup for applying Curve fitting formula from Cost vs bandwidth to obtain a
formula for MOS vs bandwidth cost
Analysis of the Cost 399
By taking different values of cost we can get different values of bandwidth and if
we replace bandwidth from Eq. (6) we can get different values for cost and MOS.
The MATLAB code for MOS vs Bandwidth cost is shown in Fig. 2. Bottleneck
capacity is given by Eq. (5) below:
bandwidth ∗ 1000000
C= (5)
12000
MOS and Bandwidth. From Eq. (1), (2) and (5) we obtained relationship
between MOS and bandwidth which is Eq. (6):
11851.851
−44∗
0.01( BW.1000 )2 +2( BW.1000
M OS = 1.46 ∗ e 12 12 )+100
11851.851 (6)
−2.9∗
0.01( BW.1000 )2 +2( BW.1000
+ 4.14 ∗ e 12 12 )+100
MOS and Bandwidth Cost. If we put bandwidth cost in the place of band-
width by help of Eq. (4) we obtain the following relationship which is Eq. (7):
400 A. A. Arnab et al.
If we keep changing buffer length it is very difficult to evaluate MOS change with
the effects of bandwidth change. So by keeping the same buffer length, Q which
is 10, we can obtain a MOS vs bandwidth curve. So it is quite evident from Fig.
4 that when we increase bandwidth, MOS is also increasing proportionally band-
width. As a sample when bandwidth is 50 Mbps, MOS is 3.5 (approximately),
when bandwidth is 100 Mbps MOS is 4.5 (approximately) and when bandwidth
is 150 Mbps MOS is close to 5.
We took some different parameters for bandwidth cost in the UK. If we look
at the Fig. 5 we can see initially the cost increases within a limited range of
bandwidth. In this case even for 20 Mbps the bandwidth price is somewhere
close to 35£ per month. The rate increases until 100 Mbps, from the graph
customer has to pay nearly 43£ for 100 Mbps per month. Beyond 100 Mbps the
cost increment is very slow. From 100 Mbps to 1000 Mbps cost only increases
from 43£ to 46£ per month which is incredibly low. So we can draw conclusions
about how much it is worth spending on bandwidth before we run into a law of
diminishing returns.
402 A. A. Arnab et al.
likely used which also experiences a rise in MOS which provides good quality
within a price range but that is less than that seen in big organizations.
Prior work in the field has shown that QoE has a complex relationship with QoS
factors like packet loss probability (PLP), delay or delay jitter. Furthermore,
contemporary analyses have indicated that the relationship between these QoS
factors and QoE can significantly vary from application to application. In this
paper we take prior analyses of the relationship between the key QoS metric of
packet loss probability and QoE and target an internet access scenario. We use
this relationship to show how QoE (measured as MOS) varies as more money
is spent on the bandwidth of the internet access link. Our results target two
different scenarios – a small number of multiplexed TCP sources and a relatively
large number of multiplexed TCP sources. We show that increase in MOS is not
a linear with increasing spent on bandwidth at all, and considering these two
different we are able to resolve the highly non-linear fashion in which MOS does
increase with spend on bandwidth.
References
1. Quality of Experience Paradigm in Multimedia Services. ScienceDirect (2017)
2. Streijl, R.C., Winkler, S., Hands, D.S.: Mean opinion score (MOS) revisited: meth-
ods and applications, limitations and alternatives. SpringerLink (2014)
3. Ahmad, N., Schormans, J., Wahab, A.: Variation in QoE of passive gaming video
streaming for different packet loss ratios. In: QoMEX 2020- The 12th International
Conference on Quality of Multimedia Experience, Athlone (2020)
4. Ahmad, N., Wahab, A., Schormans, J.: Importance of cross-correlation of QoS
metrics in network emulators to evaluate QoE of video streaming applications.
Bordeaux (2020)
Analysis of the Cost 405
5. Vasant, P., Zelinka, I., Weber, G. W.: Intelligent computing & optimization. in
SpringerLink (2018)
6. Camp, L.J., Gideon, Carolyn.: Limits to Certainty in QoS Pricing and Bandwidth.
22 July 2002. https://dspace.mit.edu/handle/1721.1/1512. Accessed 2 May 2020
7. Xiao, X.: Technical, commercial and regulatory challenges of QoS. Amsterdam
[u.a.]: Elsevier/Morgan Kaufmann, p.30. Zarki, M. (2019). QoS and QoE (2008)
8. Roshan, M., Schormans, J., Ogilvie, R.: Video-on-demand QoE evaluation across
different age-groups and its significance for network capacity. EAI Endorsed Trans-
actions on Mobile Communications and Applications, no. EAI (2018)
9. Gideon, C. (2002). Limits To Certainty in QoS Pricing and bandwidth. http://hdl.
handle.net/1721.1/1512. Accessed 11 Aug 2019
10. Internet Cost Structures and Interconnection Agreements. The journal of electronic
publishing (1995)
11. Joutsensalo, J., Hämäläinen, T.D., Siltanen, J., Luostarinen, K.: Delay guaran-
tee and bandwidth allocation for network services. In: Next Generation Internet
Networks, 2005 (2005)
12. Aragon, J.C.: Analysis of the correlation between packet loss and network delay
and their impact in the performance of surgical training applications. In: Semantic
Scholar (2006)
13. Ofcom (2017). International Communications Market Report 2017. London.
https://www.ofcom.org.uk/research-and-data/multi-sector-research/cmr/cmr-
2017/international. Accessed 7 Aug 2019
14. Expert System (2017). What is Machine Learning? A definition - Expert System.
Expertsystem.com. https://www.expertsystem.com/machine-learning-definition/.
Accessed 6 Aug 2019
15. Bisio, I., Marchese, M.: Analytical expression and performance evaluation of TCP
packet loss probability over geostationary satellite. IEEE Commun. Lett. 8(4),
232–234 (2004)
16. Cable.co.uk (2019). Best Broadband Deals August 2019 — Compare Broadband
Offers - Cable.co.uk. Cable. https://www.cable.co.uk/broadband/. Accessed 16
Aug 2019
Application of Customized Term
Frequency-Inverse Document Frequency
for Vietnamese Document Classification
in Place of Lemmatization
1 Introduction
Vietnamese is a special language whether how it was formed and developed, or how it
is expressed and used. The language itself is ambiguous, but we can find various
reasons as why Vietnamese is not a simple language to be processed by a typical NLP
model such as deep learning model [1], thus we need customization.
As a historical result, Vietnamese vocabulary is majorly structured from words
derived from Chinese, notably words in scientific or politic domains, namely Han-Viet
(Vietnamese words taken from Chinese since the Han-Dynasty era), which existence
takes roughly 70% [2]. Moreover, because of French colonization, Vietnamese had
taken in a numerous loanword from French language in multiple form. Recently, many
words from English have joined the Vietnamese vocabulary family, either being
translated word-by-word or being used directly as a part of Vietnamese [2].
NLP researches in Vietnam have been greatly delayed compare to what it could
have reached, as they have been made without a common, academic direction, there are
some notable projects but still lack the systematic approach to the some important
issues (text summarization, context based intelligent search, etc.). The lack of
Table 1. Reasons why Vietnamese are relatively easy for human to learn, but cause difficulties
for Machine Learning.
For human For NLP efficiency
Short words Easy to learn Repetitive structure, hence,
many synonyms and
homophones
No gender Easy to learn Inefficient lemmatization
No plural, no cases Easy to speak, hard to No lemmatization needed,
listen thus no lemmatization
No articles Easy to speak, hard to No lemmatization needed
listen
No verb conjugation Easy to learn Difficult to differentiate
between tenses or nouns’
genres
Simple verb tenses Easy to speak, hard to Shift the complicating task(s)
comprehend onto the listening side instead
of the speaking side, thus
increasing NLP difficulties
Tenses are optional Easy to speak, hard to Shift the complicating task(s)
comprehend onto the listening side instead
of the speaking side, thus
increasing NLP difficulties
No agreement Faster to learn, less No explicit relationship in
knowledge to terms of data between words
remember in a sentence
Single syllable words Easy to understand Increase the difficulties in
compound words if one distinguishing if a word is a
understands each compound word or it is
element multiple separated words
Homophones: due to single- Hard to distinguish Hard to distinguish
syllable characteristics, in
Vietnamese, homophones also
mean words that are written
identically
Vietnamese NLP experts have been working to address these steps with different
approaches. This project does not have the ambition to overcome all these challenges
by itself, we will implement some works of other experts before adding our contri-
bution. Our work concern in the differentiating potential of synonyms and classifiers in
Vietnamese.
Ho Le et al. conducted a systematic research using Vectorized IF-TDF. The authors
prepared their dataset for training and archived an average impressive result of 86,7%,
and peak result of 92.8%. We follow their team’s footsteps [4] while conducting this
paper. Using several different libraries, our average result provides comparable per-
formance with faster processing speed, and minor accuracy rate improvement.
The following sections are ordered as: Sect. 2 describes our methodology, Sect. 3
explains and describes the pre-processing step; we provide some case studies in
Sect. 4; and finally, Sect. 5 is reserved for conclusions and what to do next.
2 Methodology
Looking further into the characteristic of these cases in [2] Sect. 2.1.2, we gathered
the statistic of these cases by Parts of Speech (POS) as presented in Table 3.
Lastly, we reordered these data onto files to represent the following classes:
• Western fantasy novel (Lord of the ring)
• Western fairy tale (Grimm stories)
• Vietnamese fairy tale and stories for kids (Diary of a Cricket, Vietnamese fables)
• World’s politic
• Vietnam’s showbiz
• Vietnam’s economics
• Others
It would be acceptable if in our test run, we find out that a text can exist in more
than one category, for example, Vietnam’s economics matters may coerce with World’s
politic event(s) in one way or another, and a predict in either one of these categories can
be considered acceptable.
4 Data Training
To train the collected data, first we perform sentences tokenizing, using the sentence-
boundary-detection based on Naïve Bayes algorithm by Vuong Quoc Binh and Vu Anh
[4]. The result of this step is one text file for each document, with each sentence in a
separated line, for example, with the story “Diary of a Cricket” we have a chunk of
result as sampled in Table 4.
Noted that we will perform this step for all included words, including proper noun
for an individual person, place, or organization. These are valuable data that helps with
document classification, for e.g., multiple presidents’ name will help concluding the
said document is politic related. Moreover, special characters are also POS-tagged, as
dialogs keep their respectively quotation marks, the end of sentences can be a period or
a quotation mark depending on the nature of the sentence itself. This reservation helps
with indicating if a certain category will include a lot of dialogues by counting quo-
tation marks, or we indicate that brackets are mostly used in scientific articles and
books, or sometime used in novels.
Next, we transform these data into a combined state, where each item consists of a
word and its POS-tag, which we address as CWP (combined word with POS-tag).
Following what we have been studying, this combined item effectively differentiate
itself with its homophones as effective as 77.31% or better. The output of this step
should look like the sample presented in Table 6.
Application of Customized Term Frequency-Inverse Document Frequency 413
Table 6. Sample output of POS-tag and Word combining step performed on “Diary of a
Cricket”.
…
tôi__P là__V em_út__N,__CH bé__N nhất__A nên__C được__V mẹ__N tôi__P sau__N
khi__N dắt__V vào__E hang__N,__CH lại__R bỏ__V theo__V một_ít__L ngọn__Nc cỏ__N
non__A trước__E cửa__N,__CH để__E tôi__P nếu__C có__V bỡ_ngỡ__A,__CH thì__C
đã__R có__V ít__A thức_ăn__N sẵn__A trong__E vài__L ngày__N.__CH
rồi__C mẹ__N tôi__P trở_về__V.__CH
tôi__P cũng__R không__R buồn__V.__CH
trái_lại__N,__CH còn__C thấy__V làm__V khoan_khoái__N vì__E được__V ở__V
một_mình__X nơi__N thoáng_đãng__V,__CH mát_mẻ__N.__CH
tôi__P vừa__R thầm__A cảm_ơn__V mẹ__N,__CH vừa__R sạo_sục__V thăm__V tất_cả__P
các__L hang__N mẹ__N đưa__V đến__V ở__V.__CH
khi__N đã__R xem_xét__V cẩn_thận__A rồi__T,__CH tôi__P ra__V đứng__V ở__E ngoài__E
cửa__N và__C ngửng_mặt__V lên__V trời__N.__CH
qua__V những__L ngọn__Nc cỏ__N ấu__N nhọn__A và__C sắc__V,__CH tôi__P thấy__V
màu__N trời__N trong__E xanh__A.__CH
tôi__P dọn_giọng__V,__CH vỗ__V đôi__M cánh__N nhỏ__A tới__E nách__N,__CH rồi__C
cao_hứng__V gáy__V lên__V mấy__L tiếng__N rõ__A to__A.__CH
từ__E đây__P,__CH tôi__P bắt_đầu__V vào__E cuộc_đời__N của__E tôi__P.__CH
…
Since Vietnamese compound words are multiple single-syllable words being written
next to each other, to present these words, we use underscore special character “_” to
separate its units and its POS-tag. Thus, the noun “thức ăn” (food) will be presented as
“thức_ăn__N”. The final output of this step is a list of all sentences being tokenized,
with a label for each of them, stating which category it belongs to. For the sake of
simplicity, we have written these data to a separate file for each book or each article.
Fourth step is where we merge the output files of third step into two large datasets:
one with CWP and one without. This way we can compare the difference between bare-
bone training and CWP method. These datasets are written onto separate files, but with
the same structure as following: a csv file with two columns: Encoded Label and
Processed sentence. Encoded Labels are used to differentiate between classes to be
classified, and to improve processing performance. We have used a sample Encoded
Labels set as in Table 7.
Step five, we perform TD-IDF vectorizer by building up weight matrix. This step
evaluates the relevance of each CWP toward its sentence and its Encoded Label. We
will perform this step on top of the training dataset, consisting 80% number of rows in
the combined dataset in step four. The output of this step is presented in Table 8.
The TF-IDF vocabulary can now be accessed, with top rows being as in Table 9 as
sampling data.
We can now look at some cases where CWP effectively differentiate homophones
as sampled in Table 10.
Application of Customized Term Frequency-Inverse Document Frequency 415
Afterward, we train and test with Support Vector Machine (SVM). The reason for
this choice is because it is a very universal learner, and it is independent of the
dimensionality of the feature space according to Thorsten Joachims [14]. It is proven to
be effective and can be used to generalize even with many features involved, which is
ideal in our case study. In practice, with some tuning, we achieve the following result in
Table 11.
The highest accuracy rate we received is 91.075% with CWP, and 90.179%
without CWP. Based on these data, we concluded that CWP does not always give
better results, but with tuning it can give significant improvement over without.
For comparison purpose, we also perform Naïve-Bayes Classifier algorithm train-
ing model (arguably providing among one of the fastest processing speeds out of the
box) with a slight improvement as stated in Table 12.
The best we were able to achieve without overfitting is 88.329% with CWP, and
87.551% without CWP. The trend followed through as we needed proper tuning for
this result. It is also safe to conclude that in general, even though SVM leads the
accuracy rate in most of cases, Naïve-Bayes Classifier still process faster with rea-
sonably accuracy rate.
Comparing with results from “underthesea group”, who has been developing and
maintaining the popular NLP python toolkit for Vietnamese language processing, that
they have also published in their official website [2] and quoted as in Table 13, we
conclude that after skipping lemmatizing step, CWP provide a modestly comparable
result. This proves that CWP not only saves processing power, but also remain com-
petitive in accuracy front.
Table 13. underthesea group’s work result – version 1.1.17 - 2020 [2].
TfidfVectorizer(ngram_range = (1, 2), max_df = 0.5) 92.8
CountVectorizer(ngram_range = (1, 3), max_df = 0.7) 89.3
TfidfVectorizer(max_df = 0.8) 89.0
CountVectorizer(ngram_range = (1, 3) 88.9
TfidfVectorizer(ngram_range = (1, 3)) 86.8
CountVectorizer(max_df = 0.7) 85.5
This paper proposes the combination of a word with its POS to give it a unique ID, thus
effectively distinguish between homophones while preserving the difference between
synonyms. The dataset is collected and (semi-)manually processed from various fields,
then a sequence of techniques for customizing and training learning model is used. The
classification model could contribute positively to many Vietnamese text-analyzing
based businesses, such as social network, e-commerce, etc. The work results prove that
Application of Customized Term Frequency-Inverse Document Frequency 417
CWP not only saves processing power but also competes for accuracy with the best-in-
class results in the same field.
For future work, we shall apply studied method into sentiment analysis of social
network posts and comments, e-commerce reviews analysis for data mining purpose.
The study also can give a reference to other fields, in general, any text processing
field that requires synonyms difference preservation such as social network analysis,
sentiment analysis [15, 16].
References
1. Cai, J., Li, J., Li, W., Wang, J.: Deeplearning model used in text classification. In:
Proceedings 15th International Computer Conference on Wavelet Active Media Technology
and Information Processing (ICCWAMTIP), Chengdu, China, pp. 123–126 (2018)
2. Le, H., Hoang, T., Toan, D.M.: Homophone and polysemous words in Vietnamese (in
comparison with modern Han language) - Đồng âm và đa nghĩa trong tiếng Việt (Đối chiếu
với tiếng Hán hiện đại) (Vietnamese) (2011)
3. Halpern, J.: Is Vietnamese a hard language? http://www.kanji.org/kanji/jack/vietnamese/is_
VN_hard_sum_EN_VN.pdf
4. Le, H. et al.: Vietnamese NLP Toolkit https://underthesea.readthedocs.io
5. Institute of Linguistics of Vietnam: Vietnamese dictionary - Republish 12th ed, (2006)
6. Thuat, D.T.: Vietnamese phonetic (Ngữ âm tiếng Việt (Vietnamese)), Hanoi, p. 89 (1977)
7. Tolkien, J.R.R.: Lord of the Ring - Book 1: The Fellowship of the Ring, Vietnamese
Literature publisher, translated by Yen. N.T.T, Viet, D.T. (2013)
8. A collection of Grim fairy tales, Dan Tri publisher, translated by various members (2008)
9. Hoai, T.: Diary of a Cricket, Kim Dong publisher (2014)
10. 109 modern Vietnamese fables – Various artists, Hong Duc publisher (2018)
11. Various newspaper published during 2019–2020 at https://vnexpress.net
12. Various newspaper published during 2019–2020 at https://thanhnien.vn
13. Various newspaper published during 2019–2020 at https://dantri.com.vn
14. Thorsten, J.: Text Categorization with Support Vector Machines: Learning with Many
Relevant Features (2005)
15. Hung, P.D., Giang, T.M., Nam, L.H., Duong, P.M., Van Thang, H., Diep, V.T.: Smarthome
control unit using Vietnamese speech command. In: Vasant, P., Zelinka, I., Weber, G.W.
(eds.) Intelligent Computing and Optimization. ICO 2019. Advances in Intelligent Systems
and Computing, vol 1072. Springer, Cham (2019)
16. Tram, N.N., Hung, P.D.: Analyzing hot Facebook users posts’ sentiment. In: Emerging
Technologies in Data Mining and Information Security Proceedings of IEMIS (2020)
A New Topological Sorting Algorithm
with Reduced Time Complexity
1 Introduction
Topological sorting for a directed acyclic graph (DAG) means the linear ordering of the
vertices of a graph. In this ordering for every directed edge (u, v) from vertex u to
vertex v, u comes before v. In a graph, vertex represents task which has to perform and
edge indicates constraints that the task performed one after another, then a valid
sequence for the tasks is the topological ordering of the tasks. Topological ordering is
possible only for DAGs. If at least one cycle exists in a graph then the topological
ordering of this graph is not possible. Every DAG has at least one topological ordering.
The first topological sorting algorithm was given by Kahn (1969). In Kahn’s
algorithm, for topological sorting first have to take the vertices from the graph which do
not have any incoming edges. If the graph is acyclic then the set must contain at least
one vertex that does not have any incoming edges. From this set of vertices, graph
traversing will begin. Then have to find the vertices of outgoing edges and have to
remove them. For other incoming edges this vertex have to check again, if there are no
incoming edges then this vertex has to insert into the set of sorted elements. So, Kahn’s
algorithm uses a breadth first search technique. In this algorithm, we have to check all
the vertices and edges for the start vertex. Then have to start check every elements over
again. So, the overall time complexity of Kahn’s algorithm is O(|V| + |E|).
Another popular topological sorting algorithm is DFS topological sorting algo-
rithm, which is based on DFS graph traversal technique. In DFS topological sorting
algorithm, graph traversing started from the start vertex and traversing goes through
down from the start vertex. When a vertex is found which contain more incoming
vertices then by backtracking procedure a previous level vertex is traversed. The time
complexity of this DFS topological sorting algorithm is also O(|V| + |E|).
In the proposed algorithm traversing of the full graph is not needed. Just list the
empty vertices (vertices contain no outgoing edge) in a set and non-empty vertices
(vertices contain at least one outgoing edge) in another set, then order the vertices by
removing empty vertices fromP non-empty vertices set. In our proposed algorithm time
jvj
complexity reduced to O
ðNE Þ , because of using two sets and here also
i¼1 i
elements will decrease in each step. The space complexity of our proposed algorithm is
O(|V| + |E|). So, the proposed algorithm is completely new that finds the topological
ordering of a graph in an optimum time. The proposed algorithm also can detect cycle
in a graph. This is a completely new algorithm for topological sorting which reduces
the time complexity for topological sorting. This algorithm avoids using any graph
traversing algorithm like BFS or DFS. This is a new approach for simply implemen-
tation of topological sorting algorithm. This algorithm returns false, if any cycle pre-
sents in the graph that means, then not any topological sorting is possible. So, this new
topological sorting algorithm can be used for detecting cycle in the graphs. This
algorithm implementation is very simple.
Topological sorting is a popular algorithm for scheduling tasks. From a long period
of time researchers continuing research over this topic for improving this algorithm and
they also working on application of topological sorting. So, there is a large space for
researching on this algorithm. The application of topological sorting algorithm is huge.
It can be used for complex database tables that have dependencies. It is also useful for
ordering object in a machine, course prerequisite system and many more. Topological
ordering is mostly used for planning and scheduling.
2 Related Works
case of complex and large system this algorithm increases the complexity. In a food
supply chain, for checking the adulterant nodes, a new algorithm is proposed in paper
[4]. A novel optimization algorithm is proposed in paper [5] to find the solution of
many diverse topological sorts which implies many cut-sets for a directed acyclic
graph. For enhancing grid observation, the proposed algorithm is very effective. For
large graphs, another topological sorting algorithm is given in paper [6]. For topo-
logical sorting, they proposed an I/O-efficient algorithm named IterTs. But, this
algorithm is inefficient for worst case. A new algorithm is given in paper [7] that gives
the topological sorting of sparse digraphs with better time complexity than all other
previous algorithms. This topological sorting algorithm is dynamically maintained.
They have given an experimental comparison result of topological sorting for large
randomly generated directed acyclic graphs. In paper [8], P. Woelfel solved the
topological sorting with O(log2|V|) OBDD operations. For a fundamental graph
problem, it is the first true runtime analysis of a symbolic OBDD algorithm. In paper
[9], they proposed some algorithms to solve the problem of single-source shortest path,
computing a directed ear decomposition and topological sorting of a planar directed
acyclic graph in O(sort(N)) I/Os, here sort(N) is the number of I/Os needed to sort N
elements. All topological ordering of a directed acyclic graph is found in paper [10].
They proposed an extended algorithm which is capable of finding all topological
solutions of a DAG. To implement this algorithm they have used backtracking, iter-
ation in the place of recursion, data structures and many more. In paper [11], they
proposed a new algorithm by using parallel computation approach. To find topological
sorting of a directed acyclic graph, this algorithm is very simple. In this algorithm, they
have traversed the nodes parallel from a node and marked as visited. After visiting all
the source nodes, all other nodes in the graph is marked as visited. For acyclic graph the
parallel traversing will terminate. For an SIMD machine the implementation of this
algorithm is discussed in this paper. Time complexity of this proposed parallel algo-
rithm is ordered for longest distance between a source and a sink node for a directed
acyclic graph. Large networks topological sorting is given in paper [12]. The general
topological algorithms are not efficient to find topological ordering for a large network.
So, A. B. Kahn proposed this procedure to find the topological ordering of large
networks efficiently. By using this algorithm a PERT (Program Evaluation Review
Technique) network has been maintained which contains 30,000 activities which can
be ordered in less than one hour of machine time. This is an efficient method because
the location of any item is known so that no need to searching an item and for correct
sequence of network there only needed a single iteration, so the method go in single
step for this case. If all events are traversed or any cycle found in the network then the
network is terminated. This method is more efficient than all other topological algo-
rithm for large network. In paper [13], T. Ahammad et al. proposed a new dynamic
programming algorithm to solve job sequencing problem that reduces the time com-
plexity of job sequencing problem from O(n2) to O(mn). They have used the tabulation
method of dynamic programming. They have showed the step by step simulation of
their algorithm working procedure with experimental analysis. In paper [14], J.F.
Beetem proposed an efficient algorithm for hierarchical topological sorting. This
algorithm is also applicable in the presence of apparent loops. Its application is also
described in this paper.
A New Topological Sorting Algorithm with Reduced Time Complexity 421
In this proposed algorithm, from a graph we put the empty vertices means those
vertices which contains no outgoing edges in a set E and the non-empty vertices means
those vertices which contains at least one outgoing edge in another set NE. In each step,
we will take an empty vertex and remove this vertex from all non-empty vertices and
also remove this particular vertex from set E. In this procedure, if any vertex becomes
empty then we will put it in set E. The topological ordering is the set T, the sequence in
which we choose the empty vertices from the set E. When the set E becomes empty
then the algorithm terminated and if T contains all vertices of the graph then no cycle
exists and the graph is a DAG and T is the correct topological order. But if T doesn’t
contain all vertices of the graph then the graph contains cycle and the graph is not
DAG, then any topological ordering is not possible (Fig. 1).
4 Algorithm
4.1 Pseudocode
Here G is the set of vertices of the graph which divided into two set E (contains empty
vertices) and NE (contains non-empty vertices). Empty vertex contains no outgoing
edges and non-empty vertex contains outgoing edges.
When the set of empty vertices E becomes empty then the ordering is completed. If
the size of T is not equal to the size of G then the graph has at least one cycle and the
graph is not DAG. So, the topological sorting of this graph is not possible. Otherwise if
the size of T is equal to the size of G then topological ordering is possible and the
elements of set T in reverse order is in topological ordered. From this algorithm, we can
get the correct topological ordering of a graph in set T after reversing T. If topological
ordering is not possible or the algorithm detect any cycle then the algorithm return false
otherwise the algorithm return set T. T is correct topological ordering of graph G. By
this procedure this algorithm also detect cycle in a graph.
4.2 Example
Let a directed acyclic graph (DAG) is given (see Fig. 2). Have to find a correct
topological ordering of this graph.
A New Topological Sorting Algorithm with Reduced Time Complexity 423
The graph has 14 vertices numbered as (0−13) and 16 edges (Fig. 2). So the
adjacency list of this graph-
0->
1->9
2->0
3->2
4->5, 6, 7
5->
6->7
7->8
8->
9->4
10->0, 1, 12
11->13
12->0, 11
13->5
Since the graph is a directed acyclic graph (DAG), so the graph has at least one
topological ordering exist. Here the empty vertices (vertex with no outgoing edges) are
(0, 5, 8). We can take those empty vertices in any order. Each order is valid for a
particular step. After taking those empty vertices the procedure begins. If the graph is a
DAG then in the first step at least one empty vertex will exists.
Here showed adjacency list simulation for finding a topological ordering of this
graph. Here first table represents the adjacency list of the following graph (Fig. 3). By
removing its empty vertices, the size reduces which is shown below-
424 T. Ahammad et al.
Fig. 3. Step by step visualization of proposed algorithm working procedure with an example.
Step 1: Find the empty vertices (0, 5, 8) and insert these vertices in a set T (0, 5, 8).
Then remove these vertices from adjacency list and also remove these from all other
vertices. Now the empty vertices are (2, 7, 13).
Step 2: Find the empty vertices (2, 7, 13) and insert these vertices in the set T(0, 5, 8,
2, 7, 13). Then remove these vertices from adjacency list and also remove these from
all other vertices. Now the empty vertices are (3, 6, 11).
Step 3: Find the empty vertices (3, 6, 11) and insert these vertices in the set T(0, 5, 8,
2, 7, 13, 3, 6, 11). Then remove these vertices from adjacency list and also remove
these from all other vertices. Now the empty vertices are (4, 12).
Step 4: Find the empty vertices (4, 12) and insert these vertices in the set T(0, 5, 8, 2,
7, 13, 3, 6, 11, 4, 12). Then remove these vertices from adjacency list and also remove
these from all other vertices. Now the empty vertex is 9 only.
Step 5: Find the empty vertex 9 and insert this vertex in the set T(0, 5, 8, 2, 7, 13, 3, 6,
11, 4, 12, 9). Then remove this vertex from adjacency list and also remove this from all
other vertices. Now the empty vertex is 1 only.
Step 6: Find the empty vertex 1 and insert this vertex in the set T(0, 5, 8, 2, 7, 13, 3, 6,
11, 4, 12, 9, 1). Then remove this vertex from adjacency list and also remove this from
all other vertices. Now the empty vertex is 10 only.
A New Topological Sorting Algorithm with Reduced Time Complexity 425
Step 7: Insert the last empty vertex 10 in the set T(0, 5, 8, 2, 7, 13, 3, 6, 11, 4, 12, 9, 1, 10).
Then remove this vertex from adjacency list. Now there is no such empty vertex remaining.
After reversing T, we will get the topological ordering of the graph. The topological
ordering of the graph by proposed algorithm is-
(10->1->9->12->4->11->6->3->13->7->2->8->5->0)
By Kahn’s algorithm, the topological ordering is-
(3->10->2->1->12->9->0->11->4->13->6->5->7->8)
By DFS toposort algorithm, the topological ordering is-
(10->12->11->13->3->2->1->9->4->6->7->8->5->0)
The proposed topological sorting algorithm is less complex than Kahn’s and DFS
topological sorting algorithms.
5 Analysis
P
The time complexity of this proposed algorithm is O
j vj ðNE Þ and the space
i¼1 i
complexity of this proposed algorithm is O(|V| + |E|). The calculation process of the
time and space complexity is given in the next sections.
: :
: :
: :
i ¼ jV j. . .ðNE Þjvj
426 T. Ahammad et al.
Total complexity will be counted for the summation of non-empty vertex elements of
each step. So, for 1 to |V|
Compexity ¼ ðNE Þ1 þ ðNE Þ2 þ ðNE Þ3 þ ðNE Þ4 þ ðNE Þ5 þ ðNE Þ6 þ . . . þ ðNE Þjvj
Xjvj
¼ ðNE Þ ; where i ¼ 1; 2; 3; 4; 5; 6; . . .; jV j
i¼1 i
ð1Þ
X
jmj
jðNEÞi j
i¼1
The time complexity for best case, average case and worst case is depends on the
jðNEÞi j term.
X
jmj
jðNEÞi j
i¼1
If the graph has not any edge then empty vertex set E will be full and non-empty
vertex set NE will be empty. Then only the outer loop will execute |V| times. Here,
jðNEÞi j is the total number of elements in non-empty vertex set where (i) is step
number. The value of jðNEÞi j will 1 for each step means-
|(NE)1| = |(NE)2| = |(NE)3| = |(NE)4| = |(NE)5| = |(NE)6| = 1
Equation (1) becomes-
Total complexity ¼ 1 þ 1 þ 1 þ . . . þ 1
¼ jV j
So, the time complexity for the best case of this algorithm is X(|V|), where |V| is the
total number of vertices of the graph.
X
jmj
jðNEÞi j
i¼1
Time complexity for the average case depends on the jðNEÞi j term. The outer loop
of the algorithm will execute for |V| number of times and the inner loop will be execute
for |NE| number of times.
A New Topological Sorting Algorithm with Reduced Time Complexity 427
P
jvj
So, the average case of this algorithm is H i¼1ðNE Þ .
i
Which depends on the jðNEÞi j term. jðNEÞi j is the total number of elements in non-
empty vertex set in each step(i).
X
jmj
jðNEÞi j
i¼1
Here, the complexity depends on the jðNEÞi j term. In worst case the |NE| will be
greater than average case. Here also the outer loop of the algorithm will execute for |V|
number of times and the inner loop will be execute for |NE| number of times.
Equation (1) becomes-
Compexity ¼ ðNE Þ1 þ ðNE Þ2 þ ðNE Þ3 þ ðNE Þ4 þ ðNE Þ5 þ ðNE Þ6 þ . . . þ ðNE Þjvj
Xjvj
¼ ðNE Þ ; where i ¼ 1; 2; 3; 4; 5; 6; . . .; jV j
i¼1 i
P
jvj
So, time complexity for the worst case of this algorithm is O i¼1ðNE Þ .i
Which depends on the jðNEÞi j term. jðNEÞi j is the total number of elements in non-
empty vertex set in each step(i).
It is clear that the proposed algorithm’s time complexity is less than all other topo-
logical sorting algorithms (Table 1).
In this paper, we proposed a completely new algorithm in simplest way where no such
graph traversing technique is needed and the complexity has reduced. Time complexity
for the best case becomes X(|V|), here |V| is
P the total number of vertices of the graph.
jvj
Time complexity for the average case is H
i¼1 ðNE Þi , here |NE| is the number of
P
jvj
non-empty edges. Time complexity for the worst case is O i¼1 ðNE Þi . Normally
the time complexity of general topological sorting algorithm is O(|V| + |E|). In this
paper, we have showed that our algorithm is better which gives results faster than any
other topological sorting algorithm. As in future, we will again try to reduce the
average case and worst case time complexity. We will also try to do further research
about the application of this proposed algorithm.
References
1. Zhou, J., Müller, M.: Depth-first discovery algorithm for incremental topological sorting of
directed acyclic graphs. Inf. Process. Lett. 88, 195–200 (2003)
2. Kalvin, A.D., Varol, Y.L.: On the generation of all topological sorting. J. Algorithm 4, 150–
162 (1983)
3. Liu, R.: A low complexity topological sorting algorithm for directed acyclic graph. Int.
J. Mach. Learn. Comput. 4(2) (2014)
A New Topological Sorting Algorithm with Reduced Time Complexity 429
4. Barman, A., Namtirtha, A., Dutta, A., Dutta, B.: Food safety network for detecting
adulteration in unsealed food products using topologcal ordering. Intell. Inf. Database Syst.
12034, 451–463 (2020)
5. Beiranvand, A., Cuffe, P.: A topological sorting approach to identify coherent cut-sets within
power grids. IEEE Trans. Power Syst. 35(1), 721–730 (2019)
6. Ajwani, D., Lozano, A.C., Zeh, N.: A topological sorting algorithm for large graphs.
ACM J. Exp. Algorithmic 17(3) (2012)
7. Pearce, D.J., Kelly, P.H.J.: A dynamic topological sort algorithm for directed acyclic graphs.
ACM J. Exp. Algorithmic 11(1.7) (2006)
8. Woelfel, P.: Symbolic topological sorting with OBDDs. J. Discrete Algorithms 4(1), 51–71
(2006)
9. Arge, L., Toma, L., Zeh, N.: I/O-efficient topological sorting of planar DAGs. In:
Proceedings of the Fifteenth Annual ACM Symposium on Parallel Algorithms and
Architectures, pp. 85–93 (2003)
10. Knuth, D.E., Szwarcfiter, J.L.: A structured program to generate all topological sorting
arrangements. Inf. Process. Lett. 2(6), 153–157 (1974)
11. Er, M.C.: A parallel computation approach to topological sorting. Comput. J. 26(4) (1983)
12. Kahn, A.B.: Topological sorting of large networks. Commun. ACM 5, 558–562 (1962)
13. Ahammad, T., Hasan, M., Hasan, M., Sabir Hossain, M., Hoque, A., Rashid, M.M.: A new
approach to solve job sequencing problem using dynamic programming with reduced time
complexity. In: Chaubey, N., Parikh, S., Amin, K. (eds.) Computing Science, Communi-
cation and Security. COMS2 2020. Communications in Computer and Information Science,
1235. Springer, Singapore (2020). https://doi.org/10.1007/978-981-15-6648-6_25
14. Beetem, J.F.: Hierarchical topological sorting of apparent loops via partitioning. IEEE Trans.
Comput. Aided Des. Integr. Circuits Syst. 11(5), 607–619 (1992)
15. Intelligent Computing & Optimization, Conference proceedings ICO 2018, Springer, Cham,
ISBN 978-3-030-00978-6
16. Intelligent Computing and Optimization, Proceedings of the 2nd International Conference on
Intelligent Computing and Optimization 2019 (ICO 2019), Springer International Publish-
ing, ISBN 978-3-030-33585-4
Modeling and Analysis of Framework
for the Implementation of a Virtual Workplace
in Nigerian Universities Using
Coloured Petri Nets
1 Introduction
The sudden emergence of the COVID-19 pandemic has impacted the workplace.
Several nations of the world have introduced lockdown measures and social distancing
in a bid to curtail the spread of the novel coronavirus. This has led to the mass
implementation of virtual workplace by organizations in an attempt to ensure business
continuity [1]. In Nigeria, only a few organizations have adopted the virtual workplace
before the global outbreak of the COVID-19 Pandemic. Unfortunately, Nigerian
Universities are among those that are yet to embrace the concept of virtual workplace
[2] even though the recent technological advancement in information and communi-
cation technology has altered the workspace, and has ushered a new form of work
known as virtual workplace [3, 4]. Although the impact of the COVID-19 in Nigerian
Please note that the AISC Editorial assumes that all authors have used the western naming
convention, with given names preceding surnames. This determines the structure of the names in the
running heads and the author index.
universities are enormous due to the closure of schools [2, 5], this crisis offers a great
opportunity for policymakers in the educational sector to explore ways on how to
reposition Nigerian universities by adopting the necessary technologies needed for e-
learning, and also creating a virtual workplace for staff.
In a virtual workplace, employees work remotely away from the central office,
communicate with colleagues, and managers through electronic means [6]. Research
suggests that the implementation of a virtual workplace comes with a lot of benefits
such as enabling employees to work wherever it is convenient for them to carry out
their work [4], reduction in traffic congestion and vehicular emissions [7–10], greater
talent pool, less unnecessary meetings, reduction in commuting time, increased
employees productivity [3], reduction in vehicle miles traveled [11], provides oppor-
tunities for people with disabilities to work from home, savings to the organization in
terms of cost of real estate [3, 12], decrease energy consumption, and ultimately leads
to a significant reduction in air pollution [10].
Despite the enormous benefits that can be derived from the implementation of a
virtual workplace, there is currently minute theoretical and empirical research as
regards the development of a framework for the implementation of a virtual workplace
in developing countries, and more specifically Nigeria. Although there are few pub-
lications and blog posts about how best to implement a virtual workplace, however,
many of these materials are not research-based [13]. According to [14] organizations
still lack much-needed theoretical and empirical support as regards the planning and
implementation of a virtual workplace.
The objective of this study was to transform the phases and activities of the
framework for the implementation of a virtual workplace in Nigerian Universities into a
CPN model and to analyze the behaviour of the model. Interestingly, Coloured Petri
Nets have been used by several authors in modeling workflow systems and business
processes, manufacturing systems, computer networks, and embedded systems [15–17,
19]. This motivated us to used coloured Petri Nets to capture the various phases and
activities of the framework for the implementation of a virtual workplace in Nigerian
Universities into a CPN model.
The advantage of using Coloured Petri Nets to model the framework for the
implementation of a virtual workplace is that being a graphical model and also an
executable representation [16, 18], the CPN model will help to provide a better
understanding of the implementation framework than textual description. Also, simu-
lation from the CPN model will provide a more accurate view of the framework for the
implementation of a virtual workplace, and could be used to demonstrate to manage-
ment and other stakeholders on the best approach to implementing a virtual workplace.
The choice of coloured Petri Nets was based on fact that it is a graphical modeling
language and executable [19], and the analysis methods [15, 19].
The rest of this paper is organized as follows: Sect. 2 shows the proposed imple-
mentation framework for virtual workplace. Section 3 presents the developed CPN
model of the virtual workplace. Section 4 describes the state space analysis. Section 5
concludes the paper.
432 J. Okpor and S. T. Apeh
This section gives an overview of the developed framework for the implementation of a
virtual workplace in Nigerian universities. The framework for the implementation of a
virtual workplace was developed based on a comprehensive literature review, and data
obtained from selected organizations in Nigeria that have successfully implemented
virtual workplace in their respective organizations. Although the framework was
developed for Nigerian universities, it can be adopted by other organizations in Nigeria
that intend to implement the concept of a virtual workplace. The developed framework,
when adopted, will assist universities and other organizations in the successful
implementation of a virtual workplace in Nigeria. The developed virtual workplace
framework is composed of four phases, namely conceptual phase, pre-implementation
phase, implementation phase, and post-implementation phase as shown in Fig. 1.
In the conceptual phase, the university set up a committee that will be empowered
with the responsibility of identifying the benefits, barriers, and risks associated with
virtual workplace implementation. The implementation of a virtual workplace should
only be considered after an adequate evaluation of university needs for the virtual
Modeling and Analysis of Framework 433
The Hierarchical CPN model of the virtual workplace was developed based on the
framework for the implementation of a virtual workplace in Fig. 1, where the various
phases and activities involved in the implementation framework were used to develop a
CPN model of the virtual workplace, and is shown in Fig. 2. The top-level module of
the Virtual Workplace CPN model in Fig. 2 consists of 10 places and 4 substitution
transitions (drawn as rectangular boxes with double-line borders) representing the
conceptual phase, Pre-implementation phase, implementation phase, and post-
implementation phase. Each of the substitution transition has a substitution tag (sub-
module) beneath it, that model a more detailed behaviour of the conceptual phase, Pre-
implementation phase, implementation phase, and post-implementation phase.
The relationship between the submodule and their functions is described as follows:
The conceptual phase module in Fig. 3 represents the initial phase of the virtual work-
place implementation process. The benefits, barriers, and risks are identified in this
module, and thereafter a decision is made either to accept or abandon the virtual work-
place implementation. Once a decision is made to accept a virtual workplace, the next
phase (pre-implementation phase) involves series of activities such as the identification of
university needs for the virtual workplace, job suitability, cost analysis, determination of
the impact, and the identification of success factors as indicated in the pre-implementation
phase module in Fig. 4. Also, some critical decisions such as the selection of appropriate
434 J. Okpor and S. T. Apeh
Port place Suitable Job serves as an interface between the Job suitability module
and the cost Analysis module. The cost analysis module estimates the total cost and the
expected savings. The transition Determine Impact model the expected impact of the
implementation of a virtual workplace in the university while transition Identify success
factor identifies the critical success factor to successful virtual workplace implemen-
tation shown in Fig. 4.
Transition conduct IMP Assmnt model all activities relating to the conduct of the
implementation assessment. When transition conduct IMP Assmnt is fired, Places Imp
Assment Pilot, Imp Assment Phased, and Imp Assment full will receive tokens con-
currently for the conduct of implementation assessment as defined by the output arcs
expression with function PI_Assmnt(Assmnt,SSP), PH_Assmnt(Assmnt,GS), and
FU_Assmnt(Assmnt,CAES). Transition Analyze Assmnt result immediately become
enable as Place Imp Assment Pilot, Imp Assment Phased, and Imp Assment full receive
a valid token from transition Conduct Imp Assmnt. Transition Analyze Assmnt result
ensures that the strength and weakness of the virtual workplace program are critically
examined.
438 J. Okpor and S. T. Apeh
Simulation and state space are the two methods of analysis provided by CPN Tools
[19]. We conducted several simulations on the module and submodule of the developed
virtual workplace CPN model using CPN Tools installed on an Intel (R) Pentium
(R) CPU P6200 @2.13 GHz PC to ascertain whether the module and submodule work
as expected. The simulation shows that token(s) were received/consume in the correct
order, and the model always terminates in the desired state. State space was then
applied to explore the standard behavioral properties of the virtual workplace CPN
model. The analysis of the state space was achieved using the state space report. The
state space report revealed important information about the Boundedness properties,
home properties, liveness properties, and fairness properties of the CPN model.
Analysis of the state space report for virtual workplace CPN model shows that there are
12889 nodes, 30400 arcs, and was generated in 36 s while the SCC graph has 12889
nodes, 30400 arcs, and was generated in 2 s. The state space report shows that the SCC
graph and state space have the same number of nodes and arcs. It is obvious from the
report that no infinite occurrence exists, and hence we can conclude the model ter-
minate appropriately. The liveness properties show that are 24 dead markings in the
developed CPN model. A query was used to investigate these dead states and to
determine whether these dead markings represent the terminal state. The query result
proves that these dead markings are end states in the model. That is the state at which
the CPN model will terminate. The fact that these dead markings in the CPN model are
end states shows that the virtual workplace CPN model will produce the desired result
when it terminates. The state space report also shows that there are no dead transitions
in the CPN model. A transition is dead if there are no reliable markings in which it is
enabled [19].
5 Conclusions
References
1. International Labour Organization: Teleworking during the COVID-19 pandemic and
beyond A Practical Guide, pp. 1–48 (2020)
2. Pauline, E.O., Abiodun, A., Hindu, J.A.: Telecommuting: a panacea to COVID-19 spread in
Nigerian Universities. Int. J. Innov. Econ. Devt. 6(1), 47–60 (2020)
3. PwC S.A.: HR Quarterly. 1–16 (2015)
4. Richard, E., Carl, J.K., Christopher, G.P., Antony, I.S.: The new workplace: are you ready?
How to capture business value. IBM Global Serv. 1–12 (2011)
5. Ogunode, N.J., Abigeal, I., Lydia, A.E.: Impact of COVID-19 on the higher institutions
development in Nigeria. Elect. Res. J. Soc. Sci. Hum. 2(2), 126–135 (2020)
6. Cascio, W.F.: Managing a virtual workplace. Acad. Mgt. Exec. 14(3), 81–90 (2000)
7. Caldow J..: Working outside the box: a study of the growing momentum in telework. Inst.
for Elect. Govt. IBM Corp. 1–14 (2009)
8. Obisi, C.: The empirical validity of the adjustment to virtual work arrangement by business
organisations in Anambra State. Nigeria. Int. J. Sci. Res. Edu. 9(3), 173–181 (2016)
9. Koemg, B.E., Henderson, D.K., Mokhtanan, P.L.: The travel and emissions impacts of
telecommuting for the state of California telecommuting pilot project. Transpn. Res-C. 4(1),
13–32 (1996)
10. Choo, S., Mokhtarian, L.P., Ilan, S.: Does telecommuting reduce vehicle-miles traveled? An
aggregate time series analysis for the U.S. Transportation 32(1), 37–64 (2005)
11. Henderson K.D., Koening E.B., Mokhtarian L.P.: Travel diary-based emissions analysis of
telecommuting for the puget sound demonstration project. Res. Report UCD-ITS-RR-94–26,
1–54 (1994)
12. Thompson C., Caputo P.: The Reality of Virtual Work: is your organization Ready?, AON
consulting. 1–12 (2009)
13. Onpoint Consulting: Success factors of top performing virtual teams. Res. report. 1–14
(2015)
14. Madsen, S.R.: The benefits, challenges, and implications of teleworking: a Literature
Review. Cult. Relig. J. 1(1), 148–158 (2011)
15. vander Aalst, W.M.P.: Modelling and analysis of production systems using a Petri Net based
approach. In: Boucher, T.O., Jafari, M.A., Elsayed, E.A. (eds.) Proceedings of the
Conference on Computer Integrated Manufacturing in the Process Industry, pp. 179–193
(1994)
16. Kristensen, L.M., Mitchell, B., Zhang, L., Billington, J.: Modelling and initial analysis of
operational planning processes using coloured Petri nets. In: Proceedings of Workshop on
Formal Meth. Applied to Defence System, Australian Computing Society in Conference in
Research and Practice in Information Technology (12), 105–114 (2002)
17. Hafilah, D.L., Cakravastia, A., Lafdail, Y., Rakoto, N.: Modeling and simulation of air
france baggage handling system with colored Petri Nets. IFAC PapersOnLine. 2443–2448
(2019)
18. Jensen, K., Kristensen, L.M., Wells, L.: Coloured Petri Nets and CPN tools for modelling
and validation of concurrent systems. Int. J. Softw. Tools Techno. Trans. (STTT) 9(3–4).
213–254 (2007)
19. Jensen, K., Kristensen, L.M.: Coloured Petri Nets. Springer-Verlag, Modelling and
Validation of Concurrent Systems (2009)
Modeling and Experimental Verification
of Air - Thermal and Microwave - Convective
Presowing Seed Treatment
1 Introduction
In order to study the issues related to the implementability of grain presowing treatment
in such processing lines, modeling of the processes air-thermal and convective-
microwave seed treatment processes were performed. The basis of simulation model
was the following system of equations and transfer functions making it possible to
calculate heat-and-moisture exchange in elementary grain layer [12, 13]:
1
T ðpÞ ¼ T0 ðpÞ ð1 þ peps1 eps1 Þ A1 hðpÞp A2 W ðpÞp; ð1Þ
p
1 1
DðpÞ ¼ D0 ðpÞ ð1 þ peps1 eps1 Þ WðpÞp; ð2Þ
p A3
T ð pÞ1 1
hð pÞ ¼ T0 ð pÞ ðð1 A4 Þ ð1 eps1 Þ þ peps1 Þ; ð3Þ
A4 pA4
1 1
h ð p Þ ¼ ð h0 T ð pÞ Þ þ h0 þ A5 W ð pÞp þ A5 W0 þ A6 Qv ; ð3:1Þ
T1 p þ 1 p
442 A. A. Vasiliev et al.
K
WðpÞ ¼ Weq ; ð4Þ
pþK
0:435
lnð1 F Þ
Weq ¼ ; ð5Þ
5:47106 ðT þ 273Þ
745D
F ¼ : ð7Þ
ð622 þ DÞeð0:622 þ 238 þ T Þ
7:5T
cg cg cg r 0 eca aq sv
where A1 ¼ eca ca ; A2 ¼ 100ca ca e; A3 ¼ 103 cg ; A4 ¼ eca ca ; A5 ¼ r 0 cge ; A6 ¼ cg1q .
g
be equal to 0.7 m/s, 0.5 m/s and 0.7 m/s, for the first, second and third zone,
respectively. Simulink software was used for developing computer model. The diagram
of the developed model is shown in Fig. 1.
Fig. 1. Simulink model for air-thermal presowing treatment of seeds in bins with radial air
distribution.
Units ‘Interpeted MATLAB Fcn’ [14] (F = f(tau), Ta = f(tau)) serve for definition
of temperature and relative humidity change profile of ambient air, in correspondence
with the statistically processed data by meteorological services. Unit ‘Q0’ is designed
for setting the initial value of grain temperature. The dependence of relative humidity
of air fed into the grain layer on the capacity of electric heater is set with the help of
unit ‘Interpeted MATLAB Fcn’ (F = f(Pk)). Unit ‘C’ is used to set the temperature
level to which ambient air has to be heated in order to provide the required seed
presowing processing algorithm. Oscillograph units (F, W, Q, T) make it possible to
monitor the change of parameters of seeds and air over the layer thickness, in the
course of treatment.
In the process of modeling the process of presowing treatment with the use of
heated air the requirements of process control algorithm for aerated bins were con-
sidered [15]. In accordance with this algorithm the following operations have to be
performed:
– to measure the initial value of grain temperature,
– to heat ambient air in the input to the value equal to 1.3 of initial grain temperature
with the help of electric heater,
– to blow heated air through the grain layer during 1.1 h,
– to cool seeds with the help of ambient air during 1.1 h,
– to blow through the layer during 1.1 h with the help of air heated to the temperature
value equal to 1.8 of initial grain temperature,
– to cool seeds with the help of ambient air.
It is assumed that the initial grain temperature value is 20. The results of modeling
enable to monitor the change of seed temperature, that of air blowing through the grain
layer, as well as moisture content of seeds. These results of modeling are presented in
Figs. 2, 3 and 4.
444 A. A. Vasiliev et al.
It means that the requirements for temperature mode of seed presowing treatment
are complied with. A short-term air temperature drop within inter-seed spaces can be
explained by water transfer from the areas of grain layer located close to the air inlet
towards its downstream areas. This can be seen from the time dependences of grain
moisture content presented in Fig. 3.
Variations of moisture content of seeds do not exceed 3%. Besides, the values of
their moisture content upon completing the presowing process are identical to those in
the beginning. It is an essential feature since seeds have not to be exposed to
overdrying resulting in loss of their internal water.
Modeling results for temperature change in seeds under presowing treatment are
presented in Fig. 4. Diagrams shown in Fig. 4 give the reason to conclude that the
requirements for control algorithm of seed presowing treatment in aerated bins are fully
complied with, and such processing lines can be applied for the purposes under this study.
Modeling and Experimental Verification of Air 445
Fig. 3. Results of modeling the change of moisture content in grain layer during presowing
treatment with heated air.
Fig. 4. Temperature change of seeds during air-thermal presowing treatment in aerated bins.
446 A. A. Vasiliev et al.
Fig. 5. Block diagram of computer model for convective-microwave presowing seed treatment
This computer model enables to set values of input parameters of air (temperature
and relative humidity). For this purpose, model units Tvh and Fvh. Magnetron oper-
ation control is performed with the use of two-step action controller that is represented
Modeling and Experimental Verification of Air 447
From the diagrams presented in Fig. 6, different layers are heated to different
temperatures. However, this heterogeneity disappears and the grain temperature
becomes uniform over a thick layer as the grain passes through the entire dryer.
Because when moving inside the dryer, the grain is mixed and processed by a
microwave field of different intensities. Magnetron operation control ensures the
required modes of presowing grain treatment without exceeding set values of seed
temperature.
Diagrams for moisture content of seeds in each layer (see Fig. 7) show that no
water loss occurs in seeds in the process of presowing treatment. Variations of grain
moisture content in the process of treatment do not exceed 0.2% and they will not affect
sowing quality of seeds.
448 A. A. Vasiliev et al.
Fig. 7. Change of grain moisture content in the process of presowing treatment with the use of
microwave fields
The data presented in Table make it possible to evaluate the effectiveness of each
presowing treatment technology for particular purposes. For instance, air-thermal
treatment is preferable for winter crops cultivation because it ensures better conditions
450 A. A. Vasiliev et al.
4 Conclusion
The results of both computer simulations and experimental research afford a strong
ground to claim that technological equipment designed for grain drying with the
application of forced ventilation method can be effectively applied in presowing
treatment technologies of seeds. Aerated bins make it possible to implement air-thermal
presowing processing technique. In convective-microwave grain driers, the required
conditions of electrophysical processes are implementable, as well.
The results of field experiments have shown that application of various external
factors in the process of presowing treatment of seeds is an effective tool to control the
structure of grain yield. For example, application of air-thermal method may be
advisable for winter crops in which case development of secondary roots is essentially
important. Convective-microwave technique makes it possible to stimulate develop-
ment of multiple productive stems and ears per one plant.
References
1. Isaeva, A.O., Kirilin, G.M., Iskakov, A.Y., Giniyatulina, A.K.: Effects of vernalizing and
seeding method on output yield of carrot. In: Selection of Collected Papers: Scientific
Society of the 21th Century Students. Technical Science. E-print Selection of Collected
Papers Based on Materials of the 67th Inteernational Students Scientific-practical
Conference, pp. 119–123 (2018). (Russ.)
2. Shabanov, N.I., Ksenz, N.V., Gazalov, V.S., et al.: The substantiation of dose for presowing
treatment of cereal seeds in electromagnetic field of industrial frequency. Ambient Sci. 5(2),
20–24 (2018)
3. Dukic, V., Miladinov, Z., Dozet, G., et al.: Pulsed electro-magnetic field as a cultivation
practice used to increase soybean seed germination and yield. Zemdirbyste Agric. 104(4),
345–352 (2017)
4. Badridze, G., Kacharava, N., Chkhubianishvili, E., et al.: Effect of UV radiation and artificial
acid rain on productivity of wheat. Russ. J. Ecol. 47(2), 158–166 (2016)
5. Gilani, M.M., Irfan, A., Farooq, T.H., et al.: Effects of pre-sowing treatments on seed
germination and morphological growth of acacia nilotica and faidherbia albida. Scientia
Forestalis 47(122), 374–382 (2019)
6. Luna, B., Chamorro, D., Perez, B.: Effect of heat on seed germination and viability in species
of Cistaceae. Plant Ecol. Divers. 12(2), 151–158 (2019)
Modeling and Experimental Verification of Air 451
7. Mildaziene, V., Aleknaviciute, V., Zukiene, R., et al.: Treatment of common sun-flower
(Helianthus annuus L.) seeds with radio-frequency electromagnetic field and cold plasma
induces changes in seed phytohormone balance, seedling development and leaf protein
expression. Sci. Rep. 9(1), 1-12 (2019) Article Number: 6437
8. Nurieva, K.O.: Electrophysical factors for treating seeds. In: Collected Papers ‘The Youth:
Education, Science, Creating – 2019’ Collected Papers Based on Materials of Regional
scientific-practical Conference, pp. 122–124 (2019) (Russ.)
9. Budnikov, D.A., Vasiliev, A.N., Vasilyev, A.A., Morenko, K.S., Mohamed, S.I., Belov, F.:
Application of electrophysical effects in the processing of agricultural Materials. In:
Advanced Agro-Engineering Technologies for Rural Business Development. Valeriy
Kharchenko (Federal Scientific Agroengineering Center VIM, Russia) and Pandian Vasant
(Universiti Teknologi PETRONAS, Malaysia), pp. 1–27 (2019) 10.4018 / 978–1–5225–
7573–3.ch001.
10. Vasiliev, A.N., Ospanov, A.B., Budnikov, D.A.: Controlling reactions of biological objects
of agricultural production with the use of electrotechnology. Int. J. Pharm. Technol. 8(4),
26855–26869 (2016)
11. Wang, S., Wang, J., Guo, Y.: Microwave irradiation enhances the germination rate of tartary
buckwheat and content of some compounds in its sprouts. Pol. J. Food Nutr. Sci. 68(3), 195–
205 (2018)
12. Ospanov, F.B., Vasilyev, A.N., Budnikov, D.A., Karmanov, D.K., Vasilyev, A.A., et al.:
Improvement of grain drying and disinfection process in microwave field. Almaty Nur-Print
155 (2017)
13. Vasilyev, A.A., Vasilyev, A.N., Dzhanibekov, A.K., Samarin, G.N., Normov, D.A.:
Theoretical and experimental research on pre-sowing seed treatment. IOP Conf. Ser.: Mater.
Sci. Eng. 791(1), 012078 (2020) https://doi.org/10.1088/1757-899X/791/1/012078
14. Dabney, J., Harman, T.L.: Mastering Simulink 4. 412 p. Prentice-Hall, Upper Saddle River
(2001)
15. Udintsova, N.M.: Mathematical substantiation of electric heater capacity for presowing
treatment of seeds. Deposited manuscript All-Russian Institute of Scientific and Technical
Information No. 990-B2004 10.06.2004. (Russ.)
16. Sabirov, D.C.: Improvement of seed presowing treatment effectiveness. In: Collected Papers
‘Electric Equipment and Technologies in Agriculture’. Collected papers based on materials
of the 4th International Scientific-practical Conference, pp. 222–225 (2019) (Russ.)
17. G. A. Burlaka, E. V. Pertseva. Effect of presowing treatment of seeds on germination
readiness and germination ability of spring wheat. In: Collected Papers ‘Innovative
Activities of Science and Education in Agricultural Production. Proceedings of International
scientific-practical Conference, pp. 301–305 (2019) (Russ.)
Modeling of Aluminum Profile Extrusion
Yield: Pre-cut Billet Sizes
1 Introduction
Most manufacturing processes or operations can be optimized to save time and cost.
Yet theory and practice are sometimes far apart. Most practitioners in industry deem
optimization as too idealistic, cannot be done in practice. Thus, many manufacturing
processes are not yet optimized to the most effective manner. In this research an attempt
has been made to apply optimization technique to aluminum profile extrusion process.
Aluminum extrusion process is used to manufacture long aluminum profiles that are
widely used in construction, mechanical and electrical appliances. Obvious examples are
window frames, heat sink and steps on SUVs. Global aluminum extrusion market value
was $78.9 billion in 2018 [1] and is still growing. Aluminum extrusion process typically
involves 3 main steps. First, aluminum billets are cut into specific sizes, heated and then
loaded into a container, one billet at a time. Then each billet will be pushed through a die
by hydraulic press. Aluminum profile will be formed at the other side of the die. Once the
profile reaches prespecified extrusion length, it will be cut by the saw in front of the die.
The profiles will be straightened by a stretcher and cut to customer-specified lengths
afterward. Figure 1 illustrates an example of extrusion plan where one billet is extruded
three times, i.e. three identical extrusion lengths. Due to stretcher capacity, this extrusion
length is constrained to within a certain range, such as from Rmin to Rmax.
Given a customer order, consisting of product length L and order quantity n, process
planner needs to select available pre-cut billet sizes bi and extrusion lengths LE1 to meet
individual demand whilst resulting in minimal scrap. Many short billets need to be
extruded consecutively until the extrusion length exceeds Rmin before being cut by the
saw. This type of configuration will be called multiple billets per extrusion in this paper.
On the other hand, a large billet can be extruded and cut by the saw more than once with
each extrusion length between Rmin and Rmax. This will be called multiple extrusions per
billet, though a single extrusion per billet is also included in this category.
So far, in the literature, there is no explicit modeling that will consider both
extrusion setups simultaneously. The objective of this research is to formulate a yield
optimization model for the extrusion process using mass balance equations. Available
pre-cut billet sizes, process setup and customer order requirements are taken into
accounts in the model. It is validated with actual data from an aluminum extrusion
company. The model is then used to illustrate possible yield improvements through
case studies. Specifically, it is shown how two objective functions, namely minimizing
total billet weight used versus maximizing yield, provide different optimal solutions.
The effects of extrusion die design on yield is also studied.
454 J. Hassamontr and T. Leephaicharoen
2 Literature Review
Aluminum extrusion yield was first investigated by Tabucanon [2]. A dynamic pro-
gramming model was proposed to identify an optimal set of billets to be stocked, that
will minimize overall scraps. Masri and Warburton [3, 4] formulated yield optimization
as a billet selection problem. Using past customer orders’ data, the goal was to identify
optimal billet sizes that should be stocked. Their model considered a global problem-
multiple profile cross sections and lengths. Multiple pushes within the same billet were
not considered due to longer cycle time. Wastes from each billet size-customer order
pair was calculated a priori, though not explicitly shown. Hajeeh [5] formulated a
mathematical model to optimize cutting aluminum logs into billets to fill customer
orders. The scraps considered were from log cutting, butt scrap and profile cutting. The
goal was to minimize overall scrap. Again, wastes from each billet assignment were
calculated before optimization. Another approach to scrap reduction was introduced by
researchers in mechanical engineering discipline. Reggiani, Segatori, Donati and
Tomesani [6] used Finite Element simulation to predict charge welds between two
consecutive billets that must be scrapped. Oza and Gotowala [7] used HyperXtrude, a
commercial CAD software for extrusion, to study the effect of ram speed, billet and die
temperatures on traverse weld length. Ferras, Almeida, Silva, Correia and Silva [8]
conducted an empirical study on which process parameters contributing the most to
material scrap.
In this research, an alternative approach for extrusion yield optimization is pro-
posed. Instead of considering a global problem where multiple customer orders are
optimized simultaneously, the problem size can be greatly reduced by considering one
customer order at a time. Secondly, mass balance equations are used explicitly in the
model, allowing clear indication how billets are extruded in the optimal solution.
Thirdly, both general extrusion setups, including multiple pushes, are considered. As
will be shown later, there are cases where multiple pushes can lead to lower scrap. Last,
it is possible to identify the impact of improving hardware constraints. For example,
how die design can improve the optimal solution.
3 Research Methodology
3.1 Problem Formulation
There are three types of material scrap in aluminum extrusion process. The first is
incurred from cutting aluminum log into specific billet lengths. The second comes from
in-process extrusion, which are denoted by billets’ backend e, in kg, and aluminum
material remained in the die pot, P, also in kg, after the die is removed from the
extrusion press. These are denoted in Fig. 1. The third type of waste is induced by
profile cutting. They come in many forms as follow.
• Aluminum portion between die and cutting saw, Ls, is usually too short for cus-
tomer usage and therefore discarded.
Modeling of Aluminum Profile Extrusion Yield: Pre-cut Billet Sizes 455
• Part of the profile gripped by pullers for stretching are usually deformed and later
cut off. These are denoted by head loss h and tail loss t in Fig. 1. Generally, portion
of the profile that first comes out and gripped by puller, h, is less than Ls. Thus, the
scrap h is already included in Ls and can be considered zero in most cases.
• An allowance, d, shown in Fig. 2, is to avoid irregularities within the profiles. These
irregularities may come from stop marks while profile is being cut, or where
material from two different billets are merged within the same profile. These
irregularities may lead to profile failures when undertaking large loads. Thus,
manufacturers will cut this portion out.
• Another allowance, though not shown in both figures, is the material scrap remained
after the extruded profile cut to pre-determined number of pieces, say q1, of cus-
tomer order’s length L.
Other input variables for the planning includes billet weight per length, b, number
of die openings (how many profiles can be extruded in one time), a, and extruded
profile’s average weight per length, w.
In the case of multiple extrusions per billet, process planner must determine billet
sizes to be used as the first and consecutive billets B1 and B2 consecutively. For each
billet, how many extrusions C1 and C2 are performed and in each extrusion how many
pieces of product with length L are made within each profile, q1 and q2. The extrusion
lengths LE1 and LE2 from each billet should be constant for simplicity in profile cutting.
All C’s are at least 1. Figure 1, for example, demonstrates the case of three extrusions
per billet (C1 = 3), whereas consecutive billets are not shown.
Schematics of multiple billets per extrusion are shown in Fig. 2. Here at most three
different billet sizes, B1, B2 and B3, can be used to form a set of billets that will result in
single extrusion. The second billet B2 can be used D2 times within each billet set while
the first billet B1 and the last billet B3 are used only once. Each billet results in q1, q2,
and q3 pieces of product within each profile respectively. As it will be shown in the
456 J. Hassamontr and T. Leephaicharoen
solution procedure, the first billet in the next set, B1′, can be smaller than B1. The
extrusion length LE can be computed for each billet set.
Assumptions for the problem formulation are summarized as follows.
• Billets are pre-cut to specific sizes and available in stock with unlimited quantity.
• Extrusion length must be between Rmin and Rmax.
• Saw used for profile cutting on the runout table is at fixed position Ls from the die.
• Material losses due to cutting saw is negligible.
• Desired product length, L and minimum required quantity n are known input
variables for billet selection.
• Aluminum density w along the profile is constant.
• Process parameters such as extrusion speed, temperature, etc. are pre-determined.
No significant adjustment is made during extrusion.
• For multiple pushes per billet case, extrusion length is uniform for one billet size. At
most two billet sizes are used to meet a customer order.
• For multiple billets per extrusion case, extrusion length should be uniform for each
set of billets. At most three billet sizes are used to meet a customer order.
Step 1. Classify available billet sizes into two categories. For each billet size bi,
maximum extrudable length can be calculated from
bi e P
Lmax;i ¼ Ls ð1Þ
aw
Let U be the set of billets bu of which maximum extrudable length Lmax,u Rmin
and V the set of billets bv of which maximum extrudable length Lmax,v < Rmin.
Step 2. For billets in U, solve the following optimization model for multiple extrusions
per billet setup. First, a few more decision variables are needed. Let.
X1u = 1 if billet u is used as the first billet and zero otherwise, 8u.
X2u = 1 if billet u is used as consecutive billets and zero otherwise, 8u.
nb2 = number of consecutive billets used.
The objective function is to minimize total billet weight used. Equations (2) and (5)
are mass balance equations that can be derived from Fig. 1. Equations (2)–(7) are to
ensure that only one billet size is used as the first billet and only one for consecutive
billets. Equations (8)–(11) are to impose constraint on extrusion lengths for the first and
consecutive billets respectively. The number of consecutive billets, nb2, can be cal-
culated from (12). All decision variables are denoted in (13) and (15).
Modeling of Aluminum Profile Extrusion Yield: Pre-cut Billet Sizes 457
minimize B1 þ nb2 B2 .
Subject to
aC 2 q2 nb2 n aC 1 q1 ð12Þ
C1 ; C 2 1; integer ð13Þ
Step 3. For billets in V, solve the following optimization model for multiple billets per
extrusion case. A few more variables are needed. Let.
X1v = 1 if billet v is used as the first billet and 0 otherwise, 8v.
X2v = 1 if billet v is used as consecutive billets and 0 otherwise, 8v.
X3v = 1 if billet v is used as the last billet and 0 otherwise, 8v.
Y1v = 1 if billet v is used as the first billet for the next set and 0 otherwise, 8v.
ns = number of consecutive billet sets used.
The objective is to minimize total billet weight which consists of the first billet set
and consecutive billet sets. Equations (16), (19), (22) and (25) represent mass balance
equations that can be derived from Fig. 2. Equations (16)–(27) are used to make sure
that only one billet size is used for each billet type. The extrusion length for each billet
set is computed in (28) while its limits are imposed by (29). The total number of
consecutive billet sets is calculated by (30). Constraints on decision variables are
summarized in (31) and (32).
458 J. Hassamontr and T. Leephaicharoen
0
minimize ðB1 þ D2 B2 þ B3 Þ þ ns B1 þ D2 B2 þ B3 .
Subject to
B1 e þ P þ awfLq1 þ Ls þ hg ð16Þ
X
B1 ¼ X b
v 1v v
ð17Þ
X
v
X 1v ¼ 1 ð18Þ
P
B2 e þ P þ aw Lq2 þ d ð19Þ
aw
X
B2 ¼ X b
v 2v v
ð20Þ
X
v
X 2v ¼ 1 ð21Þ
P
B3 e þ P þ aw Ls þ t þ Lq3 þ d ð22Þ
aw
X
B3 ¼ X b
v 3v v
ð23Þ
X
v
X 3v ¼ 1 ð24Þ
0 P
B1 e þ P þ aw Lq1 þ d ð25Þ
aw
0
X
B1 ¼ Y b
v 1v v
ð26Þ
X
v
Y 3v ¼ 1 ð27Þ
D2 ; q1 ; q2 ; q3 ; X 1u ; X 2u ; X 3u ; Y 1v ; ns 0; integer ð32Þ
Step 4. Show results from steps 2 and 3 so that the better solution can be selected.
Even though both optimization models are non-linear, all independent variables are
discrete. A computer program can be developed to enumerate all possible values of
these variables and identify the optimal solutions, if any, in each model. The solution
procedure is developed using Visual Basic Application in Microsoft Excel.
Modeling of Aluminum Profile Extrusion Yield: Pre-cut Billet Sizes 459
4 Case Studies
4.1 Case Study 1
Actual operating data from an aluminum extrusion company is used to validate the
model. The input data, shown in Table 1, is used for all cases, unless specified
otherwise. In the company where the case data is taken, a human process planner
decides which billet sizes are used. Table 2 illustrates how this conventional extrusion
data compares with results from the enumeration program developed. To make 600
pieces of 3.1 m-long profile, the first billet used is 64 cm in length whereas consecutive
billets are smaller at 59 cm. Each billet is extruded once, single push per billet, to make
10 pieces of product in each profile extruded. There are two profiles extruded at the
same time (a = 2). Each billet, therefore, generates 20 pieces of product. With 29
consecutive billets, exactly 600 pieces of product are made. Total weight for all billets
used are 1185.7 kg. Since material yield is defined as a ratio of total product weight to
total billet weight, the resulting yield is 85.8%.
With the enumeration program, the first two billet sizes, 35 and 40 cm, are put into
set V, to be considered for multiple billets per extrusion. The program selects the first
billet set consisting of 40-cm billet as the first and the last billet while no billets are
extruded in between. Both billets result in 6 pieces per profile or 24 pieces per billet set.
The second set is similar to the first, except that the first billet size can be reduced to
35 cm, but still able to produce the same number of products per billet set. The
extrusion length for each billet set is 39.7 m. Making 600 pieces product requires
another 24 sets of billets besides the first set. Certainly, this extrusion setup requires
more billet weight than the conventional case. The resulting yield is 81.02%. In this
particular case, multiple billets per extrusion could not perform better than multiple
extrusions per billet.
As the program considers billets in U for multiple (or single) pushes per billet, it
selects 59 cm for the first billet size and 78 cm for consecutive billet sizes. The first
billet created 9 pieces of product per profile while consecutive billets 14 pieces. It
should be noted that extrusion lengths from the first billet is not the same as those from
consecutive billets. Here, only 21 consecutive billets are needed. The total product
quantity made is 606 pieces while total billet weight required is less than that from the
conventional case. The resulting yield, 90.65%, is significantly better than that from the
conventional solution.
The last column in Table 2 is used to demonstrate how objective function influ-
ences the optimal solution. If the objective function is to change from minimizing total
billet weight to maximizing yield, the optimal solution will look to add more pieces of
products, the nominator of yield, while keeping total billet weight as low as possible.
Thus, the optimal solution from maximizing yield objective is to use the largest billet
sizes for the first and consecutive billets. The resulting yield is 90.83% with more
products made. Selecting which objective function to use will depend on whether
customers allow over-shipment or if there exists a need to make extra pieces, for
example, to compensate for product defects.
The enumeration program is used to investigate the extrusion process in case study
1 with various number of die openings, from a = 1 to 5, assuming that extrusion press
capacity is sufficient, and the process is under control. Here, only the objective of
minimizing total billet weight is used and only the better solutions from two extrusion
setups are shown. When a = 1, 2 or 3, the optimal solutions are from multiple
extrusions per billet. As a increases, the optimal yield decreases. The optimal solutions
from a = 1 and a = 2 are very close to each other as the main difference comes from
only the first billet size used. The consecutive billets, when a = 1 or 2, will create 28
pieces of product. When a = 4 or 5, the optimal solution comes from multiple billets
per extrusion with only slightly less yield. At a = 5, all billet sizes are in U. Thus,
multiple billets per extrusion can be beneficial when companies look to improve both
yield and productivity at the same time.
5 Conclusion
Based on limited case studies from the industry, preliminary conclusions can be
drawn.
• The choice of objective function will have profound effects on the optimal solution.
Minimizing total billet weight seems to be a better objective than maximizing yield,
at least from the overproduction quantity standpoint.
• Multiple pushes, such as multiple extrusions per billet and multiple billets per
extrusion, can be beneficial for material utilization. Both setups should be con-
sidered unless there are some physical constraints on the extrusion press.
More work can be done to extend the model for online billet cutting where billets
can be cut to lengths and extruded directly. This should provide even better yield.
References
1. Grand View Research. https://www.grandviewresearch.com/industry-analysis/aluminum-
extrusion-market.
2. Tabucanon, M.T., Treewannakul, T.: Scrap reduction in the extrusion process: the case of an
aluminium production system. Appl. Math. Modelling 11, 141–145 (1987). https://doi.org/10.
1016/0307-904X(87)90158-2
3. Masri, K., Warburton, A.: Optimizing the yield of an extrusion process in the aluminum
industry. In: Tamiz, M. (ed.) Multi-Objective Programming and Goal Programming. LNE,
vol. 432, pp. 107–115. Springer, Heidelberg (1996). https://doi.org/10.1007/978-3-642-
87561-8_9
4. Masri, K., Warburton, A.: Using optimization to improve the yield of an aluminum extrusion
plant. J. Oper. Res. Socy. 49(11), 1111–1116 (1998). https://doi.org/10.1057/palgrave.jors.
2600616
5. Hajeeh, M.A.: Optimizing an aluminum extrusion process. J. Math. Stat. 9(2), 77–83 (2013).
https://doi.org/10.3844/jmssp.2013.77.83
6. Reggiani, B., Segatori, A., Donati, L., Tomesani, L.: Prediction of charge welds in hollow
profiles extrusion by FEM simulations and experimental validation. Intl. J. Adv. Manuf.
Technol 69, 1855–1872 (2013). https://doi.org/10.1007/s00170-013-5143-2
7. Oza, V.G., Gotowala, B.: Analysis and optimization of extrusion process using Hyperworks.
Int. J. Sci. Res. Dev. 2(8), 441–444 (2014)
8. Ferras, A.F., Almeida, F. De, Silva, E. C, Correia, A., Silva, F.J.G.: Scrap production of
extruded aluminum alloys by direct extrusion. Procedia Manufacturing, 38, 1731–1740
(2019). https://doi.org/10.1016/j.promfg.2020.01.100
Models for Forming Knowledge Databases
for Decision Support Systems for Recognizing
Cyberattacks
Abstract. Patterns of Bayesian networks have been developed for the com-
puting core of the decision support system in the course of threats prediction and
stages of intrusion into information and communication networks of informa-
tization objects. The proposed Bayesian networks templates allow one to operate
with a variety of random variables and determine the probability of a cyber
threat or a specific stage of an invasion under given conditions. Probabilistic
models for detecting network intrusions based on the use of dynamic Bayesian
networks have been added. The training of Bayesian networks parameters based
on the EM-algorithm was carried out. In contrast to existing solutions, the
proposed approach makes it possible not only to take into account the main
stages of intrusions but also to make more reasonable decisions based on the use
of both typical intrusion patterns and newly synthesized patterns. All templates
and models make up the decision support system computing core for intrusion
detection. The effectiveness of the developed models was tested on test samples
that were not previously used in training.
1 Introduction
integrate intelligent decision support systems (DSS) into their products. Note that the
basis of modern DSS is formed by various models and methods, which together form
its computational core [2, 3].
In difficult situations of guaranteed information security of various objects of
informatization, the decision-making process will take place under the condition of
active interaction of information security means and cybersecurity systems with
experts. The above circumstances determined the relevance of our research in the field
of synthesis of models for the computational core of the decision support systems as
part of the I intrusion detection system.
As noted in [4], a promising direction of research in this area has become studies on the
development of methods, models, and software systems for DSS [5] and expert systems
(ES) [6] in the field of information security.
In studies [6, 7, 24], Data Mining technologies in information security problems
were considered. The emphasis in these studies was made on the task of identifying the
situation evolution pattern, which is associated with the provision of informatization
objects information security. The considered works had no practical implementation, in
the form of a software system. The studies [8, 9] analyzed the methodology of intel-
ligent modeling in the problems of informatization objects information security. The
methodology proposed by the authors is intended to provide analysis and decision-
making under different scenarios of informatization objects intrusion implementation.
However, these studies have not been brought down to hardware or software
implementation.
The appearance of new classes makes it difficult to analyze and support decision-
making in information security tasks due to poorly amenable formalization and
structuring of the task themselves for providing information security [10, 11]. In such
cases, the parameters of the informatization objects information security state can be
represented by qualitative indicators. The latter is not always advisable.
According to the authors of [12, 13], the analysis of the security level of
informatization objects and the development of plans to counter targeted cyber-attacks
should be preceded by the stage of identifying the main threats. The authors of [14]
point out that it is problematic to qualitatively solve such a problem without the
appropriate DSS. The researchers did not describe the practical results of their
development.
In studies [15, 16], it was shown that aspects of building information security
means can be taken into account in the approach based on the applied Bayesian
networks (hereinafter BN). Compared to existing data analysis methods, they provide a
clear and intuitive explanation of their findings. Also, BN (often dynamic BN - DBN)
implies a logical interpretation and modification of the structure of the relations among
the variables of the problem. The representation of the BN in the form of a graph makes
it a convenient tool for solving the problem of assessing the probabilities of the pos-
sible cyber threat occurrence, for example, for the information and communication
networks of informatization objects.
Models for Forming Knowledge Databases for Decision 465
The computing core of the developed decision support system is based on proba-
bilistic models that make up the knowledge base. Such probabilistic models are capable
of describing processes under conditions when the data is poorly structured or insuf-
ficient for statistical analysis. Problems that can be solved based on the use of
knowledge bases probabilistic models should also use the procedures of probabilistic
inference. Various processes take place in information and communication networks of
informatization objects to generate their own data sequences. Accordingly, these data
sequences should be reflected in the description of the course of each process. The
above analysis showed that research in the field of new model development for the
formation of decision support systems knowledge bases in the process of recognizing
cyber-attacks is relevant.
The research aims to develop models based on Bayesian networks for the knowledge
base of the computing core of the decision support system in the course of identifying
complex cybernetic attacks.
design a DSS and fill its knowledge base with appropriate Bayesian network patterns,
in this subtask it is better to use a dynamic Bayesian network - DBN. Dynamic
Bayesian networks are ordinary Bayesian networks that are related by variables at
adjacent time steps [21]. In fact, the dynamic Bayesian network is a Markov process,
which can be attributed to the 1st order.
The designed dynamic Bayesian network will consist of two networks. Let's des-
ignate the first BN as ðB1Þ; respectively, the second as ðB2Þ: The network ðB1Þ will be
the original one. Transit network ðB2Þ: The original network ðB1Þ defines the prior
distributions of the available model ðpðzð1ÞÞÞ: The transit model ðB2Þ will determine
the probabilities of transitions between time slices.
Then, taking into account the above written, one can hold the dynamic Bayesian
networks for the variables marked in lines 2–4, 23, 26–28, 34 of Table 1. Also, one can
add a variable type to the DBN, which, will display the moment of transitions between
networks B1 and B2. For the design of dynamic Bayesian networks was used editor
GeNIe Modeler.
The probabilistic model of the designed dynamic Bayesian network for transitions
between the vertices of the joint distribution graph of available models is presented as
follows:
Y
N
pðZt jZt1 Þ ¼ p p Zti jPa Zti ; ð1Þ
t¼1
N Y
Y N
pðZ1:t Þ ¼ p Zti jPa Zti ; ð2Þ
t¼1 i¼1
where Z t - is the BN slice for the time point t; Z it - BN node at the moment in time t;
i
Pa Z t - a set of parent nodes for a BN node; the number of nodes to cut the BN.
Expression (2) will describe the probabilities of transitions between BN nodes.
Models for Forming Knowledge Databases for Decision 467
X
i¼k
H ðU Þ ¼ PðU ¼ ui Þ log2 ðPðU ¼ ui ÞÞ: ð3Þ
i¼1
X
j¼l
H ðUjV Þ ¼ PðV ¼ vi Þ H ðUjV ¼ vi Þ: ð4Þ
j¼1
Because the calculation was IðV; UÞ performed for discrete values, the values from
the set of fu1 ; :::; uk g and fv1 ; :::; vk g had to be discretized. In the course of test
experiments, the method of equal frequency intervals was used for sampling. Following
the method value space attribute value should be pitched be an arbitrary number of
partitions. Each section will contain the same number of points described by the
corresponding data. The application of this method allowed us to achieve a better
classification of data values.
The information amplification factor IðV; UÞ will depend on the number of values
of the corresponding attributes. Therefore, as the number of values increases, the
entropy value for the attribute decreases.
More informative attributes can be selected using the gain [23]:
I ðU; V Þ H ðU Þ H ðUjV Þ
GðUjV Þ ¼ ¼ : ð5Þ
H ðV Þ H ðV Þ
The information amplification factor GðUjVÞ will take into account not only its
amount, which is necessary to represent the result in the DSS but also HðVÞ that is
necessary to separate the information by the current attribute ðVÞ:
Consider the following example of designing a dynamic Bayesian network for a
knowledge base decision support system. We will proceed with the following
468 V. Lakhno et al.
Fig. 1. General DBN model for the inva- Fig. 2. An example of the relationship between
sion of S, E, B the observed parameters (lines 3–6, 12, 25, 29,
30, 38, 39 of Table 1)
Models for Forming Knowledge Databases for Decision 469
Y
3
PðV ðnÞjPaðV ðnÞÞÞ ¼ Pðvi ðnÞjPaðvi ðnÞÞÞ; ð6Þ
i¼1
Y
3
Pðvi ðnÞjPaðvi ðnÞÞÞ
i¼1 ð7Þ
¼ PðsðnÞjsðn 1ÞÞ PðeðnÞjeðn 1Þ; sðn 1ÞÞ PðbðnÞjbðn 1Þ; eðn 1ÞÞ:
For example, for the Bayesian network variant, which is shown in Fig. 2, the model,
which describes the relationship between the observed variables for network traffic and
the probabilities of transition from state to the state of the graph vertices, will look like
following:
Y
11
PðU ðnÞjPaðU ðnÞÞÞ ¼ P uNj ðnÞjPa uNj ðnÞ ; ð8Þ
j¼1
Y
11 Y
4
P uNj ðnÞjPa uNj ðnÞ ¼ P uNj ðnÞjPaðsðnÞÞ
j¼1 j¼1
ð9Þ
Y
8 Y11
P uNj ðnÞjPaðeðnÞÞ P uNj ðnÞjPaðbðnÞÞ ;
j¼5 j¼9
P uNj ðnÞjPaðsðnÞÞ ¼ P u31 ðnÞjsðnÞ P u42 ðnÞjsðnÞ
ð10Þ
P u53 ðnÞjsðnÞ P u64 ðnÞjsðnÞ ;
26
P uNj ðnÞjPaðeðnÞÞ ¼ P u255 ðnÞjeðnÞ P u6 ðnÞjeðnÞ
39 ð11Þ
7 ðnÞjeðnÞ P u8 ðnÞjeðnÞ ;
P u38
Y
11
P uNj ðnÞjPaðbðnÞÞ :
j¼9 ð12Þ
29 30
9 ðnÞjbðnÞ P u10 ðnÞjbðnÞ P u11 ðnÞjbðnÞ ;
¼ P u12
470 V. Lakhno et al.
Where j ¼ 1; . . .; 11 - is the number of variables that are observed for the network
state graph (see Fig. 2), N ¼ 1; . . .; 41 the corresponding property (traffic parameter)
from Table 1.
Therefore, for the previously selected informative signs of a DDoS attack, lines 2–
4, 23, 26–28, 34 of Table 1, the dynamic Bayesian networks will look like this, see
Fig. 3.
Fig. 3. An example of the relationship between the observed parameters (lines 2 - 4, 23, 26–28,
34 of Table 1)
Accordingly, for a given attack pattern, the model that describes the relationship
between the observed variables for network traffic and the probabilities of transition
from state to the state of the graph vertices will look like following:
Y
8
PðU ðnÞjPaðU ðnÞÞÞ ¼ P uNj ðnÞjPa uNj ðnÞ ; ð13Þ
j¼1
Y
8 Y
3
P uNj ðnÞjPa uNj ðnÞ ¼ P uNj ðnÞjPaðsðnÞÞ
j¼1 j¼1
ð14Þ
Y
6
P uNj ðnÞjPaðeðnÞÞ P u76 ðnÞjPaðbðnÞÞ :
j¼4
Models for Forming Knowledge Databases for Decision 471
P uNj ðnÞjPaðsðnÞÞ ¼ P u21 ðnÞjsðnÞ P u32 ðnÞjsðnÞ P u43 ðnÞjsðnÞ ; ð15Þ
26
P uNj ðnÞjPaðeðnÞÞ ¼ P u23 4 ðnÞjeðnÞ P u5 ðnÞjeðnÞ
28 ð16Þ
P u27
6 ðnÞjeðnÞ P u7 ðnÞjeðnÞ ; :
P uNj ðnÞjPaðbðnÞÞ ¼ P u34
7 ðnÞjbðnÞ : ð17Þ
5 Computational Experiments
Below we consider the results of testing Bayesian networks templates for the decision
support system knowledge base, see Fig. 2 and 3. Test samples included 30 entries for
each template. Testing of samples was implemented on a test network, see Fig. 1.
The PC and EM algorithms were used [15, 16, 21] to train the networks.
The experimental results are shown in Fig. 4 and 5. The graphs show the results of
modeling the probabilities of the correct definition and interpretation of Dos / DDos
attacks (errors of the 1st kind, respectively) for networks that were trained: using the
PC algorithm (line 1) and the EM algorithm (line 2).
And the analysis of the simulation results for the obtained Bayesian networks (see
Fig. 2, 3) was carried out in the direction of estimating errors of the 1st and 2nd kind.
Of the 60 records that were used to support decision-making and did not take part in
training the BN, the first 30 records are correct data to check whether a test attack was
missed (type 2 error). The remaining 30 entries were used to check for a false positive
(type 1 error).
472 V. Lakhno et al.
Figure 4 and Fig. 5 shows errors of the 1st and 2nd kind, respectively. The PC and EM
algorithms [15, 16, 21] have confirmed their effectiveness. BN testing showed that the
developed templates correctly interpret the attack with a probability of 95–96%.
Like the previous series of experiments that were described above, the experiments
carried out confirmed that Bayesian reasoning makes it possible to determine the likely
representation for each component in the diagram. Some laboriousness of drawing up
templates is compensated by the possibility of their repeated use without significant
modification if all the developed templates become part of the decision support system
knowledge base.
Hypothetically, a Bayesian network can be constructed using a simple enumeration
to construct the sets of all possible models, if they are not cyclical. However, note that,
as shown in [15] and [16], this approach is not optimal. This is because, with the
number of more than 7 vertices, a full search requires significant computational
resources and takes quite a long time. Therefore, for the developed decision support
system and its knowledge base, the preferable option is the preliminary creation of
attack patterns, minimization of hidden variables, and the elimination of uninformative
variables that do not have a determining effect on the accuracy of intrusion detection.
Bayesian networks training consists of incorrectly adjusting the parameters of
individual nodes for a specific task, for example, to identify a certain type of network
attack.
The prospect for the development of this research is the software implementation of
the decision support system in high-level algorithmic languages and the subsequent
testing of this decision support system and its knowledge base on the segments of real
informatization objects networks.
Models for Forming Knowledge Databases for Decision 473
7 Further Research
Further development of the proposed method has two standalone directions. The first
direction of the research will be sharpened on the development of specific thread
simulation tools that will work under an adjustable environment with most likely bio-
behavior. As soon as the proposed model is an artificial emulation it really cannot
mirror the psychology and unpredictability of real threads. However, the development
of these tools will provide an unlimited thread pool that could/will possibly be made
against the system and therefore provide silicon datasets for models to train how to
detect those threads during their lifecycle without actually being under real attack. The
second direction will be to create a hardware supply for a thread detection system that
can work in conjunction with a prediction system and will be able to block suspicious
threads. Hardware will be installed right after the main gateway transferring all the data
through itself. As a result of the described directions, we are planning to implement this
system on the real working network.
8 Conclusions
Patterns of Bayesian networks have been developed for the computational core of the
decision support system in the course of predicting threats and stages of intrusion into
the OBI of information and communication networks. The constructed BN templates
allow you to operate with a variety of random variables and determine the probability
of a cyber-threat or a specific stage of an invasion under given conditions. To improve
the efficiency of intrusion forecasting, the network parameters were trained. The EM
algorithm and the PC algorithm were used, as well as the available statistics for the test
network. Probabilistic models for detecting network intrusions based on the use of BN
are described. In contrast to the existing models, the proposed approach makes it
possible not only to take into account the main stages of intrusions but also to make
more reasonable decisions based on the use of both typical intrusion patterns and newly
synthesized patterns. All templates and models make up the decision support system
computing core for intrusion detection.
The effectiveness of the developed models was tested on test samples that were not
previously used in training. The results obtained indicate the feasibility of using the EM
algorithm to obtain a high-quality result of the recognition of cyber threats to infor-
mation and communication networks.
References
1. Elshoush, H.T., Osman, I.M.: Alert correlation in collaborative intelligent intrusion detection
systems–a survey. Appl. Soft Comput. 11(7), 4349–4365 (2011)
2. Shenfield, A., Day, D., Ayesh, A.: Intelligent intrusion detection systems using artificial
neural networks. ICT Express 4(2), 95–99 (2018)
474 V. Lakhno et al.
3. Rees, L.P., Deane, J.K., Rakes, T.R., Baker, W.H.: Decision support for Cybersecurity risk
planning. Decis. Support Syst. 51(3), 493–505 (2011)
4. Akhmetov, B., Lakhno, V., Boiko, Y., & Mishchenko, A.: Designing a decision support
system for the weakly formalized problems in the provision of cybersecurity. Eastern-Eur.
J. Enterp. Technol. (1(2)), 4–15 (2017)
5. Fielder, A., Panaousis, E., Malacaria, P., Hankin, C., Smeraldi, F.: Decision support
approaches for cybersecurity investment. Decis. Support Syst. 86, 13–23 (2016)
6. Atymtayeva, L., Kozhakhmet, K., Bortsova, G.: Building a knowledge base for expert
system in information security. In: Chapter Soft Computing in Artificial Intelligence of the
series Advances in Intelligent Systems and Computing, vol. 270, pp. 57–76 (2014)
7. Dua S., Du, X.: Data Mining and Machine Learning in Cybersecurity, p. 225. CRC Press
(2016)
8. Buczak, A.L., Guven, E.: A Survey of data mining and machine learning methods for cyber
security intrusion detection. IEEE Commun. Surv. Tutor. 18(2), 1153–1176 (2016)
9. Zhang, L., Yao, Y., Peng, J., Chen, H., Du, Y.: Intelligent information security risk
assessment based on a decision tree algorithm. J. Tsinghua Univ. Sci. Technol. 51(10),
1236–1239 (2011)
10. Ben-Asher, N., Gonzalez, C.: Effects of cybersecurity knowledge on attack detection.
Comput. Hum. Behav. 48, 51–61 (2015)
11. Goztepe, K.: Designing fuzzy rule based expert system for cyber security. Int. J. Inf. Secur.
Sci. 1(1), 13–19 (2012)
12. Gamal, M.M., Hasan, B., Hegazy, A.F.: A Security analysis framework powered by an
expert system. Int. J. Comput. Sci. Secur. (IJCSS) 4(6), 505–527 (2011)
13. Chang, L.-Y., Lee, Z.-J.: Applying fuzzy expert system to information security risk
Assessment – a case study on an attendance system. In: International Conference on Fuzzy
Theory and Its Applications (iFUZZY), pp. 346–351 (2013)
14. Kanatov, M., Atymtayeva, L., Yagaliyeva, B.: Expert systems for information security
management and audit, Implementation phase issues, Soft Computing and Intelligent
Systems (SCIS). In: Joint 7th International Conference on and Advanced Intelligent Systems
(ISIS), pp. 896–900 (2014)
15. Lakhno, V.A., Lakhno, M.V., Sauanova, K.T., Sagyndykova, S.N., Adilzhanova, S.A.:
Decision support system on optimization of information protection tools placement. Int.
J. Adv. Trends Comput. Sci. Eng. 9(4), 4457–4464 (2020)
16. Xie, P., Li, J. H., Ou, X., Liu, P., Levy, R.: Using Bayesian networks for cybersecurity
analysis. In: 2010 IEEE/IFIP International Conference on Dependable Systems & Networks
(DSN), pp. 211–220. IEEE, June 2010
17. Shin, J., Son, H., Heo, G.: Development of a cybersecurity risk model using Bayesian
networks. Reliab. Eng. Syst. Saf. 134, 208–217 (2015)
18. Özgür, A., Erdem, H.: A review of KDD99 dataset usage in intrusion detection and machine
learning between 2010 and 2015. PeerJ Preprints, 4, e1954v1 (2016)
19. Elkan, C.: Results of the KDD’99 classifier learning. ACM SIGKDD Explorat. Newsl. 1(2),
63–64 (2000)
20. Lakhno, V.A., Kravchuk, P.U., Malyukov, V.P., Domrachev, V.N., Myrutenko, L.V., Piven,
O.S.: Developing of the cybersecurity system based on clustering and formation of control
deviation signs. J. Theor. Appl. Inf. Technol. 95(21), 5778–5786 (2017)
Models for Forming Knowledge Databases for Decision 475
21. Lakhno, V.A., Hrabariev, A.V., Petrov, O.S., Ivanchenko, Y.V., Beketova, G.S.: Improving
of information transport security under the conditions of destructive influence on the
information-communication system. J. Theort. Appl. Inf. Technol. 89(2), 352–361 (2016)
22. Heckerman, D.: A tutorial on learning with bayesian networks, Tecnical report, Redmond:
Microsoft Research (1995). 58 p.
23. Raileanu, L.E., Stoffel, K.: Theoretical comparison between the gini index and information
gain criteria. Ann. Math. Artif. Intell. 41(1), 77–93 (2004)
24. Alhendawi, K.M., Al-Janabi, A.A.: An intelligent expert system for management information
system failure diagnosis. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.) Intelligent Computing
& Optimization. ICO 2018. Advances in Intelligent Systems and Computing, vol. 866.
Springer, Cham (2019). https://doi.org/10.1007/978-3-030-00979-3_26
Developing an Intelligent System
for Recommending Products
1 Introduction
With the ever-increasing growth of the internet, there have been a lot of websites
containing a continuously huge number of user-generated texts all around the world.
Depending on the types of websites, a large amount of those texts contain reviews
belonging to various types of products. If processed properly, these texts can be used in
order for understanding public opinion on different products so that manufacturing
companies can adjust their products accordingly. On the other hand, this can also help
customers make proper decisions, when it comes to buying products. While it is easier
to figure out the utilization of a product from learning the average rating, it becomes
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 476–490, 2021.
https://doi.org/10.1007/978-3-030-68154-8_43
Developing an Intelligent System for Recommending Products 477
difficult when ratings are not provided on the website. Finding the best product
involves reading a plenty number of reviews on several items and finally making
decision for getting the best product. In either case, performing sentiment analysis on
online reviews for various products can help both the companies and customers for
saving their valuable time. Supervised learning is a learning technique which requires
suitable labels for efficient result. However, polarity of individual text review is highly
unlikely to be available with texts themselves. Even if they are available, these are
either machine generated or human viewed, and the main problem for both cases is that
they do not reflect the reviewer’s original intention. Moreover, ratings provided with
the reviews can be considered a valid label since a rating is the originally intended
quantitative representation of the reviews. Hence, in this paper, we have proposed an
intelligent system in order to predict ratings for different products and also evaluate
which product is best depending on the user’s reviews. After predicting ratings with
different machine learning models, various performance measurements were checked
for the comparison and evaluation of the overall efficiency.
The rest of the paper is organized as follows: In Sect. 2, Literature review related to our
work is shown. A detail description of our proposed system is shown in Sect. 3. In Sect. 4,
Experimental result is shown. Finally, a conclusion section is provided in Sect. 5.
2 Literature Review
There are some researches which have previously been done on similar topics. For
example, Ahmed et al. Have proposed a framework in their paper [1] about deep
learning in order to predict ratings from online reviews from Yelp and Amazon
Datasets. In their paper, they have managed to achieve a very good performance using
various models. However, it is seemed that they had enough lacking using the linguistic
rules during the analysis of review text.
Due to the ever-growing availability of online reviews, it becomes difficult to draw
chronological insights from them. For this reason, in the paper [1] by Murtadha et al.,
chronological sentiment analysis using dynamic and temporal clustering was presented.
In their research, they have also considered consistency in order to assess the perfor-
mance of their algorithms. However, they have utilized unified feature set of window
sequential clustering in order to enhance the ensemble learning. Murtadha et al. Have
also proposed in another paper [2] an automated unsupervised method of sentiment
analysis. Their method involves contextual analysis and unsupervised ensemble
learning. SentiWordNet has been deployed in both phases. The research datasets
include Australian Airlines and HomeBuilders reviews for domain performance and
algorithm seems to be the most accurate for some datasets even though Support Vector
Machine (SVM) seems to be the most accurate on average. Applying deeper contextual
analysis might have resulted in a higher accuracy score.
Long Mai and Bac Le have taken a slightly different approach towards rating
prediction of product reviews. In their paper [3], they have proposed a unique
framework for automatic collection and processing of YouTube comments on various
products which are then used for sentiment analysis of those products. Various models
478 Md. S. Islam et al.
have been used on their reviews performance with accuracy ranging from 83%–92%.
However, the domain which has been used for the dataset needs to be expanded in
order to make their models more effective in the real world.
In [4], they presented a novel approach that uses aspect level sentiment detection.
They have tested their model on Amazon customer reviews and their acquired cross-
validation score seems to be somewhere near 87%, which is 10% less than the score
has been claimed in the paper. More challenging areas like spam, sarcasm etc. have not
been touched in their current research.
Rishith et al. Introduces in their paper [5], a sentiment analysis model that uses
Long Short Term Memory for analyzing movie reviews. Their model seems to have a
very high accuracy on both the train and the test data. However, it might be able to
achieve a better performance if it was applied on more varied datasets.
This paper [6] reviews the foremost later considers approximately that have utilized
profound learning to disentangle conclusion examination issues, such as suspicion
limit. It is superior to combine significant learning techniques with word embeddings
than with Term Frequency-Inverse Document Frequency (TF-IDF) when performing a
supposition examination. Eventually, a comparative take into account has been con-
ducted on the searching comes more or less gotten for the differing models and input
highlights.
In [7], the authors used applied machine learning algorithms to detect polarity of
the reviews. Their models seem having a very high accuracy with Naïve Bayes.
However, their algorithm cannot detect from review why people like or dislike a
product. In [8], they used Isolation Forest for sentiment analysis on data recovered from
Amazon on products of different brands. While their project flow diagram shows
sentiment detection process, no prediction measurement score is present within their
paper. The use of convolutional neural network is also common in text-based pre-
diction. In [9], they proposed an approach for predicting the helpfulness of different
online reviews in the form of helpfulness score by using Convolutional Neural Net-
work (CNN). However, in this paper, their dataset is limited to Amazon and Snapdeal
only, and it has seemed that they never use any model for getting metadata information
either.
Mamtesh et al. Analyzed customer reviews for a number of movies using K-
Nearest Neighbors (KNN), Logistic Regression and Naïve Bayes in their paper [10].
They were able to predict the sentiment of the reviews with very high accuracy. They
might be able to achieve higher accuracy with a hybrid approach.
In [11], the author analyzes user sentiments of Windows Phone applications and
classifies their polarity by using Naïve Bayes model. The model seems to be able to
achieve a high accuracy. But it can classify in only 2 categories: positive and negative.
Rintyarna et al. Proposed a method in order to extract features from texts by
focusing on the words’ semantics in their paper [12]. To achieve this, they used an
extended Word Sense Disambiguation method. From their experiment, it seems that
their proposed method boosts the performance of the ML algorithms sufficiently.
In [13], they have proposed a method for data pre-processing within three machine
learning techniques Support Vector Machine (SVM), Naïve Bayes Multinomial
(NBM) and C4.5 for the sentiment analysis for both English and Turkish language.
Developing an Intelligent System for Recommending Products 479
Finally, they concluded that although SVM worked well for the English language,
NBM performed well for the Turkish language. In [14], they have proposed an
appropriate and accurate algorithm for finding movie reviews among some of machine
learning based algorithms.
In [15], they proposed a hybrid sentiment analysis method for analyzing Amazon
Canon Camera Review in order to classify them into 2 polarity classes, positive
&negative. Compared to SVM, their approaches seem to be having very high score in
different performance measurements including accuracy, precision, recall, and others.
In [16], Haque et al., they used supervised learning method on Amazon Product
Review on large scale. The model used in their algorithms has performance around
86% to 94% varying on different models and different datasets. Their main limitation is
the scarcity of standard dataset required with 10-fold cross-validation. Sentiment
analysis is a demanding task comprising of natural language processing, data mining,
and machine learning. To tackle this challenge furthermore, deep learning models are
often merged with them. In [17], they have emphasized on the execution of different
DL techniques in Sentiment Analysis.
In [18], the assessment was done by utilizing 10-Fold Cross Validation and the
estimation precision is done by Confusion Matrix and Receiver Operating Character-
istic (ROC) bend. The result appeared an expanding in precision SVM of 82.00% to
94.50% results of testing the model can be concluded that the Support Vector Machine-
based Particle Swarm Optimization provide solutions to problems of classification
more accurate smartphone product reviews.
3 Proposed Methodology
The proposed methodology consists of the following steps for polarity prediction
through sentiment analysis. Figure 1 shows a block diagram for the entire process.
3.2 Pre-processing
This step involves preparing data from the dataset for training model.
There are several steps to this.
3.2.2.4. Lemmatization
Lemmatization involves resolving words to their dictionary form. Hence, dictionary
search is usually used in order to proper lemmatize the words into their dictionary-
accurate base forms. The purpose of lemmatization is to make sure that the algorithm
can distinguish a word in its different forms. This also reduces the word index of the
tokenizer and hence makes the result more efficient.
3.3 Training
In this step, K-Nearest Neighbors (KNN) classifier, Random Forest Classifier (RFC) &
Support Vector Machine (SVM) classifier are trained using padded training data as
features and ratings as labels. Moreover, a deep learning classifier is also developed for
the same purpose. For the second dataset, rather than being trained separately, the
Machine Learning (ML) models are trained on 75% of the entire combined dataset to
Battery reviews and Tablet reviews for better generalization.
Storage Module
Unlocked Amazon
Phone Product Dataset Splitting
Reviews Reviews
Training Set
Test Set
Eliminating Unnecessary
Reviews
Tokenization
Pre-Processing
Sentence Splitting
Text Sequencing
Stopword Removal
Padding
Special Character
Removal
Training Testing
Normalization
Rating Prediction
Lemmatization
Recommendation
Generation
4 Experimental Results
The dataset & number of reviews used for machine learning for different types of
products is shown in Table 1.
Table 2. Accuracy, precision and recall for several classifiers on different products
Classifiers Products Unlocked Phones Batteries Tablets
KNN Accuracy 82.75% 69.18% 84.7%
Precision 78.67% 47.26% 67.54%
Recall 76.37% 41.12% 73.3%
RFC Accuracy 86.63% 77.5% 88.86%
Precision 90.35% 77.92% 95.81%
Recall 77.44% 40.12% 73.57%
SVM Accuracy 81.80% 78.66% 89.52%
Precision 91.93% 89.48% 97.26%
Recall 72.27% 36.17% 74.57%
DL Model Accuracy 86.90% 77.80% 87.21%
Precision 80.86% 53.28% 78.34%
Recall 79.54% 47.22% 77.38%
From Table 1, we can see that 40000 reviews on unlocked phone reviews from the
first dataset have been used. On the other hand, Reviews for batteries and tablets both
come from the second dataset and contain 12071 and 14396 reviews respectively. The
different score types used for measuring performance of different classifiers are accu-
racy, precision, and recall.
TP
Precision ¼
TP þ FP
TP
Recall ¼
TP þ FN
Here, TP, FP & FN denote True Positive, False Positive & False Negative
respectively.
The overall precision for a multi-class classification is the average of the precision
scores of every individual class. The same goes for the overall recall score as well.
Table 2 shows different performance scores for certain classifiers on several types of
products.
Figure 2 shows a graphical representation of accuracy, precision and recall for
several classifiers on different product types. From this figure, it can be seen that Deep
Learning based classifier has both the highest accuracy and the highest recall score for
unlocked phone reviews, with an accuracy score of 86.90% and a recall score of
79.54%. However, Support Vector Machine (SVM) has the highest precision score of
91.93% for the same reviews.
484 Md. S. Islam et al.
The second type of products is battery. For this product’s reviews, Support Vector
Machine (SVM) seems to be having both the highest accuracy score and the highest
precision with accuracy score of 78.66% and precision score of 89.48%. However, the
Deep Learning (DL) based classifier has the highest recall value of 47.22%. Finally,
tablets are the last product type. For the selected tablet reviews, Support Vector
Machine again has the highest accuracy score and the highest precision score of
89.52% and 97.26% respectively. However, Deep Learning (DL) based model again
seems to have the highest recall score with a score of 77.38%.
Developing an Intelligent System for Recommending Products 485
Table 3 shows the training time for all 4 of the classifiers. From here, it can be seen
that the classifiers sorted in their ascending order of their training time are: KNN, RFC,
SVM and DL for either dataset; with KNN’s training time being less than 1second in
both cases and DL model’s training being greater than 300seconds for both cases. It
can be noted that training time is higher on the 2nd dataset for all classifiers; despite the
2nd dataset being smaller. This is because the 1st dataset is very large and that is why,
compared to the 2nd dataset, less training was required for better generalization.
Figure 3 shows the top 5 phones according to the average of their individual
ratings. This can used as a benchmark for comparing the effectiveness recommendation
systems implemented with the ML models.
Fig. 3. Graphical presentation of the average of the actual ratings of Top 5 phones
From Fig. 4, we can see that the DL (Deep Learning) model managed to predict 4
items from the top 5 Phones based on the average of the average ratings (Fig. 3) with
mostly the same order.
486 Md. S. Islam et al.
Fig. 4. Graphical presentation of average of predicted scores for Top 5 phones recommended by
different machine learning models
Similar to Fig. 3, Fig. 5 also shows the average of the actual ratings, but for the
batteries. From Fig. 6, we can see that all four machine learning models predicted the
order correctly. However, both KNN and DL models were able to predict the average
close to the actual values.
Developing an Intelligent System for Recommending Products 487
Fig. 5. Graphical presentation of the average of the actual ratings of the types of batteries
Fig. 6. Graphical presentation of average of predicted scores for the types of batteries
Fig. 7. Graphical presentation of the average of the actual ratings of Top 5 tablets
From Fig. 8, it can be seen that KNN model was able to predict 4 items from
Fig. 7, with the order and the values also being mostly accurate.
Fig. 8. Graphical presentation of average of predicted scores for top 5 tablets recommended by
different machine learning models
Developing an Intelligent System for Recommending Products 489
5 Conclusion
In this paper, we have developed an intelligent system which predicts ratings from
different product reviews and provides recommendation to the customers. From the
experimental results, it has been found that SVM gives better result based on the
performance scores. However, during training time for each classifier, it seems that
Random Forest (RF) is more efficient than SVM and the performance is too close to
SVM for the three product types. Overall, performance scores for all four classifiers are
pretty good. Bag of Words has the tendency of increasing the dimensionality of the
matrix with the increase of vocabulary size. Moreover, since the context of the words
cannot be recovered from Bag of Words approach, semantic meaning is often not
preserved in this method. In order to address these issues, we have decided to post-
truncate and post-pad the review texts with a maximum word length of 15 and 30 for
Dataset 1 and Dataset 2 respectively. In future, we will try to improve the performance
of our system by using some rule-based pre-processing techniques.
References
1. Al-sharuee, M.T.: Sentiment analysis : dynamic and temporal clustering of product reviews
(2020)
2. Al-sharuee, M.T., Liu, F., Pratama, M.: Sentiment analysis: an automatic contextual analysis
and ensemble clustering approach and comparison. Data Knowl. Eng. (2018)
3. Mai, L., Le, B.: Joint sentence and aspect-level sentiment analysis of product comments.
Ann. Oper. Res. (2020)
4. Nandal, N., Tanwar, R., Pruthi, J.: Machine learning based aspect level sentiment analysis
for Amazon products. Spat. Inf. Res. (2020)
5. Rishith, U.T., Rangaraju, G.: Usense: a user sentiment analysis model for movie reviews by
applying LSTM. Int. J. Res. Eng. Appl. Manag. 01, 369–372 (2020)
6. Dang, N.C., Moreno-García, M.N., De la Prieta, F.: Sentiment analysis based on deep
learning: a comparative study. Electron. 9(3) (2020)
7. Jagdale, R.S., Shirsat, V.S., Deshmukh, S.N.: Sentiment Analysis on Product Reviews Using
Machine Learning Techniques : Proceeding of CISC 2017 Sentiment Analysis on Product
Reviews Using Machine Learning Techniques, no. January. Springer Singapore (2019)
8. Salmiah, S., Sudrajat, D., Nasrul, N., Agustin, T., Harani, N.H., Nguyen, P.T.: Sentiment
Analysis for Amazon Products using Isolation Forest (6), 894–897 (2019)
9. Saumya, S., Singh, J.P., Dwivedi, Y.K.: Predicting the helpfulness score of online reviews
using convolutional neural network. Soft Comput., no. BrightLocal 2016 (2019)
10. Mamtesh, M., Mehla, S.: Sentiment analysis of movie reviews using machine learning
classifiers. Int. J. Comput. Appl. 182(50), 25–28 (2019)
11. Normah, N.: Naïve Bayes algorithm for sentiment analysis windows phone store application
reviews. SinkrOn 3(2), 13 (2019)
12. Rintyarna, B.S., Sarno, R., Fatichah, C.: Semantic features for optimizing supervised
approach of sentiment analysis on product reviews. Computers 8(3), 55 (2019)
13. Parlar, T., Özel, S.A., Song, F.: Analysis of data pre-processing methods for sentiment
analysis of reviews. Comput. Sci. 20(1), 123–141 (2019)
14. Dwivedi, S., Patel, H., Sharma, S.: Movie reviews classification using sentiment analysis.
Indian J. Sci. Technol. 12(41), 1–6 (2019)
490 Md. S. Islam et al.
15. Chhabra, I.K., Prajapati, G.L.: Sentiment analysis of Amazon canon camera review using
hybrid method. Int. J. Comput. Appl. 182(5), 25–28 (2018)
16. Haque, T.U., Saber, N.N., Shah, F.M.: Sentiment analysis on large scale amazon product
reviews. In: 2018 IEEE International Conference Innovation Research Deviation, no. May,
pp. 1–6 (2018)
17. Kalaivani, A., Thenmozhi, D.: Sentimental Analysis using Deep Learning Techniques,
pp. 600–606 (2018)
18. Wahyudi, M., Kristiyanti, D.A.: Sentiment analysis of smartphone product review using
support vector machine algorithm-based particle swarm optimization. J. Theor. Appl. Inf.
Technol. 91(1), 189–201 (2016)
19. Amazon Reviews: Unlocked Mobile Phones | Kaggle. https://www.kaggle.com/
PromptCloudHQ/amazon-reviews-unlocked-mobile-phones
20. Consumer Reviews of Amazon Products | Kaggle. https://www.kaggle.com/datafiniti/
consumer-reviews-of-amazon-products
Branch Cut and Free Algorithm
for the General Linear Integer Problem
Elias Munapo(&)
Abstract. The paper presents a branch, cut and free algorithm for the general
linear integer problem. This proposed algorithm, like most of the other exact
algorithms for the general linear integer problem also relies on the usual strategy
of relaxing the linear integer problem and then solve it to obtain a continuous
optimal solution. If solution is integer then the optimal solution is found else the
largest basic variable in the continuous optimal solution is selected and freed of
integral restrictions and then the branch and cut algorithm is used to search for
the optimal integer solution. The main and obvious challenge with the branch
and bound related algorithms is that the numbers of nodes generated to verify
the optimal solution can sometimes explode to unmanageable levels. Freeing a
selected variable of integral restrictions which is proposed in this paper can
significantly reduce the complexity of the general linear integer problem.
1 Introduction
The general linear integer programming (LIP) problem has so many important appli-
cations in life. These applications include, interactive multimedia systems [1], home
energy management [15], cognitive radio networks [7], mining operations [22], relay
selection in secure cooperative wireless communication [11], electrical power alloca-
tion management [18, 20], production planning [4], selection of renovation actions [2],
waste management, formulation and solution method for tour conducting and opti-
mization of content delivery networks [16].
The main challenge with the branch and bound related algorithms is that the
numbers of nodes that are required to verify the optimal solution can sometimes
explode to unmanageable levels. Freeing a variable of all its integral restrictions which
is proposed in this paper can significantly reduce the complexity of the general linear
integer problem. In fact the LIP problem is NP hard, thus it is very difficult to solve and
heuristics are still being used up to now to approximate optimal solutions in reasonable
times. An efficient consistent exact solution for the general LIP is still at large. The
coefficient matrix for the general linear integer problem is not unimodular [14].
A branch, cut and free algorithm which combines the branch and cut algorithm and
freeing of a selected variable is proposed in this paper. This algorithm is similar to [17]
but differs in that it can be used for any linear integer model with any number of
variables and linear constraints. On the other hand the method proposed in [17] requires
calculation of variable sum limits before solving and it works for a single constrained
linear integer problem and no variables are freed in the solving process.
Maximize Z ¼ c1 x1 þ c2 x2 þ . . . þ cn xn ,
Such that:
Minimize Z ¼ c1 x1 þ c2 x2 þ . . . þ cn xn ,
Such that:
Given n variables in an equality constraint such as that given in (2), then one of the
n variables can be freed of integral restrictions. In other words one out of the n vari-
ables is not supposed to be restricted to integer.
a1 x1 þ a2 x2 þ . . . þ aj xj þ . . . þ an xn ¼ b: ð2Þ
Where aj and b are integral constants and the unknown variables are integer
8j ¼ 1; 2; . . .; n:
If all the variables are integers then one variable ðxj Þ out of n variables is not
supposed to be restricted integer.
Branch Cut and Free Algorithm 493
Proof
From (2) the variable ðxj Þ can be made the subject as given in (3).
aj xj ¼ b ða1 x1 þ a2 x2 þ . . . þ an xn Þ: ð3Þ
Since this is the continuous optimal tableau then (4) and (5) are valid.
x1 ; x2 ; x3 ; . . .; xm 0: ð4Þ
b1 ; b2 ; b3 ; . . .; bm 0: ð5Þ
In addition c is the objective value and aij is a constant. Both c and aij can either be
negative or positive. In this paper Zcopt is the continuous optimal solution. The other
examples of general continuous optimal solutions are given in Table 2 and Table 3.
The specific continuous optimal solutions for numerical illustrations are given in
Table 4 and Table 5.
5 Surrogate Constraint
The surrogate constraint or clique constraint is obtained by adding all the rows of
original variables that are basic at optimality as given in (6).
494 E. Munapo
Table 2. Sub-Problem 1.
Table 3. Sub-Problem 2.
x1 þ x2 þ x3 þ . . . þ xm þ k1 s1 þ k2 s2 þ k3 s3 þ . . . þ km sm ¼ bc : ð7Þ
Branch Cut and Free Algorithm 495
Where
bc ¼ b1 þ b2 þ b3 þ . . . þ bm : ð9Þ
bc ¼ I þ f : ð10Þ
s1 ¼ s2 ¼ s3 ¼ . . . ¼ sm ¼ 0: ð11Þ
x1 þ x2 þ x3 þ . . . þ xm I: ð12Þ
Sub-Problem b:
x1 þ x2 þ x3 þ . . . þ xm I þ 1: ð13Þ
x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm þ xn þ 1 ¼ I: ð14Þ
x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm xn þ 2 ¼ I þ 1: ð15Þ
As a diagram, the original problem and the two sub-problems are related as shown
in Fig. 1.
x1 + x2 + x3 + ... + x j x1 + x2 + x3 + ... + x j
+... + xm + xn +1 = I . +... + xm − xn + 2 = I + 1.
a b
Where Node 0 is the original problem, Node a is given by Table 2 which is the
original problem after adding constraint (14) and Node (b) is given by Table 3 which is
obtained after adding constraint (15) to original problem.
The unrestricted variable is selected from that basic variable satisfying (16).
b‘ ¼ Maxfb1 ; b2 ; b3 ; . . .; bm g: ð16Þ
Justification: The larger the variable range, the more branches we are likely to obtain,
so it makes sense not to subject such a variable to integral restriction. The equality
constraint (14) is added to continuous optimal tableau to obtain Sub-Problem a. Sim-
ilarly (15) is added to the continuous optimal tableau to obtain Sub-Problem b.
Where variable xj is now unrestricted by using the variable integral restriction
freeing theorem. In this paper Zopt is the optimal integer solution.
The percentage complexity reduction ðqÞ when using the proposed algorithm is given
as (17).
ðR rÞ
q¼ 100%: ð17Þ
R
Where R is the number of branch and bound nodes before using the proposed algorithm
and r is the number of nodes after freeing a selected variable.
Solving (18) directly by the automated branch and bound algorithm it takes 459
nodes to verify the optimal solution given in (20).
Using the simplex method it gives (21) as the continuous optimal solution.
x1 þ x3 þ x4 þ x5 ¼ 64:3039: ð22Þ
The largest value 31.5636 comes from variable x1 which implies that this variable is
to be freed of the integral restriction.
498 E. Munapo
x1 þ x3 þ x4 þ x5 þ x6 ¼ 64: ð24Þ
x1 þ x3 þ x4 þ x5 x7 ¼ 65: ð25Þ
Note that we consider only the original variables ðx1 ; x3 ; x4 &x5 Þ that are in the
optimal basis. Where x6 and x7 are the additional variables which are also restricted to
integers. In diagram form the original problem and the two sub-problems (a) and
(b) are related as shown in Fig. 2.
x1 + x3 + x4 + x5 + x6 = 64 x1 + x3 + x4 + x5 − x7 = 65
a b
Solving Sub-Problem (a) using the automated branch and bound algorithm, it takes
21 nodes (from Sub-Problem a) to verify the optimal solution given in (20).
Solving Sub-Problem 2 using the automated branch and bound algorithm, it takes 1
node to verify infeasibility. The complexity of this problem is reduced from 459 to just
(21 + 1) = 22 nodes i.e. complexity reduction is 95.2 as given in (27).
ð459 22Þ
q¼ 100% ¼ 95:2%: ð27Þ
459
Mere addition of the cut given in (28) for Sub-problem (a) and the cut given in (29) for
Sub-problem (b) does not reduce the total number of nodes required to verify an
optimal solution by using the automated branch and bound algorithm.
x1 þ x3 þ x4 þ x5 64: ð28Þ
x1 þ x3 þ x4 þ x5 65: ð29Þ
In fact it takes 461 nodes for (Sub-Problem a) and 1 node for (Sub-Problem b) which
gives a total of 462 nodes to verify optimality.
Branch Cut and Free Algorithm 499
A set of 100 randomly generated linear integer problems have shown that com-
plexity decreases significantly if the largest basic variable in the continuous optimal
solution is freed of integral restrictions. As a result of this, freeing of variables is now
combined with cuts to form what is proposed in this paper as the branch, cut and free
algorithm for the general linear integer problem.
Previously cuts have been used to enhance the performance of the branch and bound
algorithm [3, 17, 22] to form what is now known as the branch and cut algorithm [6, 8,
15, 19]. In addition pricing has been used in a branch and bound setting to come up
with the well-known branch and price algorithm [5, 21]. The branch and cut and branch
and price are well-known ideas and have been combined to form a hybrid which is now
called the branch, cut and price [9, 10]. With all these wonderful and impressive ideas
available, the general linear integer problem is still NP hard and very difficult to solve
up to now. Heuristics are still being used to this day to approximate optimal solutions
to this difficult problem [11–13]. Freeing a variable from integral restriction is now
being used in the context of a branch and cut to give birth to the branch, cut and free
algorithm for the general integer problem.
x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm ¼ I i; i ¼ 0; 1; 2. . . ð30Þ
x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm ¼ I þ 1 þ i; i ¼ 0; 1; 2. . . ð31Þ
Where
I. ‘copt is when hyperplane x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm is at Zcopt ,
II. ‘opt is when hyperplane x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm is at Zopt ,
III. ‘s is when hyperplane x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm has the smallest value,
IV. ‘m is when hyperplane x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm has the largest value,
V. c1 x1 þ c2 x2 þ . . . þ cn xn is the green objective plane and it moves from the
continuous optimal point ðZcopt Þ to the optimal integer point ðZopt Þ.
From Fig. 3, it can be noted that there are no other integer points in the shaded
region i.e. in the region between the continuous optimal point ðZcopt Þ and the optimal
integer point ðZopt Þ besides that optimal integer point.
x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm ‘opta 1;
ð34Þ
c1 x1 þ c2 x2 þ . . . þ cn xn Zopta 1:
Let Zoptb be the optimal solution coming from Sub-Problem b. Optimality in this case
can be verified by adding constraint set (35) to the Sub-problem b.
x1 þ x2 þ x3 þ . . . þ xj þ . . . þ xm ‘optb þ 1;
ð35Þ
c1 x1 þ c2 x2 þ . . . þ cn xn Zoptb 1:
The optimal integer solution Zopt is optimal to the original ILP if both (34) and (35)
are infeasible.
Mixed Linear Integer Problem
In the case of a mixed linear integer problem, only those variables that are supposed to
be integer in the original mixed linear integer problem can be freed of integral
restriction.
The variable x3 is the only original variable that is in the optimal basis. This
variable must be freed from integral restrictions. The continuous optimal solution for
Numerical Illustration 2 is given in (37).
The largest value 568.8788 comes from variable x3 which implies that this variable is to
be freed of the integral restriction.
x3 ≤ 568 x3 ≥ 569
a b
x3 = 568, x6 = 4.8333,
x3 = 569, Z = 14894. (fathomed)
Z = 14792.1667.
The branch and cut algorithm is now used to search Sub-Problem a for integer
points.
x3 ¼ 568 i; i ¼ 0; 1; 2. . .
i ¼ 0 : Add x3 ¼ 568 to problem where x3 is a free variable. Solving by branch and cut
we have (39).
x3 568 1 ¼ 567;
ð42Þ
162x1 þ 38x2 þ 26x3 þ 301x4 þ 87x5 þ 5x6 þ 137x7 14792:
Solving Sub-Problem a using the branch ad cut algorithm with x3 as a free variable we
obtain an infeasible solution which shows that there are no other integer points on the
left hand side of ‘copt besides (41).
Similarly we add (43) to Sub-Problem b.
x3 569 þ 1 ¼ 570;
ð43Þ
162x1 þ 38x2 þ 26x3 þ 301x4 þ 87x5 þ 5x6 þ 137x7 14792:
Adding (43) to Sub-Problem b results in infeasibility and this verifies that (39) is
optimal.
8 Computational Experience
One hundred randomly generated pure LIPs with variables ranging from 10 to 110
were used in the computational analysis. The branch and cut algorithm was compared
with the proposed branch, cut and free algorithm. The same number of cuts were used
for each of these two algorithms. What emerged from the computations is that freeing a
basic variable in a branch and cut algorithm is more effective than the plain branch and
cut algorithm in solving pure LIPs.
9 Conclusions
In the paper we presented a way of selecting the variable to be freed. We also presented
a way of dividing the problem into simpler parts as given in (51) and (52). It is easier to
search the separate divisions ði ¼ 0; 1; 2. . .Þ of Sub-Problem a or ði ¼ 1; 2; 3; . . .Þ of
Sub-Problem b than the original ILP problem as a whole. In addution we presented
optimality verification of the optimal integer solution ðZopt Þ: This is a new avenue for
research that will attract the attention of many researchers in the area of linear integer
programming. A lot has been done for linear integer programming in terms of exact
methods such as branch and cut, branch and price and the hybrid branch, cut and price.
We were not aware of variables that have been freed of integral restrictions before and
the concept used in solving the linear integer problem. Variable freeing may provide
the answer to the difficult general linear integer problem. Large numbers of branches
are prevented as the variables with large variable ranges can be identified and not
subjected to integral restrictions. In the paper only one variable was freed. There is a
need to explore ways to free more than one variable. Variable freeing is an area in its
early stages of development.
504 E. Munapo
References
1. Abdel-Basset, M., El-Shahat, D., Faris, H., Mirjalili, S.: A binary multi-verse optimizer for
0-1 multidimensional knapsack problems with application in interactive multimedia systems.
Comput. Ind. Eng. 132, 187–206 (2019)
2. Alanne, A.: Selection of renovation actions using multi-criteria “knapsack” model. Autom.
Constr. 13, 377–391 (2004)
3. Alrabeeah, M., Kumar, S., Al-Hasani A., Munapo, E., Eberhard, A.: Computational
enhancement in the application of the branch and bound method for linear integer programs
and related models. Int. J. Math. Eng. Manage. Sci. 4(5), 1140–1153 (2019) https://doi.org/
10.33889/IJMEMS.2019.4.5-090
4. Amiri, A.: A Lagrangean based solution algorithm for the knapsack problem with setups,
Expert Syst. Appl. 1431 (2020)
5. Barnhart, C., Johnson, E.L., Nemhauser, G.L., Savelsbergh, M.W.P., Vance, P.H.: Branch
and price column generation for solving huge integer programs. Oper. Res. 46, 316–329
(1998)
6. Brunetta, L., Conforti, M., Rinaldi, G.: A branch and cut algorithm for the equicut problem.
Math. Program. 78, 243–263 (1997)
7. Dahmani, I., Hifi, M., Saadi, T., Yousef, L.: A swarm optimization-based search algorithm
for the quadratic knapsack problem with conflict Graphs, Expert Syst. Appl. 14815 (2020)
8. Fomeni, F.D., Kaparis, K., Letchford, A.N.: A cut-and-branch algorithm for the Quadratic
Knapsack Problem. Discrete Optimization (2020)
9. Fukasawa, R., Longo, H., Lysgaard, J., Poggi de Aragao, M., Uchoa, E., Werneck, R.F.:
Robust branch-and-cut-price for the Capacitated vehicle routing problem. Math. Program.
Series A 106, 491–511 (2006)
10. Ladányi, L., T.K. Ralphs, L.E.: Branch, cut and price: sequential and parallel. In:
Computational Combinatorial Optimization, Naddef, N., Jüenger, M., eds, Springer, Berlin (
2001)
11. Lahyani, R., Chebil, K., Khemakhem, M., Coelho, L.C.: Matheuristics for solving the
multiple knapsack problem with setup. Comput. Ind. Eng. Vol. 129, 76–89 (2019)
12. Lai, X., Hao, J.K., Fu, Z.H., Yue, Y.: Diversity-preserving quantum particle swarm
optimization for the multidimensional knapsack problem, Expert Syst. Appl., 1491 (2020)
13. Lai, X., Jin-Kao Hao, Fu, Z.H., Yue, D.: Diversity-preserving quantum particle swarm
optimization for the multidimensional knapsack problem, Expert Syst. Appl. Vol.
1491 (2020)
14. Micheli, G., Weger, V.: On rectangular unimodular matrices over the algebraic integers.
SIAM J. Discr. Math. 33(1), 425–437 (2019)
15. Mitchell, J.E.: Branch and cut algorithms for integer programming. In: Floudas, C.A.,
Pardalos, P.M., (Eds.), Encyclopedia of Optimization, Kluwer Academic Publishers (2001)
16. Munapo, E.: Network reconstruction – a new approach to the traveling salesman problem
and complexity. In: Intelligent Computing and Optimization Proceedings of the 2nd
International Conference on Intelligent Computing and Optimization 2019 (ICO 2019),
pp. 260–272 (2020)
17. Munapo, E.: Improvement of the branch and bound algorithm for solving the knapsack linear
integer problem. Eastern-Euro. J. Enter. Technol. 2(4), 59–69 (2020)
Branch Cut and Free Algorithm 505
18. Munapo, E.: Improving the optimality verification and the parallel processing of the general
knapsack linear integer problem. In: Research Advancements in Smart Technology,
Optimization, and Renewable Energy (2020)
19. Oprea, S.V., Bâra, A., Ifrim, G.A., Coroianu, L.: Day-ahead electricity consumption
optimization algorithms for smart homes. Comput. Ind. Eng. 135, 382–401 (2019)
20. Padberg, M., Rinaldi, G.: A branch and cut algorithm for the resolution of large-scale
symmetric traveling salesman problems. SIAM Rev. 33(1), 60–100 (1991)
21. Salvelsbergh, M.W.P.: A branch and price algorithm to solve the generalized assignment
problem. Oper. Res. 45, 381–841 (1997)
22. Taha, H.A.: Operations Research: An Introduction, Pearson Educators, 10th Edition (2017)
Resilience in Healthcare Supply Chains
1 Introduction
Many solutions have been proposed by different societies in order to improve
the way they carry out their activities and to adapt measures that allow them
to prepare for and mitigate disasters or catastrophes that could happen at any
time. It is important to consider that historical data and previous events are
very relevant to study in order to mitigate disasters [7]. For example, it could
be considered a disaster or a catastrophe, in the interruption of daily activities,
whether due to fires, earthquakes and floods. Therefore, governments seek to
be prepared to minimize the damage that could occur through proper disaster
management [1,3]. It should be noted that these disasters affect a country, a
government, or a society in a specific manner, which can be controlled, and the
effects mitigated in short periods. As well as other countries can support the
affected country, either by sending medical supplies, rescue support personnel
and basic necessities, among other things.
Resilience plays an extremely important role, since interruptions in the health
sector are a reality and something very common. Considering that not only a
pandemic can affect, but also natural disasters and social, economic and political
conflicts, among other things, for which a rapid response is necessary.
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 506–519, 2021.
https://doi.org/10.1007/978-3-030-68154-8_45
Resilience in Healthcare Supply Chains 507
2 Literature Review
The new way of managing knowledge is based on Digital Twins, and one of
the main fields to which it is being applied is in health. Since it is important
to take into consideration that they allow to give impartial conclusions of the
patients. That’s why healthcare supply chains have currently been considered
different from the usual supply chains due to their high level of complexity, the
presence of high value medical materials and, finally, the fact that they deal with
human lives. On the other hand, it is important to mention that international
health systems are under constant and increasing pressure to reduce waste and
eliminate unnecessary costs while improving the quality and consistency of care
provided to the patient. As well as providing 100% service level, to avoid short-
ages in hospital wards. That’s why authors have been developing different ways
to manage medical knowledge. MyHealthAvatar and Avatar Health [5] are two
clear examples of how health knowledge can be created, which aim to collect
and track people’s lifestyle and health data. MyHealthAvatar is a project whose
main feature is data analysis and presentation with which doctors can visualize
it. Because results are presented in a 3d avatar, health status dashboard, disease
risks, clock view, daily events. While Avatar Health collects information through
health monitoring devices, to check that the patient’s habits and parameters are
correct. So, it would act as an equivalent of a human, considering its physical
state, condition of life and habits. These would allow the use of this information
not only individually, but also collectively, which could be very useful for Digital
Twins focused on health issues. Because diseases could be predicted, drugs that
the population needed, emergencies, diseases, epidemics. That’s why with the
development of big data, the cloud and the Internet of things, the use of the
digital twins as a precision simulation technology of reality has been enhanced.
508 J. A. Marmolejo-Saucedo and M. S. Hartmann-González
Unlike what happens today, because the world population has been affected by
an infectious and highly contagious disease better known as COVID-19. So, each
country is trying to serve its population and safeguard its own people. Therefore,
the ability to support in such a disastrous situation makes decision-making by
top management, the interaction of societies and the activities of supply chains
impossible. So far, the effect of COVID-19 is unknown and its research very
expensive, so it is considered highly dangerous once contracted by the person,
because the contagion capacity is very high. It is sad but very true that many
countries are faced with the fact that they do not have the necessary medical
and human resources to be able to combat this virus, and even less considering
the outbreak and contagion rate of this disease. So the health care supply chain
is being severely affected, because “everyone” needs it at the same time and
does not have such responsiveness. Therefore, it is important that governments
look for the correct way to prioritize health personnel in order to provide bet-
ter service to the community and find the best way to manage the health care
supply chain to avoid its interruption and mitigate harm to the population in
the best possible way. The proposed model has as its main objective the use
of mitigation activities to reduce the effects of COVID-19, that is, reduce the
interruptions in the health care supply chains and provide better service to com-
munities that do not have access to health care. Several different solutions have
been described by [6], among which are increasing the flexibility of companies
and supply chains, such as: the postponement of production; the implementation
of strategic stocks to supply various demands; the use of a flexible supplier base
to be able to react more quickly; the use of the make or buy approach; planning
of transportation alternatives; and the active management of income and prices,
directing consumption to products with greater availability. Figure 1 shows the
possible combinations to design a supply chain considering the most important
factors that occur in the industry.
Figure 2 presents the process diagram for designing resilient supply chains,
which consists of a prior analysis of possible disruptions.
510 J. A. Marmolejo-Saucedo and M. S. Hartmann-González
4 Mathematical Model
In this section, we apply the supply chain design model for a resilient supply
network. We consider the model into a generalized network. The model is a
mixed-integer linear problem.
Let K be the set of manufacturing plants. An element k ∈ K identifies a
specific plant of the company. Let I be the set of the potential warehouse. An
element i ∈ I is a specific warehouse. Finally, let J be the set of current distri-
bution centers, a specific distribution center is any j ∈ J. Let Z denote the set
of integers numbers {0, 1}.
Resilience in Healthcare Supply Chains 511
4.1 Parameters
Qk = Capacity of plant k.
βi = Capacity of warehouse i.
Fi = Fixed cost of opening warehouse in location i.
Gki = Transportation cost per unit of the product from the plant k to the ware-
house i.
Cij = Cost of shipping the product from the warehouse i to the distribution
center (CeDis) j.
dj = Demand of the distribution center j.
We have the following sets of binary variables to make the decisions about the
opening of the distribution center, and the distribution for the cross-docking
warehouse to the distribution center.
1 If location i is used as a warehouse,
Yi =
0 otherwise,
1 If warehouse i supplies the demand of CeDis j,
Xij =
0 otherwise,
Wki = The amount of product sent from plant k to the warehouse i is repre-
sented by continuous variables
We can now state the mathematical model as a (P) problem based on [2].
min Z= Gki Wki + Fi Yi + Cij dj Xij (1)
Wki ,Yi ,Xij
k∈K i∈I i∈I i∈I j∈J
Subject to constraints:
Capacity of the plant
Wki ≤ Qk , ∀k ∈ K (2)
i∈I
Balance of product
dj Xij = Wki , ∀i ∈ I (3)
j∈J k∈K
Warehouse capacity
dj Xij ≤ βi Yi , ∀i ∈ I (5)
j∈J
512 J. A. Marmolejo-Saucedo and M. S. Hartmann-González
Demand of items
pYi ≤ Wki , ∀i ∈ I (6)
k∈K
p = min{dj } (7)
Wki ≥ 0, ∀i ∈ I, ∀k ∈ K (8)
Yi ∈ Z, ∀i ∈ I (9)
Xij ∈ Z, ∀i ∈ I, ∀j ∈ J (10)
The objective function (1) considers in the first term the cost of shipping
the product from the plant k to the warehouse i. The second term contains the
fix cost to open and operate the warehouse i. The last term incorporates the
cost of fulfilling the demand of the distribution center j. Constraint (2) implies
that the output of plant k does not violate the capacity of plant k. Balance
constraint (3) ensures that the amount of products that arrive to a distribution
center j is the same as the products sent from the plant k. The demand of each
distribution center j will be satisfied by a single warehouse i, this is achieved by
constraint (4). Constraint (5) bounds the amount of products that can be sent to
a distribution center j from an opened cross-docking warehouse i. Constraint (6)
guarantees that any opened warehouse i receives at least the minimum amount
of demand requested by a given distribution center j. Constraint (7) ensures
that the minimum demand of each distribution center j is considered. Finally,
constraints (8), (9) and (10) are the non-negative and integrality conditions.
5 Case Study
The case study considers a company that manufactures medical supplies that
has its base of operations in central Mexico, and wants to locate new warehouses
of finished product to reduce delivery times. Likewise, it is intended to use a mul-
tistore strategy that allows increasing the resilience of the supply chain. To sim-
ulate the case of an epidemic or pandemic, a hypothetical situation is presented
where the various potential warehouse locations will be affected depending on
the distribution of the disease. In other words, the government establishes differ-
ent times for closing operations depending on the geographic area. Decisions are
required to locate new facilities for the production and distribution of injectable
products. Likewise, various operating scenarios (under disruption conditions) of
the supply, manufacturing, inventory and product distribution process are mod-
eled and analyzed, see Fig. 3.
Resilience in Healthcare Supply Chains 513
This section will seek to compare the performance of several inventory policies
that allow the design of supply chain networks, in order to determine which is the
best option according to what is needed. The first approach consider disruption
scenarios during supply chain design decisions, among which are location and
allocation decisions. Whereas, the second approach makes decisions about the
design of the supply chain without considering the disruption scenarios. There-
fore, each one will give their own result. In this way, when comparing the total
514 J. A. Marmolejo-Saucedo and M. S. Hartmann-González
profits under these two approaches, the benefits of considering the disruptions
in the supply chain design model are implied. Therefore, it can be concluded
that not only the location costs of the facilities come to affect the location deci-
sions, but also the rates of disruption in the facilities are important factors in
determining and deciding where the warehouses will be located. For example,
despite the low cost of installation location, a decision may be made not to open
a warehouse at a candidate site due to high disruption rates. Additionally, the
number of clients to be served and client assignments to distribution centers
may depend on disruption rates in warehouses. Therefore, a significant increase
in total profit can be achieved by considering facility disruptions in the supply
chain design model.
The Inventory Control Policy was created in order to define inventory levels,
a policy (s, S) with Safety Stock is modeled, which is better known as Min-Max
policy with Safety Stock. Which assumes that both the time between orders and
the order quantity are variable, where the latter varies between the order level S
and the Reorder Point (ROP) s. Considering this, the parameters to be defined
are the safety stock (SS), the minimum inventory value (s) and the maximum
level (S): √
SS = z · σ · LT (11)
s = d · (LT ) + SS (12)
S =2·s (13)
Where z is the z-value obtained from the tables of the normal distribution,
σ is the standard deviation of the demand, LT is the delivery time of the supply
and d is the demand. Also, weekly consumption has been considered for the
calculation of the inventory parameters. As well as a service level equal to 99.9%
has been established for all classes of products; thus, z is set to 3.5. The initial
stock when the simulation starts has been set equal to the maximum value (S).
Considering the suppliers, it can be assumed that they have sufficient capacity
Resilience in Healthcare Supply Chains 515
to always satisfy the demand, without any problem. In other words, they have
a very intense production rhythm and can produce for the national and inter-
national market. Therefore, their inventory levels are modeled to be infinite.
The model is implemented in Anylogistix software (ALX) and simulated with a
one-year period (Jan 1, 20, Dec 31, 20) considering different scenarios, see [4].
First, the starting point is analyzed without any disruption. Next, a disruption
will be inserted into the model and then supply chain performances with and
without different recovery policies are evaluated to mitigate the impact of the
disruption. Considering the disruption, the model introduces a complete closure
of the supplier. Recovery policies are based on different strategies. The first one
requires to increase the inventory levels in the main warehouse, for all prod-
ucts with high demand. The second consists in activating lateral transhipment
between facilities (plants, warehouses and distribution centers. However, in order
to be more effective, in some cases this action requires also an increase in the
inventory levels.
6 Computational Results
From the analysis of the demand for products, it was sought to establish inven-
tory policies in order to be able to offer the best level of service to customers,
which is intended to be greater than 90%. Different inventory policies were tested,
among which are:
• Min-max policy: Products are ordered when the inventory level falls below
a fixed replenishment point (s). The ordered quantity is set to such a value that
the resulting inventory quantity equals S. • Min-max policy with safety stock:
Products are ordered when the inventory level falls below a fixed replenishment
point (s safety stock). The ordered quantity is set to such a value that the result-
ing inventory quantity equals S safety stock. • RQ policy: When the inventory
level falls below a fixed replenishment point (R), the fixed replenishment quan-
tity (Q) of products is ordered. The time horizon considered is 1 year of customer
order. The model was simulated several times, in order to observe different sce-
narios and be able to choose the best possible option. The policy that obtained
the best result from the different KPIs analyzed is the Min-Max policy with
different ranges in both minimums and maximums for products. For the mini-
mum of Carsilaza the quantity of 40,000 pieces and a maximum of 60,000 pieces
was stipulated. While for Nuverasa the minimum amount of 15,000 pieces was
stipulated and a maximum of 25,000 pieces. Obtaining the following KPIs as a
result (Fig. 6):
516 J. A. Marmolejo-Saucedo and M. S. Hartmann-González
When evaluating another scenario with the Min-Max policy with safety stock,
it showed a total cost of $ 3, 273,810 USD, but the level of service offered was 90%
and a Profit of $ 321, 112,234 USD. So it was decided to choose the previously
presented scenario because the profit was above 20M USD and the service offered
was 95%. It is necessary to consider that the higher the service level is, the higher
the costs will be, since these are directly proportional together with the inventory
level.
The following results obtained show the service level, the average available
inventory and the lead time for a recovery strategy based on Min-Max inventory
levels for a disruptive scenario, see Figs. 7, 8, 9, 10, 11 and 12. Likewise, a
comparison of KPIs is presented, corroborating that the proposed inventory
strategy substantially improves profit and other indicators.
7 Conclusions
In this paper, the effect of different disruptions on the resilient design of a health-
care supply chain is studied. The disruptions considered model those that can
occur in cases of global pandemics such as COVID-19.
The design of resilient healthcare chains is proposed through the mathemat-
ical modeling of the problem and the dynamic simulation of various inventory
policies. Likewise, different KPIs are used to evaluate the performance of the
proposals in the different disruption scenarios.
Specialized software for supply chain analysis is used to implement the pro-
posals and the alternatives for resilient designs are contrasted for different levels
of service.
The presented study offers an alternative to reduce the severe impacts in the
supply chains dedicated to health care.
Resilience in Healthcare Supply Chains 519
References
1. Acar, M., Kaya, O.: A healthcare network design model with mobile hospitals for
disaster preparedness: a case study for Istanbul earthquake. Transp. Res. Part E
Logist. Transp. Rev. 130, 273–292 (2019). https://doi.org/10.1016/j.tre.2019.09.
007. http://www.sciencedirect.com/science/article/pii/S136655451930314X
2. Marmolejo, J., Rodrı́guez, R., Cruz-Mejia, O., Saucedo, J.: Design of a distribution
network using primal-dual decomposition. Math. Probl. Eng. 2016, 9 (2016)
3. Rezaei-Malek, M., Tavakkoli-Moghaddam, R., Cheikhrouhou, N., Taheri-
Moghaddam, A.: An approximation approach to a trade-off among efficiency, effi-
cacy, and balance for relief pre-positioning in disaster management. Transp. Res.
Part E Logist. Transp. Rev. 93, 485–509 (2016). https://doi.org/10.1016/j.tre.2016.
07.003. http://www.sciencedirect.com/science/article/pii/S136655451630134X
4. anyLogistix supply chain software: supply chain digital twins, February 2020.
https://www.anylogistix.com/resources/white-papers/supply-chain-digital-twins/
5. Spanakis, E.G., Kafetzopoulos, D., Yang, P., Marias, K., Deng, Z., Tsiknakis, M.,
Sakkalis, V., Dong, F.: myhealthavatar: personalized and empowerment health ser-
vices through Internet of Things technologies. In: 2014 4th International Confer-
ence on Wireless Mobile Communication and Healthcare - Transforming Healthcare
Through Innovations in Mobile and Wireless Technologies (MOBIHEALTH), pp.
331–334 (2014). https://doi.org/10.1109/MOBIHEALTH.2014.7015978
6. Tang, C.S.: Robust strategies for mitigating supply chain disruptions. Int. J. Logist.
Res. Appl. 9(1), 33–45 (2006). https://doi.org/10.1080/13675560500405584
7. Yan, Y., Hong, L., He, X., Ouyang, M., Peeta, S., Chen, X.: Pre-disaster investment
decisions for strengthening the Chinese railway system under earthquakes. Transp.
Res. Part E Logist. Transp. Rev. 105, 39–59 (2017). https://doi.org/10.1016/j.tre.
2017.07.001. http://www.sciencedirect.com/science/article/pii/S1366554516306913
A Comprehensive Evaluation
of Environmental Projects Through
a Multiparadigm Modeling Approach
1 Introduction
The impact of human activities on the environment has become more relevant in recent
years worldwide. This has generated the need to evaluate technical proposals that allow
reducing the environmental impact derived mainly from production and consumption,
it has been chosen to reduce the generation of waste, increase recycling and produce
through clean energy. In this transition process, the United Nations determined a set of
sustainable development objectives with a horizon to 2030, among these objectives it is
worth highlighting the objectives related to energy and caring for the environment. In
this regard, it is sought to be able to have clean energy sources that allow access
worldwide to the entire population as well as maintain the systemic balance and ensure
the care of natural resources for future generations.
In this transformation process towards more sustainable production and con-
sumption, it is necessary to have the collaboration of public and private institutions. For
the design of public policies and proposals that allow achieving the objectives set. An
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 520–529, 2021.
https://doi.org/10.1007/978-3-030-68154-8_46
A Comprehensive Evaluation of Environmental Projects 521
essential point in this stage of transition towards more sustainable approaches is the
evaluation of the proposals from the public and private point of view. Since in most
cases the investments are made from the private initiative with public support, once the
technical evaluation stage is approved, the financial feasibility of the environmental
projects is evaluated. In this area where the need to innovate through the application of
new approaches is observed to be able to define if an environmental project is viable,
taking into account both financial and environmental results metrics.
The classic methodologies for evaluating projects to access financing consider the
financial profitability of the project and the period of recovery of the investments,
leaving the evaluation of environmental results in second place, considering this stage
as part of an ex-post evaluation of the project. However, through different quantitative
tools such as discrete and dynamic simulation models, the integration of project
uncertainty through real options, as well as the definition of cost-effectiveness criteria
in ex-ante evaluations, it is possible to project and evaluate the results of an envi-
ronmental project in a feasibility study. The accelerated depletion of renewable and
non-renewable natural resources in recent years has generated the need to implement
policies focused on sustainable development, which allow guaranteeing access to
natural resources for future generations.
As part of the implementation of these policies, the need to develop financing by
the specific characteristics of projects focused on environmental conservation.
Emphasis has been placed on promoting the generation of clean energy and projects
focused on reducing the environmental impact of the activity productive [1]. A fun-
damental factor in evaluating the financial viability of these projects is to determine the
degree of effectiveness of the desired environmental results, which is not necessarily
compatible with classic investment project evaluation approaches, in many cases, the
presence of negative net benefits is justifiable when intangible benefits are taken into
account. It is necessary to take into account the intangible results in problems
intertemporal related to the management and conservation of natural resources [2]. The
need arises to establish a framework according to the characteristics of environmental
projects and above all to emphasize the evaluation of the expected results that cannot
be measured through the profitability of the investment alone. It is necessary to have
comprehensive methods for evaluating environmental projects that allow evaluating
financial viability, operational feasibility, as well as the expected environmental impact.
The integration of environmental aspects as requirements in the operation of contem-
porary companies, as well as the development of sustainable companies, have gener-
ated the need to objectively assess the technical, financial, and especially
environmental viability of the proposals. Until now, there are evaluation proposals, but
most of them are focused on a single aspect of the evaluation, likewise, the classic
approaches of project evaluation are not adapted efficiently in the case of environ-
mental projects due to the particularities and objectives of the same [3, 4].
At the national and international level, schemes have been adopted to promote
sustainable development through policies to reduce environmental impact, specialized
regulation, and generation of financing schemes for clean energy projects; as well as a
decrease in the environmental impact of productive activity [5, 6]. Consistent with
these policies, it is necessary to have project evaluation schemes beyond traditional
financial or cost-benefit evaluations. Since an Environmental project can be beneficial
522 R. Rodriguez-Aguilar et al.
in environmental terms, despite operating with negative net results. An additional factor
to consider is the treatment of uncertainty since in the case of environmental projects
multiple factors can impact the feasibility of the project. Until now, the approach to
evaluating environmental projects has been analytical, focusing on specific project
segments, which is why there is a need for an eclectic approach based on robust
methods that allow support for decision-making.
The work is structured as follows, in section two the methodological framework of
the proposal is presented, briefly describing the quantitative tools that will be used.
Section three addresses the application of the proposed methodology in a case study
and finally the conclusions and recommendations are presented.
2 Methodology
In the specific case of environmental projects, the use of discrete simulation will
allow evaluating the feasibility of the process to be implemented as well as the
expected results according to different scenarios considered. One of the advantages of
using discrete simulation is that it allows considering the uncertainty in the behavior of
the system to be simulated.
A Comprehensive Evaluation of Environmental Projects 523
dy
¼ f ð x; yÞwithyðx0 Þ ¼ y0 : ð1Þ
dx
The objective is to identify the expected trajectory of a state variable of the dif-
ferential equation. Since not all differential equations have a procedure to find the
analytical solution, it is necessary to resort to the application of numerical methods for
their solution. In the case of the application of dynamic systems to systems related to
human activity, the approach known as System Dynamics was developed, which is a
methodology for the analysis and temporal modeling in complex environments [8]
(Fig. 2).
The Systems Dynamics methodology considers the behavior over time of a level
variable, flows, and the related environment. The application of this methodology will
allow considering the interactions in a system to be modeled in a defined time horizon
allowing to generate environmental variables of result for the financial evaluation and
cost-effectiveness of the project.
Costoftheintervention
Cost effectivenessratio ¼ : ð2Þ
Measureofeffectiveness
The results are shown in what is known as the cost-effectiveness plan, which allows
locating the alternatives with the trade-off between costs and effectiveness of the
intervention (Fig. 3).
The desired results are those located in plane II, being the most effective and least
expensive options.
A Comprehensive Evaluation of Environmental Projects 525
The results of each evaluation stage will be focused on an element of the project
and each result generated will be integrated into the final evaluation that seeks to
evaluate the results and costs of the alternatives on a cost-effectiveness level.
3 Case Study
These general parameters were used as inputs to evaluate the project with the
multiparadigm approach. The elaboration of each model is not detailed because the
objective of the study is to present the methodological structure of the multi-paradigm
evaluation.
The discrete simulation model is based on the daily operations of the company, the
objective of this model is to evaluate the feasibility in the change of technology when
going from the use of gas heaters to solar heaters. Company operations were simulated
for 8 business hours six days a week. Due to the intermittency of the photovoltaic
generation, the simultaneous operation of gas and solar heaters is evaluated. Seeking to
minimize the use of heaters as much as possible, which will reduce operating costs and
reduce the emissions generated by the combustion of this fuel. For its part, the dynamic
model will allow modeling based on the operation of the company with the integration
of photovoltaic energy the emissions generated. It is important to consider that in this
case the emissions generated will be compared concerning the status quo and in the
same way, in the cost-effectiveness evaluation, the new technology will be compared
with the status quo. Table 2 shows the main results of the discrete and dynamic
simulation models, in a simulation horizon of one year, it is important to highlight that
in the case of the dynamic model the horizon is long term.
With the information generated through the simulation models, the financial
evaluation will be carried out using real options. In this case, three scenarios will be
considered within the life of the project, the expansion, contraction of the project, and
A Comprehensive Evaluation of Environmental Projects 527
closure are considered. The evaluation will be carried out in year 10 with the following
considerations:
a) Increase the investment amount by 30%, with expenses of 30%.
b) Reduce the investment amount by 25%, with 28% savings.
c) Settle the project with a recovery value of 50%.
Once the probability that the initial Net Present Value (NPV) of the project will rise
or fall is calculated, using the binomial up-down model, the formula corresponding to
each scenario is applied to each probability of year ðnÞ and then all periods are dis-
counted, eliminating the effect of the probability of increase and decrease (up and
down) and is brought to present value, removing the effect of the interest rate. From the
amount obtained, the price of the exercise or initial investment is subtracted and the
present value of the project is taken with real options (Table 3).
The financial evaluation of the project in the three scenarios generated the fol-
lowing results (Table 4).
With these results, it is concluded that, in year 10, it is not convenient to either
reduce the size of the project or liquidate it, but rather expand it, since not only does it
present a more valuable NPV, but also the value of the option is positive, indicating
that it should be exercised.
Once the results measures and the costs incurred to achieve environmental benefits
in the project have been determined, it is necessary to evaluate whether these actions
are cost-effective, not only with the initial situation but also with other options available
528 R. Rodriguez-Aguilar et al.
in the market, to ensure that the choice made is the best for the company. To determine
which option is the most cost-effective, we must determine the costs associated with the
measurement, but also the elements related to the measurement of results, which is the
emission of Greenhouse gas (GHG), its savings, and the costs of producing them. In the
case of the cost-effectiveness study, different proposals are evaluated to have a com-
parator (Table 5).
Two providers of the technology to be implemented were evaluated and the costs of
each option were estimated to evaluate which is the most cost-effective option, addi-
tionally, the outcome measures were considered for each alternative. Equipment A is
cheaper than equipment B, also, it generates less GHG emissions. What makes it more
cost-effective is that the cost of avoided emissions is also lower, so the company will
make sure that choosing supplier A is better than supplier B.
4 Conclusions
The world economies are gradually moving towards more environmentally friendly
production models as a result of trying to harmonize the economic, social, and envi-
ronmental axes to achieve sustainable growth. The integration of environmental
objectives as millennium development goals is a great advance towards the protection
of the environment and the pursuit of sustainable development. One of the great
challenges for the countries in the evaluation and selection of environmental projects
with the greatest possible impact, for this up to now only financial profitability criteria
were considered, which contemplated the recovery of the investment in a defined time,
especially in those projects where there is greater participation of private capital.
However, this approach limits the environmental impact of many projects that could be
profitable in environmental and not necessarily financial benefits.
The use of a multi-paradigm evaluation approach allows a comprehensive evalu-
ation of various aspects of the project, from an operational, technical, environmental,
and financial perspective. The integration of a discrete and dynamic simulation method,
as well as the financial evaluation of the uncertainty of the project through real options,
allow having information of greater added value for decision-making on environmental
projects. The integration of a results evaluation approach in terms of cost-effectiveness
makes it possible to weigh the environmental results on the investment made in the
projects.
A Comprehensive Evaluation of Environmental Projects 529
The case study presented shows the application of the proposed methodology
through the evaluation of a reference environmental project that allows integrating the
methodological approaches addressed. The results show that the integration of simu-
lation methodologies, the use of real options as well as the cost-effectiveness approach,
allow having robust and reliable information for making decisions. This is a first
methodological proposal that seeks to build comprehensive approaches to the evalu-
ation of environmental projects, and above all to prioritize environmental criteria over
financial ones.
References
1. Panwar, N.L., Kaushik, S.C., Surendra, K.: Role of renewable energy sources in
environmental protection: a review. Renew. Sustain. Energy Rev. 15(3), 1513–1524 (2011)
2. Kronbak, L.G., Vestergaard, N.: Enviromental cost-efectiveness analysis in intertemporal
natural resource policy: evaluation of selective fishing gear. J. Environ. Manage (2013)
3. Manzini, F., Islas, J., Macías, P.: Model for evaluating the environmental sustainability of
energy projects. Technol. Forecast. Soc. Chang. 78(6), 931–944 (2011)
4. Torres-Machi, C., Chamorro, A., Yepes, V., Pellicer, E.: Current models an practicies of
econmic and enviromental evaluation of sustainable network-level pavement management.
J. Constr. 13(2), 49–56 (2014)
5. SEMARNAT: Guía de Programas de Fomento a la Generación de Energía con Recusos
Renovables (2015). Disponible en https://www.gob.mx/cms/uploads/attachment/file/47854/
Guia_de_programas_de_fomento.pdf
6. SENER: Prospectiva de Energías Renovables 2016–2030 (2016). Disponible en https://
www.gob.mx/cms/uploads/attachment/file/177622/Prospectiva_de_Energ_as_Renovables_
2016-2030.pdf
7. Schriber, T.J., Brunner, D.T.: How Discrete‐Event Simulation Software Works. In: Banks,
J. (Ed.) Handbook of Simulation (2007)
8. Rodríguez Ulloa, R., Paucar-Caceres, A.: Soft system dynamics methodology: combining
soft systems methodology and system dynamics. Syst. Pract. Action Res. 18(3), (2015)
9. Calle, A., Tamayo, V.: Decisiones de inversión a través de opciones reales. Estudios Gerenc.
25(111), 7–26 (2009)
10. Finnveden, G., et al.: Recent developments in Life Cycle Assessment. J. Environ. Manage.
91(1), 1–21 (2009)
11. Uchida, E., Rozelle, S.: Grain of green: cost-effectiveness and sustainability of China's
conservation set-aside program. Land Econ. 81(2), 247–264 (2005)
Plant Leaf Disease Recognition Using
Histogram Based Gradient Boosting
Classifier
1 Introduction
Diagnosis and detection of various plant diseases through leaves’ symptoms are
complicated for farmers and agronomists. It creates complexity due to various
symptoms in the same plant and similar symptoms in different plant diseases.
This complicated task may cause misleading to conclude the status of plants and
their proper treatments. Automated plant diagnosis using the mobile application
through the real field’s capturing image helps the agronomist and farmers make
better decisions on plant health monitoring. Due to the growth of the Graph-
ical Processing Unit(GPU) embedded processors, Machine Learning, Artificial
Intelligence makes it possible to incorporates new models and methods to detect
the appropriate ROIs and hence, identify the plant diseases correctly. However,
memory space(number of parameters) is still in consideration for mobile-based
PLD recognition.
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 530–545, 2021.
https://doi.org/10.1007/978-3-030-68154-8_47
Plant Leaf Disease Recognition Using Histogram 531
2 Related Work
3.1 Dataset
In the experiment, 403 images of size 256 × 256 pixels comprising eight differ-
ent plants, such as rice, corn, potato, pepper, grape, apple, mango, and cherry,
and twelve diseases are used to train the proposed framework. The images are
534 S. Md. Minhaz Hossain and K. Deb
collected from the PlantVillage dataset1 except rice disease images. Rice dis-
ease images are gathered from the Rice diseases image dataset in Kaggle2 , the
International Rice Research Institute (IRRI)3 and Bangladesh Rice Research
Institute (BRRI)4 .
We vary the image backgrounds among natural, plain, and complex to trace
a disease properly in different backgrounds. Our framework includes six fungal
diseases, two bacterial diseases, two diseases having both fungal and bacterial
symptoms, one viral disease, and another one from a different category. Further,
the framework considers various symptoms, such as small, massive, isolated,
and spread. Twelve samples of eight plants are represented, considering different
symptoms and image backgrounds, as shown in Fig. 2. For generalization, 235
independent (excluding the training dataset) images from twelve different classes
are used during the test phase. Complete information regarding the plant leaf
disease dataset is described in Table 1.
If images are not captured precisely due to hostile conditions, image enhance-
ment is needed to increase the PLD image quality. The enhancement is optional
as it depends on the magnitude of degradation. Two enhancement conditions
have been used here using statistical features such as mean(µ), median(x ), and
mode(M0 ) of a plant leaf image. The first condition for image enhancement is
devised as in Eq. 1.
µ < x < M0 (1)
1
https://www.kaggle.com/emmarex/plantdisease.
2
https://www.kaggle.com/minhhuy2810/rice-diseases-image-dataset.
3
https://www.irri.org/.
4
http://www.brri.gov.bd/.
Plant Leaf Disease Recognition Using Histogram 535
Fig. 2. Samples of plant leaf disease images under numerous health conditions in var-
ious backgrounds and having different symptoms: (a) Rice Sheath-rot(natural back-
ground, spread symptoms), (b) Rice Tungro(natural background, spread symptoms),
(c) Rice Bacterial leaf-blight(complex background, spread symptoms), (d) Rice blast
(complex background, isolated, small symptoms), (e) Potato Early-blight(plain back-
ground, isolated small symptoms), (f) Potato Late-blight(plain background, isolated
small symptoms), (g) Pepper Bacterial-spot(plain background, small symptoms), (h)
Grape Black-measles(plain background, small symptoms), (i) Corn Northern Leaf-
blight(plain background, spread, spot symptoms), (j) Apple Black-rot(plain back-
ground, small symptoms), (k) Mango Sooty-mould(natural background, spread symp-
toms) and (l) Cherry Powdery-mildew(natural background, small symptoms).
The modified adaptive centroid-based segmentation (ACS) has been applied once
the PLD image quality has been enhanced. In the beginning, the RGB (PLD)
image space is converted to L * a * b color space for better perceptual linearity in
differentiating colors. Conversion from RGB space to L * a * b color space signifi-
cantly increases K-means clustering performance, especially when narrow distin-
guishes among symptoms colors in different plant leaf disorders. Differentiating
among the color intensities having identical ROI color and background is non-
trivial. Another challenge is distinguishing the basic color of ROIs in the same
sunlight shade and shadowing the background. To overcome these challenges, we
536 S. Md. Minhaz Hossain and K. Deb
Where, Xi , C, and N stand for data points, the centroid of data points, and the
total number of data points.
The automatic initialization of K defined using ACS can effectively
detect image characteristics for different orientations and illuminations. ACS,
also,increases the scalability of the proposed segmentation technique as shown
in Fig. 3(c2 , c4 ) and Fig. 3(c1 , c3 ). A few examples under different circumstances,
such as in the same colored reflection on ROIs, in the presence of the shadow
behind the ROIs, overlapped blur images, the orientation of leaf images such as
shrunk ROI and rotation, are as shown in Fig. 4(b1 –b5 ).
Once the PLD image’s ROIs has been traced, the RGB segments are converted
to grayscale images. Then HLTP has been applied to extract the features of
leaf disease. We perform two approaches of feature extraction; namely HLTP-
1 (8 pixels with radius 1) and HLTP-2 (8 pixels with radius 2). Firstly, four
neighboring points are determined using Eq. 7–Eq. 10. Other four points have
been calculated by using linear interpolation coefficient for 45◦ in both HLTPs
formulated using Eq. 11–Eq. 14.
√
a=r− r (4)
b=1−a (5)
f (n + a) = a ∗ f (n + 1) + b ∗ f (n) (6)
d0 = A(r0 , c0 − r) − I (7)
d2 = A(r0 , c0 + r) − I (8)
d4 = A(r0 − r, c0 ) − I (9)
Plant Leaf Disease Recognition Using Histogram 537
d6 = A(r0 + r, c0 ) − I (10)
d1 = a ∗ A(r0 + r − 1, c0 − r + 1) + b ∗ A(r0 + r, c0 − r) − I (11)
d3 = a ∗ A(r0 + r − 1, c0 + r − 1) + b ∗ A(r0 + r, c0 + r) − I (12)
d5 = a ∗ A(r0 − r + 1, c0 + r − 1) + b ∗ A(r0 − r, c0 + r) − I (13)
d7 = a ∗ A(r0 − r + 1, c0 − r + 1) + b ∗ A(r0 − r, c0 − r) − I (14)
Where, a and b are interpolation coefficients, and r is the radius. A(r0 , c0 )
stands for the matrix of PLD gray image I considering each neighbor of position
(r0 , c0 ). In Eq. 6, f(n+a) is the unknown pixel, f(n), and f(n+1) are two known
pixels. Unknown pixels, as shown in Eq. 11–Eq. 14, are formulated by Eq. 6 using
Eq. 7–Eq. 10. In Eq. 7–Eq. 14, d0 , d1 , d2 , d3 , d4 , d5 , d6 , and d7 are all neighboring
pixels’ derivatives. These derivatives are then put into 1 × 8 vector, d. 1 × 8
vector for each pixel Pi ; where, i= 0,1,2,3,...., i.e total (m × n) × 8 matrix is
found; where, m is he width of the plant leaf disease image and n is the height
of plant leaf images.Then, mean threshold(MT) for each pixel Pi is determined
using the surrounding eight pixels of this pixel. Then we get two values; one
contains the lower pattern values and another contains the upper pattern values
formulated in [12]. From that using histogram, we get two vectors of 1 × 256;
one from lower values and another from upper values.
Traditional LTP has the limitation of uneven illumination and orientation
in leaf image. In our modified HLTP, the mean threshold(MT) in [12] has been
considered instead of a fixed threshold to overcome LTP’s drawback. It handles
the variation of the gray level of neighbors and makes it invariant to illumination.
Using linear interpolation in determining directives helps increase the ability
to extract features from different oriented plant leaf images. It outperforms, as
shown in Fig. 3(d2 −e2 ) and Fig. 3(d4 −e4 ) compared to traditional LTP, as shown
in Fig. 3(d1 − e1 ) and Fig. 3(d3 − e3 ). Our modified HLTP functions effectively in
the same colored reflection on ROIs, in the shadow behind the ROIs, overlapped
blur images, and the orientation of leaf images such as shrunk ROI and rotation,
as shown in Fig. 4(c1 − f5 ).
Table 2. Parameters used in histogram based gradient boosting classifier for plant leaf
disease recognition.
Parameters Value(s)
Loss function Categorical cross-entropy
Max iterations 100
Minimum samples in leaf node 30
Maximum leaf nodes 30
Max depth 20
Max bins 255
Learning rate 0.2
Environment. The experiments for recognizing plant leaf disease are executed
on Intel(R) Core i5 7200U 2.5 GHz with 4 GB RAM. The proposed framework
is implemented in Python with packages sklearn and MATLAB.
Dataset for Training and Test. In this experiment, 403 images of eight plants
of size 256 × 256 pixels, are used to train and 235 PLD images are used to test
for twelve classes from different sources.The statistics of different PLD train and
test images is shown in Table 3.
the same color background, due to higher mode as shown in Fig. 3(c2 ) and in a
shadow of the leaf image on the background due to its higher median than other
two statistical values, as shown in Fig. 3(c4 ).
Fig. 3. Effect of image enhancement on recognizing plant leaf disease on critical situ-
ations: (a1 ) rice blast disease image and (a2 ) apple black rot disease image. (b1 ), and
(b2 ) is the leaf image histogram of a1 and a2 , respectively. (c1 ), and (c3 ) is the color
segmentation results of a1 and a2 respectively in traditional K-means clustering having
extra noise without image enhancement, and (c2 ), and (c4 ) is the segmentation results
of a1 and a2 respectively in our modified color segmentation algorithm with image
enhancement. (d1 ), (d3 ) and (e1 ), (e3 ) are the lower and upper features of traditional
LTP respectively. (d2 ), (d4 ) and (e2 ), (e4 ) are the lower and upper features of modified
HLTP respectively.
Fig. 4. The processing examples of rice images in our proposed PLD framework under
different critical environments: (a1 − a5 ) are the RGB PLD samples. (b) Segmented
ROIs after implementation of adaptive centroid-based segmentation. (c) HLTP-1 lower
features. (d) HLTP-1 upper features. (e) HLTP-2 lower features and (f) HLTP-2 upper
features.
Fig. 5. ROC curve of each plant leaf Fig. 6. Confusion matrix for recognizing plant
diseases recognition of our framework. leaf diseases.
space than CNN. So, it is useful and reliable for recognizing PLDs using the
mobile application.
Performance Analysis. Two hundred thirty-five(235) plant leaf disease
images of twelve classes are used to evaluate our PLD recognition framework’s
542 S. Md. Minhaz Hossain and K. Deb
Table 5. Performance evaluation of each classes using our proposed plant leaf disease
recognition framework.
Fig. 7. Some misclassified images: (a), (b), (c) are some false positive rice sheath-rot
images. (d) is rice bacterial leaf blight.(e) and (f) are some false positive potato late
blight images. (g) is false positive apple black-rot image and (h) is false positive corn
northern leaf-blight.
References
1. Vasilyev, A.A., Vasilyev, G.N.S.: Processing plants for post-harvest disinfection of
grain. In: Proceedings of the 2nd International Conference on Intelligent Comput-
ing and Optimization (ICO 2019) , Advances in Intelligent Systems and Computing
1072, 501–505 (2019)
2. Borse, K., Agnihotri, P.G.: Prediction of crop yields based on fuzzy rule-based
system (FRBS) using the takagi sugeno-kang approach. In: Proceedings of the
International Conference on Intelligent Computing and Optimization (ICO 2018),
Advances in Intelligent Systems and Computing 866, 438–447 (2018)
Plant Leaf Disease Recognition Using Histogram 545
3. Boulent, J., Foucher, S., Théau, J., St-Charles, P.L.: Convolutional neural networks
for the automatic identification of plant diseases. Frontiers in Plant Science 10
(2019)
4. Brahimi, M., Mahmoudi, S., Boukhalfa, K., Moussaoui, A.: Deep interpretable
architecture for plant diseases classification. In: Signal Processing: Algorithms,
Architectures, Arrangements, and Applications (SPA), pp. 111–116. IEEE (2019)
5. Ferentinos, K.P.: Deep learning models for plant disease detection and diagnosis.
Comput. Electron. Agriculture 145, 311–318 (2018)
6. Ke, G., Meng, Q., Finey, T., Wang, T., Chen, Ma, W., Ye, Q., Liu, T.Y.: Lightgbm:
a highly efficient gradient boosting decision tree. In: 31st Conference on Neural
Information Processing Systems (NIPS 2017), Long Beach, CA, USA. pp. 1–3
(2017)
7. Liang, W.J., Zhang, H., Zhang, G.F., Cao, H.X.: Rice blast disease recognition
using a deep convolutional neural network. Scientific Reports 9(1), 1–10 (2019)
8. Mohanty, S.P., Hughes, D.P., Salathé, M.: Using deep learning for image-based
plant disease detection. Front. Plant Sci. 7, 1419 (2016)
9. Pantazi, X., Moshou, D., Tamouridou, W.: Automated leaf disease detection in dif-
ferent crop species through image feature analysis and one class classifiers. Comput.
Electron. Agric. 156, 96–104 (2019)
10. Sharma, P., Berwal, Y.P.S., Ghai, W.: Performance analysis of deep learning cnn
models for disease detection in plants using image segmentation. Information Pro-
cessing in Agriculture (2019)
11. Singh, V., Misra, A.: Detection of plant leaf diseases using image segmentation and
soft computing techniques. Inf. Process. Agric. 4, 41–49 (2017)
12. Taha H., Rassem, B.E.K.: Completed local ternary pattern for rotation invariant
texture classification. The Scientific World Journal, p. 10 (2014)
13. Too, E.C., Yujian, L., Njuki, S., Yingchun, L.: A comparative study of fine-tuning
deep learning models for plant disease identification. Comput. Electron. Agric.
161, 272–279 (2019)
Exploring the Machine Learning
Algorithms to Find the Best Features
for Predicting the Breast Cancer
and Its Recurrence
Abstract. Every year around one million women are diagnosed with
breast cancer. Conventionally it seems like a disease of the developed
countries, but the fatality rate in low and middle-income countries is
preeminent. Early detection of breast cancers turns out to be beneficial
for clinical and survival outcomes. Machine Learning Algorithms have
been effective in detecting breast cancer. In the first step, four distinct
machine learning algorithms (SVM, KNN, Naive Bayes, Random for-
est) were implemented to show how their performance varies on different
datasets having different set of attributes or features by keeping the same
number of data instances, for predicting breast cancer and it’s recurrence.
In the second step, analyzed different sets of attributes that are related
to the performance of different machine learning classification algorithms
to select cost-effective attributes. As outcomes, the most desirable per-
formance was observed by KNN in breast cancer prediction and SVM
in recurrence of breast cancer. Again, Random Forest predicts better for
recurrence of breast cancer and KNN for breast cancer prediction, while
the less number of attributes were considered in both the cases.
1 Introduction
Breast cancer is the first and most common cancer for females. Around 10% of
females are affected by breast cancer at any stage of their life. Again, among
the cancer affected women, 34.3% are affected by breast cancer and showed a
high mortality around the world [4]. Most breast cancers begin in the cells that
line the ducts, while fewer breast cancers start in the cells lining the lobules [4].
Breast cancer’s causes are multi-factorial which involves family history, obesity,
hormones, radiation therapy, and even reproductive factors [1]. The diagnosis
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 546–558, 2021.
https://doi.org/10.1007/978-3-030-68154-8_48
Prediction of Breast Cancer and Its Recurrence 547
of breast cancer is carried out by classifying the tumor. Tumors can be either
benign or malignant. In a benign tumor, the cells grow abnormally and form
a lump and do not spread to other parts of the body [5]. Malignant tumors
are more harmful than benign. Unfortunately, not all physicians are experts in
distinguishing between benign and malignant tumors, while the classification of
tumor cells may take up to two days [3].
Information and Communication Technologies (ICT) can play potential roles
in cancer care. For example, data mining approaches applied to medical science
topics rise rapidly due to their high performance in predicting outcomes, reduc-
ing costs of medicine, promoting patients’ health, improving healthcare value
and quality, and making real-time decisions to save people’s lives [2]. Again,
Machine Learning (ML) algorithms gain insight from labeled sample and make
prediction for unknown samples; which greatly being used in health informat-
ics [11,14–16], predicting autism [24], and the likes. Again, ML becomes very
impactful in several fields like diseases diagnosis, disease prediction, biomedical,
and other engineering fields [25,26]. It is an application of artificial intelligence
(AI) that utilizes the creation and evaluation of algorithms that facilitate pre-
diction, pattern recognition, and classification [9]. Machine learning algorithms
like Support Vector Machine (SVM), Random Forest (RF), Naive Bayes (NB),
and k Nearest Neighbours (KNN) are frequently used in medical applications
such as the detection of the type of cancerous cells. The performance of each
algorithm is generally varied in terms of accuracy, sensitivity, specificity, preci-
sion, and recall for predicting the possibility of being affected by diseases like
diabetes, breast cancer, etc.
Over the last few years, a number of studies have been conducted on the
prediction of breast cancer. Those studies have focused on either the existence
of the cancer or the recurrence of the cancer. Therefore, this study has the
following objectives: first, explore the performance of different classifier models
considering different sets of data having and equal number of data objects for
predicting breast cancer and its recurrence. Second, to analyze the performance
of the predictors, while considering different sets of attributes.
This paper is organized into five separate sections as follows. Section 2 high-
lights the published literature focusing on breast cancer prediction models using
data mining techniques. Section 3, explains the detailed description of data,
various prediction algorithms, and measures their performance. The prediction
results of all the classification and regression algorithms along with the accu-
racy, sensitivity, and specificity are presented in Sect. 4. Section 5 concludes with
a summary of results and future directions.
2 Literature Review
Machine learning algorithms have been used for a long time in the case of pre-
dicting and diagnosing breast cancer. Several ML related studies were conducted
using breast cancer wisconsin (Original) dataset. For example, Asri et al. [2]
compared four different algorithms namely, SVM, Decision tree, Naive Bayes
548 A. I. Aishwarja et al.
and KNN to find out the best-performed classifiers for breast cancer prediction.
As outcome, SVM shows the best accuracy of 97.13%. Karabatak and Ince [13]
proposed an automatic diagnosis system for detecting breast cancer using Associ-
ation Rules (AR) and Neural Network (NN) and then compared with the Neural
Network Model. In the test stage, they used 3 fold cross-validation method and
the classification rate was 95.6%. Islam et al. [12] proposed a model based on
SVM and KNN then compare with other existing SVM-based models namely ST-
SVM, LPS-SVM, LP-SVM, LSVM, SSVM, NSVM; and found that their model
shows the highest accuracy in term of accuracy, sensitivity, specificity. In addi-
tion to this dataset, Huang et al. [10] used another dataset having 102294 data
samples with 117 different attributes to compare the performance of SVM clas-
sifiers and SVM ensembles. They constructed the SVM classifiers using kernel
functions (i.e., linear, polynomial, and RBF) and SVM ensembles using bagging
and boosting algorithms. They found that for both datasets SVM ensembles
performed better than classifiers. Khourdif and Bahaj [18] implemented Ran-
dom Forest, Naive Bayes, SVM, KNN, Multilayer Perceptron (MLP) in WEKA
to select the most effective algorithm with and without the Fast Correlation-
Based Feature selection (FCBF). This method was used to filter irrelevant and
redundant characteristics to improve the quality of cancer classification. The
experimental results showed that SVM provided the highest accuracy of 97.9%
without FCBF. Khourdif and Bahaj [17] used several machine-learning algo-
rithms including Random Forest, Naive Bayes, Support Vector Machines, and
K-Nearest Neighbors for determining the best algorithm for the diagnosis and
prediction of breast cancer using WEKA tools. Among 699 cases they have taken
30 attributes with 569 cases and found the SVM as the most accurate classifier
with accuracy of 97.9%.
Again, to classify recurrent or non-recurrent cases of breast cancer Ojha
and Goel et al. [27] used the Wisconsin Prognostic Breast Cancer dataset (194
cases with 32 attributes). Four clustering algorithms K-means, Expectation-
Maximization(EM), Partitioning Around Medoids(PAM) and Fuzzy c-means,
and four classification algorithms (SVM, C5.0, Naive Bayes, and KNN) were
used. The study showed SVM and C5.0 achieved the best 81% accuracy; and
the classification algorithms are better predictors than the clustering algorithms.
Li et al. [20] detected the risk factors of breast cancer and multiple common risk
factors adopting the Implementing Association Rule (IAR) algorithm and N-IAR
algorithm (for n-item) using a dataset of Chinese women having 83 attributes
and 2966 samples. Experimental results showed that the model based on the
ML algorithm is more suitable than the classic Gail model. Stark et al. [25]
proposed a new model with Logistic Regression, Gaussian Naive Bayes, Deci-
sion Tree, Linear Discriminant Analysis, SVM, and Artificial Neural Network
(ANN) and trained using a dataset of 78,215 women (50–78 age). These mod-
els predicted better than previously used BCRAT (Gail model). Delen et al. [7]
applied Artificial Neural Network(ANN), Decision Tree, Logistic Regression to
predict the survivability rate of breast cancer patients in WEKA using the SEER
Prediction of Breast Cancer and Its Recurrence 549
3 Methodology
At the first stage, four datasets were selected from the UCI machine learn-
ing repository [8]. “Breast Cancer Wisconsin (Original) dataset” as dataset 1
having 11 attributes and 699 instances and “Breast Cancer Wisconsin (Diag-
nostic) dataset” as dataset 2 having 32 attributes and 569 instances were used
Prediction of Breast Cancer and Its Recurrence 551
for predicting the breast cancer. Both of these datasets include two classes,
namely Benign (B) and Malignant (M). As those datasets have different sets
of attributes and a different number of instances, the least number of instances
569 were considered also from the dataset 1. Similarly “Breast Cancer Wiscon-
sin (Prognostic) dataset” was used as dataset 3 and “Breast Cancer Data Set”
was used as dataset 4 for predicting the recurrence of breast cancer. Dataset 3
has 33 attributes and 198 instances while dataset 4 has 10 attributes and
286 instances. Dataset 3 and dataset 4 include two classes, namely recurrence
and non-recurrence. There are 30 common attributes between dataset 2 and
dataset 3. However, some instances were deleted from dataset 4 due to noisy
data and considered 180 instances. Similarly, 180 instances were considered also
from dataset 3 (see Table 2).
a test set, since ML models may be biased towards a particular class, if there
doesn’t exist an equal number of class instances in the training data [19]. To
solve this problem Synthetic Minority Over-sampling Technique (SMOTE) [6]
was used, which is capable to remove class imbalance issues.
3.3.2 KNN
Among the supervised machine learning algorithms, K-nearest neighbors (KNN)
is one of the most effective techniques. It performs classification on certain data
points [23]. The KNN algorithm is a type of supervised ML algorithm that can
be used for both classifications as well as regression predictive problems. It uses
‘attribute similarity’ to predict the values of new data-points and then the new
data point will be assigned a value based on how closely it matches the points
in the training set. This algorithm has been applied based on different sets of
Prediction of Breast Cancer and Its Recurrence 553
attributes of different datasets and the results are shown in Fig. 3. The high-
est accuracy was acquired by dataset 1 (for benign-malignant class attribute)
and dataset 3 (for recurrence-non recurrence class attribute). Moreover, the best
performance was obtained from dataset 1 as the outcome of accuracy, precision,
sensitivity, and specificity were 95.9, 96, 98.5, and 93.42 respectively.
3.3.3 SVM
The SVMs are a set of related supervised learning methods that analyze data
to recognize patterns and are used for classification and regression analysis [22].
SVM is an algorithm that attempts to find a linear separator (hyper-plane)
between the data points of two classes in multidimensional space. The result
of adopting SVM on predefined classes of attributes is shown in Fig. 4. The
highest accuracy was found for dataset 2 (for benign-malignant class attribute)
and dataset 4 (for recurrence-non recurrence class attribute). But overall best
performance was observed for dataset 2 as the outcome of accuracy, precision,
sensitivity, and specificity were 97.2, 97.3, 99.07, and 95.2 respectively.
4 Discussion
The analysis and result discussed in Sect. 3 highlight the diverse performance of
the attributes used in different datasets, keeping the same number of data and
by applying different machine learning classification algorithms.
The summary of this finding is shown in Table 3. Here, dataset 1 has 11
attributes and dataset 2 has 32 attributes, but both of them have 569 levels of
instances. Different levels of performance were observed as shown in Fig. 6 while
they have differences only in the number of attributes. However, the highest accu-
racy for dataset 1 and dataset 2 was obtained by KNN and SVM respectively.
Similarly, for predicting the recurrence-non recurrence of breast cancer, dataset 3
and dataset 4 showed different performance as shown in Fig. 6. Dataset 3 con-
sists of 33 levels of attributes and dataset 4 consists of 10 attributes. Similarly,
the different levels of performance were observed due to the different sets of
attributes, while the highest accuracy for dataset 3 and dataset 2 was acquired
by KNN and Random Forest respectively.
Prediction of Breast Cancer and Its Recurrence 555
Again, the study results indicated that dataset 2 having 32 attributes (for
benign-malignant cancer) and dataset 3 having 33 attributes (for recurrence-non
recurrence) showed better performance for different algorithms. Hence, it can
be said that attributes used in dataset 2 are better for benign-malignant breast
cancer prediction while the attributes used in dataset 3 are better for recurrence-
non recurrence of breast cancer prediction. Thus the attributes of dataset 2 are
observed to be the best for predicting benign-malignant breast cancer prediction.
Similarly, the attributes of dataset 3 are the best for predicting the recurrence
of breast cancer.
Furthermore, KNN and SVM performed best for dataset 1 and dataset 2
respectively. In this case, the difference in accuracy gained by KNN and SVM was
below 2% while the attribute of dataset 1 and dataset 2 is 11 and 32 respectively.
Similarly, KNN and Random Forest performed best for dataset 3 and dataset 4
respectively. Here, the difference of accuracy was below 3% while the attributes
of dataset 3 and dataset 4 are 33 and 10 respectively. In both cases, increasing
the number of attributes doesn’t make a huge difference while the collection of
attributes increases diagnosis or medical cost. From a cost-effective point of view,
in both cases less number of attributes can be used, 11 attributes (dataset 1)
can be used for predicting breast cancer adopting the KNN. On the other hand,
10 attributes (dataset 4) can be used for predicting recurrence of breast cancer
adopting the Random Forest algorithm.
556 A. I. Aishwarja et al.
5 Conclusion
In this research, different datasets with various attributes were implemented
using four algorithms. Aiming to find which attributes and algorithms tend to
be more effective.
Machine learning approaches have been increasing rapidly in the medical
field due to their monumental performance in predicting and classifying disease.
Research on which algorithm and attributes perform better in breast cancer
prediction has been done before but the reasons for their performing better were
not explored. In this research, considering the performance of algorithms, the
best accuracy was observed by KNN in breast cancer prediction and SVM in
the recurrence of breast cancer. This result indicated that while detecting breast
cancer, a dataset with diverse attributes tends to be more accurate in prediction.
One of the limitations of this work is, this study considers datasets having
limited number of instances. In future, larger datasets can be considered for
exploring the different insight. Besides, more algorithms can be implemented to
validate and generalize the outcomes of this research.
References
1. Aaltonen, L.A., Salovaara, R., Kristo, P., Canzian, F., Hemminki, A., Peltomäki,
P., Chadwick, R.B., Kääriäinen, H., Eskelinen, M., Järvinen, H., et al.: Incidence of
hereditary nonpolyposis colorectal cancer and the feasibility of molecular screening
for the disease. N. Engl. J. Med. 338(21), 1481–1487 (1998)
2. Asri, H., Mousannif, H., Al Moatassime, H., Noel, T.: Using machine learning
algorithms for breast cancer risk prediction and diagnosis. Procedia Comput. Sci.
83, 1064–1069 (2016)
3. Bharat, A., Pooja, N., Reddy, R.A.: Using machine learning algorithms for breast
cancer risk prediction and diagnosis. In: 2018 3rd International Conference on
Circuits, Control, Communication and Computing (I4C), pp. 1–4. IEEE (2018)
4. Chaurasia, V., Pal, S.: Data mining techniques: to predict and resolve breast cancer
survivability. Int. J. Comput. Sci. Mob. Comput. IJCSMC 3(1), 10–22 (2014)
5. Chaurasia, V., Pal, S., Tiwari, B.: Prediction of benign and malignant breast cancer
using data mining techniques. J. Algorithms Comput. Technol. 12(2), 119–126
(2018)
6. Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: Smote: synthetic
minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002)
7. Delen, D., Walker, G., Kadam, A.: Predicting breast cancer survivability: a com-
parison of three data mining methods. Artif. Intell. Med. 34(2), 113–127 (2005)
8. Frank, A., Asuncion, A., et al.: UCI machine learning repository (2010), 15, 22
(2011). http://archive.ics.uci.edu/ml
9. Gokhale, S.: Ultrasound characterization of breast masses. Indian J. Radiol. Imag-
ing 19(3), 242 (2009)
10. Huang, M.W., Chen, C.W., Lin, W.C., Ke, S.W., Tsai, C.F.: SVM and SVM
ensembles in breast cancer prediction. PLoS ONE 12(1), e0161501 (2017)
11. Inan, T.T., Samia, M.B.R., Tulin, I.T., Islam, M.N.: A decision support model to
predict ICU readmission through data mining approach. In: Pacific ASIA Confer-
ence on Information Systems (PACIS), p. 218 (2018)
Prediction of Breast Cancer and Its Recurrence 557
12. Islam, M.M., Iqbal, H., Haque, M.R., Hasan, M.K.: Prediction of breast cancer
using support vector machine and k-nearest neighbors. In: 2017 IEEE Region 10
Humanitarian Technology Conference (R10-HTC), pp. 226–229. IEEE (2017)
13. Karabatak, M., Ince, M.C.: An expert system for detection of breast cancer based
on association rules and neural network. Expert Syst. Appl. 36(2), 3465–3469
(2009)
14. Khan, N.S., Muaz, M.H., Kabir, A., Islam, M.N.: Diabetes predicting mHealth
application using machine learning. In: 2017 IEEE International WIE Conference
on Electrical and Computer Engineering (WIECON-ECE), pp. 237–240. IEEE
(2017)
15. Khan, N.S., Muaz, M.H., Kabir, A., Islam, M.N.: A machine learning-based intel-
ligent system for predicting diabetes. Int. J. Big Data Anal. Healthcare (IJBDAH)
4(2), 1–20 (2019)
16. Khan, N.I., Mahmud, T., Islam, M.N., Mustafina, S.N.: Prediction of cesarean
childbirth using ensemble machine learning methods. In: 22nd International Con-
ference on Information Integration and Web-Based Applications Services (IIWAS
2020) (2020)
17. Khourdifi, Y., Bahaj, M.: Applying best machine learning algorithms for breast
cancer prediction and classification. In: 2018 International Conference on Elec-
tronics, Control, Optimization and Computer Science (ICECOCS), pp. 1–5. IEEE
(2018)
18. Khourdifi, Y., Bahaj, M.: Feature selection with fast correlation-based filter for
breast cancer prediction and classification learning learning algorithms. In: 2018
International Symposium on Advanced Electrical and Communication Technolo-
gies (ISAECT), pp. 1–6. IEEE (2018)
19. Kotsiantis, S., Kanellopoulos, D., Pintelas, P.: Handling imbalanced datasets: a
review. Gests Int. Trans. Comput. Sci. Eng. 30, 25–36 (2006). Synthetic Oversam-
pling of Instances Using Clustering
20. Li, A., Liu, L., Ullah, A., Wang, R., Ma, J., Huang, R., Yu, Z., Ning, H.: Association
rule-based breast cancer prevention and control system. IEEE Trans. Comput. Soc.
Syst. 6(5), 1106–1114 (2019)
21. Liu, Y.Q., Wang, C., Zhang, L.: Decision tree based predictive models for breast
cancer survivability on imbalanced data. In: 2009 3rd International Conference on
Bioinformatics and Biomedical Engineering, pp. 1–4. IEEE (2009)
22. Mangasarian, O.L., Musicant, D.R.: Lagrangian support vector machines. J. Mach.
Learn. Res1, 161–177 (2001)
23. Miah, Y., Prima, C.N.E., Seema, S.J., Mahmud, M., Kaiser, M.S.: Performance
comparison of machine learning techniques in identifying dementia from open
access clinical datasets. In: Advances on Smart and Soft Computing, pp. 79–89.
Springer (2020)
24. Omar, K.S., Mondal, P., Khan, N.S., Rizvi, M.R.K., Islam, M.N.: A machine learn-
ing approach to predict autism spectrum disorder. In: 2019 International Confer-
ence on Electrical, Computer and Communication Engineering (ECCE), pp. 1–6.
IEEE (2019)
25. Stark, G.F., Hart, G.R., Nartowt, B.J., Deng, J.: Predicting breast cancer risk using
personal health data and machine learning models. PLoS ONE 14(12), e0226765
(2019)
558 A. I. Aishwarja et al.
26. Vasant, P., Zelinka, I., Weber, G.W. (eds.): Intelligent Computing and Optimiza-
tion. Springer International Publishing (2020). https://doi.org/10.1007/978-3-030-
33585-4
27. Yarabarla, M.S., Ravi, L.K., Sivasangari, A.: Breast cancer prediction via machine
learning. In: 2019 3rd International Conference on Trends in Electronics and Infor-
matics (ICOEI), pp. 121–124. IEEE (2019)
Exploring the Machine Learning
Algorithms to Find the Best Features
for Predicting the Risk of Cardiovascular
Diseases
1 Introduction
Cardiovascular diseases have been one of the note-worthy reasons for mortality
all over the world. According to the reports of WHO, every year more than 18
million people die because of cardiovascular diseases which covers almost 31%
of global death [23]. Damages in parts or all of the heart, coronary artery, or
inadequate supply of nutrients and oxygen to this organ result in cardiovascu-
lar disease. Several lifestyle choices can increase the risk of heart disease that
include, for example, high blood pressure and cholesterol, smoking, overweight
and obesity, and diabetes.
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 559–569, 2021.
https://doi.org/10.1007/978-3-030-68154-8_49
560 M. M. Jalal et al.
2 Literature Review
Several studies have been conducted using the UCI machine learning repository
heart disease dataset to classify the presence of cardiovascular disease in the
human body. This multivariate dataset involves 303 instances along with 75 fea-
tures. For example, Rajyalakshmi et al. [22] explored different machine learning
algorithms including Decision Tree (DT), Support Vector Machine (SVM), Ran-
dom Forest (RF), and a hybrid model named HRFLM that combines Random
Forest and Linear method for predicting cardiovascular diseases. The hybrid
model showed the best accuracy among the implemented algorithms. Similarly,
Sen et al. [25] implemented the Naı̈ve Bayes, SVM, Decision Tree, and KNN
algorithms and found that SVM shows the best accuracy (83%). Again, Dinesh
et al. [6] predicted the possibility of cardiovascular diseases using six different
algorithms, that includes Logistic Regression, Random Forest, SVM, Gradient
Boosting, and an ensemble model, while Logistic regression showed the best accu-
racy (91%). In another study, Maji et al. [17] proposed the use of hybridization
Predicting the Risk of Cardiovascular Diseases 561
techniques to reduce the test cases to predict the outcome using ANN, C4.5,
and hybrid Decision Tree algorithms. The result also showed that the hybrid
Decision tree performed better in terms of accuracy.
Prediction of cardiovascular diseases has also been conducted using differ-
ent datasets to bring out the best accuracy like UCI Statlog Dataset having
270 instances along with 13 features. Dwivedi et al. [7] proposed an auto-
matic medicinal system using advanced data mining techniques like the mul-
tilayer perceptron model and applied several classification techniques like Naı̈ve
Bayes, SVM, logistic Regression, and KNN. Another study conducted by Georga
et al. [9] explained AI methods(Random Forest, Logistic Regression, FRS, GAM,
GBT algorithms) to find out the most effective predictors to detect Coronary
Heart Disease (CAD). In this study, a more coherent and clustered dataset along
with hybridization were thought to be incorporated to find out the best result.
UCI Cleveland data has also been used in classification to predict the pres-
ence of cardiovascular disease in the human body in many studies. For example,
Pouriyeh et al. [21] investigated and compared the performance of the accu-
racy of different classification algorithms including Decision Tree, Naı̈ve Bayes,
SVM, MLP, KNN, Single Conjugative Rule Learner, and Radial Basis Function.
Here, the hybrid model of SVM and MLP produced the best result. Latha [16]
suggested a comparative analytical approach to determine the performance accu-
racy of ensemble techniques and found that the ensemble was a good strategy
to improve accuracy. Again, Amin et al. [2] identified significant features and
improve the prediction accuracy by using different algorithms including, KNN,
Decision Tree, Naı̈ve Bayes, Logistic Regression, SVM, Neural Network, and an
ensemble model that combines Naı̈ve Bayes and Logistic regression. It used the
most impactful 9 features to calculate the prediction and the Naı̈ve Bayes gave
the best accuracy.
Alaa et al. [1] used UK Biobank data to develop an algorithmic tool using
AutoPrognosis that automatically selects features and tune ensemble models.
Five ML algorithms were used in this study, namely, SVM, Random Forest,
Neural Network, AdaBoost, and Gradient Boosting. The AutoPrognosis showed
a higher AUC-ROC compared to all other standard ML models.
In sum, the literature review showed that most of the studies focusing on car-
diovascular diseases and ML were striving to explore the algorithm that showed
the best performance in predicting the possibility of cardiovascular diseases. The
summary of the literature review is shown in Table 1. Again, though different
studies are conducted using different sets of datasets having a different number
of features, no study has been conducted to explore how the performance accu-
racy of different algorithms are varying due to different set of features used in
different datasets. Thus, this study focus on this issue.
3 Methodology
This section describes the overall working procedure used to obtain the research
objectives. The overall methodology of this study is shown in Fig. 1
562 M. M. Jalal et al.
Table 1. (continued)
At first, we acquired data from various sources namely UCI machine learning
repository [8] and Kaggle. Four different datasets were used in this research
having different number of instances and the features of these datasets are not
similar. As Dataset 1, we have used the “Kaggle Cardiovascular Disease Dataset”
while having 13 features and 70000 instances. The “Cleveland dataset” from
kaggle having 14 attribues and 303 instances as Dataset 2. Likewise, we used
“Kaggle Integrated Heart Disease dataset” as Dataset 3 that contains 76 features
and 899 instances. As Dataset 4, we used “UCI Framingham Dataset” having
14 features and 4240 instances. However, we have considered 11 and 14 features
for computational purpose from dataset 3 and dataset 4 respectively. There
are 8 common features in each of the datasets.
564 M. M. Jalal et al.
The objectives of the analysis are to find the optimum set of features and an
algorithm for the prediction of cardiovascular diseases. For this, different machine
learning algorithms which used mostly for predicting cardiovascular disease were
chosen that includes, Logistic Regression, Support Vector Machine, K-th Nearest
Neighbor, Naı̈ve Bayes, Random Forest and Gradient Boosting. These algorithms
were applied on the selected four datasets by splitting the dataset into 70% as
training set and 30% as testing set.
4 Discussion
The study result highlights that with different dataset al.ong with different fea-
tures show diversity in the result. Datasets used here for computation purposes
consist of different numbers of features. As they have different features, this
could be one of the reasons of varying the prediction accuracy.
Again, different datasets show difference in accuracy as shown in Table 2.
For dataset 1, 2, 3 and 4 comparatively better accuracy was obtained using
Gradient Boosting; SVM, Naı̈ve Bayes and Random Forest; Random Forest;
Random Forest and Gradient Boosting respectively. The results thus indicate
that Random Forest shows the best accuracy in most of the datasets.
Each of the algorithm applied in each particular dataset, the best perfor-
mance was observed for dataset 3 having 11 features. Thus, the results indicated
that for predicting cardiovascular disease in the human body, features used in
the dataset 3 is most likely to be the best recommended attributes. The features
considered in Dataset 3 are Age, Sex, Chest Pain Type, Resting Blood Pressure,
Smoking Year, Fasting Blood Sugar, Diabetes History, Family History Coronary,
ECG, Pulse Rate and Presence of Disease.
5 Conclusion
In this research, six different ML algorithms were implemented on four different
datasets having different set of features to predict cardiovascular disease. The
result showed that 11 features of dataset 3 are the most efficient features while
the Random Forest showed the best accuracy for most of the datasets with
different set of features.
While the existing work have primarily focused on the implementation of
several algorithms in a particular dataset and then compared their performance;
This study demonstrated performance comparison among multiple datasets hav-
ing different set of features along with evaluating multiple machine learning
algorithms on them.
One of the limitations of this work is, this study considers only the traditional
and ensemble ML algorithms. Hybrid and different ensemble models could have
been developed for a different insight. In future, an app or tool can be developed
using ML algorithms to detect cardiovascular disease. Besides, algorithms can
also be applied on new datasets to validate and generalize the outcomes of this
research related to the best features to predict cardiovascular diseases.
References
1. Alaa, A.M., Bolton, T., Di Angelantonio, E., Rudd, J.H., van der Schaar, M.: Car-
diovascular disease risk prediction using automated machine learning: a prospective
study of 423,604 uk biobank participants. PLoS ONE 14(5), e0213653 (2019)
2. Amin, M.S., Chiam, Y.K., Varathan, K.D.: Identification of significant features
and data mining techniques in predicting heart disease. Telematics Inform. 36,
82–93 (2019)
568 M. M. Jalal et al.
3. Bhatt, A., Dubey, S.K., Bhatt, A.K.: Analytical study on cardiovascular health
issues prediction using decision model-based predictive analytic techniques. In:
Soft Computing: Theories and Applications, pp. 289–299. Springer (2018)
4. Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: Smote: synthetic
minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002)
5. Dangare, C.S., Apte, S.S.: Improved study of heart disease prediction system using
data mining classification techniques. Int. J. Comput. Appl. 47(10), 44–48 (2012)
6. Dinesh, K.G., Arumugaraj, K., Santhosh, K.D., Mareeswari, V.: Prediction of car-
diovascular disease using machine learning algorithms. In: 2018 International Con-
ference on Current Trends towards Converging Technologies (ICCTCT), pp. 1–7.
IEEE (2018)
7. Dwivedi, A.K.: Performance evaluation of different machine learning techniques
for prediction of heart disease. Neural Comput. Appl. 29(10), 685–693 (2018)
8. Frank, A., Asuncion, A., et al.: UCI machine learning repository, 2010, vol. 15, p.
22 (2011). http://archive.ics.uci.edu/ml
9. Georga, E.I., Tachos, N.S., Sakellarios, A.I., Kigka, V.I., Exarchos, T.P., Pelosi,
G., Parodi, O., Michalis, L.K., Fotiadis, D.I.: Artificial intelligence and data min-
ing methods for cardiovascular risk prediction. In: Cardiovascular Computing–
Methodologies and Clinical Applications, pp. 279–301. Springer (2019)
10. Inan, T.T., Samia, M.B.R., Tulin, I.T., Islam, M.N.: A decision support model to
predict ICU readmission through data mining approach. In: PACIS, p. 218 (2018)
11. Khan, N.S., Muaz, M.H., Kabir, A., Islam, M.N.: Diabetes predicting mhealth
application using machine learning. In: 2017 IEEE International WIE Conference
on Electrical and Computer Engineering (WIECON-ECE), pp. 237–240. IEEE
(2017)
12. Khan, N.S., Muaz, M.H., Kabir, A., Islam, M.N.: A machine learning-based intel-
ligent system for predicting diabetes. Int. J. Big Data Anal. Healthcare (IJBDAH)
4(2), 1–20 (2019)
13. Kotsiantis, S., Kanellopoulos, D., Pintelas, P.: Handling imbalanced datasets: a
review, gests international transactions on computer science and engineering. Syn-
thetic Oversampling Instances Using Clustering 30, 25–36 (2006)
14. Krak, I., Barmak, O., Manziuk, E., Kulias, A.: Data classification based on the
features reduction and piecewise linear separation. In: International Conference on
Intelligent Computing & Optimization, pp. 282–289. Springer (2019)
15. Kumar, G.R., Ramachandra, G., Nagamani, K.: An efficient prediction of breast
cancer data using data mining techniques. Int. J. Innov. Eng. Technol. (IJIET)
2(4), 139 (2013)
16. Latha, C.B.C., Jeeva, S.C.: Improving the accuracy of prediction of heart dis-
ease risk based on ensemble classification techniques. Informat. Med. Unlocked 16,
100203 (2019)
17. Maji, S., Arora, S.: Decision tree algorithms for prediction of heart disease. In:
Information and Communication Technology for Competitive Strategies, pp. 447–
454. Springer (2019)
18. Miah, Y., Prima, C.N.E., Seema, S.J., Mahmud, M., Kaiser, M.S.: Performance
comparison of machine learning techniques in identifying dementia from open
access clinical datasets. In: Advances on Smart and Soft Computing, pp. 79–89.
Springer (2020)
19. Omar, K.S., Mondal, P., Khan, N.S., Rizvi, M.R.K., Islam, M.N.: A machine learn-
ing approach to predict autism spectrum disorder. In: 2019 International Confer-
ence on Electrical, Computer and Communication Engineering (ECCE), pp. 1–6.
IEEE (2019)
Predicting the Risk of Cardiovascular Diseases 569
20. Patel, J., TejalUpadhyay, D., Patel, S.: Heart disease prediction using machine
learning and data mining technique. Heart Disease 7(1), 129–137 (2015)
21. Pouriyeh, S., Vahid, S., Sannino, G., De Pietro, G., Arabnia, H., Gutierrez, J.: A
comprehensive investigation and comparison of machine learning techniques in the
domain of heart disease. In: 2017 IEEE Symposium on Computers and Communi-
cations (ISCC), pp. 204–207. IEEE (2017)
22. Rajyalakshmi, P., Reddy, G.S., Priyanka, K.G., Sai, V.L.B.S., Anveshini, D.: Pre-
diction of cardiovascular disease using machine learning. Entropy 23, 24
23. Ramalingam, V., Dandapath, A., Raja, M.K.: Heart disease prediction using
machine learning techniques: a survey. Int. J. Eng. Technol. 7(2.8), 684–687 (2018)
24. Rani, K.U.: Analysis of heart diseases dataset using neural network approach.
arXiv preprint arXiv:1110.2626 (2011)
25. Sen, S.K.: Predicting and diagnosing of heart disease using machine learning algo-
rithms. Int. J. Eng. Comput. Sci 6(6) (2017)
26. Soni, J., Ansari, U., Sharma, D., Soni, S.: Predictive data mining for medical
diagnosis: an overview of heart disease prediction. Int. J. Comput. Appl. 17(8),
43–48 (2011)
27. Tahmooresi, M., Afshar, A., Rad, B.B., Nowshath, K., Bamiah, M.: Early detec-
tion of breast cancer using machine learning techniques. J. Telecommun. Electr.
Comput. Eng. (JTEC) 10(3–2), 21–27 (2018)
28. Vasant, P., Zelinka, I., Weber, G.W. (eds.): Intelligent Computing and Optimiza-
tion. Springer International Publishing (2020). https://doi.org/10.1007/978-3-030-
33585-4, https://doi.org/10.1007%2F978-3-030-33585-4
Searching Process Using Boyer Moore
Algorithm in Digital Library
Abstract. Reading helps to learn more vocabularies and their usages. People
should read books to get knowledge. Nowadays, digital libraries are used to
search for various books. If the internet connection is available, digital libraries
can be accessed easily via computers, laptops, and mobile phones. Students and
readers can search for the desired books from this digital library by typing sub-
strings of a book title that matches with parts of the original book title. To be
able to make such a searching process effectively, string searching algorithms
can be applied. There are many sophisticated and efficient algorithms for string
searching. In this paper, a Boyer Moore (BM) string-searching algorithm is
applied to search the desired books effectively in the digital library. And then the
BM algorithm is presented by comparing it with Horspool’s algorithm to prove
the performance efficiency of the BM algorithm for the searching process.
1 Introduction
Reading helps students and readers to gain knowledge. Users can use digital libraries to
get knowledge or information by reading educational or non-educational books. Most
frequently users find or track via computers or mobile phones the desired books that are
available in the library. The effectiveness of libraries is based on the results of how they
make users easier in finding books. A good computerized library system is to help users
for easy and effective use. To achieve this, string searching algorithms are required.
String searching sometimes called string matching is the act of checking the existence
of a substring of m characters called the pattern in a string of n characters called the text
(where m n) and finding its location in that text [1–3].
String searching algorithms are the basic components of existing applications for
text processing, intrusion detection, information retrieval, and computational biology
[4]. Better string-matching algorithms can improve the applications’ efficiencies more
effectively. So, a fast string matching algorithm is an important area of research. There
are several more sophisticated and more efficient algorithms for string searching. The
most widely known of them is the BM algorithm suggested by R. Boyer and J. Moore.
The BM algorithm is the most efficient string-matching algorithm because a lot of
character comparisons are ignored during the process of searching. It performs the
pattern’s character comparisons in reverse order from right to left and, in the case of
mismatched, the entire pattern to be searched is not required to compare [5, 6]. This
paper aims to recommend an efficient BM algorithm for the string processing of digital
libraries.
The remainder of this paper is organized as follows. Section 2 presents related
work focusing on using the BM algorithm for various related applications. Section 3
explains a background theory about the BM algorithm and the Horspool’s algorithm in
detail. Section 4 highlights the searching processes of BM and Horspool algorithms
and comparative analysis between them. Finally, the conclusion of this paper is
described in Sect. 5.
2 Related Work
The authors [7] developed a detection method to detect web application vulnerabilities
by using the BM string-matching algorithm. Their proposed method performed well in
terms of the ability to accurately detect vulnerabilities based on false-negative and have
no false positive with low processing time.
Using the BM algorithm as a web-based search, the authors [8] implemented a
dictionary application for midwifery. The authors [9] applied the BM algorithm for the
application of a baby name dictionary based on Android. The authors [10] proposed a
new algorithm as a regular expression pattern matching algorithm. This new algorithm
was derived from the BM algorithm, and it was an efficient generalized Boyer-Moore-
type pattern matching algorithm for regular languages.
The authors [11] studied different kinds of string matching algorithms and observed
their time and space complexities. The performance of these algorithms is tested with
biological sequences such as DNA and Proteins. Based on their study, the BM algo-
rithm is extremely fast for large sequences, it avoids lots of needless comparisons by
pattern relative to the text, and its best-case running complexity is sub-linear.
The researcher [12] analyzed the efficiency of four string-matching algorithms
based on different pattern lengths and pattern placement. The researcher showed that
among the four algorithms, the Horspool algorithm which is a simplified version of the
BM algorithm is the fastest algorithm without considering the pattern length and
pattern placement.
3 Background Theory
In this section, the BM algorithm and the Horspool’s algorithm are presented in detail.
by moving from right to left [1, 2, 5, 15]. It determines the size of the shift by using a
bad-symbol shift table and a good-suffix shift table.
txtðcÞ r: ð1Þ
where txt(c) is an entry in the bad-symbol shift table, and r is the number of matched
characters.
This table is a list of possible characters that can be found in the text. Texts may
contain punctuation marks, space, and other special characters. The txt(c) is computed
by the formula:
8
>
>
Length m of the pattern P; if character c in the text is not among
>
< P's first ðm - 1Þ characters;
txtðcÞ ¼ ð2Þ
>
> Distance from the rightmost character c amidst P's first ðm - 1Þ
>
:
characters to its final character; otherwise:
Let us build a bad-symbol shift table for searching a pattern PKSPRS in some text.
It is shown in Table 1, all entries of the table are equal to 6, except for K, P, R, and S,
which are 4, 2, 1, and 3, respectively.
An example of shifting the pattern using the above Table 1 is shown in Fig. 1. In
this example, the pattern PKSPRS is searched in some text. Before the failure of
comparison on the text’s character Z, the comparisons of the last two characters were
matched. So, the pattern can be moved to the right 4 positions because the shift size is
txt(Z) − 2 = 6 − 2 = 4.
P K S P R S
P K S P R S
Bad-symbol shift dist1 is calculated by txt(c) − r if this value is greater than zero
and by 1 if it is less than or equal to zero. This is described as a formula:
corresponding dist2 from the good-suffix shift table. Move P to the right according
to the number of positions calculated by the formula:
dist1 if r ¼ 0;
dist ¼ wheredist1 ¼ maxftxtðcÞ r; 1g: ð4Þ
maxfdist1 ; dist2 g if r [ 0;
In this section, the searching process for the BM algorithm is illustrated by comparing
it with Horspool’s algorithm which is a simplified version of the BM algorithm.
Consider the problem of searching for the desired book in a digital library. A book
title is represented by a text that comprises English letters and spaces, and the book title
or the segment of the book title is the pattern.
Text: VIP_AZWIN_ ZAWZAZA
Pattern: ZAWZAZ
Searching Process Using Boyer Moore Algorithm in Digital Library 575
V I P _ A Z W I N _ Z A W Z A Z A
Z A W Z A Z
V I P _ A Z W I N _ Z A W Z A Z A
Z A W Z A Z
Z A W Z A Z
As shown in Fig. 2, first, the pattern string is aligned with the starting characters of
the text string.
As shown in Fig. 3, after matching with two pairs of characters, the pattern’s
character ‘Z’ fails to match its partner ‘_’ in the text. So, the algorithm retrieves txt(_) =
6 from bad symbol table to compute dist1 = txt(_) – 2 = 4 and also retrieves dist2 = 5
from good-suffix shift table. And then the pattern is moved to the right by max {dist1,
dist2} = max {4, 5} = 5.
576 L. L. Lin and M. T. Soe
V I P _ A Z W I N _ Z A W Z A Z A
Z A W Z A Z
Z A W Z A Z
As shown in Fig. 4, after matching one pair of Z’s and failing the next comparison
on the text’s space character, the algorithm fetches txt(_) = 6 from bad-symbol table
to compute dist1 = 6 − 1 = 5 and also fetches dist2 = 2 from good-suffix shift table.
And then the pattern is moved to the right by max {dist1, dist2} = max {5, 2} = 5.
Lastly, after matching all the pattern’s characters with their partners in the text, a
matching substring is found in the text string. Here, the total number of character
comparisons is 11.
V I P _ A Z W I N _ Z A W Z A Z A
Z A W Z A Z
Z A W Z A Z
V I P _ A Z W I N _ Z A W Z A Z A
Z A W Z A Z
Z A W Z A Z
V I P _ A Z W I N _ Z A W Z A Z A
Z A W Z A Z
Z A W Z A Z
In the next step, as shown in Fig. 7, after failing the second comparison on the
character ‘W’ in the text, the algorithm fetches txt(Z) = 2 from the bad-symbol shift
table and shifts the pattern by 2 positions to the right along with the text.
Finally, after matching all the pattern’s characters with their partners in the text, a
matched substring is found in the text. Here, the total number of character comparisons
is 12.
300
250
200
150
100
50
0
3 5 7 9 11 13 15
Pattern Lengths
Fig. 8. Comparison between BM and Horspool algorithms based on the number of character
comparisons
578 L. L. Lin and M. T. Soe
According to the results in Fig. 8, the BM algorithm produces less comparison time
than Horspool’s algorithm. So, if the BM algorithm is used in the searching process of
the digital library system, the performance efficiency of this library system will be
improved effectively. A further experiment about the searching time and the accuracy
of the searching process in the digital library will be carried out.
5 Conclusion
In the text processing areas, finding the specific string in the least time is the most basic
factor. String-searching algorithms play a critical role in such a manner. Most of the
other string-searching algorithms use the basic concepts of the BM algorithm because it
has the least time complexity. In this paper, the BM algorithm is used for the process of
finding the desired books from the digital library. The result also based on the number
of comparisons shows that the BM algorithm is more efficient than the Horspool’s
algorithm. Therefore, if the text processing applications, such as the digital library
system, use the BM algorithm, the searching time will be fast, and their performance
will be significantly increased.
References
1. Rawan, A.A.: An algorithm for string searching. In: International Journal of Computer
Applications, vol. 177, no: 10, pp. 0975–8887 (2019)
2. Anany, L.: Introduction to the Design and Analysis of Algorithms. Villanova University,
Philadelphia (2012)
3. Robert, S., Kevin, W.: Algorithms. Princeton University, Princeton (2011)
4. Bi, K., Gu, N.J., Tu, K., Liu, X.H., Liu, G.: A practical distributed string-matching algorithm
architecture and implementation. In: Proceedings of World Academy Of Science,
Engineering and Technology, vol. 1, no: 10, pp. 3261–3265 (2007)
5. Robert, S.B., Strother, M.J.: A fast string searching algorithm. Assoc. Comput. Mach., 20
(10), 762–772 (1977)
6. Abdulellah, A.A., Abdullah, H.A., Abdulatif, M.A.: Analysis of parallel Boyer-Moore string
search algorithm. Glob. J. Comput. Sci. Technol. Hardware Compu. 13, 43–47 (2013)
7. Ain, Z.M.S., Nur, A.R., Alya, G.B., Kamarularifin, A.J., Fakariah, H.M.A., Teh, F.A.R.: A
method for web application vulnerabilities detection by using Boyer-Moore string matching
algorithm. In: 3rd Information Systems International Conference, vol. 72, pp. 112–121
(2015)
8. Rizky, I.D., Anif, H.S., Arini, A.: Implementasi Algoritma Boyer Moore Pada Aplikasi
Kamus Istilah Kebidanan Berbasis Web. Query: J. Inf. Syst. 2, 53–62 (2018)
9. Ayu, P.S., Mesran, M.: Implementasi algoritma boyer moore pada aplikasi kamus nama bayi
beserta maknanya berbasis android. Pelita informatika: informasi dan informatika 17, 97–
101 (2018)
10. Bruce, W.W., Richard, E.W.: A Boyer-Moore-style algorithm for regular expression pattern
matching. Sci. Comput. Program. 48, 99–117 (2003)
Searching Process Using Boyer Moore Algorithm in Digital Library 579
11. Pandiselvam, P., Marimuthu, T, Lawrance. R.: A comparative study on string matching
algorithm of biological sequences. In: International Conference on Intelligent Computing
(2014)
12. DU, V.: A comparative analysis of various string-matching algorithms. In: 8th International
Research Conference, KDU (2015)
13. Robbi, R., Ansari, S.A., Ayu, P.A., Dicky, N.: Visual approach of searching process using
Boyer-Moore algorithm. In: Journal of Physics, vol. 930 (2017)
14. Mulyati, I.A.: Searching process using Bayer Moore algorithm in medical information
media. In: International Journal of Recent Technology and Engineering (IJRTE), vol.
8 (2019)
15. Michael, T.G., Roberto T.: Algorithm Design and Applications. John Wiley and Sons,
Hoboken (2015)
16. Abd, M.A., Zeki, A., Zamani, M., Chuprat, S., El-Qawasmeh, E. (eds.): Informatics
Engineering and Information Science. New York (2011)
17. Yi, C.L.: A survey of software-based string matching algorithms for forensic analysis. In:
Annual ADFSL Conference on Digital Forensics, Security and Law (2015)
Application of Machine Learning
and Artificial Intelligence Technology
Gender Classification from Inertial
Sensor-Based Gait Dataset
1 Introduction
Gender is one of the understandable and straightforward human information, yet that
can opens up the entryway to the collection of facts used in various pragmatic oper-
ations. In the process of gender determination, the individual’s gender is determined by
assessing the diverse particularities of femaleness and maleness [1]. Automatic human
gender categorization is an interesting subject in pattern recognition, since gender
includes very important and rich knowledge about the social activities of individuals
The remaining of the paper is prepared in different sections. The related work is
discussed in section two. The methodology of the study is described in section three.
Section four discusses the result of the experiment. The final section demonstrates the
conclusion and the future work.
2 Related Work
Many useful techniques or methods are used to find out the gender of a person by using
the gait data.
Kanij Mehtanin Khabir et al. [10] explored twelve types of time-domain features:
average, median, maximum, minimum, variance, root mean square, mean absolute
deviation, standard error of mean, standard deviation, skewness, kurtosis, vector sum
from the inertial sensor dataset. They measured 88 features which are suitable for
classification of the dataset and regression problems. SVM provides the highest
accuracy among other classifiers. This proposed model has some over fitting problems
because training set of the data with compare to test set has higher accuracy difference.
The statistical features such as global minimum, global maximum, step duration,
step length, mean, root mean square, standard deviation, entropy, energy and amplitude
from different components of accelerations and angular velocities [11]. They estimated
total 50 features for every single step. The variety of features was too small, so this
experiment only works as proof of concept for now.
Makihara et al. [12] describes gender identification by apply a Video-based gait
feature analysis with the help of a multi-view gait database. In this research, they did a
deep machine learning process to predict gender using their own created multi-view
gait database.
Tim Van hamme et al. [13] explored the best solution of gender information using
IMU Sensor-Based Gait Traces data. They compared distinctive component engi-
neering and machine learning algorithms, including both conventional and profound
machine learning techniques.
ThanhTrung Ngo et al. [14] organized a challenging competition on gender
detection using the OU-ISIR inertial sensor dataset. Several processing and feature
extraction are done by using deep learning method, conventional classification methods
and sensor orientation handling methods. The amounts of features are not enough to
generate real time usable model.
Ankita Jain et al. [15] used accelerometer and gyroscope sensor readings for gender
identification using Smartphone. They combined the collected data from accelerometer
and gyroscope sensor for betterment of the experimental performance. Bagging clas-
sifier gives the best accuracy in this experiment.
Rosa Andrie Asmara et al. [17] used Gait Energy Image (GEI) and Gait Infor-
mation Image (GII) processes for gender recognition. The Gait Information Image
(GII) method performed better than Gait Energy Image (GEI) using SVM. The accu-
racy of those works is too low because the shortage of features.
586 R. K. Pathan et al.
Jang-HeeYoo et al. [16] is used a sequential set of 2D stick facts to characterize the
gait signature. Each gait signature discriminated by the 2D sticks and joint angles of the
hip, knee and ankle. The support vector machine (SVM) is used to classify the gender
recognition of this method.
Another researcher take out 2D and 3D gait features founded on silhouette and
shape descriptors and combined them for gender classification [18]. This combined
feature gives the higher accuracy for the Dgait dataset. They used kernel SVM for this
experiment.
In [13, 14] they have used deep and shallow architecture for the classification of
gender. Deep learning consists of several representational levels, while the shallow has
few levels. In [15] they used behavioral biometric gait information for gender classi-
fication on smartphones. They collected data from 42 subjects. In [16] they extracted
motion based feature and joint angle feature from gait data. Their dataset is only limited
in medical purpose. In [17] the data was collected only from 20 subjects. For this
reason the dataset accuracy are low. In [18] they have collected their gait data from 53
subjects walking in different direction. These data sets clearly show us the distortion of
the age and gender ratio.
3 Dataset Description
The University of Osaka has developed OU-ISIR gait dataset, inertial dataset. The
dataset is relatively well developed and is the biggest sensor-based inertial gait database
sensor [19]. These dataset used 3 IMU (inertial measurement unit) sensors known as
IMUZ which is used to capture gait signals. The IMUZ comprises of a 3-axial
accelerometer and a 3-axial gyroscope sensor. These three sensors are on the right, one
on the left and the one at the center-back of the belt. These three sensors had been
mounted to the belt with various orientations (90° for center left, center right and 180°
for left-right pair).Gait data was collected for five days from 745 visitors. Each visitor
entered and departed only once from the designated data capture tool. The dataset
included equal number of gender (384 males and 384 females).In each IMUZ sensor,
triaxial accelerometer and triaxial gyroscope sequence of signals are captured. There-
fore 6D data are collected from each signal. Five activities data have been collected
known as: slope-up walk, slope-down walk, level walk, step-up walk and step-down
walk. For each subject, data were extracted for only level walk, slope down and slope
up walk. The data has four labels namely: ID, Age, gender and Activity. Figure 1 is an
example of sequence of signals for accelerometer data and gyroscope data.
Gender Classification from Inertial Sensor-Based Gait Dataset 587
Gryosocope(rad/s)
1
Gx
0
Gy
-1
Gz
-2
Time
1
Acceleration signals
0 Ax
Ay
-1
Az
-2
Time
4 Methodology
Table 2 shows the some of the feature value which is extracted from the gait data.
To prevent a discontinuity of the LOO classification error, the soft cost function of
the neighbor assignments in the transformed space can be used. In the transformed
space, the probability pij that point j is as the nearest point I can be described as
2
exp Axi Axj
pij ¼ P
exp Axi Axj 2
k6¼j
590 R. K. Pathan et al.
Where pii ¼ 0.
Transformation matrix A can be achieved by correctly classifying the expected
number of points
2
XX exp Axi Axj
A ¼ max P 2 kjj AjjF
2
i j2Ci k6¼j exp Axi Axj
Where k parameter is used for maximizes the NCA probability and minimizes the
Frobenius norm. To solve the optimization problem, the conjugate method can be used.
If A is confined to diagonal matrix, the diagonal inputs represent the weight of the
every input feature. Therefore, the selection of the feature can be obtained on the basis
of the importance of the weights.
Matlab statistics and machine learning Toolbox functions fscnca is used for NCA
feature selection process. After applying the feature selection method NCA, the
numbers of features are now 84.
Once model training is completed, the learned models are used to classify the gender of
people. We have also demonstrated the k-fold effects of cross-validation So that the
model didn’t overfit the training dataset. Different parameters for each model and the
maximum accuracy of each model have been observed. First, we construct the models
using 104 features derived from sensor data.
To minimize the number of features, we have used a feature selection method. Our
primary objective is to select the most discriminative time-domain and frequency-
domain feature for gender classification and also find out the best accuracy among the
four classification algorithms. For this purpose, we have used four well-known clas-
sifiers KNN (K-Nearest Neighbor), SVM (Support Vector Machine), Bagging, and
Boosting algorithm.
For the experimental purpose, we have used accuracy, MAE, and RMSE to measure
the performance of the algorithms and we also compared classification model perfor-
mance with three metrics [27]: precision, Recall, and F1-score.
Matlab 2018 software was used for calculating all the result. For Feature selection
Matlab Statistics and Machine learning toolbox was used and for classification
problem.
Table 3 depicts the results from the comparison of the classification accuracy
before and after feature selection process. 1,556 samples have been used for these
classifications where each class contains the same number of samples. It can be seen
from the table that the accuracy of the all classifier are higher than 80%.Bagging
algorithm offers the best possible result with 87.5% compare to the other classifier.
Because bagging reduces the variance of one estimate, since several estimates from
various models are combined. Support vector machine shows comparatively lower
accuracy. The results from the table show that the accuracy of the each algorithm is
increased by 3% when Neighborhood component analysis (NCA) dimension reduction
method is applied to its original feature. This is because NCA method has decreased the
dimensionality of the data. Therefore the NCA method for reducing dimensionality
makes the classification process simpler and increases the classification accuracy rate.
592 R. K. Pathan et al.
100
87.854
86.091
84.105
95
83.898
83.615
81.456
81.425
Accuracy in %
79.661
90
85
80
75
70
65
60
Acc before selecƟon Acc aŌer selecƟon
From Table 4, it has been seen that the Bagging algorithm shows better perfor-
mance among the four algorithms. It also gives the lowest MAE and RMSE. Bagging
algorithm has the highest F1-score, Precision, and Recall.
Gender Classification from Inertial Sensor-Based Gait Dataset 593
Figure 4 shows the ROC curve for all classifier of male class and it can be noticed
that Bagging, Boosting and KNN classifier curve converge quickly where SVM
classifier curve converge slowly.
Figure 5 illustrates the example of a confusion matrix in the Bagging classifier. The
input of this classification is based on data after the process of feature selection (NCA-
based features). The confusion matrix is carried out between two classes namely Male
and Female classes. The table row represents the real class and the column represents
the class that classifiers know. The value in the table indicates the probability of the
actual class being recognized. The Table shows that the male class is correctly rec-
ognized with 87.4% and female class are correctly recognized with 85.3%.
In this research, we have tried to identify the best machine learning algorithm to
classify gender from the inertial sensor-based gait dataset. The largest sensor-based
inertial OU-ISIR gait dataset has been analyzed for this experiment. The use of time-
domain and the frequency-domain features are the essential part of our paper, and we
also select features which are most important for the classification of gender. These
extracted features are used successfully to train our selected classification models. From
the result, it has been observed that after selecting the features from the 104 features,
the accuracy of the classifier is increased, which in turn could be used to ensure the
safety and security of the human beings In the future, we will use some deep learning
methods for gender classification [28–31] and also apply some methodology to remove
uncertainty [32–37]. However, in this study, we have used only 1100 data for the
training set. In the future, we will use more data for the training phase and we will also
use other feature selection method like PCA and t-SNE for training our dataset.
References
1. Wu, Y., Zhuang, Y., Long, X., Lin, F., Xu, W.: Human gender classification: a review. arXiv
preprint arXiv:1507.05122,2015
2. Udry, J.R.: The nature of gender. Demography, 31(4), 561–573 (1994)
3. Murray, M.P., Drought, A.B., Kory, R.C.: Walking patterns of normal men. JBJS 46(2),
335–360 (1964)
4. Murray, M.P.: Gait as a total pattern of movement: including a bibliography on gait. Am.
J. Phys. Med. Rehabilitation 46(1), 290–333 (1967)
5. Xiao, Q.: Technology review-biometrics-technology, application, challenge, and computa-
tional intelligence solutions. IEEE Comput. Intell. Magazine, 2(2), 5–25 (2007)
6. Wong, K.Y.E., Sainarayanan, G., Chekima, A.: Palmprint based biometric system: a
comparative study on discrete cosine transform energy, wavelet transform energy, and
sobelcode methods (< Special Issue > BIOMETRICS AND ITS APPLICATIONS). Int.
J. Biomed. Soft Comput. Human Sci. Official J. Biomed. Fuzzy Syst. Assoc. 14(1), 11–19
(2009)
7. Hanmandlu, M., Gupta, R.B., Sayeed, F., Ansari, A.Q.: An experimental study of different
features for face recognition. In: 2011 International Conference on Communication Systems
and Network Technologies, pp. 567–571. IEEE, June 2011
8. Sasaki, J.E., Hickey, A., Staudenmayer, J., John, D., Kent, J.A., Freedson, P.S.: Performance
of activity classification algorithms in free-living older adults. Med. Sci. Sports and Exercise,
48(5), 941 (2016)
Gender Classification from Inertial Sensor-Based Gait Dataset 595
9. Ngo, T.T., Makihara, Y., Nagahara, H., Mukaigawa, Y., Yagi, Y.: The largest inertial
sensor-based gait database and performance evaluation of gait-based personal authentication.
Pattern Recogn. 47(1), 228–237 (2014)
10. Khabir, K.M., Siraj, M.S., Ahmed, M., Ahmed, M.U.: Prediction of gender and age from
inertial sensor-based gait dataset. In: 2019 Joint 8th International Conference on Informatics,
Electronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision &
Pattern Recognition (icIVPR), pp. 371–376. IEEE, May 2019
11. Riaz, Q., Vögele, A., Krüger, B., Weber, A.: One small step for a man: estimation of gender,
age and height from recordings of one step by a single inertial sensor. Sensors, 15(12),
31999–32019 (2015)
12. Makihara, Y., Mannami, H., Yagi, Y.: Gait analysis of gender and age using a large-scale
multi-view gait database. In: Asian Conference on Computer Vision, pp. 440–451. Springer,
Heidelberg, November 2010
13. Garofalo, G., ArgonesRúa, E., Preuveneers, D., Joosen, W.: A systematic comparison of age
and gender prediction on imu sensor-based gait traces. Sensors, 19(13), 2945 (2019)
14. Ngo, T.T., Ahad, M.A.R., Antar, A.D., Ahmed, M., Muramatsu, D., Makihara, Y., Hattori,
Y.: OU-ISIR wearable sensor-based gait challenge: age and gender. In: Proceedings of the
12th IAPR International Conference on Biometrics, ICB (2019)
15. Jain, A., Kanhangad, V.: Investigating gender recognition in smartphones using accelerom-
eter and gyroscope sensor readings. In: 2016 International Conference on Computational
Techniques in Information and Communication Technologies (ICCTICT), pp. 597–602.
IEEE, March 2016
16. Yoo, J.H., Hwang, D., Nixon, M.S.: Gender classification in human gait using support vector
machine. In: International Conference on Advanced Concepts for Intelligent Vision Systems,
pp. 138–145. Springer, Heidelberg, September 2005
17. Asmara, R.A., Masruri, I., Rahmad, C., Siradjuddin, I., Rohadi, E., Ronilaya, F., Hasanah,
Q.: Comparative study of gait gender identification using gait energy image (GEI) and gait
information image (GII). In: MATEC Web of Conferences, vol. 197, p. 15006. EDP
Sciences (2018)
18. Borràs, R., Lapedriza, A., Igual, L.: Depth information in human gait analysis: an
experimental study on gender recognition. In: International Conference Image Analysis and
Recognition, pp. 98–105. Springer, Heidelberg, June 2012
19. Lu, J., Tan, Y.-P.: Gait-based human age estimation, IEEE Transactions
20. Trung, N.T., Makihara, Y., Nagahara, H., Mukaigawa, Y., Yagi, Y.: Performance evaluation
of gait recognition using the largest inertial sensor-based gait database. In: 2012 5th IAPR
International Conference on Biometrics (ICB). IEEE, pp. 360–366 (2012)
21. Burges, C.J.: A tutorial on support vector machines for pattern recognition. Data Min.
Knowl. Discovery 2(2), 121–167 (1998)
22. Szegedy, V.V.O.: Processing images using deep neural networks. USA Patent 9,715,642, 25
July 2017
23. Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)
24. Johnson, R.W.: An introduction to the bootstrap. Teach. Stat. 23(2), 49–54 (2001)
25. Rahman, A., Verma, B.: Ensemble classifier generation using non-uniform layered clustering
and Genetic Algorithm. Knowledge-Based Syst. 43, 30–42 (2013)
26. Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: International
Conference on Machine Learning, pp. 148–156 (1996)
27. Sokolova, M., Lapalme, G.: A systematic analysis of performance measures for classification
tasks. Inf. Process. Manage. 45(4), 427–437 (2009)
596 R. K. Pathan et al.
28. Chowdhury, R.R., Hossain, M.S., ul Islam, R., Andersson, K., Hossain, S.: Bangla
handwritten character recognition using convolutional neural network with data augmen-
tation. In: 2019 Joint 8th International Conference on Informatics, Electronics & Vision
(ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition
(icIVPR), pp. 318–323. IEEE, May 2019
29. Ahmed, T.U., Hossain, M.S., Alam, M.J., Andersson, K.: An integrated CNN-RNN
framework to assess road crack. In: 2019 22nd International Conference on Computer and
Information Technology (ICCIT), pp. 1–6. IEEE, December 2019
30. Ahmed, T.U., Hossain, S., Hossain, M.S., ul Islam, R., Andersson, K.: Facial expression
recognition using convolutional neural network with data augmentation. In: 2019 Joint 8th
International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd
International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 336–341.
IEEE, May 2019
31. Islam, M.Z., Hossain, M.S., ul Islam, R., Andersson, K.: Static hand gesture recognition
using convolutional neural network with data augmentation, May 2019
32. Biswas, M., Chowdhury, S.U., Nahar, N., Hossain, M.S., Andersson, K.: A belief rule base
expert system for staging non-small cell lung cancer under uncertainty. In: 2019 IEEE
International Conference on Biomedical Engineering, Computer and Information Technol-
ogy for Health (BECITHCON), pp. 47–52. IEEE, November 2019
33. Kabir, S., Islam, R.U., Hossain, M.S., Andersson, K.: An integrated approach of belief rule
base and deep learning to predict air pollution. Sensors 20(7), 1956 (2020)
34. Monrat, A.A., Islam, R.U., Hossain, M.S., Andersson, K.: A belief rule based flood risk
assessment expert system using real time sensor data streaming. In: 2018 IEEE 43rd
Conference on Local Computer Networks Workshops (LCN Workshops), pp. 38–45. IEEE,
October 2018
35. Karim, R., Hossain, M.S., Khalid, M.S., Mustafa, R., Bhuiyan, T.A.: A belief rule-based
expert system to assess bronchiolitis suspicion from signs and symptoms under uncertainty.
In: Proceedings of SAI Intelligent Systems Conference, pp. 331–343. Springer, Cham,
September 2016
36. Hossain, M.S., Monrat, A.A., Hasan, M., Karim, R., Bhuiyan, T.A., Khalid, M.S.: A belief
rule-based expert system to assess mental disorder under uncertainty. In: 2016 5th
International Conference on Informatics, Electronics and Vision (ICIEV), pp. 1089–1094.
IEEE, May 2016
37. Hossain, M.S., Habib, I.B., Andersson, K.: A belief rule based expert system to diagnose
dengue fever under uncertainty. In: 2017 Computing Conference, pp. 179–186. IEEE, July
2017
Lévy-Flight Intensified Current Search
for Multimodal Function Minimization
1 Introduction
Based on modern optimization, many metaheuristic algorithms have been launched for
solving several real-world optimization problems under complex constraints [1, 2].
From literature reviews, the current search (CuS) is one of the most interesting
trajectory-based metaheuristic algorithms [3]. The CuS development scenario begins in
2012 once it was firstly proposed as an optimizer to solve the optimization problems
[3]. The CuS algorithm mimics the principle of an electric current behavior in the
electric circuits and networks. It performed superior search performance to the genetic
algorithm (GA), tabu search (TS) and particle swarm optimization (PSO) [3]. The CuS
was successfully applied to control engineering [4] and signal processing [5]. During
2013-2014, the adaptive current search (ACuS) was launched [6] as a modified version
of the conventional CuS. The ACuS consists of the memory list (ML) used to escape
from local entrapment caused by any local solution and the adaptive radius
(AR) conducted to speed up the search process. The ACuS was successfully applied to
industrial engineering [6] and energy resource management [7]. For some particular
problems, both the CuS and ACuS are trapped by local optima and consume much
search time. In 2014, the intensified current search (ICuS) was proposed to improve its
search performance [8]. The ICuS algorithm consists of the ML, AR and the adaptive
neighborhood (AN) mechanisms. The ML regarded as the exploration strategy is used
to store the ranked initial solutions at the beginning of search process, record the
solution found along each search direction, and contain all local solutions found at the
end of each search direction. The ML is also applied to escape the local entrapments
caused by local optima. The AR and AN mechanisms regarded as the exploitation
strategy are together conducted to speed up the search process. The ICuS was suc-
cessfully applied to many control engineering problems including single-objective and
multi-objective optimization problems [8–11]. For some optimization problems espe-
cially in large-space multimodal problems, the ICuS might be trapped by local optima.
This is properly because the random number with the uniform distribution used in the
ICuS algorithm is not efficient enough for such the problems. Thus, it needs to be
modified to enhance its search performance and to speed up the search process.
In this paper, the new trajectory-based metaheuristic algorithm called the Lévy-
flight intensified current search (LFICuS) is peoposed. The proposed LFICuS is the
newest modified version of the ICuS. The random number with the Lévy-flight dis-
tribution and the adjustable search radius mechanism are conducted to improve the
search performance. This paper consists of five sections. An introduction is given in
Sect. 1. The ICuS algorithm is briefly described and the proposed LFICuS is illustrated
in Sect. 2. Ten selected standard multimodal benchmark functions used in this paper
are detailed in Sect. 3. Results and discussions of the performance evaluation of the
ICuS and LFICuS are provided in Sect. 4. Finally, conclusions are followed in Sect. 5.
The ICuS algorithm is briefly described in this section. Then, the proposed LFICuS
algorithm is elaborately illustrated as follows.
(2) and c is the scale parameter. The random with Lévy-flight distribution has an
infinite mean and infinite variance [12]. Then, it is more efficient than the random with
uniform distribution. Many metaheuristics including the cuckoo search (CS) [13] and
flower pollination algorithm (FPA) [14] utilize the random with Lévy-flight distribution
for exploring the feasible solutions.
8
< 1
for a x b;
f ðxÞjuniform ¼ b a ð1Þ
:
0 for x\a or x [ b
rffiffiffiffiffiffi
e2ðxlÞ
c
c
f ðxÞjLevy ¼ ð2Þ
2p ðx lÞ3=2
Lévy-Flight Intensified Current Search 601
The proposed LFICuS algorithm uses the random number drawn from the Lévy-
flight distribution to generate the neighborhood members as feasible solutions in each
search iteration. Once applying in the proposed LFICuS algorithm, the Lévy-flight
distribution L can be approximated by (3) [14], where s is step length, k is an index and
C(k) is the Gamma function as expressed in (4), when C(n) = (n−1)!, k = n (an integer).
The Lévy-flight distribution works well for large space. Therefore, the adjustable
search radius (AR) mechanism is also conducted by setting the initial search radius
R = X (search space). The proposed LFICuS algorithm can be described by the pseudo
code shown in Fig. 2.
kCðkÞ sinðpk=2Þ 1
L 1þk ð3Þ
p s
Z1
CðkÞ ¼ tk1 et dt ð4Þ
0
3 Benchmark Functions
1500
1000
f(x1, x2)
500
0
500
500
x2 0 0x
1
-500 -500
Function names Functions, search space, optimal solution and optimal function value
XD pffiffiffiffiffiffi
(1) Schwefel function (SF) f1 ðxÞ ¼ 418:9829D x sinð jxi jÞ; 500 xi 500; i ¼ 1; 2; . . .; D;
i¼1 i
D ¼ 2; x ¼ ð420:9687; 420:9687; . . .; 420:9687Þ; f1 ðxÞ ¼ 0; f1 max ¼ 1 104
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
(2) Ackley function (AF) 0:02 D1
PD 2
x 1
PD
i¼1 i cosð2pxi Þ
f2 ðxÞ ¼ 20e eD i¼1 þ 20 þ e; 35 xi 35;
i ¼ 1; 2; . . .; D; D ¼ 2; x ¼ ð0; 0; . . .; 0Þ; f2 ðxÞ ¼ 0; f2 max ¼ 1 105
W. Romsai et al.
(3) Bohachevsky function (BF) f3 ðxÞ ¼ x21 þ 2x22 0:3 cosð3px1 Þ 0:4 cosð4px2 Þ þ 0:7; 100 xi 100;
i ¼ 1; 2; x ¼ ð0; 0Þ; f3 ðxÞ ¼ 0; f3 max ¼ 1 109
pffiffiffiffiffiffiffiffiffiffiffiffiffiffi
(4) Drop-Wave function (DWF) 1 þ cos 12 x21 þ x22
f4 ðxÞ ¼ ; 5:2 xi 5:2; i ¼ 1; 2;
0:5ðx21 þ x22 Þ þ 2
x ¼ ð0; 0Þ; f4 ðxÞ ¼ 1; f4 max ¼ 0:9999
(5) Egg-Crate function (ECF) f5 ðxÞ ¼ x21 þ x22 þ 25½sin2 ðx1 Þ þ sin2 ðx2 Þ; 5 xi 5; i ¼ 1; 2;
x ¼ ð0; 0Þ; f5 ðxÞ ¼ 0; f5 max ¼ 1 105
2 2 0:5
(6) Pen-Holder function (PHF) =p 1
f6 ðxÞ ¼ exp½j cosðx1 Þ cosðx2 Þej1½ðx1 þ x2 Þ j ; 11 xi 11; i ¼ 1; 2;
x ¼ ð 9:6462; 9:6462Þ; f6 ðxÞ ¼ 0:9635; f6 max ¼ 0:9634
XD
(7) Rastrigin function (RF) f7 ðxÞ ¼ 10D þ ½x2 10 cosð2pxi Þ; 5:12 xi 5:12; i ¼ 1; 2; . . .; D;
i¼1 i
D ¼ 2; x ¼ ð0; 0; . . .; 0Þ; f7 ðxÞ ¼ 0; f7 max ¼ 1 109
(8) Styblinski-Tang function (STF) 1 XD
f8 ðxÞ ¼ ðx4 16x2i þ 5xi Þ; 5 xi 5; i ¼ 1; 2; . . .; D;
i¼1 i
2
D ¼ 2; x ¼ ð2:9035; . . .; 2:9035Þ; f8 ðxÞ ¼ 39:1660D; f8 max ¼ 78:3319
XD PD 2
(9) Yang-2 function (Y2F)
f9 ðxÞ ¼ jx j e i¼1 sinðxi Þ ; 2p xi 2p; i ¼ 1; 2; . . .; D;
i¼1 i
The proposed LFICuS algorithm was coded by MATLAB version 2017b (License No.
#40637337) run on Intel(R) Core(TM) i7-10510U CPU@1.80 GHz, 2.30 GHz,
16.0 GB RAM for function minimization tests against ten selected multimodal
benchmark functions. Searching parameters of the proposed LFICuS algorithm are set
from the preliminary studies with different ranges of parameters, i.e., step length s,
index k, number of state of AR and AN mechanisms activation h, number of initial
neighborhood members n, number of search directions N. By varying s = 0.0001,
0.001, 0.01 and 0.1, k = 0.0000001, 0.000001, 0.00001, 0.0001, 0.001, 0.01, 0.1, 0.2,
0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1.0, 1.1, 1.2, 1.3, 1.4, 1.5, 1.6, 1.7, 1.8, 1.9 and 2.0, h =
2, 3, 4 and 5, n = 25, 50, 100, 150, 200 and 300 and N = 10, 25, 50, 75, 100, 125, 150
and 200, it was found that the best parameters for most benchmark functions are: s =
0.001 to 0.01, k = 0.01 to 1.3, h = 2 to 3, n = 50 to 100 and N = 50 to 75. For this
performance evaluation tests, s = 0.01, k = 0.3, h = 2, n = 100 and N = 50 are set for all
selected functions. Each search direction is terminated at 1000 iterations. 1000-trial
runs are conducted for each algorithm. All algorithms will be terminated once two
termination criteria (TC) are satisfied, i.e. (1) the function values are less than fmax in
Table 1 of each function or (2) the search meets the maximum search directions N = 50.
The former criterion implies that the search is success, while the later means that the
search is not success. For comparison with the ICuS, searching parameters of the ICuS
are fairly set as the same values to those of LFICuS.
The obtained results are summarized in Table 2. Data represented in Table 2 are in
the form of AE ± SD(SR%), where the AE is the average number (mean) of function
evaluations, the SD is the standard deviation and the SR is the success rate. The AE
value implies the searching time consumed. The SD value implies the robustness of the
algorithm. Referring to Table 2, the proposed LFICuS algorithm provides greater SR
and less AE and SD values than the ICuS. This can be noticed that the proposed
LFICuS algorithm performs much more efficient in function minimization than the
ICuS algorithm. The convergent rates of the SF function proceeded by the ICuS and
proposed LFICuS algorithms are depicted in Fig. 4. Those of other functions are
omitted because they have a similar form to that of SF function shown in Fig. 4.
Fig. 4. Convergent rates of the SF proceeded by the ICuS and proposed LFICuS algorithms.
Lévy-Flight Intensified Current Search 605
5 Conclusions
The Lévy-flight intensified current search (or LFICuS) algorithm has been developed
and proposed in this paper for multimodal function minimization. As the novel
trajectory-based metaheuristic algorithm, the proposed LFICuS is based on the ICuS
algorithm. In the proposed LFICuS, the random number drawn from the Lévy-flight
distribution and the adjustable AR mechanism have been conducted to improve the
search performance. The proposed LFICuS has been tested against ten selected stan-
dard multimodal benchmark functions for minimization. As results, the proposed
LFICuS algorithm has provided greater success rate, less number of function evalua-
tions and less standard deviation values than the ICuS. This can be concluded that the
proposed LFICuS algorithm is one of the most efficient trajectory-based metaheuristic
algorithms for function minimization. For future research, the proposed LFICuS
algorithm can be probably conducted to any real-world engineering applications and
other optimization problems including finance, forecasting, meteorology and so all.
References
1. Talbi, E.G.: Metaheuristics Forn Design to Implementation. John Wiley & Sons, Hoboken
(2009)
2. Yang, X.S.: Engineering Optimization: An Introduction with Metaheuristic Applications,
John Wiley & Sons (2010)
3. Sukulin, A., Puangdownreong, D.: A novel metaheuristic optimization algorithm: current
search. In: The 11th WSEAS International Conference on Artificial Intelligence, Knowledge
Engineering and Data Bases (AIKED 2012), pp. 125–130 (2012)
4. Puangdownreong, D.: Application of current search to optimum PIDA controller design.
Intell. Control Autom. 3(4), 303–312 (2012)
5. Puangdownreong, D., Sukulin, A.: Current search and applications in analog filter design
problems. J. Commun. Comput. 9(9), 1083–1096 (2012)
6. Suwannarongsri, S., Bunnag, T., Klinbun, W.: Energy resource management of assembly
line balancing problem using modified current search method. Int. J. Intell. Syst. Appl. 6(3),
1–11 (2014)
7. Suwannarongsri, S., Bunnag, T., Klinbun, W.: Optimization of energy resource management
for assembly line balancing using adaptive current search. Am. J. Oper. Res. 4(1), 8–21
(2014)
8. Nawikavatan, A., Tunyasrirut, S., Puangdownreong, D.: Application of Intensified Current
Search to Optimum PID Controller Design in AVR System. Lecture Notes in Computer
Science, pp. 255–266 (2014)
9. Nawikavatan, A., Tunyasrirut, S., Puangdownreong, D.: Optimal PID controller design for
three-phase induction motor speed control by intensified current search. In: The 19th
International Annual Symposium on Computational Science and Engineering (ANSCSE19),
pp. 104–109 (2015)
10. Thammarat, C., Puangdownreong, D., Nawikavatan, A., Tunyasrirut, S.: multiobjective
optimization of PID controller of three-phase induction motor speed control using intensified
current search. In: Global Engineering & Applied Science Conference, pp. 82–90 (2015)
606 W. Romsai et al.
11. Nawikavatan, A., Tunyasrirut, S., Puangdownreong, D.: Application of intensified current
search to multiobjective PID controller optimization. Int. J. Intell. Syst. Appl. 8(11), 51–60
(2016)
12. Pavlyukevich, I.: Lévy flights, non-local search and simulated annealing. J. Comput. Phys.
226(9), 1830–1844 (2007)
13. Yang, X.S., Deb, S.: Engineering optimisation by cuckoo search. Int. J. Math. Model. Num.
Opt. 1(4), 330–343 (2010)
14. Yang, X.S.: Flower pollination algorithm for global optimization. Unconventional Compu-
tation and Natural Computation, Lecture Notes in Computer Science 7445, 240–249 (2012)
15. Ali, M.M., Khompatraporn, C., Zabinsky, Z.B.: A numerical evaluation of several stochastic
algorithms on selected continuous global optimization test problems. J. Global Opt. 31, 635–
672 (2005)
16. Jamil, M., Yang, X.S., Zepernick, H-.J.: Test Functions for Global Optimization: a
Aomprehensive Survey. Swarm Intelligence and Bio-Inspired Computation Theory and
Applications, Elsevier Inc., pp. 193–222 (2013)
Cancer Cell Segmentation Based
on Unsupervised Clustering
and Deep Learning
1 Introduction
Human bodies are rapidly producing trillions of cells that form tissues and organs such
as muscles, bones, the lungs, and the liver. Genes of each cell instruct it when to grow,
work, divide, and die. Sometimes these instructions get mixed up and go out of control;
then, the cells become abnormal as they grow and divide but do not die in time. These
uncontrolled and abnormal cells can form a lump in the body known as a tumor.
Tumors are of two types, such as non-cancerous (benign) and cancerous (malignant).
When cancer starts in the lungs, it is known as lung cancer, and cancer that has spread
to the lungs from another part of the body might be called secondary lung cancer.
Leukemia is a cancer of white blood cells, which cannot fight infection and develops
bone marrow. To classify the MRI image to detect the tumor category in the patient's
brain is the purpose of brain tumor classification. Many test types can be used for brain
tumor detection, such as MRI, Biopsy, and Computed Tomography scan. However, in
comparison to CT scan images, MRI images are safe, and they also provide higher
contrast. Moreover, MRI images do not affect the human body. It is easy to detect and
classify the brain through MRI images as they have a fair resolution for different brain
tissues.
Image segmentation always plays a significant part in cancer diagnosis. The actual
meaning of segmentation is splitting an image into several regions and extracting
meaningful information from these regions [1]. There are several segmentation tech-
niques have been applied to the MRI images. For medical image analysis, edge
detection, threshold-based, region-growing, model-based, and clustering-based seg-
mentation techniques are the most widely used methods. All the segmentation tech-
niques have their advantages and disadvantages, and thus, it depends on the user's
choice. Besides the segmentation and detection of cancer cells, it is necessary to
classify the types of cancer after segmentation to help the specialists provide the correct
direction to the patient for early emergence. There have various classification methods
among them multi-SVM (Multi-support vector machine) is applied in our proposed
method because it is a linear learning algorithm used for classification and is a powerful
supervised classifier and accurate learning technique. The Convolutional Neural Net-
work models have been very successful for a large number of visual tasks [2]. In a
word, this paper approaches benign, malignant, Leukemia, and Lung cancer detection
using partial contrast stretching, Otsu, K-means, median filtering, and performs the
classification of the benign, malignant Leukemia and lung using M-SVM and CNN
classifier.
This system uses MRI, RGB and CT images of Brain MRI Images for Brain Tumor
Detection dataset for both Benign and Malignant, SN-AM dataset for leukemia [3–6],
LIDC database for lung cancer [7].
This paper is organized as follows. Section 2 summarizes the previous works
completed in the fields of cancer cell detection. Section 3 gives a brief description of
the methodology that has been proposed for the different types of cancer cell detection,
feature extraction, classification and calculation of the detected area of cancer cell and
conclusions is in Sect. 4.
2 Literature Review
Many researchers have applied different methods to segment, detect, and classify
images of cancer cells. Some of the existing methods are described here.
Cancer Cell Segmentation Based on Unsupervised Clustering and Deep Learning 609
Wang et al. [1] has presented a novel cell detection method that utilizes both
intensity and shape information of a cell to improve the segmentation. Upon obtaining
the cell shape information with the binarization process, they have used both intensity
and shape information for local maxima generation. Lastly, the nuclei pixels are per-
mitted to move inside the gradient vector field. The pixels will finally join at these local
maxima, and the detected cells are then segmented through the seeded watershed
algorithm.
Purohit and Joshi [8] introduced a new efficient approach towards the K-means
clustering algorithm. They proposed a new method that produces the cluster center by
reducing the mean square final cluster's error without a considerable increment in the
execution time. Many judgements have been done and it can conclude that accuracy is
more for dense dataset rather than small dataset.
Alan Jose, S. Ravi, and M. Sambath [9] have proposed a methodology to segment
Brain Tumor using K-means Clustering and Fuzzy C-means Algorithm and its' area
calculation. They split the process into several parts. First of all, preprocessing is
implemented by using the filter to improve the quality of the image. After that, the
advanced K-means algorithm is applied, followed by Fuzzy c-means to cluster the
image. Then the resulted segment image is used for the feature extraction of the region
of interest. They have used MRI images for the analysis and calculated the size of the
extracted image's tumor region.
Selvaraj Damodharan and Dhanasekaran Raghavan [10] have proposed a neural
network-based procedure for brain tumor detection and classification. In their
methodology, the quality rate is formed separately for segmentation of WM, GM, CSF,
and tumor region and claimed accuracy of 83% using the neural network-based
classifier.
Nihad Mesanovic, Mislav Grgic, Haris Huseinagic [11] have proposed region
growing algorithm for segmentation of CT scan images of the lung. This algorithm
starts with a seed pixel which also checks other pixels that surround it. It determines the
most alike one. If it meets confident criteria, it will include in the region. The region is
established by examining all unallocated adjacent pixels to the region.
Kanitkar, Thombare, and Lokhande [12] proposed a methodology that takes a CT
scan image of the lung, then preprocesses the input image for smoothing using the
Gaussian filter and, then enhanced using Gabor filter. Then thresholding and marker-
controlled watershed segmentation methods are used to segment the processed image.
They used a marker-based watershed segmentation technique to overcome the draw-
backs of the watershed segmentation technique, i.e., over-segmentation. Then feature
extraction and classification of cancer stages are done.
610 J. Sikder et al.
3 Proposed Methodology
Figure. 1 shows the proposed method. This method proposes the preprocessing, seg-
mentation, detection, feature extraction and classification of cancer cells using Otsu, k-
means, multi support vector machine and convolutional neural network. As the initial
step, an input image is given to the system.
Input Image
Preprocessing
Segmentation Stage
Median Filtering
Equal
?
Yes No
Detected Region
Calculate Area
Fig. 2. (a) Original Brain test image-Benign; (b) Original Brain test image- Malignant;
(c) Original Leukemia test image; (d) Original Lung Cancer test image.
Cancer Cell Segmentation Based on Unsupervised Clustering and Deep Learning 611
3.1 Preprocessing
RGB2GRAY
In the preprocessing stage firstly the input image checked either RGB or gray. If the
testing image becomes a RGB image it converts the true color RGB image into gray
scale by eliminating the hue and saturation information while retaining the luminance.
Fig. 3. Gray image- (a) Benign; (b) Malignant; (c) Leukemia; (d) Lung Cancer.
Remove Noise
Gaussian filtering method used to smooth the testing images so that it looks more
meaning full for segmentation. Filtering method helps to improve certain parameters of
testing images such as improving the signal-to-noise ratio, enhancing the visual
appearance of testing image, removing the irrelevant noise.
Fig. 4. Removed Noise- (a) Benign; (b) Malignant; (c) Leukemia; (d) Lung Cancer.
612 J. Sikder et al.
Fig. 5. Partial Contrast Stretched- (a) Benign; (b) Malignant; (c) Leukemia; (d) Lung Cancer.
Fig. 6. Enhanced using Otsu Method (a) Benign; (b) Malignant; (c) Leukemia; (d) Lung cancer.
Cancer Cell Segmentation Based on Unsupervised Clustering and Deep Learning 613
Input Image
Centroid
satisfy the No
tolerance or
error value?
Yes
Reshape the cluster pixels into image
3.2 Segmentation
K-means clustering is an unsupervised clustering method that classifies a given set of
data into k number of disjoint clusters. It is also an iterative method which is used to
partition an image into K clusters. Firstly, it takes k number of clusters and k centroid
where the value of k is defined as 2 and initial 2 centroids are chosen randomly, then it
calculates the Euclidean distance between the center and each pixel of the input image
[14]. Here Euclidean distance is used to define the distance of the nearest centroid and
assigns each point to the cluster which has nearest centroid from the respective data
point. After assigning all data points it recalculates the new centroid of each cluster and
using that centroid, a new Euclidean distance. So, K-means is an iterative algorithm, it
repeats the process and calculates new centroid in each iteration until it satisfies the
tolerance or error value. In our proposed system we divide the whole region into two
clustered. Best clustering region is taken as region of interest from two clustered e.g.
the white cancer cell cluster is taken as segmented with black background [15].
Fig. 8. K-means Segmented-(a) Benign; (b) Malignant; (c) Leukemia; (d) Lung cancer.
614 J. Sikder et al.
Fig. 9. Median Filtering of- (a) Benign; (b) Malignant; (c) Leukemia; (d) Lung cancer.
3.4 Classification
In this stage this study proposed multi SVM and Convolutional Neural Network
(CNN) for classification. Multi-SVM has two steps. Firstly, relevant features have
extracted from the filtered image. Secondly, the extracted features have feed the multi-
SVM to classify the test image. The process of two classifier is described below.
Feature Extraction
Collecting higher level information of an image such as shape, texture, color and
contrast is known as feature extraction. The analyzing methods that have been done so
far has used the values of pixel intensities, pixel coordinates and some other statistic
features namely mean, variance or median, which have error in determination process,
low precision and low efficiency in classification [10]. Gray Level Co-occurrence
Matrix (GLCM) is one of the most widely used image analysis applications. This
procedure follows two steps so that it can extract features from the medical images. For
this purpose, firstly the GLCM is computed, and in the other step, the texture features
based on the GLCM are calculated. In the proposed system the GLCM texture features
chosen are contrast, correlation, energy, homogeneity and features like mean, standard
Cancer Cell Segmentation Based on Unsupervised Clustering and Deep Learning 615
deviation, RMS (Root Mean Square), entropy, variance, smoothness, kurtosis, skew-
ness, IDM. After extracting the features from segmented image, all features fed to the
multi-SVM and CNN classifier to classify the segmented image. Those features are
extracted from datasets to train the multi-SVM and also extracted from test image to
classify it. Again, CNN features are extracted using fc1000 layer from both datasets
and test image, while the test image is being classified using CNN.
Classification the Test Image
Multi-SVM is used to classify the input image as Benign, Malignant, Leukemia or
Lung cancer, a systematic technique for four class problems [17]. It needs two stages,
training and testing. The Multi-SVMs can be trained by features given as an input to its
learning algorithm and while training, it finds the suitable margins between two classes.
When extracted features are feat to the multi-SVM, it performs the steps to classify. In
this methodology, multi SVM works as a two class SVM repeatedly. When an input
image features are fed to the multi-SVM, it classifies the input image from first two
class. After classifying from two class, it takes the classified class and take the next
third class two classify the input image. It also gives a classified class. Lastly, the
classified class and fourth class is used to classify again. From this last classification,
we get the correct classified class where the input image belongs to. In testing phase,
the multi-SVM produces 90.75% to 95.55% accuracy.
Convolutional neural networks (CNN) are important tools for deep learning, and also
especially useful for image classification, object detection, and recognition tasks. Deep
learning uses convolutional neural networks to learn useful representations of data
directly from images [18]. Neural networks combine diverse nonlinear processing layers,
using simple elements operating in parallel, biological systems and deep learning models
are trained using a large set of labeled data [19]. Again, neural network architectures that
contain many layers, usually include some convolutional layers. The cell images are fed
into a CNNs model to extract deep-learning features. The CNN based classification is
divided into two phases such as training and testing phases. The number of images is
divided into different category by using labels name such as benign, malignant, leukemia
and lung. In the training and testing phase, preprocessing is done for image resizing.
Further, passing the input dataset to all the layers mentioned the classification network
undertakes training. The convolution neural network is used for automatic cancer cell
classification and for forecasting the class label of a new test image.
Firstly, cancer cell database is loaded which create an image data store to manage
the data. Then it calculated images and the category labels associated with each image.
The labels are automatically assigned from the folder names of the image files and
adjust unequal number of images per category so that the number of images in the
training set is balanced. Using the resnet-50 function from Neural Network, we Load
``ResNet-50'' pretrained network. To avoid biasing the results, randomize the split and
the training and test sets processed by the CNN model. Then prepare training and
testing set for CNN. It can easily extract features from one of the deeper layers
using the activations method and extract training features using ‘fc1000′ layer. CNN
image features used to train a multiclass SVM classifier. To extract image features
from test set, repeat the procedure used earlier and test features can then be passed to
the classifier to measure the accuracy of the trained classifier. Finally, newly trained
616 J. Sikder et al.
classifier used to classify the desired segmented image. A general block diagram of the
Convolutional neural networks (CNNs) is shown in Fig. 10.
Load database
Evaluate Accuracy
incorrect predictions of each class given by the classifier. It provides use the errors and
types of errors being made by the classifier. The CNN classifier produces approxi-
mately 87.5% to 100% accuracy.
TP þ TN
Accuarcy ¼ ð1Þ
TP þ TN þ FP þ FN
Fig. 12. Confusion matrix-(a) Benign; (b) Malignant; (c) Leukemia; (d) Lung cancer.
Fig. 13. Detected region of-(a) Benign; (b) Malignant; (c) Leukemia; (d) Lung cancer.
X
n;m
Area ¼ Pði; jÞ ð2Þ
i¼1;j¼1
The area of detected area is obtained by counting the number of affected regions
pixels. The ratio of cancer cells area was calculated using the formula:
Here, the benign tumor has a 7.4% affected region where the test image has 49295
pixels area, and the detected benign tumor region is 3663 pixels. The malignant has a
4.6% affected region where the test image has a total of 51790 pixels area and detected
malignant region is 2399 pixels. Leukemia has a 31.4% affected region where the test
image has 65536 pixels area, and the detected leukemia region is 20561 pixels. The
lung has a 1.5% affected region where the test image has 54231 pixels area, and the
detected lung region is 814 pixels.
Cancer Cell Segmentation Based on Unsupervised Clustering and Deep Learning 619
4 Conclusions
This research developed a methodology that can identify and classify different types of
cancer cell using unsupervised clustering method and deep learning more specially
multi SVM and Convolutional Neural Network and can also be able to segment per-
fectly compare to the results from current methodologies. In comparison with found
results to those of current studies, it must be pointed out that this methodology is not
limited to a fixed. Moreover, the current methods recognized on a specific type of class
or a specific type of cancer dataset. This experimental results indicate that the proposed
approach is an improved method and it can significantly generate accurate and auto-
matic detection of cancer cells than existing methods. This research has shown that the
overall accuracy of the system is approximately 93%. One of the primary challenges of
this research is the expected centroid of clusters, need initial knowledge of the image
and more awareness to train dataset.
In future research, it will give some directions such as developing better segmen-
tation technique, selecting better feature extraction, clssification process when optimal
learning algorithms are applied on different feature descriptors. This research will open
a pathway to implement and analyze different areas of image segmentation in future.
References
1. Wang, M., Zhou, X., Li, F., Huckins, J., King, R.W., Wrong, S.T.C.: Novel cell
segmentation and online learning algorithms for cell phase identification in automated line-
lapse microscopy. biomedical imaging from Nano to Macro. In: ISBI 2007 4th IEEE
International Symposium, pp. 65–6 (2007)
2. Nandagopalan, S., Kumar, P.K.: Deep convolutional network based saliency prediction for
retrieval of natural images. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.) Intelligent
Computing & Optimization. ICO 2018. Advances in Intelligent Systems and Computing,
vol. 866. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-00979-3_51
3. Gupta, R., Mallick, P., Duggal, R., Gupta, A., Sharma, O.: Stain color normalization and
segmentation of plasma cells in microscopic images as a prelude to development of
computer assisted automated disease diagnostic tool in multiple Myeloma. In: 16th
International Myeloma Workshop (IMW), India, March 2017
4. Duggal, R., Gupta, A., Gupta, R., Wadhwa, M., Ahuja, C.: Overlapping cell nuclei
segmentation in microscopic images using deep belief networks. In: Indian Conference on
Computer Vision, Graphics and Image Processing (ICVGIP), India, December 2016
5. Duggal, R., Gupta, A., Gupta, R.: Segmentation of overlapping/touching white blood cell
nuclei using artificial neural networks. CME Series on Hemato-Oncopathology, All India
Institute of Medical Sciences (AIIMS), New Delhi, India, July 2016
6. Duggal, R., Gupta, A., Gupta, R., Mallick, P.: SD-Layer: stain deconvolutional layer for
CNNs in medical microscopic imaging. In: Medical Image Computing and Computer-
Assisted Intervention − MICCAI 2017, MICCAI 2017. Lecture Notes in Computer Science,
Part III, LNCS 10435, pp. 435–443. Springer, Cham. https://doi.org/https://doi.org/10.1007/
978-3-319-66179-7_50
7. Armato, I.I.I., Samuel, G., McLennan, G., Bidaut, L., McNitt-Gray, M.F., Meyer, C.R.,
Reeves, A.P., Clarke, L.P.: Data from LIDC-IDRI. Cancer Imaging Archive (2015). https://
doi.org/10.7937/K9/TCIA.2015.LO9QL9SXArchive
620 J. Sikder et al.
8. Purohit, P., Joshi, R.: A new efficient approach towards k-means clustering algorithm. Int.
J. Comput. Appl. 65(11), 0975–8887 (2013)
9. Jose, A., Ravi, S., Sambath, M.: Brain segmentation using k-means clustering and fuzzy C-
means algorithm and its area calculation. Int. J. Innov. Res. Comput. Commun. Eng. 2(2)
(2014)
10. Damodharan, S., Raghavan, D.: Combining tissue segmentation and neural network for
brain detection. Int. Arab J. Inf. Technol. 12(1) (2015)
11. Mesanovic, N., Grgic, M., Huseinagic, H., Males, M., Skejic, E., Smajlovic, M.:
Automatic CT image segmentation of the lungs with region growing algorithm
12. Kanitkar, S.S., Thombare, N.D., Lokhande, S.S.: Detection of lung cancer using marker -
controlled watershed transform. In: proceedings of the IEEE Conference on Pervasive
Computing, 978–1–4799–6272–3/15 (2015)
13. Abdullah, A.S., Rosline, H.: Improving color leukemia images using contrast enhancement
techniques. In: 2010 IEEE EMBS Conference on Biomedical Engineering and
Sciences IECBES 2010, Kuala Lumpur, Malaysia (2010)
14. Abdul Nasir, A.S., Mashor, M.Y., Harun, N.H., Abdullah, A.A., Rosline, H.: Improving
colour image segmentation on acute myelogenous leukaemia images using contrast
enhancement techniques. In: 2010 IEEE EMBS Conference on Biomedical Engineering &
Sciences (IECBES 2010), Kuala Lumpur, Malaysia (2010)
15. Dhanachandra, N., Manglem, K., Chanu, Y.J.: Image segmentation using K-means
clustering algorithm and subtractive clustering algorithm. Article in Procedia Computer
Science, December 2015. https://doi.org/10.1016/j.procs.2015.06.090
16. Abd Halim, N.H., Mashor, M.Y., Abdul Nasir, A.S., Mustafa, N., Hassan, R.: Color image
segmentation using unsupervised clustering technique for acute leukemia images. In:
International Conference on Mathematics, Engineering and Industrial Applications 2014,
Proceedings 1660, 090038–1–090038–10. https://doi.org/10.1063/1.4915882,2015
17. Rejintal, A., Aswini, N.: Image processing based leukemia cancer cell detection. In: IEEE
International Conference on Recent Trends in Electronics Information Communication
Technology, 978–1–5090–0774–5/16 (2016)
18. Joshua Thomas, J., Pillai, N.: A deep learning framework on generation of image
descriptions with bidirectional recurrent neural networks. In: Vasant, P., Zelinka, I., Weber,
G.W. (eds.) Intelligent Computing & Optimization. ICO 2018. Advances in Intelligent
Systems and Computing, vol. 866. Springer, Cham (2019). https://doi.org/10.1007/978-3-
030-00979-3_22
19. Ghoneim, A., Muhammad, G., Hossain, M.S.: Cervical cancer classification using
convolutional neural networks and extreme learning machines. Future Generation Comput.
Syst. 102 (2020) 643–649 (2019)
Automated Student Attendance Monitoring
System Using Face Recognition
1 Introduction
technique was used for personal identification, access control, and security manage-
ment [9]. This technique is also suitable for taking attendance of students in an edu-
cational institution. The method is also applicable for taking attendance of the
employee in government institution, military, bank, garment industry, private organi-
zation, shop, etc.
Furthermore, real-world examples such as auto-tagging on Facebook and screen
lock in the mobile application are using face recognition [10, 11]. Computer vision,
artificial intelligence and its sub field machine learning and deep learning have
achieved a great achievement in this regard.
However, currently available systems are not time efficient, cost-effective and much
correct. Due to these problems in the conventional method and feeling the necessity of
face recognition for person identification, we were motivated to propose the face
recognition-based attendance monitoring system. The system uses a HOG to extract
distinctive features and facial landmark estimation due to pose and projection variation.
Then, we use CNN for face encoding and SVM for face recognition.
The aim of this work is to contribute to the development of the smart classroom.
Thus, an automated student attendance monitoring system has been proposed. The
main purpose of this work is to create such a system that will be accurate alongside
time and cost-efficient and can be implemented in real-time. The other key goals of this
work are as follows: first, is to detect the face for feature extraction so that we can get
some important characteristics for further processing. Secondly, we use CNN for face
encoding so that we can recognize the face for taking attendance. Thirdly, we use LFW
database for determining the accuracy and our own dataset. Finally, we will make our
proposed system for real-time application.
The rest work of this paper is organized in the following way: the proposed
methodology is described in Sect. 2, then the result and discussion are discussed in
Sect. 3, finally the conclusion and future scope are given in Sect. 4.
2 Proposed Methodology
Figure 1 illustrate the abstract view of the proposed system. It comprises HOG features
extraction, facial landmark estimation, face encoding using CNN, and face recognition
using SVM. At the entering path of a classroom, there should have a low-resolution
webcam which will directly be connected to the computer or there should use a laptop
computer which have built-in webcam. This computer/laptop will be operated by the
concerned class teacher at any particular schedule. Students will enter the classroom
one by one at any particular time and look on webcam for a while. Each time our
system will process the captured images of every student in real-time to recognize their
faces to make sure his/her attendance of that class. These techniques are described in
the following sub-sections in details as follows.
Automated Student Attendance Monitoring System 623
image is being replaced by an arrow. These arrows are known as gradients, and we
show the flow from light to dark across the entire image. Finally, we get the HOG
version of our image. Figure 2 show the feature extraction of our image using HOG.
Fig. 3. (a) The 68 landmarks on face; (b) the 68 landmarks on test images.
We have performed the basic image transformations such as rotation and scaling
that keep parallel lines parallel i.e. affine transformations using these landmark points.
Figure 4 visualizes the results of the transformation for a given test face images.
By repeating this step many times for different students, the network is learned to
generate a 128-d output feature vector i.e. a list of 128 real-valued numbers which is
used to quantify the faces. If once the network has been trained, it can generate
measurement points for any face even if it has never seen before. Hence, this step needs
626 B. C. Roy et al.
only one time. We used OpenFace [18] model and adapted it to our dataset to get 128
measurement points for the face images.
W:X þ b ¼ 0 : ð1Þ
Where W is a weight vector, W = {w1, w2, w3…wn}, X is the training image, and b is
the bias term.
The decision boundary of SVM is as follows:
X
D XT ¼ yi ai Xi XT þ b0 : ð2Þ
Where yi is the class level of the Xi training image, XT is test images, ai and b0 are
numeric parameters.
If the sign of the Eq. (2) is positive, the SVM predicts, the student is registered
otherwise unregistered. Figure 6 shows liner SVM separating hyperplane.
Fig. 6. SVM separate the registered student from the unregistered student.
Our aim is to find whether a student is present or not. Hence, we use a web camera
to capture the students on the classroom to recognize them using SVM [19] classifier
that take 128-D encoding values from the test image and compares it with the training
Automated Student Attendance Monitoring System 627
image. Based on the comparison, the classifier recognizes the registered student
showing their “Name” and unregistered student showing “Unknown”. To simulate the
model, we use 450 images of 45 students (10 images/student) with multiple illumi-
nations, pose variations and expressions. Then we test the trained model by 45 images
(one image/student). Figure 7 show some sample images of our own dataset.
After training our own dataset, the system was implemented in our classroom
environment. Figure 8 illustrates the recognized output in the real-time situation.
Finally, the system automatically updates the attendance status of students in the Excel.
The result of our proposed system is to give an attendance report automatically. The
generated attendance report with the status is shown in Fig. 9. Here, ‘P’ indicates the
student was present in a class and ‘A’ indicates he/she was absent. Our proposed
system also computes the total number of days of the students who are presented in a
month or a year or an entire semester.
It also provides security such as, who is registered or not. The registered student
is recognized by his/her name and the unregistered student is identified by the
unknown, as shown in Fig. 10.
True Positive is the registered student who is correctly identified as registered. True
Negative is the unregistered student who is correctly recognized. False Positive refers
to the unregistered student identified as registered. Finally, False Negative is the reg-
istered student who is identified as unregistered. The accuracy is 97.83%, however,
precision is 100%, and recall is 97.78%. To confirm the accuracy of the proposed
system, we evaluated and compared it to LFW dataset. Table 2 show the accuracy
comparison with other methods.
Comparing our result with previous results, we found the recognition accuracy is
reasonable. As we were mainly concerned for real-time application and here we found
97.83% accuracy on classroom dataset. Thus, it reveals that the proposed system is
implementable in the real-time environment. The overall performance of the proposed
system on LFW database and our own database is illustrated in Fig. 11.
630 B. C. Roy et al.
The face recognition system is used to solve many problems in our daily activities. This
paper demonstrates that the manual attendance system can be replaced by automated
student attendance system using face recognition. It works efficiently in different
lighting conditions and recognizes the student’s faces and updates the attendance report
accurately. We were interested to use HOG rather than Haar Cascade due to much false
positive output. We use CNN for a more accurate result. For face classification, we use
linear SVM for simplicity instead of using CNN itself. We have accomplished 99.50%
and 97.83% accuracy on LFW and our own student datasets, respectively. In the
future, we will use a deep neural network for feature extraction and more accu-
rate result. Also, we want to introduce a feedback system based on facial expression.
Adding this, the system will help the teacher to improve or modify their teaching
method.
References
1. Lim, T., Sim, S., Mansor, M.: RFID based attendance system. In: IEEE Symposium on
Industrial Electronics & Applications, ISIEA 2009, vol. 2, pp. 778–782. IEEE (2009).
https://doi.org/10.1109/isiea.2009.5356360
Automated Student Attendance Monitoring System 631
2. Kadry, S., Smaili, K.: A design and implementation of a wireless iris recognition attendance
management system. Inf. Technol. Control 36(3), 323–329 (2007)
3. Bhanu, B., Tan, X.: Learned templates for feature extraction in fingerprint images. In:
Proceedings of the IEEE Conference on Computer Vision and Pattern recognition, vol. 2,
pp. 591–596 (2009). https://doi.org/10.1109/CVPR.2001.991016
4. Bhalla, V., Singla, T., Gahlot, A., Gupta, V.: Bluetooth based attendance management
system. Int. J. Innov. Eng. Technol. (IJIET) 3(1), 227–233 (2013)
5. Belhumeour, P.N., Hespanha, J.P., Kriegman, D.J.: Eigenfaces vs Fisherfaces: recognition
using class specific linear projection. IEEE Trans. Pattern Anal. Mach. Intel. 19(7), 711–720
(1997). https://doi.org/10.1109/34.598228
6. Viola, P., Jones, M.J.: Robust real-time face detection. Int. J. Comput. Vision 57(2), 137–
154 (2004). https://doi.org/10.1023/B:VISI.0000013087.49260.fb
7. Boyko, N., Basystiuk, O., Shakhovska, N.: Performance evaluation and comparison of
software for face recognition based on Dlib and Opencv library. In: IEEE Second
International Conference on Data Stream Mining & Processing, Lviv, Ukraine, 21–25
August (2018). https://doi.org/10.1109/dsmp.2018.8478556
8. Berini, D.J., Van Beek, G.A., Arnon, I., Shimek, B.J., Fevens, R.B., Bell, R.L.: Multi-
biometric enrolment kiosk including biometric enrolment and verification, face recognition
and fingerprint matching systems. US Patent 9,256,719, 9 February (2016)
9. Priya, T., Sarika, J.: IJournals: Int. J. Softw. Hardware Res. Eng. 5(9) (2017). ISSN-2347-
4890
10. Tambi, P., Jain, S., Mishra, D.K.: Person-dependent face recognition using histogram of
oriented gradients (HOG) and convolution neural network (CNN). In: International
Conference on Advanced Computing Networking and Informatics. Advances in Intelligent
Systems and Computing, Singapore (2019)
11. Bong, C.W., Xian, P.Y., Thomas, J.: Face recognition and detection using Haars features
with template matching algorithm. In: ICO 2019, AISC 1072, pp. 457–468 (2020). https://
doi.org/10.1007/978-3-030-33585-4_45
12. Navneet, D., Triggs, B.: Histograms of oriented gradients for human detection. In: IEEE
Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2005,
vol. 1, pp. 886–893. IEEE (2005). https://doi.org/10.1109/cvpr.2005.177
13. Hanamsheth, S., Rane, M.: Face recognition using histogram of oriented gradients. Int. J. 6
(1) (2018)
14. Rosebrock, A.: Facial landmarks with dlib, OpenCV, and Python [Electronic resource] -
Access mode. https://www.pyimagesearch.com/2017/04/03/facial-landmarks-dlib-opencv-
python/
15. Kazemi, V., Sullivan, J.: One millisecond face alignment with an ensemble of regression
trees. In: CVPR 2014 Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 1867–1874 (2014). https://doi.org/10.1109/cvpr.2014.241
16. Nam, N.T., Hung, P.D.: Pest detection on traps using deep convolutional neural networks.
In: Proceedings of the 2018 International Conference on Control and Computer Vision
(ICCCV 2018), pp. 33–38. ACM, New York. https://doi.org/10.1145/3232651.3232661
17. Joshua Thomas, J., Pillai, N.: A deep learning framework on generation of image
descriptions with bidirectional recurrent neural networks. In: ICO 2018, AISC 866, pp. 219–
230 (2019). https://doi.org/10.1007/978-3-030-00979-3_22
18. Brandon Amos and his team, Access mode: https://cmusatyalab.github.io/openface/
19. Timotius, I.K., Linasari, T.C., Setyawan, I., Febrianto, A.A.: Face recognition using support
vector machines and generalized discriminant analysis. In: The 6th International Conference
on Telecommunication Systems, Services, and Applications (2011). https://doi.org/10.1109/
tssa.2011.6095397
632 B. C. Roy et al.
20. Hung, P.D., Kien, N.N.: SSD-MobileNet implementation for Classifying Fish Species, ICO
2019, AISC 1072, pp. 399–408 (2020). https://doi.org/10.1007/978-3-030-33585-4_40
21. Taigman, Y., Yang, M., Ranzato, M., Wolf, L.: DeepFace: closing the gap to human-level
performance in face verification. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 1701–1708 (2014). https://doi.org/10.1109/cvpr.2014.
220
22. Schroff, F., Kalenichenko, D., Philbin, J.: FaceNet: a unified embedding for face recognition
and clustering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 815–823 (2015). https://doi.org/10.1109/cvpr.2015.7298682
23. Sun, Y., Wang, X., Tang, X.: Deeply learned face representations are sparse, selective, and
robust. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition,
pp. 2892–2900 (2015). https://doi.org/10.1109/cvpr.2015.7298907
24. Zhong, Y., Chen, J., Huang, B.: Towards end-to-end face recognition through alignment
learning arXiv:1701.0717 (2017). https://doi.org/10.1109/lsp.2017.2715076
Machine Learning Approach to Predict
the Second-Life Capacity of Discarded EV
Batteries for Microgrid Applications
Keywords: Machine learning model First life stage of battery Second life
stage of battery Charging capacity prediction
1 Introduction
Nowadays battery energy storage systems are being used widely in electric vehicles
(EV) and microgrid applications because of their continuously decreasing price due to
the advancement in technology and mass production of these batteries. EV batteries can
be utilized for around 8 years because of their lifetime limitation and after this, a new
battery is required instead of a retired one [1]. These retired batteries from EV are
known as second-life batteries. The lifetime of these retired batteries can be increased
by investigating the remaining capacity so that the requirements of other storage
applications can be fulfilled. These batteries come under electrochemical battery
technologies which are classified into three categories (see Fig. 1): Flow Batteries,
Primary Batteries, and Secondary Batteries. From these several battery technologies,
nickel-metal hydride (Ni-MH) and lithium-ion (Li-ion) are the main battery tech-
nologies used in EVs [2]. Due to the several advantages of Lithium-ion batteries such
as high power density, high energy density, high energy efficiency, high specific
energy, long life cycles, and low daily self-discharge [3–5] these batteries are preferred
by vehicle manufacturers. Among Li-ion batteries, lithium iron phosphate (LiFePO4)
batteries are the most favorable batteries used in several applications because they are
cheaper, less toxic, having flat charge/discharge voltage, comparatively better cycle
life, and more structural stability [6].
The concept of a second life battery is that how to utilize batteries that are discarded
from their first life application. After a lifetime of about 6–8 years, the owners of EVs
or plug-in hybrid electric vehicles (PHEVs) usually replace their vehicle’s batteries due
to low performance, warranty, etc. At this stage, these retired batteries can be reused
again in different applications some of which are mentioned below [7, 8]:
I. For storage of wind and solar power at small or large scale, off-grid or grid-
connected mode for household applications.
II. For peak shaving in industries to reduce power demand.
III. For EV charging to reduce power demand at the time of charging.
IV. For increasing grid capability and stability by reducing the installation of large
cables.
V. For electricity trading in the form of a battery farm for electricity companies.
A second life battery comes into existence when a new battery that is being used in
an EV, can provide only up to 80% of its actual (initial) energy. To know the char-
acteristics of these batteries some performance tests such as hybrid pulse power
characterization test (HPPC), capacity test, electrochemical impedance spectroscopy
(EIS), and open-circuit voltage (OCV) test need to be performed [9].
The transition of the new battery into the second life stage happens due to the aging
phenomenon which increases capacity fade and the resistance growth. There are two
types of battery aging: cycle aging and calendar aging [10]. Calendar aging happens
during energy storage in the battery whereas cycle aging happens due to battery
operation during charge and discharge cycles. These aging mechanisms are affected by
Machine Learning Approach to Predict the Second-Life Capacity 635
Ri
SOHR ¼ 100½% ð1Þ
R0
Cmax;i
SOHC ¼ 100½% ð2Þ
C0
Where,
Ri = battery resistance at any time interval i
R0 = battery initial resistance
Cmax;i = maximum capacity of the battery at any time interval i
C0 = battery initial capacity (nominal capacity or rated capacity)
To estimate SOH accurately and reliably, the charging curve based Gaussian
process regression (GPR) model was proposed by [19]. [20] investigated that a fuzzy
logic-based approach is also a reliable solution to determine battery SOH through
offline discharge testing of the batteries. [21] used a support vector machine (SVM) to
predict SOH for Li-ion batteries. To handle a large set of data for a complex and non-
linear system, a neural network was proposed by [22, 23]. For accurate prediction of
SOH in different working conditions, it uses a robust algorithm. A neural network-
based SOH estimation method was also proposed by [24]. They used a recurrent neural
network (RNN) for the prediction of deterioration in the performance of the Li-ion
battery. A backpropagation neural network (BPNN) was used by [25] for SOH esti-
mation. Long short-term memory (LSTM) based prediction of remaining useful life of
Li-ion cells under different operation conditions, was proposed by [26]. [23] proposed a
deep convolutional neural network (DCNN) for the assessment of SOH during a charge
636 A. Bhatt et al.
cycle based on the voltage, current, and capacity measurements. For calculating
capacity, the coulomb counting method is used to integrates the charge current over-
time for the entire charge cycle. [27] used feedforward neural network, convolutional
neural network (CNN), and LSTM for estimating SOH based on capacity measurement
from charging profiles of the battery. Similarly, the Li-ion battery’s remaining useful
life prediction by using LSTM-RNN is also proposed by [28].
This paper shows how multi-layer perceptron (MLP), a machine learning model,
can predict the second-life stage capacity of a lithium-ion battery from its first life
stage. It also shows that how prediction results can be improved by improvising input
parameters. For the proposed machine learning model, activation functions used are
leaky rectified linear unit (ReLU) and ReLU activation functions.
For the present work, we are using the datasets of commercially available lithium-ion
18650 rechargeable batteries from Prognostics Data Repository under NASA [29],
which is a public source. In this study, the B0005 Li-ion battery was considered for
obtaining experimental data from three operational tests: charging, discharging, and
impedance at room temperature of 24 °C. The test profile of the B0005 battery is
mentioned in Table 1. During the charging process, the constant current constant
voltage (CCCV) method was used where, we have to provide a constant current of 1.5
A until the voltage reaches up-to 4.2 V, and then charging at a constant voltage will
remain to continue until the charge current reduces to 20 µA. For the discharging
process, a constant current of 2 A was provided until the voltage reached 2.7 V and
2.2 V for battery B0005 and B0007 respectively.
Machine Learning Approach to Predict the Second-Life Capacity 637
For the present work, out of 3 operational test profiles, only the charging profile of
the battery was considered for creating data to train, validate, and test the machine
learning model. Based on the charging profile of batteries during several cycles, the
degradation characteristics of the battery can be analyzed. During the charging process
of battery B0005, the capacity at cycle 1 is maximum whereas at cycle 115 the capacity
is reduced to a minimum value (see Fig. 3). Due to the continuous switching between
charging and discharging cycles [30], a sudden increase in battery capacity is also
shown in battery degradation characteristics which are also known as the capacity
regeneration phenomenon [31].
This study focuses on the capacity prediction of the second life stage of the battery
which means that the SOH of the battery is almost lesser than 80% of its initial state.
For this purpose, we are considering that the parameters (input and output) associated
with the first life stage of the battery will be used for training purposes of the machine
learning model, and for testing purposes we will consider the parameters of the second
638 A. Bhatt et al.
life stage of battery. Here we are proposing two different cases of input parameters for
the proposed machine learning model and comparing the results based on the perfor-
mance indicators.
3.1 Case 1
Case 1 consists of 3 input parameters, charging voltage (V), charging current (I), and
the number of cycles (K). Based on these 3 input parameters we are trying to predict the
final charging capacity of each cycle during the second life stage of the battery. Here,
for input parameters, we are arranging our datasets into one lag (previous) cycle dataset
(K − 1) and one current cycle dataset (K), and based on these arrangements we are
predicting final charging capacity for future (next) cycle (K + 1) (see Fig. 4). By
arranging datasets in this manner we are providing a total of 6 inputs for our model.
Here, initial 60 cycles were used for creating training datasets, 61–70 cycles were used
for validation datasets and 71–115 cycles were used for testing datasets. After 70
cycles, the final SOH of B0005 is around 81% so we are considering that up to 70
cycles battery can be considered at its first life stage whereas after 70 cycles we are
considering that our battery is in the second life stage because SOH is lower than 80%.
3.2 Case 2
Case 2 also consists of 3 input parameters but here instead of the number of cycles, we
are using capacity (C) as an input. Here, we are arranging our datasets into a 5 lag
(previous) cycle dataset (K − 5 to K − 1) and one current cycle dataset (K), and based
on these arrangements we are predicting final charging capacity for future (next) cycle
(K + 1) (see Fig. 5). Due to this arrangement, the total 18 inputs we are providing to
our machine learning model. The division of training, validation, and testing datasets
are the same as mentioned in Case 1.
Machine Learning Approach to Predict the Second-Life Capacity 639
After preparing input and output data from the Prognostics Data Repository under
NASA [29], the preprocessing of data through an appropriate normalization technique
is important for making a robust and efficient machine learning algorithm. Normal-
ization of input and output data will be in the range of 0 to 1, as given by Eq. (3).
x xmin
xðnormalized Þ ¼ ð3Þ
xmax xmin
Where,
xmax = upper limit of input and output value ‘x’
xmin = lower limit of input and output value ‘x’
This study used the Multi-Layer Perceptron (MLP) model which is a feed-forward
artificial neural network (ANN) model. It maps multiple input data with the output data.
It consists of three basic layers: an input layer, hidden layer, and output layer (see
Fig. 6), where each layer consists of several nodes and is interconnected with each
other. By increasing the number of hidden layers we can convert our machine learning
model into a deep neural network. In MLP every node is known as an artificial neuron
which is linked with a nonlinear activation function such as sigmoid, tanh, linear,
rectified linear unit (ReLU), Leaky ReLU, etc.
For training the network, MLP employs a supervised learning-based, back-
propagation algorithm, which performs propagation and weight update tasks [32]. The
back-propagation is an accurate learning algorithm for nonlinear relationships between
640 A. Bhatt et al.
input and output. In this algorithm, the output values during each iteration are com-
pared with the actual value and based on the deviation in output to the actual value, an
error function is generated. By feeding back the error function, the model tries to do
some changes in the weights so that the error function can reduce to a certain level.
In this study, we are using two activation functions: ReLU and Leaky ReLU.
Table 2 shows the mathematical and graphical representations of the activation func-
tions used. For deep learning applications, the most widely used activation function is
ReLU because of its faster learning ability. In comparison with Sigmoid and tanh
activation functions, ReLU performs and generalize in a better way for deep learning
applications [33]. It can easily optimize with gradient descent methods because of its
linear properties. This function rectifies the negative input values by imposing zero
value to them and eliminates the gradient vanishing problem which makes this function
better than several other activation functions. Apart from this ReLU has a significant
limitation that sometimes it becomes fragile during training which causes some gra-
dients to die. This causes some neurons to be dead which hinders the learning process.
To overcome this issue, we can use the Leaky ReLU activation function. In Leaky
ReLU, we use the a (alpha) parameter to overcome the problem of the dead neuron.
Due to this parameter, gradients never become zero during the entire training time.
Machine Learning Approach to Predict the Second-Life Capacity 641
The selected machine learning architecture (MLP) for the simulation is designed and
coded by using Python Jupyter Notebook which is an open-source web application that
allows us to create and share documents that contain live code, equations, visualiza-
tions, and narrative text. It performs, data cleaning and transformation, numerical
simulation, statistical modeling, data visualization, machine learning, and many other
tasks. In this software, NumPy, TensorFlow, Matplotlib, and Keras libraries are
applied. (https://www.python.org) (https://jupyter.org). The software version we are
using here is Python 3.7.6. System configuration is shown in Table 3 for the present
analysis.
6 Result Analysis
For evaluating the performance of the MLP model, we are using Mean Square Error
(MSE), Root Mean Square Error (RMSE), Mean Absolute Error (MAE), and R-
Squared (R2) as the performance indicators. Based on the input parameters, we are
comparing the performance of the machine learning model for two different cases
mentioned as Case 1 and Case 2.
642 A. Bhatt et al.
Table 6. Comparative analysis of the MLP model for case 1 and case 2 based on error indicators
Error indicators Case 1 Case 2
MLP model @ Epochs 38 MLP model @ Epochs 91
Predicted Predicted Predicted Predicted
capacity @ capacity @ capacity @ capacity @
alpha = 0.06 alpha = 0.01 alpha = 0.06 alpha = 0.04
MSE 0.00007758 0.00005706 0.00003579 0.00002959
RMSE 0.00880823 0.00755415 0.00598296 0.00544033
MAE 0.00769091 0.00634489 0.00487101 0.00408353
R-Squared (R2) 0.9170 0.9390 0.9617 0.9683
7 Conclusion
This paper aims to find a suitable machine learning-based capacity prediction model for
the second life stage of the battery from its first life stage. In this regard, this work
proposes an MLP model in which feeding of input parameters being done in two
different ways. These two different ways are categorized as Case 1 and Case 2. The
result shows that Case 2 consisting of charging voltage, charging current, and capacity
as input parameters along with the arrangement of datasets in 5 lag cycle values and 1
current (present) cycle value, can predict the capacity for the second life stage of the
battery more reliably and accurately in comparison with Case 1. This analysis shows
that charging voltage and charging current are the important input parameters for
predicting the capacity for the next cycle. We also noticed that instead of taking cycle
numbers as an input parameter if we are considering capacity as an input from some
previous (lag) cycles, then the machine learning model can predict the capacity for the
future cycles more accurately and reliably. Therefore, we can conclude that the pro-
posed machine learning model can predict the capacity of the second life stage of the
battery from its first life stage up-to several extents.
References
1. Strickland, D., Chittock, L., Stone, D.A., Foster, M.P., Price, B.: Estimation of transportation
battery second life for use in electricity grid systems. IEEE Trans. Sustain. Energy 5(3), 795–
803 (2014)
2. Abdel-Monem, M., Hegazy, O., Omar, N., Trad, K., Breucker, S.D., Bossche, P.V.D.,
Mierlo, J.V.: Design and analysis of generic energy management strategy for controlling
second-life battery systems in stationary applications. Energies 9(11), 889 (2016)
3. Abdel-Monem, M., Trad, K., Omar, N., Hegazy, O., Bossche, P.V.D., Mierlo, J.V.:
Influence analysis of static and dynamic fast-charging current profiles on ageing performance
of commercial lithium-ion batteries. Energy 120, 179–191 (2017)
4. Luo, X., Wang, J., Dooner, M., Clarke, J.: Overview of current development in electrical
energy storage technologies and the application potential in power system operation. Appl.
Energy 137, 511–536 (2015)
5. Bhatt, A., Tiwari, S., Ongsakul, W.: A review on re-utilization of electric vehicle’s retired
batteries. In: Proceedings of Conference on Ind. Commercial Use Energy, ICUE, vol. 2018-
October, pp. 1–5. IEEE, Phuket, Thailand (2019)
6. Tan, Q., Lv, C., Xu, Y., Yang, J.: Mesoporous composite of LiFePO4 and carbon
microspheres as positive-electrode materials for lithium-ion batteries. Particuology 17, 106–
113 (2014)
7. Olsson, L., Fallahi, S., Schnurr, M., Diener, D., Loon, P.V.: Circular business models for
extended EV battery life. Batteries 4(4), 57 (2018)
8. Martinez-Laserna, E., Gandiaga, I., Sarasketa-Zabala, E., Badeda, J., Stroe, D., Swierczyn-
ski, M., Goikoetxea, A.: Battery second life: hype, hope or reality? a critical review of the
state of the art. Renew. Sustain. Energy Rev. 93, 701–718 (2018)
9. Abdel-Monem, M., Hegazy, O., Omar, N., Trad, K., Bossche, P.V.D., Mierlo, J.V.: Lithium-
ion batteries: comprehensive technical analysis of second-life batteries for smart grid
applications. In: 19th European Conference on Power Electronics and Applications, EPE
2017 ECCE Europe, vol. 2017-January, pp. 1–16. IEEE, Poland (2017)
Machine Learning Approach to Predict the Second-Life Capacity 645
10. Ali, M.C.: Exploring the potential of integration quality assessment system in construction
(qlassic) with ISO 9001 quality management system (QMS). Int. J. Qual. Res. 8(1), 73–86
(2014)
11. Stiaszny, B., Ziegler, J.C., Krauß, E.E., Zhang, M., Schmidt, J.P., Ivers-Tiffée, E.:
Electrochemical characterization and post-mortem analysis of aged LiMn 2O4-
NMC/graphite lithium ion batteries part II: Calendar aging. J. Power Sources 258, 61–75
(2014)
12. Waag, W., Käbitz, S., Sauer, D.U.: Experimental investigation of the lithium-ion battery
impedance characteristic at various conditions and aging states and its influence on the
application. Appl. Energy 102, 885–897 (2013)
13. Giordano, G., Klass, V., Behm, M., Lindbergh, G., Sjoberg, J.: Model-based lithium-ion
battery resistance estimation from electric vehicle operating data. IEEE Trans. Veh. Technol.
67(5), 3720–3728 (2018)
14. Thomas, J.J., Pillai, N.: A deep learning framework on generation of image descriptions with
bidirectional recurrent neural networks. In: International Conference on Intelligent
Computing & Optimization, ICO 2018, vol. 866, pp. 219–230. Springer, Pattaya (2018)
15. Nandagopalan, S., Kumar, P.K.: Deep convolutional network based saliency prediction for
retrieval of natural images. In: International Conference on Intelligent Computing &
Optimization, ICO 2018, vol. 866, pp. 487–496. Springer, Pattaya (2018)
16. Birla, S., Kohli, K., Dutta, A.: Machine learning on imbalanced data in credit risk. In: 7th
IEEE Annual Information Technology, Electronics and Mobile Communication Conference,
IEEE IEMCON 2016, pp. 1–6. IEEE, Canada (2016)
17. Rodriguez-Aguilar, R., Marmolejo-Saucedo, J.A., Vasant, P.: Machine learning applied to
the measurement of quality in health services in Mexico: the case of the social protection in
health system. In: International Conference on Intelligent Computing & Optimization, ICO
2018, vol. 866, pp. 560–572. Springer, Pattaya (2018)
18. Thomas, J.J., Ali, A.M.: Dispositional learning analytics structure integrated with recurrent
neural networks in predicting students performance. In: International Conference on
Intelligent Computing & Optimization, ICO 2019, vol. 1072, pp. 446–456. Springer, Koh
Samui (2019)
19. Yang, J., Xia, B., Huang, W., Fu, Y., Mi, C.: Online state-of-health estimation for lithium-
ion batteries using constant-voltage charging current analysis. Appl. Energy 212, 1589–1600
(2018)
20. Singh, P., Kaneria, S., Broadhead, J., Wang, X., Burdick, J.: Fuzzy logic estimation of SOH
of 125Ah VRLA batteries. In: INTELEC, International Telecommunications Energy
Conference (Proceedings), pp. 524–531. IEEE, USA (2004)
21. Nuhic, K., Terzimehic, A., Soczka-Guth, T., Buchholz, T., Dietmayer, M.: Health diagnosis
and remaining useful life prognostics of lithium-ion batteries using data-driven methods.
J. Power Sources 239, 680–688 (2013)
22. Hannan, M.A., Lipu, M.S.H., Hussain, A., Saad, M.H., Ayob, A.: Neural network approach
for estimating state of charge of lithium-ion battery using backtracking search algorithm.
IEEE Access 6, 10069–10079 (2018)
23. Hannan, M.A., Hoque, M.M., Hussain, A., Yusof, Y., Ker, P.J.: State-of-the-art and energy
management system of lithium-ion batteries in electric vehicle applications: issues and
recommendations. IEEE Access 6, 19362–19378 (2018)
24. Eddahech, A., Briat, O., Bertrand, N., Delétage, J.Y., Vinassa, J.M.: Behavior and state-of-
health monitoring of Li-ion batteries using impedance spectroscopy and recurrent neural
networks. Int. J. Electr. Power Energy Syst. 42(1), 487–494 (2012)
646 A. Bhatt et al.
25. Yang, D., Wang, Y., Pan, R., Chen, R., Chen, Z.: A neural network based state-of-health
estimation of lithium-ion battery in electric vehicles. Energy Procedia 105, 2059–2064
(2017)
26. Zhang, Y.Z., Xiong, R., He, H.W., Pecht, M.: Validation and verification of a hybrid method
for remaining useful life prediction of lithium-ion batteries. J. Clean. Prod. 212, 240–249
(2019)
27. Choi, Y., Ryu, S., Park, K., Kim, H.: Machine learning-based lithium-ion battery capacity
estimation exploiting multi-channel charging profiles. IEEE Access 7, 75143–75152 (2019)
28. Lui, Y., Zhao, G., Peng, X., Hu, C.: Lithium-ion battery remaining useful life prediction with
long short-term memory recurrent neural network. In: Annual Conference on Prognostics
and Health Management Society, vol. 1, pp. 1–7 (2017)
29. Bole, B., Kulkarni, C.S., Daigle, M.: Adaptation of an electrochemistry-based Li-ion battery
model to account for deterioration observed under randomized use. In: PHM 2014 -
Proceedings of Annual Confernce Prognostics and Health Management Society, pp. 502–
510 (2014)
30. Liu, Z., Zhao, J., Wang, H., Yang, C.: A new lithium-ion battery SOH estimation method
based on an indirect enhanced health indicator and support vector regression in PHMs.
Energies 13(4), 830 (2020)
31. Qin, T., Zeng, S., Guo, J., Skaf, Z.: A rest time-based prognostic framework for state of
health estimation of lithium-ion batteries with regeneration phenomena. Energies 9(11), 1–
18 (2016)
32. Khumprom, P., Yodo, N.: A data-driven predictive prognostic model for lithium-ion
batteries based on a deep learning algorithm. Energies 12(4), 1–21 (2019)
33. Nwankpa, C.E., Ijomah, W., Gachagan, A., Marshall, S.: Activation Functions: Comparison
of Trends in Practice and Research for Deep Learning, pp. 1–20 (2018)
Classification of Cultural Heritage Mosque
of Bangladesh Using CNN and Keras Model
1 Introduction
We stay in an age of technology and technology has been developed based on many
formed and unformed data and that data can be collected from many historical sources.
These ancient information are needed to be stored for future purposes and for a nation’s
pride. Research fields can’t be thought without historical information. Again, we have
responsibilities to increase awareness to save cultural heritage resources. Different
researchers and organizations are doing their level best to keep continuing researches to
facilitate cultural heritages.
Bangladesh is a country which is bestowed with huge cultural heritages. There are
several monuments, ancient mosques, places, properties that are part of cultural heritages.
These resources are needed to be documented and preserved for various purposes. Again,
country people have affection to know about their heritages. Study of cultural heritages
help to know about hidden cultural treasures and to increase knowledge.
Our country occupies a large number of mosques. Most of them have been built
recently or after British period. But some had been built at the time of British period or
before that period. Several organizations have recognized such mosques as cultural
heritages. Some mosques have not been recognized yet. All of these mosques are our
valuable resources. These mosques contain valuable historic information. They need
proper researches and analysis for storage. Our objective is to collect the images of
these mosques, analyze them and classify them depending on their patterns.
2 Problem Statement
Nesterov: Nesterov is one kind of momentum that facilitates convex functions with
potential or powerful convergence. It performs better than typical momentum [19].
3.1 Dataset
Our dataset contains 196 images of buildings which are classified into two types i.e. the
ancient mosques and the recent mosques including other buildings. These images are
mainly collected using Google and Wikipedia information.
Classification of Cultural Heritage Mosque of Bangladesh 651
Among the images, 176 were used for the training phase and the remaining were
used for the test phase. 20% of the training images were used for validation phase
which had been selected randomly.
All the images have been used in our experiment are color images and have been
resized using appropriate image processing tools.
Our train and test datasets are: (Fig. 2 and Fig. 3).
3.2 RStudio
There are many programming languages available to perform image classification.
Among them we have used R programming language to implement our experiment.
RStudio has been used as the editor to write R codes [22].
3.3 Methodology
A convolutional neural network is a class of artificial neural network (ANN) where
connectivity patterns between neurons resemble the structure of the biological brain
visual cortex [6, 23]. Its typical structure is composed of several layers. The primary
layers are the convolutional and the pooling layers. The end layers consist of the fully
connected layer and the output layer. Feature extraction are basically done by the
convolutional and the pooling layers and classification is performed based on the
extracted features using fully connected layer [5, 6].
Convolutional Neural Network provides ideas about sparse interaction, parameter
sharing and equivariant representations [5, 7]. It can automatically extract features from
images and classify them based on the extracted features [7].
Several approaches are needed to be considered while applying CNN to input
images i.e. data driven approach and model driven approach. Data driven approach
uses each dimension as channel and then executes 1D convolution on them. The
outputs of each channel are flattened after convolution and pooling. On the other hand,
inputs are resized to virtual 2D images by model driven approaches to adopt a 2D
convolution [7].
The more accurate version of the approach can be defined from the following
figure: (Fig. 4).
4.1 Result
Here is our fitted trained data plot given in the following figure: (Fig. 6).
Here is our predicted and actual train data given in the confusion matrix Table 2:
We have used 20 images as the test images to check our experiment. Here is the
predicted and actual train data given in the confusion matrix Table 3:
4.2 Discussion
Our objective is to classify train images and get a better accuracy to make our
experiment meaningful. From our experiment, classification of train images was
accurate enough and it was 100% accurate. It was shown in Table 1 in the previous
result section. Our model can easily classify train images based on the extracted fea-
tures. Among 176 trained images, 35 images have been classified to ancient mosques
class and 141 images have been categorized to the other buildings class. Again, the
accuracy of the classification between both classes was good and appreciable.
20 images have been used to test our experiment. Among them 15 images have
been given to ancient mosque classes and 5 images have been given to other buildings
classes for the test purpose. Among the 15 ancient mosque images, 13 images have
been correctly classified and 2 images haven’t been classified correctly. The 2 images
that have been misclassified are shown by the following Fig. 7:
Again, the 5 others buildings images have been classified correctly. We have got
86.67% accuracy for the ancient mosque class images and 100% for other buildings
class images. So far, the total accuracy is 90%.
Finally, a test has been done for negative images to verify our model. 20 images
have been provided for the test. The ancient mosque image class has been considered to
be the positive image class and another class which contains the buildings images has
656 M. A. Saadat et al.
been considered as the negative image class. 15 images have been given to positive
image class which have been actually negative and 5 images have been given to the
negative image class which have been actually negative. Among the 15 images, 14
images have been predicted as the negative class images and 1 image has been iden-
tified as the positive class image which has been also a negative class image from our
perspective. The one image that has been predicted as the positive class images and the
14 images that have been predicted negative class images have been shown by the
following figure: (Fig. 8 and Fig. 9).
It is known that CNN is a good feature extractor as well as a good classifier. So, we
didn’t need to use additional feature extractor and classifier for the entire experiment.
This is one of the reasons to choose CNN instead of others. Besides, our proposed
methodology provided better accuracy.
Although the quantity of images was small, the accuracy of the result was well
enough. We focused on the quality of the research rather than the quantity of data.
Again, the sizes of data were not that small to fulfill our requirements. We utilized
ancient mosque images of our country which are essential part of our heritages and add
a new dimension to cultural heritage related researches.
Classification of Cultural Heritage Mosque of Bangladesh 657
6 Conclusion
Our proposed methodology has performed well as per our expectations. Convolutional
neural network along with keras has made the mechanism more efficient. Again, feature
extraction has been done with that convolutional neural network without adding any
additional feature extractor. So, we can say that convolutional neural network is a
complete package for classification jobs and we are satisfied with our experimental
performance.
7 Future Scope
1. Other researchers may get motivated from the experiment.
2. Documentation and preservation of the ancient mosques name may be done based
on the current research.
References
1. Lillesand, T.M., Kiefer, R.W., Chipman, J.W.: Remote Sensing and Image Interpretation,
Open Access (2004)
2. Jaswal, D., Sowmya, V., Soman, K.P.: Image classification using convolutional neural
networks. Int. J. Sci. Eng. Res. 5(6), June 2014
3. Bengioy, Y., Courville, A., Vincent, P.: Representation learning: a review and new
perspectives. IEEE Trans. Pattern Anal. Mach. Intell. (2014)
4. Schmidhuber, J.: Deep learning in neural networks: an overview. Neural Netw. (2014)
5. LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521, 436–444 (2015)
6. Llamas, J., Lerones, P.M., Medina, R., Zalama, E., Gómez-García-Bermejo, J.: Classifica-
tion of architectural heritage images using deep learning techniques. Appl. Sci. 26 September
2017
7. Wang, J., Chen, Y., Hao, S., Peng, X., Hu, L.: Deep Learning for Sensor-based Activity
Recognition: A Survey. Elsevier (2017)
8. Du, T., Shanker, V.K.: Deep Learning for Natural Language Processing
9. Wang, J.: Deep Learning - An Artificial Intelligence Revolution, Ark Invest, June 2017
10. Convolutional Neural Networks (LeNet) – DeepLearning 0.1 documentation, LISA Lab
(2013)
11. Aghdam, H.H., Heravi, E.J.: Guide to Convolutional Neural Networks: A Practical
Application to Traffic-Sign Detection and Classification. Springer, Cham (2017)
12. Nagi, J., Ducatelle, F., Di Caro, G.A., Cireşan, D., Meier, U., Giusti, A., Nagi, F.,
Schmidhuber, J., Gambardella, L.M.: Max-pooling convolutional neural networks for vision-
based hand gesture recognition. In: IEEE International Conference on Signal and Image
Processing Applications (ICSIPA) (2011)
13. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a
simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1929–1958
(2014)
14. van den Oord, A., Dieleman, S., Schrauwen, B.: Deep Content-Based Music Recommen-
dation (2013)
658 M. A. Saadat et al.
15. Collobert, R., Weston, J.: A unified architecture for natural language processing: deep neural
networks with multitask learning. In: Proceedings of the 25th International Conference on
Machine Learning, ICML 2008, New York, NY, USA (2008)
16. Quora. https://www.quora.com/What-is-an-epoch-in-deep-learning
17. Machine Learning Mastery (2018). https://machinelearningmastery.com/difference-between-
a-batch-and-an-epoch/
18. Stack Exchange (2018). https://stats.stackexchange.com/questions/153531/what-is-batch-
size-in-neural-network
19. CS231n Convolutional Neural Networks for Visual Recognition. http://cs231n.github.io/
neural-networks-3/
20. Dishashree Gupta, “Analytics Vidhya,” 23 October 2017. https://www.analyticsvidhya.com/
blog/2017/10/fundamentals-deep-learning-activation-functions-when-to-use-them/
21. Data Science Bootcamp. https://medium.com/data-science-bootcamp/understand-the-
softmax-function-in-minutes-f3a59641e86d
22. Chollet, F., Allaire, J.J.: Deep Learning with R, Manning Publications Co.
23. Matsugu, M., Mori, K., Mitari, Y., Kaneda, Y.: Subject independent facial expression
recognition with robust face detection using a convolutional neural network. ScienceDirect
16(5–6) (2003)
Classification of Orthopedic Patients Using
Supervised Machine Learning Techniques
1 Introduction
Machine learning algorithm has been successfully used in various medical fields like
disease prediction, classification, and therapy and medicine recommendation. To
extend the service of machine learning in medical field, this paper addresses a lot of
supervised machine learning models to classify the orthopedic patients according to
their health condition. It seems easy to choose a classifying model based on their
accuracy. But in most cases accuracy is not the final metrics to judge a classifier model.
Accuracy sometimes hides the predicting capability of models. More analysis of dif-
ferent evaluation metrics needed to validate a classifier model. The purpose of this
work is to evaluate and compare the performance of popular supervised machine
algorithms which have been applied to classify orthopedic patients.
2 Background
Biomechanics is known as the study of the composition, motion of the living beings of
biological systems [1]. Biomechanics help to understand the physical condition of a
person. According to the American Academy of Orthopedic surgeons (AAOS), four out
of five adults feel lower backside pain at different point during their lives and many of
them have common intervertebral disc disorders [1]. Disc hernia and spondylolisthesis
are examples of common vertebral disease which cause intense pain. One of the major
causes of LBP is intervertebral disk herniation [2]. The condition when the gel like
material (Nucleus Pulposus) gets squeezed out through fractures in outer wall of
intervertebral disc is known as Lumbar Disc Herniation [3]. Figure 1 shows an example
of hernia where. Disk hernia causes pressure to the nerves which leads to severe pain.
Fig. 2. Spondylolisthesis
Classification of Orthopedic Patients 661
Both conditions can squeeze the spinal cord or nerve roots and cause pain [5]. But
they create different impact on human body but shows similar symptoms. Hence,
diagnosis of hernia and spondylolisthesis may cause misclassification. In this study,
different types of machine learning algorithm are applied to classify patients. The goal
of this work is to answer the following question:
“From selected algorithm which one is best for classifying orthopedic patients from
their biomechanical studies?
This study will help specialist to classify patients easily and accurately according to
their biomechanical feature.
3 Literature Review
Application of machine learning in medical field has been very essential to specialists.
Different machine learning algorithms were used in different medical fields. Recently a
study has been conducted by applying a lot of machine learning techniques named
Naive Bayes, Multilayer perceptron, Support vector machine, Logistic Regression,
Quadratic Discriminant Analysis, K-Nearest Neighbor, Adaptive Boosting, Gradient
Descent, Decision Tree, Random forest to classify orthopedic patients. The classifi-
cation was done in two stage where the best performed algorithm was decision tree as it
stood out with 99% accuracy [6].
Nureedeen et al. [7] presented an approach to classify orthopedic patients in one
stage. In this work, different evaluation metrics were evaluated and the best performed
algorithm was Naïve Bayes classifier with 83.4% accuracy, 100% specificity,16%
error. Rocha et al. [8] conducted an experiment to classify orthopedic patients by the
use of support vector machine, Multilayer Perceptron at SIMPATICO platform. And
the result showed that the classifiers are better than standalone classifier. Alwaneh et al.
[8] proposed a method to classify the hernia and normal disk from MRI image which
contains suspected region. The work was done by conducting different task like image
acquisition, image annotation, enhancement and extraction of region of Interest, feature
extraction, and image classification. S. Liao et al. [4] proposed an automatic system to
detect spondylolisthesis by learning based detector. Synthesized spondylolisthesis
image was used to train the learning based detector.
4 Methodology
Most of the algorithms in machine learning are parametric but K-Nearest Neighbor is
non-parametric. It classifies the data based on number of majority of vote it gets from
its closest neighbors [15]. This algorithm predicts the class by using the Euclidean
distance. Choosing ‘K’ as the certain value of the nearest neighbors and these neigh-
bors refers to the core of deciding the values and the value of ‘K’ can influence the final
result. The equation for selecting ‘K’ is given below:
p
K¼ n ð3Þ
Decision Tree is the most powerful and a widely used classifier. Instances classified
start from the root node and rank by their attribute values [16]. It partitions data into
Classification of Orthopedic Patients 663
subsets and the partition continues until there is no partition possible. The partition is
done using Binary Splitting. The splitting is completed when the subset at a node all
has the same value of the target variable, or when no longer splitting is possible.
Decision Tree uses the entropy function for characterizing impurity of a dataset.
Random forest classifier is a supervised classification model. Random forest, as its
name indicates, consists of a set of single decision trees which are created from ran-
domly selected subset of training set. Every single tree in the random forest classifier
makes a class prediction about the data. And the class getting the most of the votes
becomes the model’s predictions [17]. The large number of uncorrelated trees operates
as a committee to get the higher chance of accurate results. RF gives high accurate
results and it learns very fast. It is very efficient on large data sets. Random forest uses
Gini index for deciding the final class of each tree. If data set T contains examples from
n class Gini index, Gini (T) is defined as
In the era of machine learning, the problem of different statistical classification, the
confusion matrix is known as the error matrix, which is a specific table layout that
allows visualization of the performance of an algorithm. Each row of the confusion
matrix describes the instances in predicted class and each column describes the
instances in the actual class The name is originated from the idea that it may confuse
during predict the classes such as mislabeling one class as another. It is a particular
kind of contingency table, containing two dimensions like “actual” and “predicted”,
where identical sets of “classes” are in both dimensions [20].
Specificity also called the true negative rate measures the ratio of actual negatives
class that are accurately identified like the percentage of healthy people those are
664 N. Jahan et al.
correctly classified as not having the disease [21]. Specificity is defined as False
Positive Rate(FPR) represents the proportion of negative class level that are mistakenly
determined as positive, with respect to the all negative class level.
False Positive
FPR = ð6Þ
False Positive þ True Negative
Sensitivity also called the true positive rate, the recall, or probability of detection in
some fields measures the ratio of actual positives class that are correctly classified as
positive [22]. True positive rate defines the proportion of the positive samples that are
correctly determined as positive class level, with respect to the all positive class level.
False Positive
TPR = ð7Þ
False Positive þ True Positive
4.4 Dataset
In this work, the datasets are downloaded from the website [24] each containing 310
instances. Each instance contains six features named pelvic incidence, pelvic tilt
numeric, lumbar lordosis angle, sacral slope, pelvic radius and degree spondylolisthesis
[25]. The Classification is done in two stage. In first stage, patients had no spinal
vertebral issues were labeled as normal and the rests were labeled as abnormal. In
second stage patients were classified according to their disease and normal patients.
The feature of this dataset is used as the learning parameters. Table 1 and Table 2
shows the sample dataset for first analysis.
Table 3 and Table 4 shows the sample dataset for second analysis where classes are
classified into three levels named normal, disk hernia, spondylolisthesis.
5 Results
In this study, seven classifiers are used named Naive Bayes (NB), K-Nearest Neighbor
(k-NN), Support Vector Machine (SVM), Decision Tree (DT) Random Forest (RF),
Logistic Regression (LR), Multilayer Perception (MLP). Python (3.7) as well as its
related packages are used in this work for classification. The whole dataset is used for
two stage classification. For first stage classification analysis, dataset is classified into
two class normal and abnormal. For second stage classification, dataset is classified into
three class normal, hernia and spondylolisthesis. For each stage, 30% of total data is
used for test data. The six attribute obtained from the dataset have been used as the
input attribute of the system for classification. After training the system, firstly the
system was tested to evaluate whether the person is in normal state or not. After that,
the second test is conducted for classifying the patient according to their disease like
hernia, spondylolisthesis.
Test data is 30%, so 93 samples out of 310 is used for predicting the model.
Different evaluation metrics like accuracy, specificity, sensitivity, AUC, mislabeled
data points are calculated to measure the performance. An algorithm is chosen as final
classifier with higher metrics. Figure 3 shows the histogram analysis of different
evaluation metrics of first stage analysis.
666 N. Jahan et al.
From this analysis, it can be seen that different model outperformed with great
accuracy and different evaluation metrics and different model stood out with similar
accuracy, specificity, sensitivity. In this case, it is tough to find a suitable model which
can be declared as the best model. So, other metrics should be analyzed.
AUROC (Area under the Receiver Operating Characteristics). Curve may come
with a good solution in this case. It is one of the most important evaluation metrics to
check any classification model’s performance. The AUC-ROC curve is plotted with
TPR against the FPR. TPR is plotted on y-axis and FPR is plotted on the x-axis. The
dotted navy blue line represents the threshold, which is set by value .5. The more area
captured by a model, the more performance model will show. The ROC of different
model is drawn in a curve and will help to choose a model. The AUC-ROC curve is
drawn below: (Fig. 4).
By analyzing above curve, it can be seen that the most area is covered by K-Nearest
Neighbor model. It is also figured out that only MLP and KNN model show a linear
line while other lines are nonlinear. As KNN results in higher value in accuracy,
specificity, sensitivity, AUC so the chosen classifier is K-Nearest Neighbor model.
Figure 5 shows the different performance for second analysis.
From the above analysis, it can be seen that most of the algorithm outperformed
with great performance like accuracy, specificity, sensitivity. Among these, k-nearest
neighbor stood out with 83% accuracy, 84% specificity, 83% sensitivity, which is
highest among all other model. So, for second analysis k-nearest neighbor is the right
classifier to classify orthopedic patients.
6 Conclusion
In this work, several machine learning models are applied to classify orthopedic
patients according to their biomechanical features. To choose a best classifier, their
different evaluation metrics are analyzed. As there is not enough application of machine
learning in the field of classification of orthopedic patients. Six biomechanical features
from preprocessed dataset have been used as learning parameters for training the
model. For first analysis, a complexity was raised as many classifiers produced same
accuracy, specificity, sensitivity. In this case, an additional evaluation metric AUC-
ROC is drawn to determine suitable model. As the KNN model captures most of the
area in ROC curve, so it is chosen as final classifier for first analysis. In second
analysis, KNN stands with highest accuracy, specificity, sensitivity that’s why it is
chosen as final classifier. This work is more significant in this way that it is evaluated
with different metrics. This analysis will help specialist to find out the best classifier in
an accurate and faster way. In concluding remark we would like to say that with the
help of this research possible orthopedic patients will be classified and thus save many
lives.
668 N. Jahan et al.
References
1. Ghosh, S., Alomari, R., Chaudhary, V., Dhillon, G.: Composite features for automatic
diagnosis of intervertebral disc herniation from lumbar MRI. In: Annual International
Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 5068–
5071, August 2011
2. Alomari, R.S., Corso, J.J., Chaudhary, V., Dhillon, G.: Automatic diagnosis of lumbar disc
herniation using shape and appearance features from MRI. In: Proceedings of SPIE
Conference on Medical Imaging (2010). http://web.eecs.umich.edu/jjcorso/pubs/
alomarispie2010.pdf
3. Alawneh, K., Al-dwiekat, M., Alsmirat, M., Al-ayyoub, M.: Computer-aided diagnosis of
lumbar disc herniation. In: 6th International Conference on Information and Communication
System (ICICS 2013) (2013)
4. Liao, S., et al.: Automatic lumbar spondylolisthesis measurement in CT images. IEEE Trans.
Med. Imag. 35(7), 1658–1669 (2016)
5. Herath, H.M.M.G.T., Kumara, J.R.S.S., Fernando, M.A.R.M., Bandara, K.M.K.S., Serina,
I.: Comparison of supervised machine learning techniques for PD classification in generator
insulation. In: IEEE International Conference on Industrial and Information Systems (ICIIS),
Peradeniya, pp. 1–6 (2017)
6. Hasan, K., Islam, S., Samio, Md.M.R.K., Chakrabarty, A.: A machine learning approach on
classifying orthopedic patients based on their biomechanical features. In: Joint 7th
International Conference on Informatics, Electronics & Vision (ICIEV) and 2nd Interna-
tional Conference on Imaging, Vision & Pattern Recognition (icIVPR) (2018)
7. Matoug, N.A.A., AL-Jebury, H., Akyol, K., Alsarrar, M.R.: Comparison supervised learning
algorithms for spinal-column disease. Int. J. Sci. Res. 8(1), January 2019
8. Rocha Neto, A.R., Barreto, G.A.: On the application of ensembles of classifiers to the
diagnosis of pathologies of the vertebral column: a comparative analysis. IEEE Latin Am.
Trans. 7(4), 487–496 (2009)
9. Thomas, J.J., Pillai, N.: A deep learning framework on generation of image descriptions with
bidirectional recurrent neural networks. In: International Conference on Intelligent
Computing & Optimization. Springer, Cham (2018)
10. Akinsola, J.E.T.: Supervised machine learning algorithms: classification and comparison.
Int. J. Comput. Trends Technol. (IJCTT). 48, 128–138 (2017). https://doi.org/10.14445/
22312803/ijctt-v48p126
11. Rish, I.: An empirical study of the Naïve Bayes classifier. Work Empir. Methods Artif. Intell.
IJCAI (2001)
12. Campos, P.G., Oliveira, E.M.J., Ludermir, T.B., Araujo, A.F.R.: MLP networks for
classification and prediction with rule extraction mechanism. In: IEEE International Joint
Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, 2004, vol. 2,
pp. 1387–1392 (2004)
13. Chen, J., Jiao, L.: Classification mechanism of support vector machines. In: 5th International
Conference on Signal Processing Proceedings, WCC 2000 - ICSP 2000. 16th World
Computer Congress, Beijing, China, 2000, vol. 3, pp. 1556–1559 (2000)
14. Liu, L.: Research on logistic regression algorithm of breast cancer diagnose data by machine
learning. In: International Conference on Robots & Intelligent System (ICRIS), Changsha,
2018, pp. 157–160 (2018)
15. Li, Y., Cheng, B.: An improved k-nearest neighbor algorithm and its application to high
resolution remote sensing image classification. In: 2009 17th International Conference on
Geoinformatics, Fairfax, VA, pp. 1–4 (2009)
Classification of Orthopedic Patients 669
16. Tu, P., Chung, J.: A new decision-tree classification algorithm for machine learning. In:
Proceedings Fourth International Conference on Tools with Artificial Intelligence TAI 1992,
Arlington, VA, USA, pp. 370–377 (1992)
17. Patel, S.V., Jokhakar, V.N.: A random forest based machine learning approach for mild steel
defect diagnosis. In: 2016 IEEE International Conference on Computational Intelligence and
Computing Research (ICCIC), Chennai, 2016, pp. 1–8 (2016)
18. Bylinskii, Z., Judd, T., Oliva, A., Torralba, A., Durand, F.: What do different evaluation
metrics tell us about saliency models? IEEE Trans. Pattern Anal. Mach. Intell. 41(3), 740–
757 (2019)
19. Cufoglu, A., Lohi, M., Madani, K.: A comparative study of selected classification accuracy
in user profiling. In: 2008 Seventh International Conference on Machine Learning and
Applications, San Diego, CA, pp. 787–791 (2008)
20. Marom, N.D., Rokach, L., Shmilovici, A.: Using the confusion matrix for improving
ensemble classifiers. In: IEEE 26-th Convention of Electrical and Electronics Engineers in
Israel, Eliat, 2010, pp. 000555–000559 (2010)
21. Marín, N., Rivas-Gervilla, G., Sánchez, D., Yager, R.R.: Specificity measures and referential
success. IEEE Trans. Fuzzy Syst. 26(2), 859–868 (2018)
22. Bong, C.W., Xian, P.Y., Thomas, J.: Face Recognition and detection using haars features
with template matching algorithm. In: International Conference on Intelligent Computing &
Optimization. Springer, Cham (2019)
23. Keedwell, E.: An analysis of the area under the ROC curve and its use as a metric for
comparing clinical scorecards. In: 2014 IEEE International Conference on Bioinformatics
and Biomedicine (BIBM), Belfast, pp. 24–29 (2014)
24. https://archive.ics.uci.edu/ml/datasets/Vertebral Column, prepared and built by (Dr. Hen-
rique da Mota)
25. Berthonnaud, E., Dimnet, J., Roussouly, P., Labelle, H.: Analysis of the sagittal balance of
the spine and pelvis using shape and orientation parameters. J. Spinal Disorder. Techn.
(2005)
Long Short-Term Memory Networks
for Driver Drowsiness and Stress Prediction
Abstract. Road safety is crucial to prevent traffic deaths and injuries of drivers,
passengers, and pedestrians. Various regulations and policies have been pro-
posed to aim at reducing the number of traffic deaths and injuries. However,
these figures have remained steady in recent decade. There has been an
increasing number of research works on the prediction of driver status which
gives warning before undesired status, for instance drowsiness and stress. In this
paper, a long short-term memory networks is proposed for generic design of
driver drowsiness prediction and driver stress prediction models using electro-
cardiogram (ECG) signals. The proposed model achieves sensitivity, specificity,
and accuracy of 71.0–81.1%, 72.9–81.9%, and 72.2–81.5%, respectively, for
driver drowsiness prediction. They are 68.2–79.3%, 71.6–80.2%, and 70.8–
79.7%, for driver stress prediction. The results have demonstrated the feasibility
of generic model for both drowsiness and stress prediction. Future research
directions have been shared to enhance the model performance.
1 Introduction
The world has experienced an important issue that the road traffic accidents have led to
1.35 million deaths and 50 million injuries every year [1]. It ranks 8th in the global
leading causes of death. Most severe, it ranks 1st for youngsters (5–29 years old). The
deaths and injuries results in 3–5% loss in gross national product. United Nations has
defined 17 sustainable development goal and 169 targets in which the global number of
traffic deaths and injuries should achieve at least 50% reduction by 2020 [2]. However,
the total number of traffic deaths has been increased and the rate of traffic deaths has
remained nearly constant (about 18 deaths per 100,000 population). With the failure in
meeting the target, it is expected United Nation will postpone the completion date,
likely before 2030 to meet the overall vision.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 670–680, 2021.
https://doi.org/10.1007/978-3-030-68154-8_58
Long Short-Term Memory Networks 671
Various measures have been implemented for road traffic safety. For instance,
government officials have proposed mandatory inspections of vehicles [3], education
[4], stronger legislation [5], speed limit [6], and highway rest areas [7]. Nevertheless,
these measures have limited benefits on lowering the number of traffic accidents.
Monitoring and detecting the physical status of drivers has become an emergent
research trend to give warning to drivers with undesired status. A large-scale survey
collected questionnaires from 60 million drivers and concluded that 54% and 28% of
respondents had experienced drowsiness and fallen asleep on the wheel, respectively
[8]. Aggressive driving behaviors, as a result of stress, are common (90%) in traffic [9].
In this paper, attention is drawn to the generic design of prediction algorithm for driver
drowsiness and stress. In contrast, some research studies were discussing the driver
drowsiness [10, 11] and stress detection [12, 13] which current status of drivers was
examined. In reality, this may not help preventing traffic accidents as the average time
for reaction time may vary from 0.5 s [14] to 1.5 s [15] reported in different studies.
Therefore, to allow sufficient time for drivers’ response, it is required to reformulate the
driver drowsiness and stress from detection to prediction.
This section is organized as follows. Firstly, the driver drowsiness and stress databases
are presented. This is followed by data pre-processing and ECG beat segmentation. At
last, the proposed EMD based LSTM network will be discussed for driver drowsiness
and stress prediction
p1 ðtÞ ¼ xðtÞ EðtÞ: ð2Þ
674 K. T. Chui et al.
Equations (1) and (2) are repeated until the standard deviation r is less than the
threshold value, 0.05 [29]. The iterations can be expressed as:
ij ðtÞ:
pij ðtÞ ¼ piðj1Þ ðtÞ E ð3Þ
EMD completes when final residue rN ðtÞ (N is the total number of IMFs) is in any
of the following conditions (i) a constant; (ii) a function with one extreme; and (iii) a
monotonic function. The individual ECG sample is the sum of all IMFs and rN ðtÞ.
XN
xðtÞ ¼ i¼1
IMFi ðtÞ þ rN ðtÞ: ð5Þ
the extent of retained information from the upper cell. In general, the output is ranged
from 0 (completely ignore) and 1 (completely retain).
~ t ¼ tanh(Wc ½ht1 ; xt þ bc Þ:
C ð9Þ
~t:
Ct ¼ ft Ct1 þ it C ð10Þ
ht ¼ ot tanhðCt Þ: ð11Þ
The effectiveness of proposed LSTM network has been evaluated based on (i) perfor-
mance evaluation of prediction models for driver drowsiness and stress prediction
under varying times of advance prediction; (ii) performance comparison between
propose work and existing works; and (iii) future research directions.
common practice [30, 31], has been adopted for performance evaluation. Sensitivity,
specificity, and accuracy of the LSTM network are obtained by averaging the 5 sets of
results in 5-fold cross-validation.
Table 2 summarizes the results under 1–5 s advance prediction. In both driver
drowsiness prediction and driver stress prediction, the proposed EMD based LSTM
network achieve highest sensitivity, specificity, and accuracy, in which the perfor-
mance gradually decreases with the increase of time of advance prediction. The phe-
nomenon could be suggested by more unseen information with earlier advance
prediction. There is a trade-off between time of advance prediction and performance. In
this paper, the analysis was not made towards long period of advance prediction (for
example 10 s, 30 s, and 1 min in literature) because the expected performance of
prediction model is much lower than that of short period of advance prediction. Given
the fact that short period of advance prediction could help preventing severe stages of
drowsiness and stress levels and thus avoiding accident. It may also relate to the nature
of the design of features as well as prediction model. Further investigation is sought for.
Table 2. Performance evaluation of proposed LSTM network under 1–5 s advance prediction.
Driver prediction Advance prediction Results
Sensitivity Specificity Accuracy
Drowsiness 1s 81.1% 81.9% 81.5%
2s 78.8% 80.3% 79.8%
3s 76.4% 78.1% 77.5%
4s 73.6% 74.4% 74.0%
5s 71.0% 72.9% 72.2%
Stress 1s 79.3% 80.2% 79.7%
2s 75.8% 77.2% 76.6%
3s 73.4% 74.4% 73.9%
4s 70.5% 71.6% 71.2%
5s 68.2% 72.3% 70.8%
Research studies [20, 22] adopted real world dataset with number of candidates of 14
and 1, respectively, which are small sample size compared with that in our work, 126
candidate.
Input Signals. The works [17–23] rely on multiple input signals which are more
dependent on the quality of the signals. A research study [32] has suggested the
measurement stability of image and EEG signal could not be maintained in excellent
performance (59% for image and 85% EEG signal) in real world environment. Also,
multiple data sources may experience difficulty in handling data heterogeneity espe-
cially if the sampling rates do not align with each other.
Table 3. Comparisons between proposed work and existing works for driver status prediction.
Work Driver Dataset Input signal Methodology Advance Cross- Results
status prediction validation
[16] Drowsiness Simulation: Image CNN, LSTM 3–5 s No ACC: 75%
18 candidate
[17] Drowsiness Simulation: EEG; HRV; kNN, SVM, 10 s 5-fold ACC: 87.9%
16 candidate Image LR
[18] Drowsiness Simulation: EEG; EOG; FLDA 5s No ACC: 79.2%
11 candidate HRV; Image (deviation up to
40%)
[19] Drowsiness Simulation: HRV; Image; ANN 1 min 5-fold RMSE: 6 min
21 candidate Lateral and
steering
information
[20] Stress Real world: Demographic kNN, SVM, 2s 10-fold ACC: 98.92%; SEN:
14 candidate data; EDA; ANN, RF, 99.36%; SPE:
EMG; RESP; DT 98.46%
Time; Weather
[21] Stress Simulation: HRV; Weather NB 30 s 10-fold ACC: 78.3%
5 candidate (deviation up to
37%)
[22] Stress Real world: Accelerometer; LR Not 10-fold SEN: 60.9%; SPE:
1 candidate EDA, PPG specify 86.7%
[23] Stress Simulation: HRV; intensity DBN 1 min 10-fold SEN: 82.3%; SPE:
10 candidate of turning; 83.6% (deviation 25–
speed; 38%)
Proposed Drowsiness Real world: ECG EMD; LSTM 1–5 s 5-fold Drowsiness
and stress 126 SEN: 71.0–81.1%
candidate SPE: 72.9–81.9%
using CAPS ACC: 72.2–81.5%
and SRAD Stress SEN: 68.2–
databases 79.3% SPE: 71.6–
80.2% ACC: 70.8–
79.7%
Accuracy (ACC); artificial neural network (ANN); convolutional neural network (CNN); deep-belief network (DBN);
decision tree (DT); electrocardiogram (ECG); electrodermal activity (EDA); electroencephalogram (EEG); empirical mode
decomposition (EMD); electrooculography (EOG); Fisher’s linear discriminant analysis (FLDA); heart rate variability
(HRV); k-nearest neighbor (kNN); logistic regression (LR); long short-term memory (LSTM); Naive Bayes (NB);
photoplethysmogram (PPG); respiration (RESP); random forest (RF); root-mean-square error (RMSE); sensitivity (SEN);
specificity (SPE); support vector machine (SVM).
678 K. T. Chui et al.
4 Conclusion
In this paper, an EMD based LSTM network has been proposed to serve as generic
model for driver drowsiness and stress prediction. As preliminary study, we choose
Pearson correlation coefficient in feature extraction process and varying the times for
advance prediction from 1 to 5 s. The proposed model achieves sensitivity, specificity,
and accuracy of 71.0–81.1%, 72.9–81.9%, and 72.2–81.5%, respectively, for driver
drowsiness prediction. When it comes to driver stress prediction, the sensitivity,
specificity, and accuracy are 68.2–79.3%, 71.6–80.2%, and 70.8–79.7%, respectively.
This has demonstrated the feasibility of generic model for dual applications. Four future
research directions have been suggested to enhance the performance of the prediction
model. In near future, it is foreseeable that there are many more research in intelligent
applications using intelligent computing and optimization techniques [34, 35].
Long Short-Term Memory Networks 679
Acknowledgments. The work described in this paper was fully supported by the Open University
of Hong Kong Research Grant (No. 2019/1.7).
References
1. Global Status Report on Road Safety 2018, World Health Organization. https://www.who.
int/publications/i/item/global-status-report-on-road-safety-2018
2. Transforming Our World: The 2030 Agenda for Sustainable Development, United Nations.
http://sustainabledevelopment.un.org
3. Das, S., Geedipally, S.R., Dixon, K., Sun, X., Ma, C.: Measuring the effectiveness of vehicle
inspection regulations in different states of the US. Transp. Res. Rec. 2673, 208–219 (2019)
4. Alonso, F., Esteban, C., Useche, S., Colomer, N.: Effect of road safety education on road
risky behaviors of Spanish children and adolescents: findings from a national study. Int.
J. Environ. Res. Public Health 15, 2828 (2018)
5. Castillo-Manzano, J.I., Castro-Nuño, M., López-Valpuesta, L., Pedregal, D.J.: From
legislation to compliance: the power of traffic law enforcement for the case study of Spain.
Transp. Policy 75, 1–9 (2019)
6. Silvano, A.P., Koutsopoulos, H.N., Farah, H.: Free flow speed estimation: a probabilistic,
latent approach. Impact of speed limit changes and road characteristics. Transport. Res. A-
Pol. 138, 283–298 (2020)
7. Choi, J., Lee, K., Kim, H., An, S., Nam, D.: Classification of inter-urban highway drivers’
resting behavior for advanced driver-assistance system technologies using vehicle trajectory
data from car navigation systems. Sustainability 12, 5936 (2020)
8. Royal, D., Street, F., Suite, N.W.: National Survey of Distracted and Drowsy Driving
Attitudes and Behavior. Technical report, National Highway Traffic Safety Administration
(2002)
9. Pfeiffer, J.L., Pueschel, K., Seifert, D.: Interpersonal violence in road rage. Cases from the
medico-legal center for victims of violence in Hamburg. J. Forens. Leg. Med. 39, 42–45
(2016)
10. Dua, M., Singla, R., Raj, S., Jangra, A.: Deep CNN models-based ensemble approach to
driver drowsiness detection. Neural Comput. Appl. 32, 1–14 (2020)
11. Zhang, X., Wang, X., Yang, X., Xu, C., Zhu, X., Wei, J.: Driver drowsiness detection using
mixed-effect ordered logit model considering time cumulative effect. Anal. Meth. Accid.
Res. 26, 100114 (2020)
12. Chung, W.Y., Chong, T.W., Lee, B.G.: Methods to detect and reduce driver stress: a review.
Int. J. Automot. Technol. 20, 1051–1063 (2019)
13. Kyriakou, K., Resch, B., Sagl, G., Petutschnig, A., Werner, C., Niederseer, D., Liedlgruber,
M., Wilhelm, F.H., Osborne, T., Pykett, J.: Detecting moments of stress from measurements
of wearable physiological sensors. Sensors 19, 3805 (2019)
14. Dickerson, A.E., Reistetter, T.A., Burhans, S., Apple, K.: Typical brake reaction times
across the life span. Occup. Ther. Health Care 30, 115–123 (2016)
15. Arbabzadeh, N., Jafari, M., Jalayer, M., Jiang, S., Kharbeche, M.: A hybrid approach for
identifying factors affecting driver reaction time using naturalistic driving data. Transp. Res.
Part C Emerg. Technol. 100, 107–124 (2019)
16. Saurav, S., Mathur, S., Sang, I., Prasad, S.S., Singh, S.: Yawn detection for driver’s
drowsiness prediction using bi-directional LSTM with CNN features. In: International
Conference on Intelligent Human Computer Interaction, pp. 189–200. Springer, Cham
(2019)
680 K. T. Chui et al.
17. Gwak, J., Hirao, A., Shino, M.: An investigation of early detection of driver drowsiness
using ensemble machine learning based on hybrid sensing. Appl. Sci. 10, 2890 (2020)
18. Nguyen, T., Ahn, S., Jang, H., Jun, S.C., Kim, J.G.: Utilization of a combined EEG/NIRS
system to predict driver drowsiness. Sci. Rep. 7, 43933 (2017)
19. de Naurois, C.J., Bourdin, C., Bougard, C., Vercher, J.L.: Adapting artificial neural networks
to a specific driver enhances detection and prediction of drowsiness. Accid. Anal. Prev. 121,
118–128 (2018)
20. Hadi, W.E., El-Khalili, N., AlNashashibi, M., Issa, G., AlBanna, A.A.: Application of data
mining algorithms for improving stress prediction of automobile drivers: a case study in
Jordan. Comput. Biol. Med. 114, 103474 (2019)
21. Alharthi, R., Alharthi, R., Guthier, B., El Saddik, A.: CASP: context-aware stress prediction
system. Multimed. Tools Appl. 78, 9011–9031 (2019)
22. Bitkina, O.V., Kim, J., Park, J., Park, J., Kim, H.K.: Identifying traffic context using driving
stress: a longitudinal preliminary case study. Sensors 19, 2152 (2019)
23. Magana, V.C., Munoz-Organero, M.: Toward safer highways: predicting driver stress in
varying conditions on habitual routes. IEEE Veh. Technol. Mag. 12, 69–76 (2017)
24. Terzano, M.G., Parrino, L., Sherieri, A., Chervin, R., Chokroverty, S., Guilleminault, C.,
Hirshkowitz, M., Mahowald, M., Moldofsky, H., Rosa, A., et al.: Atlas, rules, and recording
techniques for the scoring of cyclic alternating pattern (CAP) in human sleep. Sleep Med. 2,
537–553 (2001)
25. Goldberger, A.L., Amaral, L.A.N., Glass, L., Hausdorff, J.M., Ivanov, P.C.H., Mark, R.G.,
Mietus, J.E., Moody, G.B., Peng, C.K., Stanley, H.E.: PhysioBank, PhysioToolkit, and
PhysioNet: components of a new research resource for complex physiologic signals.
Circulation 101, e215–e220 (2003)
26. Healey, J.A., Picard, R.W.: Detecting stress during real-world driving tasks using
physiological sensors. IEEE Trans. Intell. Transp. 6, 156–166 (2005)
27. Tompkins, W.J.: Biomedical Digital Signal Processing C-Language Examples and
Laboratory Experiments for the IBM®PC. pp. 236–264. Prentice Hall, Upper Saddle River
(2000)
28. Kohler, B.U., Hennig, C., Orglmeister, R.: The principles of software QRS detection. IEEE
Eng. Med. Biol. 21, 42–57 (2002)
29. Azbari, P.G., Abdolghaffar, M., Mohaqeqi, S., Pooyan, M., Ahmadian, A., Gashti, N.G.: A
novel approach to the extraction of fetal electrocardiogram based on empirical mode
decomposition and correlation analysis. Aust. Phys. Eng. Sci. Med. 40, 565–574 (2017)
30. Chui, K.T., Fung, D.C.L., Lytras, M.D., Lam, T.M.: Predicting at-risk university students in
a virtual learning environment via a machine learning algorithm. Comput. Hum. Behav. 107,
105584 (2020)
31. Wong, T.T., Yeh, P.Y.: Reliable accuracy estimates from k-fold cross validation. IEEE
Trans. Knowl. Data Eng. 32, 1586–1594 (2020)
32. Sun, Y., Yu, X.: An innovative nonintrusive driver assistance system for vital signal
monitoring. IEEE J. Biomed. Health Inform. 18, 1932–1939 (2014)
33. Savku, E., Weber, G.W.: A stochastic maximum principle for a Markov regime-switching
jump-diffusion model with delay and an application to finance. J. Optim. Theory Appl. 179,
696–721 (2018)
34. Vasant, P., Zelinka, I., Weber, G.W. (eds.): Intelligent Computing & Optimization, vol. 866.
Springer, Cham (2018)
35. Vasant, P., Zelinka, I., Weber, G.W. (eds.): Intelligent computing and optimization. In:
Proceedings of the 2nd International Conference on Intelligent Computing and Optimization
2019. Springer, Cham (2019)
Optimal Generation Mix of Hybrid
Renewable Energy System Employing
Hybrid Optimization Algorithm
1 Introduction
In recent times, the energy policymakers have shown a vivid inclination towards
replenishable sources of energy like Hybrid Renewable Energy Sources (HRES)
to meet escalating energy demand and lessen the impact of consumption of non-
renewable energy sources [1]. HRES allows the hybridization of renewable and
conventional energy sources and facilitates in obtaining higher efficiency than a
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 681–692, 2021.
https://doi.org/10.1007/978-3-030-68154-8_59
682 Md. A. Hossain et al.
single power source [2]. Among the numerous functional HRES technology, the
proven, environment-friendly and economically viable ones are the combinations
of Photo-Voltaic (PV) cell, Wind Turbine (WT), and Battery System (BS) [3].
1.1 Background
2 Methodology
This section briefly interprets the steps which were used to reach the optimum
result. In this study, the optimum result is defined as the minimum attainable
setup, maintenance and operational cost of the chosen components of the HRES
for a period of 20 years while maintaining an LPSP of zero throughout this
duration. All the calculations made in this study are on an hourly basis.
The HRES considered in this study consist of three main components, two of
which are renewable energy sources (Photovoltaic and wind energy), and the
third one is an energy storage unit (Battery) forming a PV-WT-battery based
HRES. These system components are connected to a 24V DC bus via power
electronic converters, as shown in Fig. 1 [10]. For this work, the typical hourly
summer domestic demand of Auckland, New Zealand has been taken from [11]
as the hourly load demand.
where PP V (t, β) represents the output of the PV array at the tth hour with
a tilt angle of β. VOC and ISC represent the open circuit voltage and short
circuit current of a PV module respectively, F F is the fill factor, KV and KI
are the open circuit voltage temperature coefficient and short circuit current
temperature coefficient respectively, G represents the global solar irradiance on
a PV module, TA is the ambient temperature and N COT is the nominal cell
operating temperature. Finally, the total output power of a PV array is expressed
in terms of the following equation.
where ηP V represents the efficiency of the PV module and its respective converter
and NS and NP denote the number of series-connected and parallel-connected
PV modules respectively. The specifications of the PV module considered for
the HRES of this study are presented in Table 1 [12].
v3
where Pr is the rated power, a = v3P−v
r
3 and b = v 3 −v 3 . The cut-in speed, rated
ci
r ci r ci
speed, and cut-out speed of the WT are denoted by vci , vr and vc0 respectively.
The wind speed,vh at a particular height, can be calculated from Eq. (4) [12].
α
h
v=v (4)
hr
where hr represents the reference height and according to [13], the hourly wind
speed at hr = 33m was considered for this study. On the other hand, the power-
law exponent,α was chosen to be 1/7 for open space [14]. The output electrical
power generated from a WT can be obtained from the following equation.
PW G = Pw AW G ηwG (5)
where ηwG represents the efficiency of the WT generator and its associated
converters and AW G is the total swept area of the WT. The specifications of the
WT generator opted for this study are presented in Table 2 [12].
Hybrid ACOR -GA in HRES 685
where SOC(t) represents the state of charge at the current hour,t and SOC(t−1)
denotes the same for the previous hour,(t − 1). According to [5], the battery
charging efficiency ηbat is taken to be 0.8, and the discharging efficiency was set
to 1. For this study, the self-discharge rate σ of the battery was assumed to be
0.2% per day. Furthermore, in Eq. (6), Cbat denotes the capacity of the battery
and Ibat (t) refers to the current from the battery at the tth hour. For the HRES
considered in this work, Ibat (t) due to the incorporation of the battery can be
determined as follows.
PP V (t) + PW G (t) − Pload (t)
Ibat (t) = (7)
Vbat (t)
where PLoad (t) refers to the load demand at the tth hour and Vbat (t) indicates
the battery voltage. The battery specifications considered for this study are
presented in Table 3 [12].
VOC (V) ISC (A) Vmax (V) Imax (A) Pmax (W) Capital Cost ($)
64.8 6.24 54.7 5.86 320 640
Power (W) hlow (m) hhigh (m) WG Capital Cost($) Tower Capital Cost ($/unit length)
100 11 35 2400 55
LPSP is a very important term that signifies the probability of unsatisfied load
demand due to a shortage of power supply or due to economic constraints [15].
Thus, the goal of this study is to optimize the number of PV modules NP V ,
the number of WT generators NW G , and the number of batteries Nbat along
with the tilt angle β of the PV modules and WT installation height h required
to achieve zero LPSP at the minimum total cost for the HRES model. The
objective function for minimizing the overall cost associated with the HRES
system is highlighted in Eq. (8) [12].
In the objective function of Eq. (8), the total cost is defined in terms of the
initial capital costs of the PV modules, WT generators, and batteries along with
the associated maintenance and operational cost over a span of 20 years [10]. In
this equation, CP V ,CW G and Cbat are the capital costs associated with a PV
module, WT generator, and battery, respectively. Whereas, the maintenance and
operational costs for a PV module, WT generator, and battery are denoted by
MP V ,MW G and Mbat , respectively. The capital cost per unit height (Ch ) of a
WT and the annual operational and maintenance cost per unit height (Mh ) of
a WT are also incorporated in the objective function. The expected number of
batteries to be replaced within the period of 20 years is represented by ybat [10].
Thus, the goal of this study is to determine the optimum values of NP V (NS ×
NP ), NW G , Nbat , β and h in order to minimize the objective function in (8) with
respect to the constraints mentioned in (9).
In this study ACOR proposed by the authors in [16] is hybridized with the aid
of conventional crossover and mutation operations which are the core part of the
Genetic Algorithm [17] to formulate the ACOR -GA algorithm. The formulated
hybrid algorithm commences by initializing the required parameters of ACOR ,
crossover, mutation, initial population (pop), initial population size, number of
iterations, and loading the required data for simulation. The initial population
is generated randomly within the adopted parameter bounds. After calculating
the solution weights (ω) and selection probabilities (p) for an individual run, the
algorithm enters the main loop which is run for a definite number of iterations.
In this study, it is assumed that the optimization algorithm will cease once the
maximum number of iterations reaches 100 for each individual run. Besides,
the number of the initial population is set to 50. After entering the main loop,
Hybrid ACOR -GA in HRES 687
mean (μ) and standard deviation (σ) of the initial population set are calculated
which, along with previously calculated (ω) and (p) influence the generation of
the new population (newpop) set in that particular iteration. Thereafter, based
on roulette wheel selection, a solution kernel is generated which governs the
generation of a new population termed as newpop. This process takes place
50 times i.e. equal to the number of initial population size and each time it
updates all the optimizing parameters. It is to be noted that all the parameters
of ACOR are calculated following the equations provided by the authors in [18].
Consequently, the algorithm performs crossover and mutation [19] on randomly
selected populations from the initial population set and generates two more new
population sets named popc and popm respectively. All the solutions, i.e.newpop,
popc, popm and initial population (pop) are then merged to generate an updated
pop of larger dimension. After merging, all the solutions are checked if any of
the solutions exceed the adopted bounds. The solutions are pulled back to the
upper boundary provided they go beyond the defined upper and lower limits.
This solution set is then sorted based on the obtained cost from the objective
function and only the best 50 solutions are kept for the next iteration. The
current best solution in a particular iteration is stored provided it supersedes the
best solution achieved in the previous iteration. And thus the above-mentioned
methodology runs for 100 times which is the maximum number of iterations.
The overall work-flow of the proposed ACOR -GA is illustrated in Fig. 2. It is to
be mentioned here that the hybrid algorithm takes a slightly higher computation
time concerning the single algorithms. But because the convergence rate is much
quicker, this difference becomes negligible.
Table 4 demonstrates that the mean value of the ACOR -GA is less than all the
other algorithms. However, only the mean value can not signify the superiority
of an algorithm. Variance and standard deviation are two other parameters that
help us further analyze the data. These parameters demonstrate the spreading of
our data around the mean. An algorithm with high variance and standard devi-
ation indicates that the result obtained in each independent run is significantly
different from the other independent runs. And thus, for this algorithm to be
reliable, it is needed to carry out more number of independent runs. On the other
hand, a smaller value of variance and standard deviation signifies that the result
obtained in each independent run is not significantly different from the other
runs. From Table 4, it is seen that both the values of standard deviation and
variance of ACOR -GA are considerably smaller than the rest of the algorithms.
Thus the value of the mean together with the values of standard deviation and
variance indicates that the ACOR -GA is more reliable than the rest of the algo-
rithms. Further, the minimum value reached by the hybrid algorithm is 37943.70
$ which is again less than the comparing algorithms. The number of iterations
signifies the convergence speed of an algorithm. In this aspect, the hybrid algo-
rithm leaves behind all the other algorithms by a very high margin. The average
number of iterations needed by the ACOR -GA is 26.3 and the algorithm closest
to this value is GA with 34 iterations. It is to be noted here that apart from
the hybrid algorithm, GA has the best convergence speed among the four algo-
rithms. Thus hybridizing GA with ACOR significantly improves the convergence
of the latter algorithm.
In this normality test, the null hypothesis states that the data is not statisti-
cally significantly different from the normal distribution. The important aspect
to realize in this particular test is that it is expected that the null hypothesis is
rejected. Because the data which are being analyzed here is the optimum costs
obtained in each independent run. Thus, since all the costs here are optimum,
it is expected that all the values will cluster to a particular point to increase
the reliability of the algorithm. However, if the null hypothesis is not rejected,
signifying that the data is normally distributed, means that for that particular
algorithm to have reasonable credibility, a lot of independent runs are required.
The value of asymptotic sigma (Sig.) determines whether the null hypothesis is
rejected or not. If its value is less than 0.05 then the null hypothesis is rejected
and vice versa. From Table 5, it is observed that ACOR -GA and ACO are not
normally distributed in both K-S and S-W tests, while PSO is marginally reject-
ing the normal distribution in the case of K-S test and the other algorithms are
accepting the null hypothesis. Thus the reliability of the hybrid algorithm is also
supported by these tests of normality.
Parameter Algorithm
ACOR -GA ACO GA PSO GOA
NP V 24 8 10 9 10
NW G 3 5 5 5 5
Nbat 2×1 2×2 2×2 2×2 2×2
β 14 4 33 24 29
h 35 29 29 28 26
Cost ($) 37943.70 37997.4 39149.70 38453.4 38873.4
Hybrid ACOR -GA in HRES 691
parameters and not the number of batteries connected in series. As can be seen
from the table, the number of batteries in series will always be two owing to the
particular setup that was considered in this study.
References
1. Bajpai, P., Dash, V.: Hybrid renewable energy systems for power generation in
stand-alone applications: a review. Renew. Sustain. Energy Rev. 16(5), 2926–2939
(2012)
2. Amer, M., Namaane, A., M’sirdi, N.: Optimization of hybrid renewable energy
systems (HRES) using PSO for cost reduction. Energy Procedia 42, 318–327 (2013)
3. Kamjoo, A., Maheri, A., Dizqah, A.M., Putrus, G.A.: Multi-objective design under
uncertainties of hybrid renewable energy system using NSGA-ii and chance con-
strained programming. Int. J. Electr. Power Energ. Syst. 74, 187–194 (2016)
4. Anand, P., Rizwan, M., Bath, S.K.: Sizing of renewable energy based hybrid system
for rural electrification using grey wolf optimisation approach. IET Energ. Syst.
Integr. 1(3), 158–172 (2019)
5. Yang, H., Zhou, W., Lu, L., Fang, Z.: Optimal sizing method for stand-alone hybrid
solar-wind system with LPSP technology by using genetic algorithm. Sol. Energy
82(4), 354–367 (2008)
6. Suhane, P., Rangnekar, S., Mittal, A., Khare, A.: Sizing and performance analy-
sis of standalone wind-photovoltaic based hybrid energy system using ant colony
optimisation. IET Renew. Power Gener. 10(7), 964–972 (2016)
7. Fathy, A., El-baksawi, O.: Grasshopper optimization algorithm for extracting maxi-
mum power from wind turbine installed in al-jouf region. J. Renew. Sustain. Energy
11(3), 033303 (2019)
8. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent Computing & Optimization, vol.
866. Springer, Cham (2018)
692 Md. A. Hossain et al.
9. Mitiku, T., Manshahia, M.S.: Fuzzy logic controller for modeling of wind energy
harvesting system for remote areas. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.)
International Conference on Intelligent Computing & Optimization, pp. 31–44.
Springer, Cham (2019)
10. Tito, M.R., Lie, T.T., Anderson, T.: Sizing optimization of wind-photovoltaic
hybrid energy systems under transient load. Int. J. Power Energy Syst. 33(4),
168–174 (2013)
11. Authority, T.E.: (2014). Available at: https://www.ea.govt.nz/dmsdocument/4755
12. Tito, S., Lie, T., Anderson, T.: Optimal sizing of a wind-photovoltaic-battery
hybrid renewable energy system considering socio-demographic factors. Sol. Energy
136, 525–532 (2016)
13. Auckland, T.K.: (2014). Available at: https://bit.ly/2Bu4Y4H
14. Borowy, B.S., Salameh, Z.M.: Methodology for optimally sizing the combination
of a battery bank and PV array in a wind/PV hybrid system. IEEE Trans. Energy
Convers. 11(2), 367–375 (1996)
15. Yang, H., Lu, L.: Study of typical meteorological years and their effect on building
energy and renewable energy simulations. ASHRAE Trans. 110, 424 (2004)
16. Socha, K., Dorigo, M.: Ant colony optimization for continuous domains. Eur. J.
Oper. Res. 185(3), 1155–1173 (2008)
17. Grefenstette, J.J.: Optimization of control parameters for genetic algorithms. IEEE
Trans. Syst. Man Cybern. 16(1), 122–128 (1986)
18. Nguyen, T.K., Lee, I.G., Kwon, O., Kim, Y.J., Hong, I.P.: Metaheuristic optimiza-
tion techniques for an electromagnetic multilayer radome design. J. Electromagn.
Eng. Sci. 19(1), 31–36 (2019)
19. Yoon, Y., Kim, Y.H.: The Roles of Crossover and Mutation in Real-Coded Genetic
Algorithms. Citeseer, London (2012)
Activity Identification from Natural Images
Using Deep CNN
1 Introduction
2 Literature Review
Human activity recognition performs an important role in medical research and security
system. HAR has been studied for years and researchers have introduced different
solutions to challenge the problem. CNN usually produces more accurate and reliable
results than the existing methods. CNN's are deep learning-based algorithm extends the
capability of traditional artificial neural network by adding additional constraints to the
earlier layers and increased the depth of the network. Recent work has focused on
tuning different deep learning-based [7] architecture to achieve maximum performance
on different human action video dataset [8].
Activity Identification from Natural Images Using Deep CNN 695
This section provides the background to the traditional CNN based Model 1, a
description of the proposed modified CNN of Model 1 (Model 2) and a brief
description of the traditional 4 Layer CNN based Model 3, basic overview of the
optimized 4 Layer CNN of Model 3 (Model 4) architectures. Besides, this section
provides the Layer operation and configuration table of all neural network that has
proposed in this paper.
3.1 Dataset
The evaluation process of our proposed method on two different human activity video
databases. The Human activity video databases were KTH [13] action database and
UCF11 [14] YouTube action dataset. The KTH dataset created by the Royal Institute of
Technology in 2004 that achieved an important milestone in the computer vision. This
dataset contains six types of human actions (walking, jogging, running, boxing, hand
696 Md. Anwar Hossain and M. A. F. M. Rashidul Hasan
waving, and hand clapping) performed several times by 25 subjects in four different
conditions. There is a total of 25 6 4 = 600 video files for each combination of 25
individuals, 6 actions, and 4 conditions. KTH dataset is an RGB color space video
dataset and the sequences from these videos provide all our requirements of proposed
human activity recognition.
This UCF11 [14] dataset was created videos from YouTube by the University of
Central Florida in 2009. This dataset contains 11 varieties of human actions (basketball
shooting, biking/cycling, diving, golf swinging, horseback riding, soccer juggling,
swinging, tennis swinging, trampoline jumping, volleyball spiking, and walking with a
dog). This dataset is a collection of 1600 video samples with large variations under
camera motion, object appearance, and pose, object scale, viewpoint, cluttered back-
ground, illumination conditions, etc. This is a more challenging task due to the above
properties of this dataset.
3.2 Preprocessing
In this preprocessing step, we extracted frames from every video in the database and
converted them from RGB to grayscale for the purpose of less memory and low
computation requirement. The depth of every image was 15 with 64 64 15
resolution. We modified the input dimension to 5D due to the requirement of input to
the 3D convolution. Table 1 indicates the dataset characteristics for human activity
recognition model.
in Fig. 3. As shown in Fig. 3, the model takes video sequences as input to the con-
volution layer which was followed by pooling and fully connected layer. Table 2
shows the parameters used in Model 3.
Videos(Images)
Videos (Images) Kernel Initializer &
Kernel Regularize Convolution ReLU
Fig. 1. Architecture of the traditional CNN Fig. 2. Architecture of the proposed CNN
based Model 1 for Human activity recognition. based Model 2 for human activity recognition.
Activity Identification from Natural Images Using Deep CNN 699
Dropout Pooling
Convolution Dropout
ReLU
Kernel Initializer &Convolution ReLU
Kernel Regularize
Convolution ReLU
Batch Normalization
Pooling Kernel con- Convolution ReLU
straint
Dropout
Pooling
Flatten Layer
Dropout
Fully Connected ReLU Flatten Layer
Dropout
Kernel con- Fully Connected ReLU
straint
4 Experimental Results
This section provides the experimental results obtained when using the Tradi-
tional CNN, Proposed CNN, Traditional 4 Layer CNN, Proposed 4 Layer CNN on
KTH and UCF11 action dataset. This experiments mainly evaluated by python Keras,
Scikit-learn libraries. The presented models of this paper evaluated on two different
human action datasets. We applied preprocessed dataset from KTH and UCF11 on four
different model that were introduced in this paper. The accuracy of testing phase is
shown in Table 3. The experiment results show that our proposed model 2 and model 4
architecture achieved higher performance than the performance obtained by the model
1 and model 3 architecture.
Table 3. Testing Accuracy for KTH and UCF11 dataset of different model
Model Representation Test Accuracy Test Accuracy
(KTH Dataset) (UCF11 Dataset)
1 Traditional CNN 67% 71%
2 Proposed CNN 70% 73%
3 Traditional 4 Layer CNN 74% 75%
4 Proposed 4 Layer CNN 80% 79%
When using the KTH dataset applied on the four models, our proposed Model 2
and Model 4 obtained accuracy 70% and 80% which were 3% and 6% higher than the
performance obtained by model 1(67%) and model 3(74%) respectively. When using
the UCF11 dataset applied on the four models, our proposed model 2 and model 4
obtained accuracy 73% and 79% which were 2% and 4% higher than the performance
obtained by model 1(71%) and model 3(75%) respectively.
Table 4. Class wise right(R) and wrong(W) cases for KTH dataset over different model
(Training samples 480 and testing samples 120)
Class No Class name Traditional Proposed Traditional 4 Proposed 4
CNN CNN (Model Layer CNN Layer CNN
(Model 1) 2) (Model 3) (Model 4)
R W R W R W R W
1 Boxing 15 2 15 2 15 2 16 1
2 Hand Clapping 12 3 12 3 14 1 14 1
3 Hand Waiving 15 5 16 4 18 2 17 3
4 Jogging 11 16 14 13 14 13 14 13
5 Running 11 12 11 12 14 9 14 9
6 Walking 16 2 15 3 15 3 16 2
Overall R/W 80 40 83 37 90 30 91 29
Activity Identification from Natural Images Using Deep CNN 701
The overall KTH dataset classification performance of the Model 2 and Model 4 are
high and Table 4 shows that class wise right and wrong prediction over four different
models. Fig. 5 depicts the confusion matrices show that all of the images which were
misclassified by Model 1 and Model 3, were correctly classified by Model 2 and Model 4.
Fig. 6 and Fig. 7 depict the model accuracy and training loss when evaluating four
different models over the KTH dataset. These figures indicate that training and testing
performance were close together during different epochs, which also indicate that the
data were not overfitting during training.
702 Md. Anwar Hossain and M. A. F. M. Rashidul Hasan
Fig. 6. Accuracy for Training and Testing Data for KTH Dataset
The overall accuracy for UCF11 dataset classification of the Model 2 and Model 4
is high and gave the significant result over Model 1 and Model 3 by optimizing the
parameters. Table 5 shows that class wise right and wrong predictions over four dif-
ferent models. Fig. 8 depicts the confusion matrices show that all of the images which
were misclassified by Model 1 and Model 3, were correctly classified by Model 2 and
Model 4.
Activity Identification from Natural Images Using Deep CNN 703
Table 5. Class wise right(R) and wrong(W) cases for UCF dataset over different model
(Training samples 1280 and testing samples 320)
Class No Class name Traditional Proposed Traditional 4 Proposed 4
CNN CNN Layer CNN Layer CNN
(Model 1) (Model 2) (Model 3) (Model 4)
R W R W R W R W
1 Basketball 24 8 25 7 25 7 26 6
2 Biking 15 13 16 12 22 6 15 13
3 Diving 27 9 31 5 29 7 31 5
4 Golf Swing 18 6 21 3 22 2 20 4
5 Horse Riding 29 7 27 9 30 6 30 6
6 Soccer Juggling 19 14 20 13 21 12 24 9
7 Swing 16 4 12 8 9 11 18 2
8 Tennis Swing 29 5 27 7 33 1 32 2
9 Trampoline Jumping 19 1 19 1 20 0 19 1
10 Volleyball Spiking 19 8 21 6 21 6 24 3
11 Walking 12 18 13 17 9 21 14 16
Overall R/W 227 93 232 88 241 79 253 67
704 Md. Anwar Hossain and M. A. F. M. Rashidul Hasan
Fig. 9 and Fig. 10 depicts the model accuracy and training loss when evaluating
four different models over the UCF11 dataset. These figures indicate that training and
testing performance were close together during different epochs, which also indicate
that the data were not overfit-ting during training.
Activity Identification from Natural Images Using Deep CNN 705
Fig. 9. Accuracy for training and testing data for UCF11 dataset
To investigate the results of all the models proposed in this paper, we observed that
the improvement in CNN model performance highly attributed to network initializa-
tion, network regularization, weight constraint which were allowed to the Proposed
Model 2 and Model 4. These models have shown higher performance than the existing
traditional model 1 and model 3 which were the similar layer except optimization
parameters.
706 Md. Anwar Hossain and M. A. F. M. Rashidul Hasan
5 Conclusion
References
1. Xiaoran, S., Yaxin, L., Feng, Z., Lei, L.: Human activity recognition based on deep learning
method. In: IEEE International Conference on Radar, IEEE (2018)
2. Saad, A., Arslan, B., Mubarak, S.: Chaotic invariants for human action recognition. In: IEEE
11th International Conference, pp. 1–8, IEEE (2007)
3. Moshe, B., Lena, G., Eli, S., Michal, I., Ronen, B.: Actions as space-time shapes. In:
Tenth IEEE International Conference, vol. 2, pp. 1395–1402. IEEE (2005)
4. Heng, W., Alexander, K., Cordelia, S., Cheng-Lin, L.: Dense trajectories and motion
boundary descriptors for action recognition. Int. J. Comput. Vision 103(1), 60–79 (2013)
5. Alani, A., Arabic, H.: Digit recognition based on restricted boltzmann machine and
convolutional neural networks. Information 8(4), 142 (2017)
6. Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional
neural networks. In: NIPS (2012)
7. Dhillon, J.K., Chandni., Kushwaha A.K.S.: A recent survey for human activity recognition
based on deep learning approach. In: Fourth International Conference on Image Information
Processing (ICIIP), Shimla (2017)
8. José, M.C., Enrique, J.C., Antonio, F.C.: A survey of video datasets for human action and
activity recognition. Comput. Vis. Image Underst. 117(6), 633–659 (2013)
9. Qingchang, Z., Zhenghua, C., Yeng, C.S.: A novel semisupervised deep learning method for
human activity recognition. IEEE Trans. Ind. Inform. 15(7), (2019)
10. Wen, H.C., Carlos, A.B.B., Chih, H.T.: LSTM-RNNs combined with scene information for
human activity recognition. In: IEEE 19th International Conference on e-Health Networking,
Applications and Services (Healthcom), IEEE (2017)
11. Shuiwang, J., Wei, X., Ming, Y., Kai, Y.: 3D convolutional neural networks for human
action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221–231 (2013)
12. Krishanu, S., Mohamed, M., Saeid, B., Shihao, J.: Towards robust human activity
recognition from RGB video stream with limited labeled data. In: 17th IEEE International
Conference on Machine Learning and Applications (ICMLA) (2018)
13. Laptev, I., Caputo, B.: Recognition of human actions. https://www.nada.kth.se/cvap/actions/
(2011)
14. University of Central Florida.: UCF YouTube action dataset. https://www.cs.ucf.edu/*liujg/
YouTubeAction_dataset.html (2011)
15. Thomas, J.J., Pillai, N.: A deep learning framework on generation of image descriptions with
bidirectional recurrent neural networks. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.)
Intelligent Computing & Optimization. ICO 2018, Advances in Intelligent Systems and
Computing, vol. 866, Springer, Cham (2019). https://doi.org/10.1007/978-3-030-00979-3_
22
16. Zeng, M., Nguyen, L.T., Yu, B., Mengshoel, O.J., Zhu, J., Wu, P., Zhang, J.: Convolutional
neural networks for human activity recognition using mobile sensors. In: 6th international
conference on mobile computing, applications and services, ICST (2014).
Learning Success Prediction Model for Early
Age Children Using Educational Games
and Advanced Data Analytics
Abstract. The early years of a child’s life greatly affects education potential
and furthermore potential of educational achievements in adulthood. The brain
develops faster during early age, and missed cognitive opportunities in that
period are difficult to make up for. In this research, we have developed a
machine learning model based on the data points obtained from the educational
game, with aim to predict how many attempts are necessary for an individual
child to complete the task or assessment as part of educational game. In-game
assessments are based on the skills that the child already possess and those
developed while playing the game. Training of the machine learning model is
based on collected and processed data points (features), while model intercon-
nections are related to the factors of the child’s cognitive growing up process.
Model performance benchmarks are elaborated in results and conclusion section
of the paper as quality measures of the forecast indicators.
1 Introduction
Major improvements in computing power, the availability of large data sets, and new
algorithms have paved the way for the development of artificial intelligence. Artificial
intelligence is an interdisciplinary field of research that deals with the development of
computers capable of intelligent activity. The approach that enables artificial intelli-
gence to be an intelligent activity, is called machine learning and represents its
skeleton. This will be the basis for the practical part of this paper. The initial idea of the
work comes from PBS KIDS Measure Up!1, educational computer games for children
between the ages of three and five, where PBS Kids is the most famous educational
medium in the United States. The app helps children learn early STEM concepts by
concentrating on length, width, capacity and complexity. Anonymized data collected
from the game available on the Kaggle platform will be processed in order to train the
model for learning success prediction. In this paper, the collected game data will be
used to predict the number of attempts needed for child to pass the offered task, and in
order to discover the key relationship between the child’s experience, engagement and
learning itself, which can be crucial for new educational concepts for modern
generations.
The topic of this paper stems from the world’s largest data science competition with
a focus on the social good- the Data Science Bowl2. The competition was presented on
the Kaggle online platform. Prediction model aims for four different outcomes, as
follows: (i) the task is solved in one attempt; (ii) the task is solved in the second
attempt; (iii) the task is solved in three or more attempts; (iv) the task has never been
solved yet. A successful solution will mostly depend on defining and extracting the
relevant features (data points) of examples for learning machine learning models. The
emphasis is on decision tree developed using advanced libraries, dominantly XGBoost
and CatBoost. The reason for their use is interpretability, they work very well without
great requirements for optimizing hyperparameters. In the evaluation of the results, the
Quadratic Weighted Kappa (QWK) will be used as an official measure, given by the
organizers.
The motivation and importance of competition rely in the fact that the first five
years of a child’s life are crucial to his or her development. The child begins to learn
and nurture his own skills that anticipate his later actions in different domains and
situations. Access to high-quality and effective early learning resources is crucial
during the early childhood development period, and support for early childhood edu-
cation is also key to other factors of long-term educational success. Machine learning
will try to use information on how children aged three to five use the app. This would
determine what children currently know and learn based on experience in order to
discover the important relationships between their engagement with the educational
medium and learning itself, as well as relevant aspects of development.
Machine Learning Methods and Concepts
According to the definition of machine learning, computers are programmed in such a
way as to optimize some performance criteria based on data examples or experience.
For this purpose, a model defined with carefully selected parameters is used. In order
for a given model to learn progressively, optimization of model parameters based on
data is practiced. It is important to point out that model training is performed on seen
data, and key prediction on unseen data. The main goal of machine learning is to build
a model that generalizes well under given circumstances.
The evaluation of the model is performed using certain metrics and should be per-
formed on a separate set, called a test set. To calculate the match coefficient of the
evaluator’s “decisions”, Cohen’s Kappa statistics are used [1]. If evaluators are present,
the Cohen kappa is used, while in the case of evaluators, the Fleiss kappa is used. Since
kappa-statistics come in multiple forms, its application depends on the type of data. In
the practical part of the paper, the evaluation requires the use of a weighted cap, i.e. a
Quadratic Weighted Kappa (QWK). Thus, the QWK metric in this paper can be
710 A. Tolic et al.
interpreted as a measure of the similarity between actual values and predicted values.
The measure can take values from a segment where 1.0 is a perfect estimate (predictions
and actual values are completely corresponding) while the lowest possible score is −1.0
(predictions are furthest from actual values).
2 Research Preparation
The data used in this competition is anonymous. Each piece of data was generated
during the child’s interaction with the PBS KIDS Measure Up! App. PBS KIDS is
committed to creating a safe environment that can be enjoyed by family members of all
ages. It is important to note that the PBS KIDS Measure Up! does not collect any
personal information such as name or location. Moreover, the data was reviewed by
PRIVO (Privacy Vaults Online, Inc.), the world’s leading expert in the children’s
Internet privacy industry, to ensure that all children’s privacy requirements are met.
Data Set
By launching the application, the child can choose one of the three default worlds
through which to move. Children move around the map and perform different tasks at
different levels. Tasks can be grouped into activities, videos, games, and assessments.
The child’s learning path is represented by the sequential execution of interrelated
tasks. It is designed for the child to first watch a video on the current topic and
approach to problem solving and to explore an activity that is similar to the game, but
without a defined goal, and in which the child can train their own skills. The child then
trains their own skills in play, but with the goal of solving a specific problem.
Demonstrates previously acquired skills in an assessment that is interactive and
specifically designed to measure and test a child’s knowledge, skills, and understanding
of the current topic. However, it is not necessary for the child to follow the stated order,
and some of the previously mentioned tasks may have several levels. Based on the
given data, an attempt will be made to predict in how many attempts the child will pass
the assessment task. Any wrong answer will count as an attempt [9].
The target variable has the following construction/state: (3) The assessment was
solved in the first attempt; (2) The assessment was solved in the second attempt;
(1) The assessment is solved in three or more attempts; (0) The estimate was never
resolved.
Data Preparation
In this step, among other things, the main goal is to extract data that is in JSON format
and represents the values of the event_data variable. Since a large amount of nested
data is present, the file explicitly states to what level the nested values the algorithm
approaches, after which a subset of the obtained values is selected for further process.
An example of such data can be seen in the following figure (Fig. 1).
Learning Success Prediction Model for Early Age Children 711
In this paper, parsing is done to level one. At the end of the parsing, a unify
operation is performed with the rest of the data set. We have to point out that, as part of
the process, outcome of the last assessment is goal for prediction. This means that there
may be a history of outcomes related to the installation identification number of a
particular assessment. In other words, a child can go through the same assessment
multiple times, and only the outcome of the last game is relevant to prediction.
Visualization and Exploratory Analysis
Figure below shows a histogram of experimental groups according to individual
assessment titles. It is clear that Mushroom Sorter and Cart Balancer are easier for
children because the category marked with 3 predominates, while Chest Sorter is
somewhat heavier because the number of categories marked with 0 predominates
(Fig. 2).
Figure below shows a histogram that provides insight into the attendance of inter-
actions. The picture shows all the titles (Fig. 3).
712 A. Tolic et al.
The graph above shows that the highest activity is present at the end of September,
while the end of July has the lowest recorded game play activity. The time range of the
given data can also be seen on the graph. Thus, the time interval covers four months
(Fig. 4).
3 Model Development
New features are defined based on the resulting data set. Feature parsing was based on
free assessment taking into account computing power/hardware capacity. The history
of the given data is observed. Since the prediction will take place at the level of
installation identification numbers, the operations listed below are performed on groups
of the same. The goal is to have a unique installation identification number per
instance. Thus, after grouping, counting, summing, and arithmetic mean computation
operations are performed on the following features [10].
for_aggregation = {
'event_count' : sum,
'game_time' : ['sum','mean'],
'event_id' : 'count',
'total_duration' : 'mean',
'duration' : 'mean',
'size' : 'mean',
'level' : ['sum','mean'],
'round' : ['sum','mean'],
'correct' : ['sum','mean'],
'source' : sum,
'weight' : ['sum','mean'],
'coordinates_x' : 'mean',
'coordinates_y' : 'mean',
'coordinates_stage_width' : 'mean',
'coordinates_stage_height' : 'mean'
The motivation for constructing the previous features is as follows: it has been
estimated that the coordinates feature with the corresponding suffix could be useful as
they may indicate distractions and distractions to the child while looking at the screen.
Other extracted features could also be directly related to the child’s skills [6–8].
Model Selection and Optimization of Hyperparameters
Since this is a problem of supervised learning, with the availability of input data,
appropriate output categories are available. The histogram of the target variable is
shown in the following figure (Fig. 5).
The used data set is divided into a training data set that holds 80% of the total data
share, and a test set that holds the remaining 20% of the total data share. XGBoost and
CatBoost are trained. In addition to the data set for training and testing, a third set, the
so-called validation set, is often used. In this paper, the validation set will be 20% of the
training set. Then two of the hyperparameters for each of the models.
grid_parameters_xg = {
'n_estimators' : [50, 100, 150],
'min_child_weight' : [3, 4, 5],
'max_depth' : [2, 3],
'learning_rate' : [0.01, 0.03, 0.05],
'reg_lambda' : [0.01, 0.1, 1],
'reg_alpha' : [0.01, 0.1, 1]
}
grid_parameters_cb = {
'depth' : [1, 2],
'learning_rate' : [0.01, 0.03, 0.06, 0.08],
'iterations' : [100, 150, 200],
'l2_leaf_reg' : [0.01, 0.1, 1, 20, 50]
The following figures visualize the learning curves for XGBoost and CatBoost
(Figs. 6 and 7).
CatBoost
- depth = 2
- learning_rate = 0.08
- iterations = 200
Para>
Model Evaluation
In this part we describe concrete results that indicate how reliable and applicable the
models made are, and how well the models generalize. The choice of metrics generally
depends on the machine learning task. Taking into account the imbalance of the set of
values of the target variable, the metrics of the square-weighted cap and some addi-
tional metrics visible in the classification report were used to assess the success of the
classifiers implemented in the paper.
Additionally, two simpler models were made. Here, the goal was to outperform the
results of simpler models with more complex models. The first is a simpler model
called DummyClassifier that does random guessing. In the second case, a mod-rule
solution was created. The initial default set was divided into a training set and a test set
where the training set was 75% of the total share and the test set was 25% of the total
data share. Target values were constructed by grouping assessment titles and finding
mod-values of the same using only the training set (Tables. 1, 2, 3, 4, 5 and 6).
The following figures show the associated (confusion) matrices with comparisons of
actual and predicted XGBoost and CatBoost model classes (Figs. 8, 9, 10 and 11).
The following figures show the importance of individual features for the XGBoost
and CatBoost models. It is important to note that the selection of features works on the
built-in principle, that is, the algorithm itself finds the most important features (Figs. 12
and 13).
Learning Success Prediction Model for Early Age Children 717
Fig. 8. XGBoost - confusion matrix (test set) Fig. 9. XGBoost - confusion matrix (train-
ing set)
Fig. 10. CatBoost - confusion matrix (test set) Fig. 11. CatBoost - confusion matrix (train-
ing set)
Fig. 12. XGBoost - importance of features Fig. 13. CatBoost - importance of features
718 A. Tolic et al.
4 Conclusion
The simplicity of the game certainly does not imply the simplicity of the data.
Moreover, the data are extremely complex, so understanding them takes some time.
The dimension of the data directly affected the execution time of the algorithms. In
particular, the execution time of individual algorithms was measured in hours. In order
to prepare the data in a correct way, individual phases required the implementation of
additional algorithms. The implementation of the extraction algorithm due to its
complexity required special attention and time.
This research indicates that the implemented models, in addition to the previously
mentioned parameters, can predict the outcome of the evaluation with a square
weighted kappa value of 0.76 for XGBoost and 0.74 for CatBoost on the test set. The
values of the square-weighted cap, observing the training set, are 0.79 for XGBoost and
0.76 for CatBoost. Thus, the results related to the square-weighted kappa support the
XGBoost model. Results F1 (macro) - scores are identical for implemented XGBoost
and CatBoost models, with a value of 0.62 on the test data set. The same score on the
training set is 0.72 for XGBoost and 0.68 for CatBoost. Additionally, two simpler
models were implemented where the goal was to outperform the simpler ones with
more complex models, and the experiment proved to be successful. The more complex
XGBoost and CatBoost models significantly outperformed the simpler models.
In conclusion, the elevation of existing results could be achieved by possibly
searching a larger space of hyperparameters, extracting the values of higher levels of
the event_data feature, and more advanced feature construction, taking into account
hardware capacities. Furthermore, the PBS KIDS Measure Up! provides free access to
content that has been proven to help children learn. In the generated data of the
mentioned type of game, it was crucial to identify the relevant prediction factors. This
research, among other things, provide a basic insight into the same. Such results are
extremely important in building future education systems, and more advanced analysis
of relevant features can be of great help when it comes to a child’s development. Based
on the experiments performed, it is evident that even in a limited work environment, a
respectable prediction can be achieved that can be crucial for an understanding how
cognitive child’s growth can be measured and, hopefully, empowered.
Notes
1. https://pbskids.org/apps/pbs-kids-measure-up.html.
2. https://datasciencebowl.com/.
References
1. Ben-David, A.: Comparison of classification accuracy using Cohen’s Weighted Kappa.
Expert Syst. Appl. 34(2), 825–832 (2008). https://www.sciencedirect.com/science/article/
abs/pii/S0957417406003435
Learning Success Prediction Model for Early Age Children 719
2. Chen, T., Guestrin, C.: XGBoost: A scalable tree boosting system. In: Proceedings of the
22nd acm sigkdd international conference on knowledge discovery and data mining. ACM
(2016) https://dl.acm.org/citation.cfm?doid=2939672.2939785
3. Domingo, P.: A few useful things to know about machine learning. Commun. ACM 55(10),
78–87 (2012). https://homes.cs.washington.edu/*pedrod/papers/cacm12.pdf
4. Hastie, T., Tibshirani, R., Friedman, J.: Introduction to Statistical Learning. Berlin: Springer
(2013). http://faculty.marshall.usc.edu/gareth-james/ISL
5. Hastie, T., Tibshirani, R., Friedman, J.: The Elements of Statistical Learning. Berlin:
Springer (2009). https://web.stanford.edu/*hastie/Papers/ESLII.pdf
6. Nielsen, D.: Tree Boosting with XGBoost-Why Does XGBoost Win “Every” Machine
Learning Competition? MS thesis. Trondheim: NTNU (2016). https://ntnuopen.ntnu.no/
ntnu-xmlui/bitstream/hanle/11250/2433761/16128_FULLTEXT.pdf, prosinac. 2019
7. Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A.V., Gulin, A.: CatBoost: Unbiased
boosting with categorical features. In: Advances in neural Information Processing System.
Moskva: Yandex (2018). https://papers.nips.cc/paper/7898-catboost-unbiased-boosting-
with-categorical-features.pdf
8. Skansi, S.: Introduction to Deep Learning. Springer, London (2018)
9. Intelligent Computing & Optimization, Conference proceedings ICO 2018, Springer, Cham
(2018) ISBN 978-3-030-00978-6
10. Intelligent Computing and Optimization, Proceedings of the 2nd International Conference on
Intelligent Computing and Optimization 2019 (ICO 2019), Springer International Publishing
(2019) ISBN 978-3-030-33585-4
Advanced Analytics Techniques for Customer
Activation and Retention in Online Retail
1 Introduction
Online retail industry has a whole new business potential that has not yet been deeply
unlocked both from a market and customer perspective. The clear difference from the
offline world is that Internet companies monitor consumers' decisions in the buying
process. But like any traditional offline marketplace, the network industry has its own
set of business rules, stakeholders, business problems and solutions. Some of the
typical Internet retail challenges are Churn Prediction, Retention, and Prevention
(PRP). A possible approach towards solution for PRP is to use machine learning
techniques. The main prerequisite is the unique identification of the customer required
for anticipation and retention. The second prerequisite is a reliable relationship between
the provider and the user through the identity and consent of the customer. Forecasting
As a starting point for research we need to analyse the customer, what and which
relevant customer data we have, create and define customer journey, analyse the
behaviour of our customers, classify and segment the customer in similar groups.
Differentiating between customer groups has several advantages. In order to be able to
adequately differentiate the customers, we must assure that the data used for this
purpose meet certain requirements. The most important factors are the following: (f1)
Buying behaviour. Suitable indicators for future buying behaviour; (f2) Measurability.
Measurable and recordable with existing market research methods; (f3) Accessibility.
Ensure targeted addressing of the formed segments; (f4) Ability to act. Ensure the
targeted use of the marketing tool; (f5) Profitability. The benefit of the analytics should
be greater than the costs incurred; (f6) Temporal stability. Longer-term validity of the
information collected. A distinction is made between the segmentation criteria between
geographical, socio-demographic, psychographic and behaviour-oriented content. In
our case we would concentrate on the non-purchase behaviour because this is the
information, we need to understand for our churn case. The so-called “lifestyle seg-
mentation” divides the consumers according to their lifestyle. It contains criteria that
address individual and collective value orientations and goals. A personality is iden-
tified that, in conjunction with the segmentation criteria just mentioned, demographic,
socio-economic as well as cultural origin allow a distinction according to lifestyle. We
can distinct and segment online retail customers based on their lifestyle and put them in
several groups. Most retail customers can be put in six lifestyle groups: struggles,
action oriented, believers, experiencers, achievers, full fields which are mainly self-
explanatory [23].
result would be a loss-making business. In order to avoid this, the evaluation of the
customer with regard to his purchase and the associated earnings/costs value, Customer
Lifetime Value (CLV) calculation is used. CLV defines itself as the sum of the income
that a customer generates for the company during the business relationship, minus the
costs for customer acquisition, the sales process and the customer service. Putting this
in the perspective of the online retailer, we used order costs, value of goods, delivery
costs and other administrative costs compared with the order-related earnings values
per calendar year. The customer should be included in this calculation from the first
purchase and seasonal influences should be excluded. The result of the calculation can
either be positive or negative telling us that customers with negative earnings are
undesirable. Customers with a positive value can be divided into categories from A to
C, in which category A contains the top customers. Other customer categories are CR
(reactivated customers) and CN (new customers). A further distinction is made by
calculating total number of orders within the last 12 months, if the customer had only
one or more than three in this period. To be able fine tune this customer segmentation, a
quotient that describes the effort that was necessary to generate the earnings would be
used. The quotient calculation will differentiate less profitable from profitable cus-
tomers and how to cluster them. The yield quotient (earnings ratio) used for this is
calculated from the order-related earnings value divided by the delivery value (sum of
the sales price successfully delivered products). This short mathematical example
illustrates yield quotient (YQ) = EPO / Delivery Value. If you set this quotient
depending on the respective customer category, further segments are created. This
makes differences in efficiency within a category more clearly visible. The quotient
alone does not make any statements about the coverage of additional costs that were
not yet included in the earnings per order (EPO) calculation which can be either high or
low level. The preference of the order profit amount is higher than the order efficiency
(the yield quotient). Because the earnings ratio was opposite to the EPO, in our online
retail company, A1 customers brought an average profitability of 38.0% and A2 cus-
tomers value of 44.1%. For churn prevention, the scoring of customers is very
important since we need to calculate break-even points for each churn case individually
to know what customer we need to address and reactivate or retain. The customer
categories should be visually differentiated from each other so we can better see where
our target is. If you plot the two characteristics (EPO and yield quotient) on the axes of
a diagram, you get a cluster that shows the customer categories according to the cost-
income ratio. The clear subdivision by the earnings quotient becomes clear, but above
all the areas of the customer segments are shown with their respective reach. Cate-
gories B and C are limited to 300 EUR each in their EPO area which raises the question
of the distribution of the customers. Customers with a high EPO are forced to show
regular activity in order to achieve this earnings value at all. It is therefore unrealistic
that large parts of customer category A suddenly become inactive which makes them
excluded from churn target. Based on previous company experience, it is primarily new
customers who miss another purchase recognized as a good candidate for the customer
churn prevention program. Another group are already reactivated customers. They
form a large part of the customer base and are mostly only active again due to an
incentive. With the churn prevention process, they too need to be converted into regular
buying behaviour [24] (Fig. 1).
Advanced Analytics Techniques for Customer Activation 723
3 Research
1
Activity and inactivity of customers refers in the present use cases to orders; other activity criteria
(newsletter, online shop, …) are treated as indicators.
724 I. Matic et al.
frequency, although the customer ultimately has no loss of interest and thus results in a
process to prevent customer churn. It is therefore recommended to set a minimum
interval that must be exceeded in any case before a churn action is taken [8]. For this
churn use case an empirical value for the minimum interval is set to: ILP, min = 2
months (this value can be adapted based on each individual case). On the other hand,
for low-frequency customers, this idea can also be accepted but with inverted logic.
Exceeding a certain period for which no churn prevention has been started in the
medium term can potentially lead to a complete loss of the customer if the approach is
made too late. Formerly regularly active customers who avoid their direct qualification
for Churn due to ever increasing purchase distances, but who nevertheless have a
certain risk, cannot be included. A maximum period should be set, with which the
process must be started in order to address the customers just mentioned. This period
can be defined with typical value: ILP, max = 6 months (this value can be adapted based
on each individual case). If we summarize this in a final definition for churn risk in the
sense of the churn case, the following definition applies: Churn is always present if the
period since the last purchase of a customer is twice as long as the average order
interval. This period is limited by at least two months and by a maximum of six months
at the top [9–11].
is four times as large as the average purchasing interval (IAP). There are two limitations:
(1) High-frequency buyers cannot enter the process until after 3 months; (2) Low-
frequency buyers enter the process after 12 months at the latest. This principle is quite
similar as for churn prevention, with the difference of the time period observed and
taken into account for the customer segmentation (Fig. 2).
Fig. 2. Churn process overview (IAP = average purchasing interval; ILP = elapsed time since
last purchase)
In this comparative view we see that the main difference is in the elapsed time since
last purchase values which is much greater for the reactivation case. Using this prin-
ciple, we would segment different customer group and we would use different process
for the churn case. Also, the modelling part would be different since we would use
slightly different input parameters (Fig. 3).
If we put complete customer base in this IAP/ ILP matrix and visually present the
result we would get status view on potential target group for the churn prevention and
the reactivation process. It is to be expected that the reactivation group would be bigger
but churn prevention group could show more financial potential.
Advanced Analytics Techniques for Customer Activation 727
fees. Before applying the propensity model on the whole customer based some test
campaigns should be conducted. These tests could be applied using minor customer
segment: Testing of different voucher values, testing of different content variants (product
groups), Testing of different tonalities (considering the customer typologies) or Selection
of channel: email is much more inexpensive and more flexible than print but with print the
achievable effect size is significantly higher as compared to email.
Differential Response Analysis (Uplift Model). One of the best-known models which
can make changes in customers and their respective buying behaviour visible through
stimuli triggered by companies, is the so-called uplift or incremental model. In order to
optimally pursue its goals, lower the churn rate and the associated increase in sales
while ignoring the unprofitable customers, the uplift model can be used to optimally
differentiate customers. Before the customer scoring can be continued later in order to
split up the customer groups and then addressed by the churn process, the first step
should be to respond to a stimulus (in our churn case, this is synonymous with
advertising) like addressing the customer at different communication levels (e.g.
newsletter, letter advertising, etc.). The background is that potentially unnecessary
investments should be excluded in advance. The conceivable situations in which this
would be the case is visible in a four-field stimulus panel. Addressing customers who
buy regardless of a stimulus (see figure below) would be just as unnecessary as
addressing customers who would never order (bottom left). Most important is the
exclusion of the small group of customers, at the same time, where a stimulus would
prevent an order (bottom right). It is important to filter out the customers who fall into
the grid at the top left and seem to only become active again due to a stimulus. This
filtering can be achieved relatively simply by carrying out a campaign-dependent Ps/Po
test. Similar like in the propensity model, forecasting methods based on machine
learning, regression models and decision matrices are necessary, but in uplift model we
create a monetary assessment of the individual customer potential [13–16]. For the final
evaluation purpose, the customer groups are divided into target and control groups,
with the control group mostly accounting for almost 7% of the total group. Previous
experience has shown that this amount has always been optimal so far, on the one hand,
to make deviations visible and, on the other hand, to determine the maximum beha-
viour of the target group. Potential reactivation measures for the dispatching of
vouchers are similar with the propensity model with some additions like incentive
which is typically more “aggressive” as compared to churn prevention and rewards for
being a loyal customer. Same tests campaigns proposals as for propensity models can
be used for the reactivation model but with some further marketing campaigns opti-
misations. Differential Response Analysis usually maximizes the Marketing Campaign
effects [17]. For the churn case, optimization of the marketing campaigns is needed.
Following table shows potential effect using the different campaigns approach
(Table 1).
Advanced Analytics Techniques for Customer Activation 729
Purchase rate for the conventional campaign selections (like Campaign B) maximize
the value for the target group (8,3%) but can be high even without a stimulus (8,2%)
which gives only 0,1% campaign effect. Differential response analysis maximizes the
effects of campaigns and will propose only the best Campaigns which have maximal
points effect value, example is in the case – Campaign C with 2.7% + points.
Relevant Forecasting Methods. Typical forecasting methods used in the Machine
Learning Environment in the churn prediction and prevention are Regression models,
Decision three and Neural networks. Regression algorithm was chosen because was the
best in this churn case. Neural was too complex and did not add any gains and decision tree
did not deliver good scores. The forecasting process from the individual attributes to a
final scoring takes place in several steps and at some points allows an independent and
consequent improvement of the applied decision criteria by the procedure itself. Based on
the data sources from customer database and behavioural (clickstream) data, a first cus-
tomer profile is developed, which brings additional input variables with it and flows into
the data forecasting model. There, various decision trees are used to identify individual
customers through existing correlations (correlation filter), such as the payment method-
basket size key figure (filtered out). Logistic regression filters develop an individual
forecast for the customer as to whether the resulting uplift would turn out to be positive
and the decision-making processes of the customer (Decision Tree Learner / Predictor),
which are shown using clickstream (behavioural) data, also influence the evaluation.
Evaluation of the prediction models is based on some key figures used for the assessment
of the forecasting quality and selectivity [12] (Figs. 4 and 5).
Fig. 4. Lift chart for web tracking Fig. 5. Gains chart for web tracking
730 I. Matic et al.
4 Results
After initial analysis and necessary agreement with the key business users, final
implementation scenario represents the potential customer base for the online retail
churn processing scenario. As mentioned in status view we segment the customers
based on the purchasing behaviour and focus on the most potential customer groups.
The complete customer base is around 2 million customers. These end results tell us
which customer segment is the best suited for the churn prevention and which group is
better to approach with the reactivation measures. Here are the key characteristics of
this customer status scenario: (i) Customers enter the churn prevention process as soon
as ILP is more than twice as large as the average purchasing interval (IAP); (ii) Cus-
tomers enter the reactivation process as soon as ILP is more than four times as large as
the average purchasing interval (IAP); (iii) 18 months after the last purchase customers
are assigned the status “inactive” (independent of the average purchasing interval). We
can split the customer segment to 3 vertical models based on the last purchase period.
The main focus for our model development, training and scoring is on the most
prospective churn customers (orange coded customer groups) which is a combination
of the horizontal and vertical attributes. Following this logic, the best churn prospective
groups are the middle groups of the status view matrix (M06-M08xM13-M29 col-
umns). If we address the customers too soon (left side) or too late (right side) our churn
activities would have less chance to success (Fig. 6).
We have prepared 3-steps filtering of the customers to achieve the best potential
customer segment for our churn case. In the first step we select the customers which
ordered in the last 2-year period. Then we filtered only customers with more the two
orders in this period. In the third step, we excluded internal employees (and connected
persons), customers with some profile blocks/locks like marketing block, order block or
Advanced Analytics Techniques for Customer Activation 731
dunning/financial lock. Final number on the left represent customer segment base
which are eligible (good potential candidate) for churn prevention or reactivation
activities. In the churn case definition chapters, we defined two customer selection
principles for the proof of success: Manual selection - just picking up some potential
churn customers by “thumbs” principle and Model based - ML algorithm model (re-
gression) segment. For the customer segment selection, we selected best-score interval,
not the highest potential score customers but just below this. This was chosen because
first short test with highest score did not deliver expected results. This could be
explained with the fact that highest profile churn customers have firm reasons and
could not so easily be persuaded to switch back or to (re)activate their purchases. Fort
the main proof of the success we calculate the uplift and additional gross sales
potential. The uplift stands for the effect caused by the implementing respective mar-
keting measure compared to the manual selection or (control group) without these
measures. The basis for comparison is the customer average gross sales per each
respective group. We would also check additional gross sales impact, which would
give the additional gross sales purchase value per each targeted customer. They rep-
resent the additional sales of the model-based selection compared to the manual
selection and expresses the additional benefit that results from the model-based
selection. The additional gross sales (mainly upsell and cross-sell orders) are the results
from the uplift values of the respective campaigns. Future subsequent orders are not
considered here.
Here is a comparison to the baseline or manual selection for churn prevention
case: (i) The uplift in gross sales per customer compared to a manual selection was
between 5.40 and 7.50%; (ii) The uplift in gross sales per customer compared to a
model-based selection (best score interval) was around 18%; (iii) Additional gross sales
per customer compared to a manual selection (model-assisted) was 10.5%; (iv) Addi-
tional gross sales per customer compared to ‘no activity’ (baseline) was 18%. For the
quantification of the uplift/improvement we can use of the following formula:
Pot_Gross_Sales = additional gross sales per customer * number of customers per
wave * number of waves, where waves are the marketing waves, we have conducted
during the specific marketing campaigns. The usual number of waves was between 2
and 12.
Here is a comparison to the baseline or manual selection for churn reactivation
case: (i) The uplift in purchase probability per customer compared to a manual
selection was between 2.0% and 2.5%; (ii) The uplift in purchase probability per
customer compared to a model-based selection (best score interval) was 4.5%;
(iii) Additional purchase probability per customer compared to a manual selection
(model-assisted) was 2.0%; (iv) Additional purchase probability per customer com-
pared to ‘no activity’ (baseline) was 4.5%. For the quantification of the churn reacti-
vation improvement we can use this simple formula: Pot_Active_Cust = additional
purchase probability per customer * number of recipients per wave * number of waves.
If we put the values from the Print test campaign and calculate how many additional
active customers (compared to baseline) resulting from a regular monthly reactivation
process we have so far and use the values for the customer segment we have targeted
(60000) and the number of marketing waves we have performed during the campaign
(12) the final result is following: Num_Active_Cust = additional purchase probability
732 I. Matic et al.
per customer * number of recipients per wave * number of waves = 0.045 * 60,000 *
12 = 32,400. Additional gross sales (compared to baseline) resulting from a regular
monthly reactivation process can give us the monetary explanation of this potential.
This can be change based on the company gross sales figures and market potential but
if we put the numbers what we have so far, the expected range would be from 2–6% on
the yearly base (Table 2).
In order to calculate the complete potential, the first step is to calculate the number
of customers that could be retained by applying the churn prevention process.
Implementing the 12% churn reduction rate derived from the 19.2% baseline value, a
new customer churn rate of 16.9% is obtained. If we apply it to the target groups of
active customers in the AC, CN and CR customer categories, which corresponds to
95.3% of the total number of customers, we get additional active 32,200 customer.
Multiplying this by the average order value of all customers, excluding related costs
such as vouchers or email costs, we can achieve a final uplift value of about 3% on
annual gross sales. Quite a considerable amount, which certainly exceeds the cost of
overall implementation in a very short period of time [22].
5 Conclusion
Online retail company already has data analysis potential as essential part and must
learn how to use it in the best manner. With the ever-increasing competitive situation
among online retailers, customer churn is accelerating for several reasons. This is due,
among other things, to the freedom of choice and the ability to switch quickly between
the websites. Customer loyalty therefore drops just as quickly as the competition
grows, so that there is a need for action in order to avoid further customer churn
through the (artificial) intelligent customer segmentation and carefully planning mar-
keting prevention campaigns. Successful implementation requires an intensive con-
sideration of various influencing factors as well as the development of solution
approaches, taking into account the risks and problems that arise. If we look at the
maximum potential that can be raised by addressing the most likely customers, in the
case of our online retailer which has couple of million customers, we are talking around
Advanced Analytics Techniques for Customer Activation 733
the amount of five million euros per year and in connection with this around 32,000
customers who became active. The use of data-driven programs to identify and prevent
migration risks offers some advantages that should not be missed in the current context
of strong competition in the online retail industry because maintaining the customer
base of a company is vital task for such a company. If a large number of customers
drops out, only the much more complex and cost-intensive method of acquiring new
customers remains in the midterm perspective. If this does not succeed, there is a long-
term risk of falling below the average customer cost limit with the consequences of a
potential complete cash-flow insolvency.
References
1. Mitchell, T.: Machine Learning. McGraw Hill, New York (1997)
2. Mrsic, L., Klepac, G., Kopal, R.: Developing Churn Models Using Data Mining Techniques
and Social Network Analysis. IGI Global, Pennsylvania (2015). https://doi.org/10.4018/978-
1-4666-6288-9
3. Provost, F., Fawcett, T.: Data Science for Business. O’Reilly Media, Inc., Sebastopol
(2013).978-1-449-36132-7
4. Burkov, A.: The Hundred-Page Machine Learning Book, Leanpub, Victoria (2019) www.
themlbook.com
5. Andrew, N.G.: Machine Learning Yearning (2018) www.deeplearning.ai/machine-learning-
yearning
6. Kraljević, G., Gotovac, S.: Modeling data mining applications for prediction of prepaid
churn in telecommunication services. Automatika 51(3), 275–283 (2010)
7. Hemann, C., Burbary, K.: Digital Marketing Analytics. Que Publishing, Indianapolis (2013)
ISBN-13: 978–0–7897–5030–3
8. Farris, P., Bendle, N., Pfeifer, P., Reibstein, D.: Marketing Metrics. Pearson Education,
Upper Saddle River (2010)
9. Harvard Business Review. https://hbr.org/2014/10/the-value-of-keeping-the-right-customers
10. Ahlemeyer-Stubbe, A., Coleman, S.: Monetising Data: How to Uplift your Business. Wiley,
Newark (2018)
11. Franks, B., Bucnis, R.: Taking Your Analytics Up A Notch By Integrating Clickstream Data.
Teradata, Atlanta (2011)
12. Schellong, D., Kemper, J., Brettel, M.: Clickstream data as a source to uncover con-sumer
shopping types in a large-scale online setting, Association for Information Systems, ECIS
(2016)
13. Bishop, C.: Pattern Recognition and Machine Learning. Springer, New York (2006)
14. Strickland, J.: What are Uplift Models (2015) https://www.analyticbridge.datasciencecentral.
com/profiles/blogs/what-are-uplift-models
15. Cottineau, C., Chapron, P., Reuillon, R.: Growing models from the bottom up. an
evaluation-based incremental modelling method (EBIMM) applied to the simulation of
system of cities. J. Artif. Societies Soc. Simul. 18(4), 9 (2015)
16. Bucklin, R., et al.: Choice and the internet: from clickstream to research stream. Mark. Lett.
13, 245–258 (2002)
17. Han, J., Kamber, M., Pei, J.: Data Mining. Concepts and Techniques. Morgan Kaufmann,
Waltham (2011)
18. Davenport, T.: Big Data @ Work. Harvard Business Review Press, Boston (2014)
734 I. Matic et al.
19. Freiling, J., Kollmann, T.: Entrepreneurial Marketing. Springer, Wiesbaden (2015)
20. Few, S.: Show Me the Numbers. Analytics Press, Burlingame (2012)
21. Aaker, D.A.: Strategic Market Management, 10th edn. Wiley, Hoboken (2014)
22. Olson, D.L., Delen, D.: Advanced Data Mining Techniques. Springer, Berlin (2008)
23. Intelligent Computing and Optimization, Conference proceedings ICO 2018, Springer,
Cham, ISBN 978–3–030–00978–6
24. Intelligent Computing and Optimization. In: Proceedings of the 2nd International
Conference on Intelligent Computing and Optimization 2019, ICO 2019, Springer
International Publishing, ISBN 978–3–030–33585–4
An Approach for Detecting Pneumonia
from Chest X-Ray Image Using Convolution
Neural Network
Abstract. Pneumonia is one of the most cynical problems to human beings all
over the world and detecting the presence of pneumonia in an early stage is very
necessary to avoid Premature Death. According to the World Health Organi-
zation above 4 million sudden deaths happen each year from domiciliary air
pollution correlated diseases inclusive of pneumonia. Generally, pneumonia can
be identified using chest X-ray images that are performed by an expert radiol-
ogist. But only rely on the radiologist sometimes blocks the treatment because of
detecting diseases from the chest X-ray images which requires human effort,
experience, and time. In this case, Computer-aided diagnosis (CAD) system is
required for identifying pneumonia from chest X-ray images automatically. In
this research, a modified model is proposed using Convolution Neural Network
(CNN) model to train sample data to relegate and diagnose the presence of
pneumonia from an amassment of chest X-ray images. From the experimental
result, it is found that the proposed model performs better results (89%) com-
pared to other related existing algorithms.
1 Introduction
The threat of pneumonia is a great problem for developing nations where billion of
people facing hunger, penuriousness, and unemployment related problems. Pneumonia
is one of the drastic inhaling inflammation in which the lunch of individuals are
affected. It is a disastrous disease that causes inflammation alveoli of individuals’ lunch
and it is very painful to inhaling naturally when the alveoli is full of other fluid or pus.
According to the report of World Health Organization (WHO) above 4 million
immature deaths occur each year from domiciliary air pollution correlated diseases
inclusive of pneumonia [1]. Above 150 million individuals get affected with pneu-
monia every year particularly children below the age of 5 years [2]. In these conditions,
the complication can be further instigated due to the scarcity of medical supplies and
workforce. Looking at the situation of Africa's 57 nations, where there exists a gap of
2.3 million doctors and nurses [3, 4]. For this immense population, a very accurate and
fast diagnosis means a lot for saving those lives.
The manual process of identifying pneumonia from chest x-ray image are
accomplished by an expert radiologist. This procedure requires humongous time,
expertise, and personnel. To overcome this difficulty, an unusual but straightforward
model is required that can automatically perform the optimal relegation tasks from a
collection of sample datasets with a deep neural network model. Here in this study, a
Convolution Neural Network (CNN) model is proposed to train sample data to relegate
and diagnosis the existence of pneumonia from an amassment of sample datasets.
Consider another deep learning algorithm that depends only on transfer learning
techniques based on traditional handcrafted to attain an incredible accuracy, and to
develop a convolution neural network model that is trained from scratch with a sample
dataset in order to identify the existence pneumonia in the human body. The proposed
CNN model could help us to alleviate the reliability and explainable challenges during
working with medical imagery. Compare another deep learning relegation tasks with
ample image depository, it is arduous to obtain a substantial quantity of dataset for this
pneumonia relegation task.
It is found that the CNN-stimulated deep learning framework have become one the
usual choice for most of the recent researchers for relegation of images. Examples some
of the excellent and most commonly used architectures for medical image classification
are U-Net [5], SegNet [6], and CardiacNet [7].
2 Related Work
There are a lot of researchers have been doing their research in deep learning for image
classification in past. Kermany, Daniel S., et al. [12] uses CAD system based on a
deep-learning framework to detect blinding retinal diseases using a dataset of optical
coherence tomography images and pediatric pneumonia using chest X-ray images.
Antin, Benjamin, Joshua Kravitz, and Emil Martayan [13], developed an algorithm that
uses a 121-layer Convolutional Neural Network similar to Chex Net for the classifi-
cation of image collected chest x-ray images from the NHIS dataset to identify the
presence of pneumonia. From the experimental result, it was observed that the pro-
posed deep learning model achieved an overall accuracy of 0.684. Rajpurkar, Pranav,
et al. [14], developed an algorithm that able to identify 14 disease viruses utilizing the
Chex Net algorithm. Park et al. [15], developed a Gabor filter-based algorithm for the
detection of rib reduction on abnormal tissue by scanning the X-ray imaging. Dina A.
Ragab [16] developed a CAD system for the classification of malignant and benign
tumors from mammography. Livieris, Ioannis E., et al. [18], developed a new algorithm
based on ensemble semi-supervised learning in order to relegate the abnormalities of
lung. Choudhari, Sarika, and Seema Biday [17], developed an algorithm that uses
neural networks to detect skin cancer. Omar, H. S., and A. Babalık. [20], developed a
convolution neural network model to diagnoses the presence of pneumonia from the
sample datasets. Here the researchers observed that compared to other machine
learning algorithms their proposed model achieved higher accuracy of 87.5. Abiyev
et al. [21], also developed a deep convolutional neural network for the detection of
An Approach for Detecting Pneumonia from Chest X-Ray 737
Chest diseases. In this work, the researcher work uses 380 images of size 32 32
pixels and from the experiments it was observed that backpropagation neural networks
attained the utmost classification accuracy rate of 99.19%. Naranjo-Torres, José, et al.
[22] construct a CNN based model for processing collection of fruit images. Han,
Fenglei, et al. [23] uses a CNN modified model for processing image of underwater to
help the robot. Alazab, Moutaz, et al. [24] developed a technique inherited from CNN
model for the diagnosis of COVID-19 from original datasets. Chakraborty, S., Aich, S.,
Sim, J.S., Kim, H.C [25] presents computer aided based Convolutional Neural Network
(CNN) models to perfectly identify pneumonitis from chest X-ray image dataset, that
can be used in the real world application such medical department in order to handle
pneumonia. In order to reduce overfitting dropout regularization technique is applied in
the model. Anthimopoulos, M., Christodoulidis et al. [26] Designed Convolution
neural network model for the classification of Interstitial Lung Liseases (ILDs). To
conduct the experiment on CNN model, 14696 images were used which were taken by
120 CT scans from several OCR device and health center and achieve a remarkable
classification accuracy of 85.5%.
3 Proposed Methods
For assessing the performance of the proposed model, the experiments and evaluation
steps are presented in detail. Due to experimental purpose, the chest X-ray image
dataset was taken from [8]. As a deep learning framework, Keras is used. It is an open-
source with tensor flow backend in which the proposed convolutional neural network
model is built and trained. For the experimental purpose an HP laptop is used with the
configurations, Intel Core i5–8265 processor, 8 GB RAM, cuDNN v7.0 library, and
CUDA Toolkit 9.0. Here Python 3.7 version is used to compile the proposed model.
3.1 Dataset
For experimental purpose, the datasets were collected from [8] which consists of a total
of 5840 X-ray images of anterior-posterior chests which were picked punctiliously
from patients aged between 1 to 5 years old. The datasets were assembled by the
National Institutes of Health (NIH), and it has been carefully selected from Guangzhou
Women's Medical Center in Guangzhou. The entire chest X-ray images were selected
from the patients in the time of their scheduled checking time. Table 1 shows the
details of the dataset and Fig. 1 and Fig. 2 shows samples images one is being normal
and the other is affected by pneumonia from the dataset.
Fig.1. Sample Chest X-ray image without pneumonia affected taken from dataset
Fig.2. Sample Chest X-ray image with pneumonia affected taken from dataset
4 Experimental Result
To assess the performance and effectiveness of the suggested model same experiment is
taken several times. The parameters are tuned to get better performance of the proposed
model. During the training process the epoch number and dropout is set as 3 and 0.2
respectively. The graphics view of the accuracy and loss function at the time of the
training process are shown in Fig. 4(a) and (b).
740 S. Kar et al.
From the experiment it is noticed that the proposed algorithm perform better
compared to other machine learning and deep learning algorithm. The proposed
method accuracy is found 89%. Table 3 shows the experimental result of the proposed
model CNN against state-of-the-art supervised algorithms SMO [9], C4.5 [10], 3NN
[11], WvEnSL3 [18] and CNN [23] based on dataset of Pneumonia. The overall results
show that the proposed model exhibits better performance compare to other supervised
algorithm listed in the table. (SMO), (C4.5), (3NN) etc. are the base learner self-labeled
machine learning algorithm for object relegation task.
An Approach for Detecting Pneumonia from Chest X-Ray 741
Fig. 4. (a). The accuracy diagram during the training process (b). The loss diagram during the
training process.
Convolution Neural Network is one of the most favored and established deep learning
algorithm for image processing and image classification tasks [19]. In this research an
algorithm is developed in order to relegate and diagnosis the existence of pneumonia
from an amassment of chest X-ray images. The developed CNN model is trained and
evaluated from the chest X-ray images to get the presence of pneumonia. The proposed
CNN model is compared with other machine learning and CNN methods. The simu-
lation result shows that the proposed CNN model is capable of achieving more accurate
results than the existing algorithms. This research work can be extended for the other
chest related diseases such as effusion, pleurisy, pneumonia, infiltration, bronchitis
nodule, atelectasis, pericarditis, cardiomegaly, pneumothorax, fractures.
References
1. World Health Organization. Household Air Pollution and Health [Fact Sheet] Geneva,
Switzerland: WHO (2018). https://www.who.int/newa-room/fact-sheets/detail/household-
air-pollution-and-health.
2. Rudan, I., Tomaskovic, L., Boschi-Pinto, C., Campbell, H.: Global estimate of the incidence
of clinical pneumonia among children under five years of age. Bull. World Health Organ. 82,
85–903 (2004)
3. Narasimhan, V., Brown, H., Pablos-Mendez, A., et al.: Responding to the global human
resources crisis. Lancet 363(9419), 1469–1472 (2004). https://doi.org/10.1016/s0140-6736
(04)16108-4
4. Naicker, S., Plange-Rhule, J., Tutt, R.C., Eastwood, J.B.: Shortage of healthcare workers in
developing countries. Africa Ethn. Dis. 19, 60 (2009)
5. Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image
segmentation. In: Navab, N., Hornegger, J., Wells, W., Frangi, A. (eds.) International
Conference on Medical image computing and computer-assisted intervention. Springer,
Cham (2015)
6. Badrinarayanan, V., Kendall, A., Copolla, R.: Segnet: deep convolutional encoder-decoder
architecture for image segmentation (2015) https://arxiv.org/abs/1511.00561
7. Mortazi, A., Karim, R., Rhode, K., Burt, J., Bagci, U., Cardiacnet.: Segmentation of left
atrium and proximal pulmonary veins from MRI using multi-view CNN. In: Descoteaux, M.,
Maier-Hein, L., Franz, A., Jannin, P., Collins, D., Duchesne, S. (eds.) Medical Image
Computing and Computer-Assisted Intervention, MICCAI 2017, Springer, New York
(2017).
8. National Institutes of Health Chest X-Ray Dataset. https://www.kaggle.com/nih-chest-xrays/
datasets. Accessed 30 Aug 2020
9. Platt, J.: Advances in Kernel Methods—Support Vector Learning. MIT Press, Cambridge
(1998)
10. Quinlan, J.: C4.5: Programs for Machine Learning, Morgan Kaufmann, San Francisco
(1993)
11. Aha, D.: Lazy Learning. Kluwer Academic Publishers, Dordrecht (1997)
12. Kermany, D.S., Goldbaum, M., Cai, W. et al.: Identifying medical diagnoses and treatable
diseases by image-based deep learning. Cell 172(5), 1122–1131 (2018)
An Approach for Detecting Pneumonia from Chest X-Ray 743
13. Antin, B., Joshua, K., Martayan, E.: Detecting pneumonia in chest X-Rays with supervised
learning. Semanticscholar.Org (2017)
14. Rajpurkar, P., Irvin, J., Zhu, K. et al.: Chexnet: Radiologist-level pneumonia detection on
chest x-rays with deep learning. arXiv preprint arXiv: 1711.05225 (2017)
15. Park, M., Jin, J.S., Wilson, L.S.: Detection of abnormal texture in chest X-rays with
reduction of ribs. In: Proceedings of the Pan-Sydney area workshop on Visual information
processing (2004)
16. Ragab, D.A., Sharkas, M., Marshall, S., Ren, J.: Breast cancer detection using deep
convolutional neural networks and support vector machines. Peer 7, e6201 (2019)
17. Choudhari, S., Seema, B.: Artificial neural network for skin cancer detection. Int. J. Emerg.
Trends Technol. Comput. Sci. (IJETTCS) 3(5), 147–153 (2014)
18. Livieris, I., Kanavos, A., Tampakas, V., et al.: A weighted voting ensemble self-labeled
algorithm for the detection of lung abnormalities from X-rays. Algorithms 12(3), 64 (2019)
19. Yamashita, R., Nishio, M., Togashi, K., et al.: Convolutional neural networks: an overview
and application in radiology. Insights imaging 9(4), 611–629 (2018)
20. Omar, H.S., Babalık, A.: Detection of Pneumonia from X-Ray Images using Convolutional
Neural Network. Proceedings Book, p. 183 (2019)
21. Abiyev, R.H., Ma’aitah, M.K.S.: Deep convolutional neural networks for chest diseases
detection. J. Healthc. Eng. 2018 (2018)
22. Naranjo-Torres, J., Mora, M., Hernández-García, R., Barrientos, R.J., Fredes, C.,
Valenzuela, A.: A review of convolutional neural network applied to fruit image processing.
Appl. Sci. 10(10), 3443 (2020)
23. Han, F., Yao, J., Zhu, H., Wang, C.: Underwater Image Processing and Object Detection
Based on Deep CNN Method. J. Sensors 2020 (2020)
24. Alazab, M., Shalaginov, A., Mesleh, A., et al.: COVID-19 prediction and detection using
deep learning. Int. J. Comput. Inf. Syst. Ind. Manage. Appl. 12, 168–181 (2020)
25. Chakraborty, S., Aich, S., Sim, J.S., Kim, H.C.: Detection of pneumonia from chest x-rays
using a convolutional neural network architecture. Int. Conf. Future Inf. Commun. Eng. 11
(1), 98–102 (2019)
26. Anthimopoulos, M., Christodoulidis, S., Ebner, L., Christe, A., Mougiakakou, S.: Lung
pattern classification for interstitial lung diseases using a deep convolutional neural network.
IEEE Trans. Med. Imaging 35(5), 1207–1216 (2016)
An Analytical Study of Influencing
Factors on Consumers’ Behaviors
in Facebook Using ANN and RF
1 Introduction
– To explain the reasons behind buying something from Facebook pages and
groups.
– To construct ANN and RF decision tree models from the dataset for predic-
tion.
– Lastly, to show what factors are driving consumers to purchase more.
The rest of this paper’s sections are organized in the following order. In
Sect. 2, we include relevant works on the study of consumer behavior analy-
sis, in particular, using machine learning. Section 3 illustrates the methodology.
In Sect. 4, we describe the experiment with models. We show the results and
observations in Sect. 5 and discuss future works in Sect. 6 before we conclude.
2 Related Works
Heijden et al. [21] researched on 228 online consumers and tried to determine
consumers’ intentions to purchase online products. This research presented the
empirical study of a conceptual model from which we can understand the types
of factors effect the actions of the consumers, adapted from [8] by Fishbein et al.
The foundation of this model is a relationship between online buying behavior
and online buying intention. In this study, this model is a base for defining the
purchase intentions of online consumers. There are some issues in the sense of
competitive market and some critical factors that attract consumers to purchase
where three main factors are usability, interactivity and psychological matters
in online shops [6]. Kumar [16] gave a summary of the advertising actions on
consumers. Besides that, the elements of consumer decision-making were iden-
tified by Khaniwale [14]. The author discussed about the internal and external
factors of purchasing products.
Artificial Intelligence (AI)’s advancement helps business enterprises analyz-
ing the consumers’ behaviors to another degree. Business organizations strongly
emphasize on scrutinizing consumers’ feedbacks to make the best possible deci-
sions for their businesses. As a result, in recent years, many research works have
been carried out on digital signage [18], shopping atmosphere [7], and consumers’
746 S. Hossain et al.
reactions [1,9,15,17]. Alhendawi et al. [1] applied ANN to develop an expert sys-
tem for predicting a Management Information System (MIS)’s characteristics by
reviewing end users’ comments. Besides developing the AI-based tools, many sec-
tions of research are being conducted on framing consumers’ emotions, such as
Kumar et al. [15] analyzed the consumers’ emotions on their reviews over Ama-
zon web resources with the combination of Natural Language Processing (NLP)
and Machine Learning (ML) models.
Recently, deep learning has been applied to plenty of consumers’ reaction
investigation tasks. Gloor et al. [9] developed a system tool named Tirbefinder
by analyzing consumer behavior on Twitter. The research’s crucial contribution
is that the tool can categorize the consumers by applying word embedding and
LSTM1 RNN2 on their tweets in three macro-categories: alternative realities,
lifestyles, and recreations. Lang et al. [17] applied RNN on consumers’ actions
history data collected from Europe’s renowned fashion platform Zalando. In
this collaboration, they predicted the probability of consumers’ orders with-
out explicit feature engineering and achieved a better result than that of other
machine learning algorithms.
3 Methodology
In this section, the most significant steps of this research have been described.
First, the data have been collected, then have been engineered and divided into
training and testing data. After that, the training dataset has been fitted with
the proposed models. Finally, the testing dataset has been predicted through
the RF and ANN models. From starting to prediction on testing data using the
models step in Fig. 1.
1
LSTM: Long Short-Term Memory.
2
RNN: Recurrent Neural Network.
Influencing Factors on Consumers’ Behaviors in Facebook 747
Our samples have been collected from Facebook buying and selling groups of
authentic Facebook e-commerce consumers. The responses are collected through
Google form. The participations of 505 people have finally been recorded from
the survey. Attributes are chosen based on the parameters that require for the
extraction of consumer behavior. We choose questions for survey based on Fish-
bein et al. [8]’s conceptual model and the attributes are chosen from those ques-
tions. Some attributes are derived from the responses of participants. The total
number of dataset attributes is 22 with the number of instances being 505.
Here the symbol σ denotes the sigmoid activation function, W represents the
weights between two units of connected layers (k − 1, k) and b represents the
bias unit.
Influencing Factors on Consumers’ Behaviors in Facebook 749
In our proposed ANN model (Fig. 3) after the data pre-processing and cross-
validation, we feed the 18 featured values and a bias as the input to the ANN.
The dimension of the input layer is R19 . The hidden layer consists of 4 units that
are forwarded to the output unit with another bias unit. So, the dimension of the
hidden layer and output layer are R5 and R1 . We do experiment by using Keras,
a python deep learning library where TensorFlow works in the back-end. For
training, we run the model for 300 times with a learning rate of 0.01. Besides, to
minimize the error, we use the binary cross-entropy loss function and stochastic
gradient descent optimizer. Finally, model predicts our testing data through the
trained model.
750 S. Hossain et al.
(a) Bar chart for bought from FB (b) Bar chart for bought from FB
pages under following buy and sell FB pages under special discount and offer:
groups: real vs. prediction real vs. prediction
(c) Bar chart for bought from FB (d) Bar chart for bought from FB
pages under using of FB during study pages under the factor (not buying due
or work: real vs. prediction to lack of trust): real vs. prediction
According to Table 4, both models have acquired the same 98% accuracy. The
ANN model has performed best in precision and in specificity metrics with 99%,
whereas the RF model has attained 98% in both metrics. On the other hand,
the RF model has achieved the best score in the sensitivity metrics, which is
99%, and in the case of the ANN model, it is 98% as illustrated in Fig. 4.
After applying the ANN Model, The predicted values have been plotted using
the matplotlib, a python plotting library. In the x-axis, attributes are taken
to compare and to analyze the result. The y-axis demonstrates the count of
the value of that attribute. Figure 4a shows that some participants follow FB
752 S. Hossain et al.
buying and selling groups, and they have a positive intention to buy from FB
pages. Another observation is that some participants do not follow the FB groups
or pages, but they are buying from FB groups or pages. So, if any consumer fol-
lows any Facebook group or any Facebook page, then the probability of buying
from F-commerce is high. According to Fig. 4b, it can be noticed that some
participants have the intention to buy from Facebook pages if there is any spe-
cial discount and offer. Some consumers are still buying from Facebook pages
though there is no special discount and offer. Figure 4c, reveals that how much
participants involve in technology. From Fig. 4d it is being seen that, if any
consumer has any trust issues from any FB pages, then buying anything from
Facebook is zero. Some consumers have no issues with the reliability of any Face-
book pages, but have no intention to purchase. The same type of outcome has
been observed in the question of reliability, bad experience, and slow response
from the Facebook pages.
Although our proposed systems achieved robust results, it has some limitations
which invoke future research directions. Our research has only focuses on Face-
book users’ activities on F-commerce, which limits the influence of other social
media. In future, we would like to specify how and why some classes of the
consumers are persuaded to buy products from social media frequently.
In this research work, we have proposed two machine learning models to
predict consumers’ buying intentions from Facebook groups or pages. We have
applied RF and ANN for the prediction that has given a significant result
and improve the perceptions of purchase intentions from Facebook. We aim
to expand this research and analyze individual tastes more accurately. We hope
this research will help to open up a more sophisticated study area for other
researchers on consumers’ behaviors to achieve significant outcomes.
References
1. Alhendawi, K.M., Al-Janabi, A.A., Badwan, J.: Predicting the quality of mis char-
acteristics and end-users’ perceptions using artificial intelligence tools: expert sys-
tems and neural network. In: Proceedings of the International Conference on Intel-
ligent Computing and Optimization, pp. 18–30. Springer (2019)
2. Ali, J., Khan, R., Ahmad, N., Maqsood, I.: Random forests and decision trees. Int.
J. Comput. Sci. Issues (IJCSI) 9(5), 272 (2012)
3. Ballantyne, D., Varey, R.J.: The service-dominant logic and the future of market-
ing. J. Acad. Mark. Sci. 36(1), 11–14 (2008)
4. Bhat, S.A., Kansana, K., Khan, J.M.: A review paper on e-commerce. Asian J.
Technol. Manage. Res. [ISSN: 2249–0892] 6(1), 16–21 (2016)
5. Breiman, L.: Random forests. Mach. Learn. 45(1), 5–32 (2001)
6. Constantinides, E.: Influencing the online consumer’s behavior: the Web experi-
ence. Internet Res. 14(2), 111–126 (2004)
Influencing Factors on Consumers’ Behaviors in Facebook 753
7. Eroglu, S., Machleit, K., Davis, L.: Atmospheric qualities of online retailing: a
conceptual model and implications. J. Bus. Res. 34, 177–184 (2001)
8. Fishbein, M., Ajzen, I.: Predicting and changing behavior the reasoned action
approach (2015). OCLC: 1190691560
9. Gloor, P., Colladon, A.F., de Oliveira, J.M. and Rovelli, P.: Put your money where
your mouth is: using deep learning to identify consumer tribes from word usage.
Int. J. Inf. Manage. 51, p. 101924, April 2020
10. Mohamad, M.H.: Fundamentals of Artificial Neural Networks. MIT Press, Cam-
bridge (1995)
11. Hawkins, D.I., Mothersbaugh, D.L.: Consumer Behavior: Building Marketing
Strategy, 13th edn. McGraw-Hill Education, New York (2016)
12. Hecht-Nielsen, R.: Theory of the backpropagation neural network. In: Neural net-
works for perception, pp. 65-93. Elsevier (1992)
13. Tin Kam Ho: The random subspace method for constructing decision forests. IEEE
Trans. Pattern Anal. Mach. Intell. 20(8), 832–844 (1998)
14. Khaniwale, M.: Consumer buying behavior. Int. J. Innov. Sci. Res. 14(2), 278–286
(2015)
15. Kumar, P.K., Nandagopalan, S., Swamy, L.N.: Investigation of emotions on pur-
chased item reviews using machine learning techniques. In: Proceedings of the
International Conference on Intelligent Computing & Optimization, pp. 409–417.
Springer (2018)
16. Kumar, R.: Consumer behaviour and role of consumer research in marketing. J.
Commer. Trade 12(1), 65–76 (2017)
17. Lang, T., Rettenmeier, M.: Understanding consumer behavior with recurrent neu-
ral networks. In: Proceedings of the Workshop on Machine Learning Methods for
Recommender Systems (2017)
18. Ravnik, R., Solina, F., Zabkar, V.: Modelling in-store consumer behaviour using
machine learning and digital signage audience measurement data. In: Proceedings
of the International Workshop on Video Analytics for Audience Measurement in
Retail and Digital Signage, pp. 123–133. Springer (2014)
19. Rokach, L., Maimon, O.: Top-down induction of decision trees classifiers–a survey.
IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 35(4), 476–487 (2005)
20. Siregar, E.: Young attitudes towards f-commerce advertising. In: Proceedings of
the International Conference on Industrial Technology and Management (ICITM),
pp.218–223 (2018)
21. van der Heijden, H., Verhagen, T., Creemers, M.: Understanding online purchase
intentions: contributions from technology and trust perspectives. Eur. J. Inf. Syst.
12(1), 41–48 (2003)
Autism Spectrum Disorder Prognosis
Using Machine Learning Algorithms: A
Comparative Study
1 Introduction
Autism Spectrum Disorder (ASD) refers to the resistance in neurological as well
as genetic development leading towards difficulties in interaction and rudimen-
tary tasks. Autism patients experience several struggling with verbal communi-
cation, sensing, learning, and other chores that an average child can do easily.
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 754–765, 2021.
https://doi.org/10.1007/978-3-030-68154-8_65
ASD Prognosis Using Machine Learning Algorithms 755
2 Literature Review
In recent times, many algorithms and models using machine learning have been
introduced to predict Autism Spectrum Disorder (ASD) and achieved different
accuracy.
In [3] Raj et al. introduce six models to predict the autism disorder where
three datasets are used containing child, adolescents, and adulthood autism traits
and accuracy of 99.53%, 98.30%, 96.88% are the highest for respective stages.
Erkan et al. use three models in [4], including KNN, SVM, and Random Forest
(RF) where the accuracy of 100% is achieved for the RF-based model. Deep
embedding representation is proposed as an attested model with 99% sensitivity
and specificity in [5].
Lerthattasilp et al. in [6] illustrate multivariate logistic regression analysis
to develop the speculated model, and datasets are acquired from a standardized
diagnostic tool. In [7], Hyde et al. research the previous work done in ASD
prediction using supervised machine learning and investigate the clinical and
statistical addressing for data mining.
Machine learning is applied for the first time according to the authors of [8] to
investigate the kinematic movement traits with hand gestures in finding autism.
756 O. Jyoti et al.
3 Proposed Framework
The proposed models using neural-based algorithms e.g. Deep Neural Network,
Support Vector Machine (SVM), and K-Nearest Neighbor (KNN) are developed
in several phases, including data overview, data pre-processing, data analysis,
prediction model development, evaluating the prediction model. The flowchart
of the framework is given in Fig.1.
Fig. 1. Framework of the proposed method containing the steps e.g. ASD dataset, data
preprocessing, splitting data, build three models using three algorithm DNN, SVM and
KNN and evaluation of performance are shown in the flowchart.
ASD Prognosis Using Machine Learning Algorithms 757
Table 1. Overview of the dataset with the attributes name and attribute type.
The main focus of the research is to estimate the preferable prediction model
for Autism Spectrum Disorder. For this, three efficient machine learning models
have been developed which are trained on the same dataset. Furthermore, a
comparison study on the performance of each method regarding the prediction
and precision have been established through the process, which upholds the
effectiveness so that choosing the model for early diagnosis can be easier. Among
these models, Deep Neural Network has come up with the highest accuracy of
100%, which is, in our knowledge is the better outcome so far in DNN than
previous work.
4 Model Progression
In solving complex issues, Deep Neural Network is gaining its momentum day by
day [13]. In the prediction of ASD, after pre-processing of data, it has been split
into 70% train and 30% test data which means among 704 instances, 493 samples
have been used as train dataset, and 211 have been delegated as test dataset. The
network has been developed with two hidden layers for better function and both
layers consist of 12 layers each. For the initialization of weight Glorot Uniform
initialize has been introduced. The loss function evaluates the similarity between
the predicted scores and the ground truth labels in the training data [14]. In the
backpropagation stage, the loss has been reduced by the Batch Gradient Descent
ASD Prognosis Using Machine Learning Algorithms 759
algorithm and as a loss optimizing function, Mean Squared Error (MSE) has
been delegated. The epochs and learning rate for training the model is 10,000
and 0.01 respectively, and an activation function, Hyperbolic tangent (tanh),
Rectified linear unit (ReLu), Softplus and Swish function have been used. The
highest accuracy gained from this model is 100%. The comparison of accuracy
according to activation functions is shown in the Table 2.
distance Euclidean metric, which is the most popular method, has been used
with 6 numbers of neighbors. The Euclidean equation can be written as-
n
d(p, q) = (qi − pi )2 (2)
i=1
Fig. 2. Relationship status between the traits and variables of dataset as well as the
dependency of features are shown.
Fig. 3. Comparison graph of accuracy gain for activation functions of Deep Neural
Network (DNN) model is shown above
The estimation metrics used in this research are error rate, sensitivity, and
specificity, which can be determined through confusion matrix [19]. Each column
of confusion matrix illustrates the traits of actual class, and by row, it visualizes
the predicted class. True positive, true negative, false positive, and false negative
are the instances of matrix (Table 4).
762 O. Jyoti et al.
Fig. 4. Testing and training accuracy gained from KNN model are illustrated with
number of neighbors as x-label and accuracy as y-label in the figure.
Predicted 0 Predicted 1
Actual 0 TN FP
Actual 1 FN TP
Fig. 5. Accuracy, f1-score, specificity and sensitivity analysis of DNN, SVM and KNN
models is shown in this figure.
As per the above discussion, the proposed system has developed the predic-
tion model with supreme accuracy using DNN and converged the utmost goal.
Furthermore, the finest precision and moderate time make the model acceptable
in the field of detecting autism among adults.
6 Conclusion
Autism is considered to be one of the enlarging disorders around the globe. ASD
cannot be cured entirely but an appropriate method can demolish the sufferings
and hardships of the affected person as well as his/her family and relatives.
Keeping pace with the progression of the world, a constructive method must
be introduced to predict autism accurately. A vast area of research in machine
learning has opened up the opportunity of contribution in this field. In this paper,
the accuracy of 100% using DNN has managed to attain the contemplated result.
Delegated three models e.g. Deep Neural Network, Support Vector Machine,
and K-Nearest Neighbor have delivered an outstanding result throughout the
research by which detecting autism in an adult would be efficacious. However,
Deep Neural Network has secured the best outcome with preferable precision
and as a tool for predicting ASD successfully, it has become more effective than
others. Mostly, the work done in this paper has improved the previous result. For
future work, along with DNN other machine learning models will be analyzed
and modified so that accuracy can be refined.
764 O. Jyoti et al.
References
1. Autism spectrum disorders (2019). https://www.who.int/news-room/fact-sheets/
detail/autism-spectrum-disorders. Accessed 21 July 2020
2. Prevalence of autism spectrum disorder among children in select countries world-
wide as of 2020 (2020). https://www.statista.com/statistics/676354/autism-rate-
among-children-select-countries-worldwide/. Accessed 15 Aug 2020
3. Raj, S., Masood, S.: Analysis and detection of autism spectrum disorder using
machine learning techniques. Procedia Comput. Sci. 167, 994–1004 (2020)
4. Erkan, U., Thanh, D.N.: Autism spectrum disorder detection with machine learning
methods. Curr. Psychiatry Res. Rev. Formerly Curr. Psychiatry Rev. 15(4), 297–
308 (2019)
5. Wang, H., Li, L., Chi, L., Zhao, Z.: Autism screening using deep embedding rep-
resentation. In: International Conference on Computational Science, pp. 160–173.
Springer (2019)
6. Lerthattasilp, T., Tanprasertkul, C., Chunsuwan, I.: Development of clinical pre-
diction rule for diagnosis of autistic spectrum disorder in children. Mental Illness
(2020)
7. Hyde, K.K., Novack, M.N., LaHaye, N., Parlett-Pelleriti, C., Anden, R., Dixon,
D.R., Linstead, E.: Applications of supervised machine learning in autism spectrum
disorder research: a review. Rev. J. Autism Dev. Disord. 6(2), 128–146 (2019)
8. Li, B., Sharma, A., Meng, J., Purushwalkam, S., Gowen, E.: Applying machine
learning to identify autistic adults using imitation: an exploratory study. PloS one
12(8), e0182652 (2017)
9. Baranwal, A., Vanitha, M.: Autistic spectrum disorder screening: prediction with
machine learning models. In: 2020 International Conference on Emerging Trends
in Information Technology and Engineering (ic-ETITE), pp. 1–7. IEEE (2020)
10. Shuvo, S.B., Ghosh, J., Oyshi, A.: A data mining based approach to predict autism
spectrum disorder considering behavioral attributes. In: 2019 10th International
Conference on Computing, Communication and Networking Technologies (ICC-
CNT), pp. 1–5. IEEE (2019)
11. Geng, X. Kang, X., Wong, P.C.: Autism spectrum disorder risk prediction: a sys-
tematic review of behavioral and neural investigations (2020)
12. Omar, K.S., Mondal, P., Khan, N.S., Rizvi, M.R.K., Islam, M.N.: A machine learn-
ing approach to predict autism spectrum disorder. In: 2019 International Confer-
ence on Electrical, Computer and Communication Engineering (ECCE), pp. 1–6.
IEEE (2019)
13. Yang, X., Li, F., Liu, H.: A comparative study of dnn-based models for blind image
quality prediction. In: 2019 IEEE International Conference on Image Processing
(ICIP), pp. 1019–1023. IEEE (2019)
14. Vasant, P., Zelinka, I., Weber, G.-W.: Intelligent Computing & Optimization,
vol. 866. Springer, Cham (2018)
15. Zhang, Y.: Support vector machine classification algorithm and its application. In:
International Conference on Information Computing and Applications, pp. 179–
186. Springer (2012)
16. Guo, G., Wang, H., Bell, D., Bi, Y., Greer, K.: Knn model-based approach in
classification. In: OTM Confederated International Conferences” On the Move to
Meaningful Internet Systems”, pp. 986–996. Springer (2003)
ASD Prognosis Using Machine Learning Algorithms 765
17. Altay, O., Ulas, M.: Prediction of the autism spectrum disorder diagnosis with
linear discriminant analysis classifier and k-nearest neighbor in children. In:2018
6th International Symposium on Digital Forensic and Security (ISDFS), pp. 1–4.
IEEE (2018)
18. Niu, K., Guo, J., Pan, Y., Gao, X., Peng, X., Li, N., Li, H.: Multichannel deep
attention neural networks for the classification of autism spectrum disorder using
neuroimaging and personal characteristic data. Complexity 2020 (2020)
19. Al Diabat, M., Al-Shanableh, N.: Ensemble learning model for screening autism
in children (2019)
Multidimensional Failure Analysis Based
on Data Fusion from Various Sources Using
TextMining Techniques
1 Introduction
Failure and reliability analysis is a well-recognized subject whose results have a direct
impact on technical diagnostics. Maintenance of machinery directly affects efficiency
issues, making them an important variable for optimizing production as well as the
company’s financial result. In practice, both failure analysis is often based on advanced
mathematical methods. U. Kumar and B. Klefso (1990) considered the issue of
applying statistical distributions (with particular emphasis on the Weibull distribution)
for the purposes of modeling the residual lifetime of underground mining machinery.
The paper presents the methodology for estimating distribution parameters based on
empirical data from the machine operation process [1]. This subject was also taken up
in paper [2] with particular emphasis on the use of this method for modeling the
lifetime of haul trucks. In paper [3] the authors conducted a study of the reliability and
availability of underground loaders along with the analysis of the failure of individual
subsystems using modeling based on Markov chain theory. The method based on a
support vector machine for estimating the residual life-time of machines, taking into
account varying operational conditions has been prosposed in [4]. In [5] we can find the
method of applying a non-linear statistical model for estimating the lifetime of mining
machines. In [6] the authors described ways of using the Kaplan-Meier estimator for
the purposes of statistical analysis of reliability, availability, and maintenance of
underground haul trucks. The article [7] conducted a comparative analysis of failures
occurring among four drill rigs operating in an underground mine in Sweden. The
paper also proposes a method of estimating the time remaining to the next failure.
Continuing this topic, in [8] presented methods of using Monte Carlo simulation in the
analysis of the reliability of the underground drilling platform.
Despite many works dealing with the reliability of LHD machines and good
recognition of this issue, the main problem remains the collection of data on the failure
of machines, especially in the case of huge machinery parks (e.g. over 1200 machines).
In practice, very often the machine’s operational history is still kept manually in
notebooks. Some mine departments run individual spreadsheets. A step forward turned
out the launch of a dedicated IT system ensuring data entry via the form. Unfortunately,
despite a large number of computer stations in the underground mine with EAM
software (e.g. CMMS - Computerised Maintenance Management Systems), the data is
entered manually by service technicians due to time limitations and the amount of
maintenance and repair works carried out. As a consequence, there are huge amounts of
unstructured data in digital storage, but the automation of this data analysis or its
integration with other data sources is very limited. This is an especially important issue
for BIG DATA analytics. For this purpose, several solutions have been developed, but
only a fraction of them can be applied in relation to the Polish language. In [9] a 3 step
Text-Mining tool for the CMMS databes was presented. This tool focuses mainly on
the clustering the terms describing failures and later mining of the association rules.
Paper [10] describes 3 methods for dataset analysis using Text-Mining techniques in
order to classify failure mode for each of the records. The aspects of the use of Text-
Mining tools in different systems for the support is described in the review [11].
Authors searched the state of published evidences on the use of Text Mining and tried
to understant the issues from organizational prespective and to create a core list of Text-
Mining tools.
In this paper, the authors propose a methodology used for creation of the Text-
Mining tool used for data acquisition from CMMS-like databes containing records of
every work that was perform on evidences machines. In addition, a methods of analysis
of such gathered data are proposed.
768 M. Stachowiak et al.
2 Description of Problems
Examples of problems encountered in the CMMS database relate to: (a) one entry may
refer to several elements of the machine not related to each other, (b) individual entries
containing several independent operations very often are not separate punctuation
marks, (c) appearance of stylistic and spelling mistakes, (d) shorter forms and mine
jargon are used, (e) very often the same corrective action is called differently, this also
applies to machine components and foreign names.
TextMining is the proposed tool for extracting data from the text as well as its
subsequent processing. Text mining can rely on finding key phrases, sentences, which
are then encoded in the form of numeric variables. Later, statistics and data mining
methods are used to discover relationships between variables. After that, it is also
possible to combine data from other sources (SCADA, SAP, etc.) and further data
fusion and multidimensional analyzes using machine learning techniques. Because the
resulting variables are usually nominal, basket analysis may be particularly useful and
allows one to identify unknown regularities accompanying the failure phenomenon
(operator’s driving style, diagnostic symptoms, workplace, mine, maintenance stuff,
machine type).
In this paper, we propose a tool for extensive machine failure analysis based on data
fusion from various sources. The paper is organized as follows: a short review on the
subject and current practice in the mining will be described; some remark and
assumption will be formulated and a novel procedure will be presented and discussed;
finally, application of the method will be provided.
3 Methodology
The algorithm scheme proposed in this paper is shown in Fig. 1. The algorithm consists
of: Data preparation – which is a set of actions to clean data and prepare for further
processing, Lemmatization – a process of finding the core of a word using an additional
dictionary, Feedback Loop – actions aiming at errors correction and improvement
algorithm overall effectivity, Categorization and Grouping – which are a processes for
sorting and labeling data collection. The described algorithm allows for actions like
developing a specialized dictionary of entries, catching stylistic and spelling mistakes,
unification of CMMS database records, sorting or searching entries by specified rules,
etc. In addition, the results of the algorithm can be used to develop further analyzes
such as e.g. affinity analysis.
Data mining tool for entry data uses raw records from CMMS-like Polish database
containing information about self-propelled machines used in underground copper
mines. Each of the records consists of separate variables like a machine or department
number, and cells containing a description of failure and repair. Most of the variables
are raw copied to the output, only text variables containing event information are
processed. Data contained in those cells did not pass any form of grammatical,
punctuation, or semantic control which causes that many of them contain multiple
errors. Many of them also contain information like dates or numbers that hinder
analysis. The most useful information contained in those cells is basic information like
Multidimensional Failure Analysis Based on Data Fusion from Various Sources 769
Fig. 1. Algorithm of text data mining tool for extraction and categorization of information from
maintenance logs.
the name of the damaged component and description of its repair (like part repair or
part replaced). For lemmatization, the algorithm uses “PoliMorf” which is an open-
source morphological dictionary for Polish language. PoliMorf is a result of the
Morfologik project operation. Dictionary contains 3 main categories which are: word,
the core of a word and grammatical category. Each of these variables is used in the
process of lemmatization and later categorization. It is worth adding that a given
dictionary is a general and not a specialized one which may cause conflicts.
3.2 Lemmatization
Lemmatization is an operation of finding the core of a word from one of its changed
forms. This process is shown in Fig. 2. This task, depending on language, can be done
with either stemming or lemmatization. While stemming is a crude process of chopping
parts of a word with a set of rules to reduce it to form that cannot be influenced by
factors such as tense. Lemmatization refers to doing the same operation (reducing
770 M. Stachowiak et al.
words to basic forms) but with different tools like vocabulary and morphological
analysis. In the case of morphologically complex languages like Polish stemming
simply doesnt’t work, the error rate of this operation is too high. This is the reason for
using the lemmatization process with a polish dictionary.
In the process of lemmatization, each of the words from the record is searched in
the PoliMorf dictionary. If there will be a match in the searching process, the basic
form of a word is saved from the core column of the dictionary. The grammatical
category of a word (from the dictionary) is also stored. An exception to these opera-
tions are words consisting of one letter and numbers. If a word hits no match in
dictionary searching it is also stored for further processing. If a searched word had more
than one match, it was also stored for further decision.
There are many reasons why a word did not find its equivalent in the dictionary.
The most common are spelling errors and typos. They occur because of manual data
entry and the lack of unified form. Another reason for lacking match is that word is a
shortcut, proper name, or comes from jargon or colloquial speech. Many of such
unique words are already introduced in the dictionary, but due to the a wide range of
colloquial words and continuous evolution of speech it is impossible to predict all of
them, therefore their collection should be updated repeatedly. Many cases of repetition
of words not found in the dictionary were recorded, so it was decided to leave them in
the most common form. A different situation for missing words is the lack of space
between two words. In cases like these two words are connected and processing as one.
connected words. For each of these cases, a separate algorithm was created. As
mentioned earlier, some cases should require human verification, however, as the tool
is used, their number will gradually disappear.
Words with Errors. words containing errors like typos are similar to their correct
form (usually they do not differ by more than a few characters). Some measures allow
for measuring the similarity between words. These metrics measure “distance” between
words, and the smaller the distance, the more similar words. In the algorithm two
metrics were used, The Jaro-Winkler algorithm and the Levenshtein distance. The Jaro-
Winkler distance algorithm pays a lot of attention to the order of letters. In addition, the
speed of processing words is relatively fast. Typos and words with only a few errors
will be characterized by a small value of distance. On the other hand, the Jaro-Winkler
does not seem to care about interlaced, randomly placed, or missing characters. As long
as the target word characters are present at the correct distance, the words are similar.
The Jaro-Winkler Distance for two string of characters (s1 ; s2 ) is expressed by the
formula:
(
0; m ¼ 0
dj ¼ ð1Þ
1
3
m
js1 j þ jsm2 j þ mt
m ; m 6¼ 0
where, m – is the number of correct characters and t is half the number of transposi-
tions. Those two words are recognized as matched only when they are the same, or with
distance lesser than:
maxðjs1 j;js2 jÞ
ð2Þ
2
Where 1ðai 6¼bj Þ is the indicator function equal to 0 when ai ¼ bj , and equal to 1
otherwise, and leva;b ði; jÞ is the distance between the first i characters if a and the first j
characters of b.i and j are 1-based indices. Thanks to these two metrics the algorithm
searches the dictionary for the word that has the highest match [12].
772 M. Stachowiak et al.
Connected Words. Every word that is longer than 5 characters is taken as a possible
connection of 2 words. Algorithm recursively divides that word into two words where
both must have at least 2 characters (in the Polish language the shortest word that
carries useful information have 2 characters) and searches for matches in the dictionary
with the use of distance functions mentioned before. If the similarity between two
strings and words in the dictionary is large enough, the string is replaced by two words.
In every other case, a string is saved as an incorrect word for later human processing.
Shortcuts and Colloquial Speech. In the mining facility, the use of colloquial speech,
abbreviations, and jargon is very high. This includes also foreign-language names
which can be translated into Polish in a variety of ways (in both formal and common
speech). This problem can be solved by looking at the whole dataset. If such word
repeats multiple times it’s recognized as a proper name or shortcut. This action,
because of low fault-tolerance needs to be later accepted by a human.
Words with Multiple Meanings. In Polish and many other languages there is a
significant number of homonyms and homographs, that is, words that have multiple
meanings but the same spelling (for example: crane – as a bird or a construction
machine). In the case of such words, multiple cores can be assigned. To solve the
problem a simple method was developed. If at the stage of lemmatization it turns out
that more than one core can be assigned to a word, those cores are saved. Later, the
amounts of occurrences of given cores are compared in the entire data sample. A core
that has a greater number of repetitions is assigned to the word. If the number is the
same for both cores, the case is recorded and referred to human intervention.
3.4 Categorization
Cleaned text is subjected to exploratory analysis. The main action in order of proper
categorization is the extraction of keywords. To study those we used tag clouds. This
tool presents the frequency of a given word in the form of a cloud. The more often a
word is repeated, the larger is it’s representation in the cloud when comparing to other.
This tool works great to find nuances and relations with data. To find proper keywords
we take most frequently repeated words and form another tag clouds with two or three
words expressions that contain words from the first cloud. This can be seen in Fig. 3.
For clarity and understanding of the results, all of them have been translated from
Polish into English. We discovered that many of the keywords they mean the same as
others (e.g. propulsion engine means the same as the combustion engine). At this stage,
specialized knowledge in the field of construction and operations on the tested
machines is required.
For proper categorization, first, the data need to be cleaned again. Most of them
contain words that still doesn’t bring useful information like pronouns or sentence like
“manufacturer’s materials were used for repair”. The next step was to assign each word
its grammatical form. In this way we get:
Multidimensional Failure Analysis Based on Data Fusion from Various Sources 773
• Maintenance actions – in the Polish language most of the analyzed actions are
verbs. Exceptions like “replacement” or “repair” (in Polish they are nouns) were
added manually. In that simple way, the algorithm can isolate the action that was
performed and described in the record.
• Machine components – in Polish every component is a noun. In addition, most of
them occur with an adjective. By this the algorithm can isolate not only component
that is involved in records by also can specify which component (for example
“silnik spalinowy” (ang. combustion engine) where the word “silnik” (noun) means
engine and “spalinowy” (adjective) describe what kind of engine - combustion).
• Number of actions performed – this information was isolated from conjunctions.
Words like “and” bring the information that more than one element was involved in
the event. There were also less frequently words like “in” or “from” which contain
information about the system which were the elements belong to.
After assigning grammatical form, first, the actions are isolated. Often there was more
than one maintenance action. Next, to each action, the machine component involved were
matched. The last step was to assign to each group an adjective to specify elements.
Adjectives were selected on the basis of the quantity with which they appeared. In this
way, crude categories were isolated. The main problem at this stage is that many cate-
gories are repeated. Some of them carry the same information but are written differently. It
is caused by adding new information or changing the order of the sentence. Action
performed were always matching, but the nouns were the problem. To solve this problem,
categories were grouped. If the sentence describing the machine component has more
than 3 words, the algorithm checks how much it has common with other phrases. If there
were a category, that differs with only one word, the phrases were grouped as matching. In
the case of shorter clusters (less than 3 words), the algorithm searches for categories with
at least one word the same and matching letters in other words. In this way, proper
categories were formed (a few of them connecting 3 or 4 elements).
774 M. Stachowiak et al.
3.5 Grouping
For better data presentation and further analysis, the categories were matched to one of
the 5 main groups. This was performed manually based on the action performed. The
main groups are: repair, replacement, maintenance (preservation), inspection, correc-
tion. Grouping data allows one to analyze in a simple way what actions (without
specified details) were performed on the machine.
An example of such analysis is shown in Fig. 4. On the machine’s lifeline, groups
of individual works have been marked. Experimental data comes from underground
mines that belong to the KGHM Polska Miedź SA. Input data are entries from the
CMMS database that contain information about the work carried out on the machines
(information entered by mechanics performing a specific job). Data comes from 3
mines and a total of 24 different machines which gives a total of 13332 unique entries.
Each entry was generated on the basis of a form that contained a dozen or so fields
where information was collected by means of: text, option selection, Yes/No fields, etc.
For this type of analysis, the main column used was the one containing the scope of the
work carried out, in which the mechanic entered the activities performed on the
machine using the keyboard. This cell has no form restrictions, and any text can be
entered there.
Fig. 4. Categories of actions performed on the machine (results have been translated).
The course of maintenance and repair actions can be used in analysis with data from
the machine monitoring system for fault analysis, better tracking of diagnostic symptoms,
verification of the effectiveness of repair work, and so on. An example of such analysis is
shown in Fig. 5. The nominal exceedances of quantiles of the gearbox oil temperature
parameter (red horizontal line) have been compared with the periods of its repair and
maintenance. After the quantiles began to show overshoots, two machine inspections and
repairs were made and the overshoots stopped. Longer periods of missing data are
associated with machine downtimes caused by failure events. Some advanced algorithms
to detect anomaly in temperature data have been described in [13–15].
Multidimensional Failure Analysis Based on Data Fusion from Various Sources 775
Fig. 5. Quantiles analysis with maintenance and repair actions performed (results have been
translated). A red horizontal line indicates the threshold value for the gearbox oil temperature.
4 Conclusion
The general advantage of a text mining tool is it’s versatility in further implementation.
Once created can be applied to similar types of data with only little effort. Using a text-
mining tool with CMMS databases allows for fast keywords extraction, categorization
by almost any factor. That kind of action can be also used for automatic abstract or
report generation or connecting data from incompatible sources. In this paper we
presented a text mining tool for CMMS-like databases, that consists of several lesser
algorithms. The main goal for that tool was to process raw text data from facilities in
order of later analysis of that data. That target was achieved with a satisfactory level of
efficiency. As a result, each of the records are categorized and grouped allowing for
actions like quick search, statistical analysis, or more complex multivariate failure
analysis. The main drawback of the presented analysis are distance based operation on
short words wich can bring unexpected results. For data processed by the text mining
tool, we presented two kinds of analyses that we then applied to real industrial data.
One of them is the quantitative analysis that allows us to look at the data numerically. It
allows the creation of various types of statistics or rankings thanks to which we can
optimize work and look for hidden regularities. The result of algorithm work can be
also used for the creation of a more complex tool for analysis that consists of methods
like basket analysis.
References
1. Kumar, U., Klefsjö, B.: Reliability analysis of hydraulic systems of LHD machines using the
power law process model. Reliab. Eng. Syst. Saf. 35(3), 217–224 (1992)
2. Camelia, D.S.M., Silviu, N. M., Emil, D.: Study of reliability modeling and performance
analysis of haul trucks in quarries. In: Advances in Computer and Information Sciences and
Engineering (2015)
3. Samanta, B., Sakar, B., Mukherjee, S.K.: Reliability modelling and performance analyses of
an LHD system in mining. J. South Afr. Inst. Min. Metall. 104(1), 1–8 (2004)
776 M. Stachowiak et al.
4. Galar, D., Kumar, U., Lee, J., Zhao, W.: Remaining useful life estimation using time
trajectory tracking and support vector machines. In: Journal of physics: Conference series,
vol. 364, no. 1, p. 012063. IOP Publishing (2012)
5. Si, X.S., Wang, W., Hu, C.H., Zhou, D.H., Pecht, M.G.: Remaining useful life estimation
based on a nonlinear diffusion degradation process. IEEE Trans. Reliab. 61(1), 50–67 (2012)
6. Mouli, C., Chamarthi, S., Gȧ, R.C., Vȧ, A.K.: Reliability modeling and performance analysis
of dumper systems in mining by KME method. IJCET 2, 255–258 (2014)
7. Al-Chalabi, H.S., Lundberg, J., Wijaya, A., Ghodrati, B.: Downtime analysis of drilling
machines and suggestions for improvements. J. Qual. Maintenance Eng. 20, 306–332 (2014)
8. Al-Chalabi, H., Hoseinie, H., Lundberg, J.: Monte Carlo reliability simulation of
underground mining drilling rig. In: Current Trends in Reliability, Availability, Maintain-
ability and Safety, pp. 633–643. Springer, Cham (2016)
9. Gunay, H.B., Shen, W., Yang, C.: Text-mining building maintenance work orders for
component fault frequency. Build. Res. Inf. 47(5), 518–533 (2019)
10. Chen, L., Nayak, R.: A case study of failure mode analysis with text mining methods. In:
Ong, K.-L., Li, W., Gao, J. (eds.) Proceedings 2nd International Workshop on Integrating
Artificial Intelligence and DataMining (AIDM 2007) CRPIT, 84, pp. 49–60. Gold Coast,
QLD (2007)
11. Paynter, R.A., Bañez, L.L., Berliner, E., Erinoff, E., Lege-Matsuura, J.M., Potter, S.: Use of
text-mining tools for systematic reviews. Value in Health 19(3), A108 (2016)
12. Cohen, W.W., Ravikumar, P., Fienberg, S.E.: A comparison of string distance metrics for
name-matching tasks. In: IIWeb, vol. 2003, pp. 73–78, August 2003
13. Wodecki, J., Stefaniak, P., Michalak, A., Wyłomańska, A., Zimroz, R.: Technical condition
change detection using Anderson-Darling statistic approach for LHD machines–engine
overheating problem. Int. J. Min. Reclam. Environ. 32(6), 392–400 (2018)
14. Stefaniak, P., Śliwiński, P., Poczynek, P., Wyłomańska, A., Zimroz, R.: The automatic
method of technical condition change detection for LHD machines-engine coolant
temperature analysis. In: International Conference on Condition Monitoring of Machinery
in Non-Stationary Operation, pp. 54–63. Springer, Cham June 2018
15. Wodecki, J., Stefaniak, P., Polak, M., Zimroz, R.: Unsupervised anomaly detection for
conveyor temperature SCADA data. In: Advances in Condition Monitoring of Machinery in
Non-Stationary Operations, pp. 361–369. Springer, Cham (2018)
Road Quality Classification Adaptive
to Vehicle Speed Based on Driving Data
from Heavy Duty Mining Vehicles
1 Introduction
accidents. It should be highlighted that each year, thousands of people in the world are
involved in serious road accidents (very often lethal) resulting from a damaged road
surface [3, 4]. In a rich country like U.S. planned yearly expenditures for maintenance
of federal highways are estimated at $46 billion. The poor condition of roads is yearly
responsible for nearly 10,000 traffic fatalities. Considering the above, the major
emphasis should be taken on safety and maintenance costs. The two critical factors
influencing driving quality are: the surface roughness of the pavement. and vehicle
response to the road. A commonly used road roughness metric is IRI (the International
Roughness Index) which defines based on simulation a scale presented relation
between vehicle response regarding road quality obtained at speeds 80 km/h [5]. In the
practice, conducting continuous repairs is difficult to achieve because of expensive
manpower, various weather conditions, and heavy traffic.
Thus, what is really important here is the development of a robust monitoring
system both for early detection of road failures, tracking its evolution and support
planning the maintenance tasks in advance. So far, an inspection of the road is per-
formed via a manual process which is based on a visual approach and strongly depends
on subjective impressions of service personnel and is very time consuming and not
objective. In general, such an approach has many limitations and requires frequent
inspections. An example of standard procedures has been described in [6].
In the literature subject of automated road, surface inspection is very popular. We
can highlight: vision methods [7], LIDAR scanning [8], ground penetration radar
(GPR) [9], and techniques based on inertial sensors [10, 11]. In [12] the authours used a
quarter car model with Kalman Filter estimate road roughness index (in underground
mine) based on data from IMU, wheel speed sensors and RSSI from Wi-Fi. The most
common are accelerometer-based methods. The optimal solution is to collect real-time
driving data from smartphones in terms of service like Google Maps. As a result, it will
be possible to immediately obtain information about the condition of all the most
frequently used roads practically for free. Paper [13] presents a method of surface
monitoring based on Accelerometer Data, where Gaussian model was used for
abnormal event detection, and X-Z axis ratio filtering for event classification (like
pothole or hump). The issue of road quality assessment is well described in Interna-
tional Road Roughness Experiment (IRRE) [14], in which the authors compared many
methods of measuring this value and proved that they are compatible with each other.
The problem of road quality evaluation has also become very popular in mining.
The development of methods initiated the analysis of the factors of dynamic overloads
observed among heavy-duty mining machines, which resulted in damage to the con-
struction nodes. An inertial sensor installed on the haul truck and data from the on-
board monitoring system [15] was used for the tests. In paper [16] clarifies that the
main factors that have influence on vibration level are road conditions (shape of road,
failures, bumps, slope), driving style (e.i. adapting speed to road conditions), and
technical conditions of the machine (structure stiffness) [17]. As part of the preliminary
work, the authors proposed a simple method of road-quality assessment which comes
down to: (1) signal segmentation in order to identify haulage cycles and its particular
operation [18, 19] (loading of the cargo box, driving with a full cargo box), unloading
in dumping point and returning to mining area with empty box, (2) estimation of
Road Quality Classification Adaptive to Vehicle Speed Based on Driving Data 779
trajectory for driving with full and empty cargo box, (3) calculation of statistics for
inertial data, (4) its tristate classification and representation in the route visualization.
This paper is an extension of the original method [16]. The key is to fill the gap
related to the impact of speed on the values of recorded vibration accelerations. This is
critical for several reasons. First of all, machine operators work in the piecework
accounts mode and care for the machine moderately not using high speeds where
thoose should be avoided. Secondly, one machine is run by dozens of different oper-
ators with completely different driving styles. Thirdly, each cycle may have different
capacity of the route resulting from a traffic jam, a busy dumping point. Fourthly and
most importantly, the driving speed has a real impact on the magnitude of the recorded
vibration values. For these reasons, authors present a novel method fitted to mining
machines. This adaptive approach is based on inertial data and machine learning
techniques. The classifier model used in the procedure is based on the k-means algo-
rithm. The developed model is adapted to the case of an underground mine, but after
minor modifications, it can be applied to diagnose the state of pavement in other types
of road infrastructure. The article presents the application of the model on real
industrial data. The result data from the classification model was integrated into the
machine motion trajectory estimation results obtained from inertial navigation. In this
way, a spatial distribution of the surface condition on the GIS map was obtained. The
description of the navigation algorithm is beyond the scope of this article.
2 Input Data
3 Algorithm Assumptions
The main assumption of the algorithm is the effect of speed on machine vibration from
the road. This means that when driving on an identical road, only at different speeds,
the vibrations of the machine will be different (will transform because of the change in
speed). Therefore, in order to correctly estimate the state of road quality, the algorithm
should be separated and “immunized” against the influence of speed.
The next assumption of the algorithm is the frequency band in which vibrations
from the road are manifested. The algorithm assumes that the quality of the road is
manifested in vibration at higher frequencies, which is true e.g. passenger cars. Mining
machines, however, are quite different in many factors. These machines are much
larger and heavier than ordinary passenger cars. The speeds with which they travel are
also quite different, maximum it is up to 18 km/h where it is usually about 5–10 km/h.
The last assumption concerns the idle and the machine’s frequencies. At times
when the machine is found to be idling, a spectrum is extracted from the vibration
signal. The algorithm assumes that the spectrum largely presents the machine’s natural
frequencies. This assumption, however, can be confirmed by methods like modal
analysis, which were unavailable for the duration of the experiments.
4 Methodology
The adaptive road quality classifier can be divided into two similar algorithms: the
algorithm for creating the classifier and its further learning and algorithm for the
classification process. There is great convergence between these two algorithms,
mainly in the functions/ideas used, but they were broken due to easier implementation
(and faster operation). The main blocks of both algorithms are shown in Fig. 2.
Road Quality Classification Adaptive to Vehicle Speed Based on Driving Data 781
Fig. 2. Schematics of algorithms for the creation of the classifier and the classification.
indicate that the machine has moved. Therefore, only proposals are extruded based on
speed, while the main detection takes place using a signal from the engine (ENGRPM).
Engine speed during normal operation is spread between 1000 and 2000 rpm.
When the machine switches to idle, these rotations decrease to the level of around
800 rpm. This relationship is the main factor based on which idle is detected, while the
detection itself takes place using simple decision thresholds.
From the fragments that were considered idling, an amplitude spectrum is created
using FFT (Fast Fourier Transform). These spectra are then averaged to create one
spectrum representing the machine. This can be described by the formula (1):
1 XN
Xj ¼ Y ;
n ¼ 1 n;j
ð1Þ
N
where: Xj is new spectrum value for j frequency, N describe the number of calculated
spectra and Y n;j is value of N spectrum for j frequency . Such a spectrum should mainly
contain only native frequencies. An example spectrum obtained from fragments with a
total length of about 2 h is shown in Fig. 3. It can be seen despite the low resolution
that clear peaks are observed for some frequencies.
velocity values found in the data. At a later stage, each of these X signals is processed
separately (all further described activities are repeated for each of the signals).
Further segmentation consists of dividing the given speed signal into fragments of
the window length (2 s). The amplitude spectrum is calculated for each such fragment
using the Fourier transform. Then, for each such spectrum, the difference with the idle
spectrum is calculated (spectrum - idle spectrum). If negative values appear as a result
of this operation, they are filled with zeros.
The last step involves creating one parameter to describe the given spectrum. For
this, we used a weighted average with an additional weight system that would favor
higher frequency spectra components. Due to the sampling frequency of 50 Hz, the
useful spectrum was in the range of 0–25 Hz, the weight system for this range is shown
in Table 1.
Calculation of the weighted average from the spectrum completes the entire process
of data pretreatment. Only two values are used for the classification itself: the average
and the speed of the fragment from which it was calculated. All the activities described
up to this point can be represented by the pseudo-code visible in Fig 4.
New models can be created when speed range of the vehicle extends, while those
already created can be developed further.
Each of the models in the classifier is a separate KMeans working on one-
dimensional data. These data are weighted average spectra calculated from signal
segments for the currently processed speed. Each of the KMeans models is designed to
divide the spectrum data into 3 groups (corresponding to bad, medium, and good road
quality). This is also partly the reason for choosing this grouping algorithm. As one of
the few, it allows grouping data into a specific number of groups, and KMeans from a
group of algorithms that allow it seemed to work best.
With this classifier design, there is also the problem of assigning different markings
each time the model is started. The solution to this is to sort the assigned labels using
group centers. The lowest of all groups always get the label 0, medium one – 1, while
the highest – 2.
Figure 5 presents the classifier which we were able to obtain based on the collected
data. The colors marked the next detected road quality (and their boundaries): white
(lowest) - good, gray - medium, black (highest) - bad. Division of groups in the
classifier is not simple (as in the case of simple decision thresholds), and as the speed
increases, the acceptable vibrations of good and medium quality roads increase. After
exceeding the speed of 15 km/h, the vibrations seem to drop, however, there are too
few measurements to be able to definitely conclude. It can also be stated that the
increase in thresholds is similar in both directions (forward/backward), at least up to a
speed of 5 km/h because after that there are not enough measurements on the negative
side. Maximum vibrations are recorded for a speed of 7–9 km/h.
Fig. 5. Road Classifier model with established values, white (lowest) - good, gray (medium) –
medium, black (highest) – bad.
Road Quality Classification Adaptive to Vehicle Speed Based on Driving Data 785
5 Experimental Results
The algorithm was learned on the data from the vehicles hauling the spoil in one of the
mines owned by KGHM Polska Miedź SA. A total of 10 work shifts were used to learn
the algorithm, each lasting about 6 h (4 days of experiments, each consists of 2 shifts).
The Speed of the machine was recorded using internal measurement unit SYNAPSA,
the vibration were recorded by the accelerometer obtained from the IMU (NGIMU)
mounted on the machine. Then the algorithm was tested on one additional work shift,
from which it was decided to show one fragment. The designated section of road
against the background of the mine map with color-coded quality is shown in Fig. 6.
The correctness of the method was confirmed by the passengers of the vehicle driving
on this route.
At the moment, the described method lacks unequivocal confirmation of its cor-
rectness. Single fragments of routes and the results of the algorithm’s operation have
been confirmed by machine operators, however, this assessment is highly inaccurate
and also subjective, thus some further works in this direction are planned.
6 Summary
The article deals with the issue of assessing the condition of mining road infrastructure,
which is crucial from the point of view of efficient, sustainable and safe exploatation.
The authors proposed an algorithm based mainly on data from an inertial measuring
unit and the speed signal. A classification model based on the spectral analysis and
KMeans algorithm was developed, which allows to perform a three-state assessment of
the road surface condition adaptively to the driving speed of a mining vehicle. The
paper presents the integration of the result data from the classification model with the
GIS map and the vehicle movement path estimated from an inertial navigation. This
approach allows to achieve a holistic view of the condition of road infrastructure in the
underground mine. In the future, it is suggested to integrate the proposed solution with
an IoT platform based on thousands of low-cost sensors installed on the each of
vehicles in mine. In this way, on-line measurements can cover the entire road network
in a mine. Besides that, the more data the machines will gather, the greater that
classifier can be in terms of accuracy. At this moment, a relatively small sample of data
allowed for the creation of a classifier which performed the detection with a satisfactory
accuracy for a machine running at the most popular (average) speeds. Since machines
rarely run at higher speeds, more data is needed to obtain similar results for these
speeds. There is also the need to compare classification results with the more accurate
and tested method of the road quality detection, because the validation method pre-
sented here is highly subjective.
Acknowledgements. This work is a part of the project which has received funding from the
European Union’s Horizon 2020 research and innovation programme under grant agreement No
780883.
References
1. Bishop, R.: A survey of intelligent vehicle applications worldwide. In: Proceedings of the
IEEE Intelligent Vehicles Symposium 2000 (Cat. No. 00TH8511), pp. 25–30. IEEE,
October 2000
2. Eriksson, J., Girod, L., Hull, B., Newton, R., Madden, S., Balakrishnan, H.: The pothole
patrol: using a mobile sensor network for road surface monitoring. In: Proceedings of the 6th
International Conference on Mobile Systems, Applications, and Services, pp. 29–39, June
2008
3. U.S. Department of Transportation, Traffic Safety Facts – Crash Stats, June 2015
4. Pothole (2002). http://www.pothole.info
5. Gillespie, T.D.: Everything you always wanted to know about the IRI, but were afraid to ask.
In: Road Profile Users Group Meeting, Lincoln, Nebraska, pp. 22–24, September 1992
Road Quality Classification Adaptive to Vehicle Speed Based on Driving Data 787
6. Pierce, L.M., McGovern, G., Zimmerman, K.A.: Practical guide for quality management of
pavement condition data collection (2013)
7. Ferguson, R.A., Pratt, D.N., Turtle, P.R., MacIntyre, I.B., Moore, D.P., Kearney, P.D.,
Breen, J.E.: U.S. Patent No. 6,615,648. Washington, DC: U.S. Patent and Trademark Office
(2003)
8. Pothole Marker And More - Apps on Google Play, Google Play
9. Mahmoudzadeh, M.R., Got, J.B., Lambot, S., Grégoire, C.: Road inspection using full-wave
inversion of far-field ground-penetrating radar data. In: 2013 7th International Workshop on
Advanced Ground Penetrating Radar, pp. 1–6. IEEE, July 2013
10. Basavaraju, A., Du, J., Zhou, F., Ji, J.: A machine learning approach to road surface anomaly
assessment using smartphone sensors. IEEE Sens. J. 20(5), 2635–2647 (2019)
11. Vittorio, A., Rosolino, V., Teresa, I., Vittoria, C.M., Vincenzo, P.G.: Automated sensing
system for monitoring of road surface quality by mobile devices. Procedia-Soc. Behav. Sci.
111, 242–251 (2014)
12. Åstrand, M., Jakobsson, E., Lindfors, M., Svensson, J.: A system for under-ground road
condition monitoring. Int. J. Min. Sci. Technol. 30, 405–411 (2020)
13. Harikrishnan, P.M., Gopi, V.P.: Vehicle vibration signal processing for road sur-face
monitoring. IEEE Sens. J. 17(16), 5192–5197 (2017)
14. Sayers, M.W., Gillespie, T.D., Queiroz, C.A.V.: The international road roughness
experiment: establishing correlation and a calibration standard for measurements. University
of Michigan, Ann Arbor, Transportation Re-search Institute, January 1986
15. Stefaniak, P.K., Zimroz, R., Sliwinski, P., Andrzejewski, M., Wyłomanska, A.: Multidi-
mensional signal analysis for technical condition, operation and performance understanding
of heavy duty mining machines. In: International Conference on Condition Monitoring of
Machinery in Non-Stationary Operation, pp. 197–210. Springer, Cham, December 2014
16. Stefaniak, P., Gawelski, D., Anufriiev, S., Śliwiński, P.: Road-quality classification and
motion tracking with inertial sensors in the deep underground mine. In: Asian Conference on
Intelligent Information and Database Systems, pp. 168–178. Springer, Singapore, March
2020
17. Wodecki, J., Stefaniak, P., Michalak, A., Wyłomańska, A., Zimroz, R.: Technical condition
change detection using Anderson-Darling statistic approach for LHD machines–engine
overheating problem. Int. J. Min. Reclam. Environ. 32(6), 392–400 (2018)
18. Wodecki, J., Stefaniak, P., Śliwiński, P., Zimroz, R.: Multidimensional data segmentation
based on blind source separation and statistical analysis. In: Advances in Condition
Monitoring of Machinery in Non-Stationary Operations, pp. 353–360. Springer, Cham
(2018)
19. Stefaniak, P., Zimroz, R., Obuchowski, J., Sliwinski, P., Andrzejewski, M.: An effectiveness
indicator for a mining loader based on the pressure signal measured at a bucket’s hydraulic
cylinder. Procedia Earth and Planet. Sci. 15, 797–805 (2015)
Fabric Defect Detection System
1 Introduction
The textile industry is a rising sector. Development and advancement of the sector
normally bring to build the going through huge investment. Be that as it may, the
textile, like any other sector, industry experienced various issues. These include some
insurance to diminish the effect of misfortunes that are budgetary, client disappoint-
ment, time squandering, and so on. Fabric defects are probably the greatest test
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 788–800, 2021.
https://doi.org/10.1007/978-3-030-68154-8_68
Fabric Defect Detection System 789
confronting the textile business. Fabric is made in a day by day life utilizing fibers and
a usually utilized material. Most fabrics are delivered after passing through a series of
making stages. Various machines and methods are utilized during the making stages.
Fabrics are subjected to pressures and stresses along these lines that cause defects. As
indicated by their structures and directions, defects take various names. The textile
business has distinguished in more than 70 types of defects [1] such as laddering, end-
out, hole, and oil spot as shown in Fig. 1. Unexpected tasks might be the reason for
various defects on the fabric surface during the manufacturing of fabric [2]. The lack of
defects can diminish the cost of fabric by 50–60% [1]. The decrease in the impacts in
the production process is typical for the industrialist.
Thus, fabric manufacturing is one of the largest traditional businesses where fabric
inspection systems can play a vital role in growing the manufacturing rate. These days,
the significance of an inspection process nearly rises to with manufacturing process in
the current industrialist viewpoint. The idea of inspection process is to recognize the
happened errors or defects, on the off chance that any exist, at that point to change
argument or give alert of inspector for checking the manufacturing procedure [3]. For
the most part, fabric defects recognition utilizes two kinds of investigation models [4].
The essential one is that the human-based inspection systems as shown in Fig. 2.The
second framework is automated based inspection systems as shown in Fig. 3.
Accordingly, human-based defect detection done by specialists’ turns out to be rapidly
a mind-boggling and fussy task [5, 6]. In this manner, having proficient and automated
based frameworks nearby is a significant necessity for improving unwavering quality
and accelerating quality control, which may expand the profitability [7–10]. The
subject of automated based defect detection has been examined in a few works in the
most recent decades. In spite of the fact that there is no widespread methodology for
handling this issue, a few strategies dependent on image processing procedures have
been proposed in recent years [11–13]. These strategies were utilized to recognize
defects at the image level, so the precision rate is little and additionally, it is hard to find
790 T. Mahmud et al.
the defects precisely. In this way, they can’t be stretched out to various fabrics. As of
late, some different techniques dependent on local image-level have been proposed,
which utilize the base unit as the fundamental activity object to extract image features.
These methodologies can be ordered into four principle gatherings: Statistical,
Signal processing-based, Structural methodology, and Model-based methodology.
In the statistical approach, gray-level properties are utilized to describe the textural
property of texture image or a measure of gray-level reliance, which are called 1st-order
statistics and higher-order statistics, separately [14]. The 1st-order statistics, for
example, mean and standard deviation [15, 16], rank function [17], local integration,
can gauge the variance of gray-level intensity among different features between
defective areas and background. The higher-order statistics depends on the joint
probability distribution of pixel sets, for example, gray-level co-occurrence matrix [18]
gray-level difference strategy [15] and autocorrelation method. In any case, the
inconvenience of this strategy is that defects size is sufficiently enormous to empower a
compelling estimation of the texture property. So this methodology is feeble in han-
dling local little defects. Additionally, the calculation of higher-order statistics is
tedious [17].
In the subsequent class model-based methodology, the generally utilized strategies
are Markov random field Gaussian Markov random field [16]. The texture features of a
contemplated texture and can signify to all the more exactly spatial interrelationships
between the gray-levels in the texture. However, like the methodologies based on
second-order statistics, additionally it is tough for model-based methodology to deal
with identifying small-sized defects in light of the fact that the methodologies as a rule
require an adequately large region of the texture to assess the parameters of the models.
The structural approach generally utilized on properties of the primitives of the
defect-free fabric texture for the nearness of the flawed region, and their related
placement rules. Apparently, the practicability of this methodology is to congestion to
those textures with regular macro texture.
Fabric Defect Detection System 791
Not at all like the above methodologies which separate the defects as far as the
visual properties of the fabric texture, the signal processing based methodology extract
features by applying different signal processing procedures on the fabric image. It is
projected that the distinguishable between the defect and the non-defect can be
improved in the handled fabric image. This methodology further comprises of the
accompanying techniques: Spatial filtering, Karhunen-Loeve transform, Fourier
transform, Gabor transform, and Wavelets transform.
As a weakness of this methodology, its performance is effortlessly influenced by
the noise in the fabric image. These coefficients exemplify to optimal the defect-free
fabric image, be that as it may, not the optimal separation between the defect and the
non-defect.
They are progressively proficient in the separation of fabric defects than different
techniques that depend on the texture investigation at a single scale [19]. Contrasted
with the Gabor transform; the wavelet transform has the benefit of greater adaptability
in the decomposition of the fabric image [20]. Consequently, the wavelet transform is
seen as the most suitable way to deal with the feature extraction for fabric defect
detection.
Table 1 Illustrates the taxonomy of most recent fabric defects detection methods, in
light of their classifier, machine learning technique, and accuracy rate.
In this paper, we propose an innovative defect detection algorithm which has the
capability to cope with different types of defects. Our algorithm is based on four
phases. In the initial phase, image segmentation has been utilized on an excess of a
couple of fabric images so as to enhance the fabric image and to locate the important
data and wipe out the unusable data of the image by utilizing different edge detection
strategies. After the initial phase, morphological operations have been utilized on the
792 T. Mahmud et al.
fabric image. In the third step, feature extraction has been done through FAST (Fea-
tures from Accelerated Segment Test) extractor. After feature extraction, If PCA is
applied as it lessens the dimensions and preserves the helpful data, and characterizes
the different fabric defects through a neural network; additionally classifier has been
utilized to find the accuracy rate. The proposed framework gives high precision when
contrasted with the other framework. The investigation has been done in MATLAB
environment on real images of the TILDA database [33].
The remaining of the paper is arranged as follows: In Sect. 2, the various types of
fabric defects are presented. In Sect. 3 explains our proposed approach for defect
detection. In Sect. 4 presents the application of our system and analysis. Finally, Sect. 5
accomplishes the paper and presents our future research plans.
2 Defects in Fabric
In order to prepare various categories and forms of fabric items in the industry, fabric
materials are used. Consequently, yarn quality and/or loom defects affect the fabric
quality. Fabric defect has been estimated [34] that the price of fabrics is reduced by
45%-65% due to the presence of defects such as dye mark/dye Spot, slack warp, faulty
pattern card, holes, spirality, grease oil/ dirty stains, mispick, slub, wrong end, slack
end, and so on [1]. In a fabric, defects can occur due to: machine faults, color bleeding,
yarn problems, excessive stretching, hole, dirt spot, scratch, poor finishing, crack point,
material defects, processing defects, and so on [35, 36].
Figure 4 shows the steps of methodology, to sum up; the following steps are image
segmentation, feature extraction, PCA (Principal Component Analysis), and image
classification.
Fabric Defect Detection System 793
Accuracy
120.00%
100.00%
80.00%
60.00%
40.00%
20.00% Accuracy
0.00%
Article
The detection of faulty fabrics plays an important role in the success of any fabric
industry. The fabric industry needs a real-time quality control to find defects quickly
and efficiently. Manual control is inefficient and time-consuming that leads to heavy
loss. On the other hand, automatic quality control is considerably more proficient, in
light of the fact that it is a real-time and autonomous compared to manual productivity.
Till now all the fabric detection systems suggested by all the researchers, the accuracy
rate for detecting defective fabric is very low. However, this paper analyzed the
shortcomings of the traditional approach for fabric defect detection, and proposed an
innovative fabric defect detection technique based on a FAST (Features from Accel-
erated Segment Test) extractors and PCA (Principal Component Analysis) combined
with a neural network classification, to enhance the recognition accuracy rate that
texture fabrics cannot be effectively detected by present existing techniques. The paper
concludes that the proposed fabric defect detection technique gave better accuracy after
applied the machine learning algorithm and PCA in comparison to the other referenced
approaches. Additionally, our method notably showed its efficiency in separating
defect-free from defective areas of the fabric. Moreover, after a series of improvements,
our method exhibited better recognition performance for the fabric images. Having
successfully trained the neural network, 30 samples of each type of defect were used to
798 T. Mahmud et al.
assess the accuracy of the network classification. The defective images were then
graded with a 97.21% overall accuracy score. With a 100% accuracy score, the dye
spot defect was identified. The experimentation has been applied to the real images of
the TILDA database dataset. The implementation has been done in MATLAB software
as shown in Fig. 7 and Fig. 8. It automatically detects the fabric defect.
In the future, we will focus on sensor data-oriented systems and developing an
optimal system to match more accurately with a real system for fabric defect detection
as well as applying our machine learning algorithms for other different feature
extractors.
References
1. Stojanovic, R., Mitropulos, P., Koulamas, C., Karayiannis, Y., Koubias, S., Papadopoulos,
G.: Real-time vision-based system for textile fabric inspection. Real-Time Imaging 7, 507–
518 (2001)
2. Aasim A.: A catalogue of visual textile defects, ministry of textiles (2004)
3. Newman, T.S., Jain, A.K.: A survey of automated visual inspection. Comput. Vis. Image
Underst. 61(2), 231–262 (1995)
4. Kumar, A.: Computer-vision-based fabric defect detection: a survey. IEEE Trans. Ind.
Electron. 55(1), 348–363 (2008)
5. Huart, J., Postaire, J.G.: Integration of computer vision on to weavers for quality control in
the textile industry. In: Proceeding SPIE 2183, pp. 155–163, February 1994
6. Dorrity, J.L., Vachtsevanos, G.: On-line defect detection for weaving systems. In:
Proceeding IEEE Annual Technical Conference Textile, Fiber, and Film Industry, pp. 1–
6, May 1996
7. Ryan G.: Rosandich: Intelligent Visual Inspection, Chapman & Hall, London (U.K.) (1997)
8. Batchelor, B.G.: Lighting and viewing techniques. In: Batchelor, B.G., Hill, D.A., Hodgson,
D.C. (eds) Automated Visual Inspection. IFS and North Holland (1985)
9. Roberts, J.W., Rose, S.D., Jullian, G., Nicholas, L., Jenkins, P.T., Chamberlin, S.G.,
Maroscher, G., Mantha, R., Litwiller, D.J.: A PC-based real time defect imaging system for
high speed web inspection. In:Proceeding SPIE 1907, pp. 164–176 (1993)
10. Bayer, H.A.: Performance analysis of CCD-cameras for industrial inspection. In: Proceed-
ing SPIE 1989, pp. 40–49 (1993)
11. Cho, C., Chung, B., Park, M.: Development of real-time vision-based fabric inspection
system. IEEE Trans. Ind. Electron. 52(4), 1073–1079 (2005)
12. Kumar, A.: Computer-vision-based fabric defect detection: a survey. IEEE Trans. Ind.
Electron. 55(1), 348–363 (2008)
13. Ngana, H., Panga, G., Yung, N.: Automated fabric defect detection a review. Image Visi.
Comput. 29(7), 442–458 (2011)
14. Smith, B.: Making war on defects. IEEE Spectr. 30(9), 43–47 (1993)
15. Fernandez, C., Fernandez, S., Campoy P., Aracil R.: On-line texture analysis for flat
products inspection. neural nets implementation. In: Proceedings of 20th IEEE International
Conference on Industrial Electronics, Control and Instrumentation, vol. 2, pp. 867–872
(1994)
16. Ozdemir S., Ercil A.: Markov random fields and Karhunen-Loeve transforms for defect
inspection of textile products. In: IEEE Conference on Emerging Technologies and Factory
Automation, vol. 2, pp. 697–703 (1996)
Fabric Defect Detection System 799
17. Bodnarova A., Williams J. A., Bennamoun M., Kubik K. Optimal textural features for flaw
detection in textile materials. In: Proceedings of the IEEE TENCON 1997 Conference,
Brisbane, Australia, pp. 307–310 (1997)
18. Gong, Y.N.: Study on image analysis of fabric defects. Ph.D. dissertation, China Textile
University, Shanghai China (1999)
19. Zhang, Y.F., Bresee, R.R.: Fabric defect detection and classification using image analysis.
Text. Res. J. 65(1), 1–9 (1995)
20. Nickolay, B., Schicktanz, K., Schmalfub, H.: Automatic fabric inspection– utopia or reality.
Trans. Melliand Textilberichte 1, 33–37 (1993)
21. Habib, M.T., Rokonuzzaman, M.: A set of geometric features for neural network-based
textile defect classification, ISRN Artif. Intell. 2012, Article ID 643473, p. 16 (2012)
22. Saeidi, R.D., Latifi, M., Najar, S.S., Ghazi Saeidi, A.: Computer Vision-Aided Fabric
Inspection System For On-Circular Knitting Machine, Text. Res. J. 75(6), 492–497 (2005)
23. Islam, M.A., Akhter, S., Mursalin, T.E.: Automated textile defect recognition system using
computer vision and artificial neural networks. In: Proceedings World Academy of Science,
Engineering and Technology, vol. 13, pp. 1–7, May 2006
24. Murino, V., Bicego, M., Rossi, I.A.: Statistical classification of raw textile defects. In: 17th
International Conference on Pattern Recognition (ICPR 2004), ICPR, vol. 4, pp. 311–314
(2004)
25. Karayiannis, Y.A., Stojanovic, R., Mitropoulos, P., Koulamas, C., Stouraitis, T., Koubias,
S., Papadopoulos, G.: Defect detection and classification on web textile fabric using multi
resolution decomposition and neural networks. In: Proceedings on the 6th IEEE International
Conference on Electronics, Circuits and Systems, Pafos, Cyprus, pp. 765–768, September
1999
26. Kumar, A.: Neural network based detection of local textile defects. Pattern Recogn. 36,
1645–1659 (2003)
27. Kuo, C.F.J., Lee, C.-J.: A back-propagation neural network for recognizing fabric defects.
Text. Res. J. 73(2), 147–151 (2003)
28. Mitropoulos, P., Koulamas, C., Stojanovic, R., Koubias, S., Papadopoulos, G., Karayiannis,
G.: Real-time vision system for defect detection and neural classification of web textile
fabric. In: Proceedings SPIE, vol. 3652, San Jose, California, pp. 59–69, January 1999
29. Shady, E., Gowayed, Y., Abouiiana, M., Youssef, S., Pastore, C.: Detection and
classification of defects in knitted fabric structures. Text. Res. J. 76(4), 295–300 (2006)
30. Campbell, J.G.,. Fraley, C., Stanford, D., Murtagh, F., Raftery, A.E.: Model-based methods
for textile fault detection, Int. J. Imaging Syst. Technol. 10(4), 339–346, July 1999
31. Islam, M.A., Akhter, S., Mursalin, T.E., Amin, M.A.: A suitable neural network to detect
textile defects. Neural Inf. Process. 4233, 430–438. Springer, October 2006
32. Habib, M.T., Rokonuzzaman, M.: Distinguishing feature selection for fabric defect
classification using neural network. J. Multimedia 6 (5), 416–424, October 2011
33. TILDA Textile texture database, texture analysis working group of DFG. http://lmb.
informatik.unifreiburg.de
34. Srinivasan, K., Dastor, P. H., Radhakrishnaihan, P., Jayaraman, S.: FDAS: a knowledge-
based frame detection work for analysis of defects in woven textile structures, J. Text. Inst.
83(3), 431–447 (1992)
35. Rao Ananthavaram, R.K., Srinivasa, Rao O., Krishna P.M.H.M.: Automatic defect detection
of patterned fabric by using RB method and independent component analysis. Int.
J. Comput. Appl. 39(18), 52–56 (2012)
36. Sengottuvelan, P., Wahi, A., Shanmugam, A.: Automatic fault analysis of textile fabric using
imaging systems. Res. J. Appl. Sci. 3(1), 26–31 (2008)
800 T. Mahmud et al.
37. Abdi. H., Williams, L.J.: Principal component analysis. Wiley Interdisciplinary Rev.
Comput. Stat. 2 (4), 433–459 (2010). https://doi.org/10.1002/wics.101
38. Kumar, T., Sahoo, G.: Novel method of edge detection using cellular automata. Int.
J. Comput. Appl. 9(4), 38–44 (2010)
39. Zhu, Q.: Efficient evaluations of edge connectivity and width uniformity. Image Vis.
Comput. 14, 21–34 (1996)
40. Senthilkumaran. N., Rajesh, R.: Edge detection techniques for image segmentation – a
survey of soft computing approaches. Int. J. Recent Trends Eng. 1(2), 250–254 (2009)
41. Rizon, M., Hashim, M.F., Saad, P., Yaacob, S.: Face recognition using eigen faces and
neural networks. Am. J. Appl. Sci. 2(6), 1872–1875 (2006)
42. Rosten, E., Porter, R., Drummond,T.: FASTER and better: a machine learning approach to
corner detection, IEEE Trans Pattern Anal Mach Intell. 32, 105–119 (2010)
43. Wikipedia, Corner Detection. http://en.wikipedia.org/wiki/Corner_detection. Accessed 16
March 2011
44. Chang, J.Y., Chen, J.L.: Automated facial expression recognition system using neural
networks. J. Chin. Inst. Eng. 24(3), 345–356 (2001)
45. Jianli, L., Baoqi, Z.: Identification of fabric defects based on discrete wavelet transform and
back-propagation neural network. J. Text. Inst. 98(4), 355–362 (2007)
46. Tamnun, M.E., Fajrana, Z.E., Ahmed, R.I.: Fabric defect inspection system using neural
network and microcontroller. J. Theor. Appl. Inf. Technol. 4(7) (2008)
47. Bhanumati, P., Nasira, G.M.: Fabric inspection system using artificial neural network. Int.
J. Comput. Eng. 2(5), 20–27 May 2012
Alzheimer’s Disease Detection Using CNN
Based on Effective Dimensionality
Reduction Approach
1 Introduction
natural language processing and machine learning [21]. Liu et al. also employed
ensemble method to detect Alzheimer disease [22]. One of the limitations of those
work is high dimensionality of feature vector.
In the study, we applied machine learning and data mining methodology based on the
Alzheimer’s disease neuroimaging initiative (ADNI) dataset for classifying the various
stages of Alzheimer disease. Especially we have classified Alzheimer disease (AD) and
normal control (NC) from ADNI dataset using Random forest and Cognitive neural
network (CNN) model. To reduce the dimensionality of feature vector in this field
Davatzikos et al. and Yoon et al. employed principle component analysis (PCA) algo-
rithm [23, 24]. Moreover, the technique of the algorithm is to select parameters carefully.
To overcome the high dimensionality problem we have applied here three-dimensionality
reduction method to reduce the dimensionality of feature vector namely Principle com-
ponent analysis (PCA), Random projection (RP) and Feature agglomeration (FA) tech-
nique. We have experimented Random Forest and Cognitive Neural Network (CNN)
through the dimensionality reduction method. Finally, we found combined of Random
projection (RP) and Random forest (RF) (RPRF) produced best result.
2 Dataset
Table 1. Dataset
Age/Gender Male Female
Min 56.4 55.1
Max 89 89.6
Average Age 75.49 74.72
3 Proposed Methodology
3.1 Preprocessing
We took the ADNI dataset and load the dataset, we used Python as language and used
pyCharm as our IDE. Then the dataset was proceeded for next step. We checked for
null values in dataset using a Pandas Dataframe and find null values. If any null values
are found then we used .sum () to convert them into Boolean function. After that again .
sum () was used for converting them to binary values. Then we removed the irrelavent
data from our dataset. All the data are sorted according to their features and classes. All
the columns containing true/false values, APOES were calculated. And normalized
using one hot encoding.
TL ¼ XWL ð1Þ
Here TL is the reduced component eigen vector. In other words, PCA look like a linear
transformation, which is like Eq. (2)
t ¼ WTx ð2Þ
Alzheimer’s Disease Detection Using CNN 805
Random Projection
Random projection is a method used to minimize the dimensionality of the Euclidean
space set of points. In contrast to other methods, random projection methods are known
for their strength, simplicity, and low error rate. In random projection, the original d-
dimensional data is projected to a k-dimensional (k << d) subspace, using a random k *
d – dimensional matrix R with unit column length [26]. If Xd *N is the original set of N
d-dimension, then the projection of data in a lower k-dimension will be
RP
XkN ¼ Rkd XdN ð5Þ
Feature Agglomeration
Feature Agglomeration is a dimensionality reducing method that is Similar to
Agglomerative clustering but recursively merges features instead of samples. Let Mmn
representing n samples of dimension m that you want to cluster. In feature agglom-
eration, the algorithm clusters the transpose of the matrix, i.e., MT so it clusters m
samples of dimension n; these samples represent the features [27]. If there is 3 samples
of dimension 3 (a 3 3 matrix with three features, then this can reduce the dimension
of your dataset to 2 dimensions, the algorithm clusters together feature 1 and feature 2,
and leaves feature 3 unchanged.
3.3 Classification
The reduced data set we have splitted into the train and test data consequently in 70%
train data and 30% of test data. We applied a machine learning algorithm namely
Random forest and Convolution neural network algorithm.
We experimented with input data into different combination of models namely
(i) PCA & Random Forest, (ii) PCA & CNN, (iii) Random Projection & Random
Forest, (iv) Feature Agglomeration & Random Forest to differentiate patients with the
AD from healthy patients. We have used two classifiers for our study:
Random Forrest
Random forest algorithm builds decision trees on sample data, then gets the prediction
from each of them, and eventually selects the best solution by voting. It is an ensemble
approach that is stronger than a single decision tree, and by integrating the result, it
eliminates the overfitting [28]. The key thing to consider when doing Random Forests
806 A. S. M. Miah et al.
based on classification data is that the Gini index is often used or the method used to
determine how nodes are to be chosen on a branch of a decision tree.
Xc
Gini ¼ 1 i¼1
ð pi Þ 2 ð6Þ
This method uses the class and probability to calculate the Gini of each branch on a
node, which branch is more likely to occur. Here pi represents the relative frequency of
the class you observe in the dataset and C is the number of classes to observe.Another
way to decide how nodes split in a decision tree is by entropy.
XC
Entropy ¼ i¼1
pi log2 ðpi Þ ð7Þ
4 Experimental Result
After splitting the dataset (ADNI) into training and test, we got 157 patients as test data.
Table 2 shows the details about test data.
Evaluating this test data into our model, we have found different performance
accuracies. Here we evaluated four models; some of them are performed quite well and
gave good accuracy. To understand the accuracy of our model first, we computed a
confusion matrix. Then computed precision, recall, f1 score, and finally, the accuracy
of the proposed model as well.
In Table 3, Table 4, Table 5 and Table 6 shows confusion matrix for every model.
The confusion matrix is one of the best way to evaluate classifier performance. Table 3
shows the confusion matrix for Random forest with PCA (RFPCA) model. Table 3
shows RFPCA model with 24 subjects have correctly classified and 8 subjects have
misclassified in AD stage, best accuracy has acquired in CN stage because in this stage
misclassified only 3 subjects and 46 subjects have correctly classified finally have
classified correctly 57 subjects and misclassified 19 subjects. Table 4 shows the con-
fusion matrix of CNN with PCA (CNNPCA) model. Therefore, the model also
achieved the best accuracy at the CN stage because this stage CNNPCA has classified
correctly 47 subjects and misclassified only 2 subjects.
Table 5. Confusion matrix: Random pro- Table 6. Confusion matrix: Feature agglom-
jection + Random forest eration + Random forest
Stage AD CN MD Stage AD CN MD
AD 30 0 2 AD 23 0 9
CN 2 43 4 CN 1 45 3
MCI 1 1 74 MCI 8 2 66
Table 5 shows the confusion matrix of Random forest with the Random projection
(RFRP) model
This model has correctly classified 32 subjects and misclassified 2 subjects at the
AD stage. Moreover, at CN stage model has classified correctly 43 subjects and
misclassified six subjects. RFRP finally maximum correctly classified at the MCI stage.
In the same way, Table 6 shows the confusion matrix for Random forest with Feature
808 A. S. M. Miah et al.
The summaries accuracy of all four models shows in Table 8. The accuracy table
shows that the Random forest with PCA (RFPCA) model generated 88.00% accuracy,
same as Cognitive neural network with PCA (CNNPCA), Random forest with Random
Alzheimer’s Disease Detection Using CNN 809
References
1. Moser, A., Pike, C.J.: Obesity and sex interact in the regulation of Alzheimer’s disease.
Neurosci. Biobehavioral Rev. 67, 102–118 (2016). ISSN 0149-7634. https://doi.org/10.
1016/j.neubiorev.2015.08.021
2. Ligthart, S.A.: Cardiovascular prevention in older people: The pre DIVA trial Thesis,
Academic Medical Center – University of Amsterdam (2015). ISBN: 978-94-6169-623-6
3. Farjana, S.: World Alzheimer’s day: Let’s not forget the forgetful 11:17 AM, 21 September
2018
4. Aggarwal, N.T., Tripathi, M., Alladi, H.H., Anstey, K.S.: Trends in Alzheimer’s disease and
dementia in the Asian-pacific region. Int. J. Alzheimer’s Dis. Hindawi Publishing
Corporation VL - 2012 SN - 2090–8024. https://doi.org/10.1155/2012/171327
5. Dr. Taha, S.: World Alzheimer’s Day: Forgetting dementia in Bangladesh. Paragraph: The
Impact of Dementia in Bangladesh, 21 September 2014
6. Sneddon, R., Shankle, W.R., Hara, J., Rodriquez, A., Hoffman, D., Saha, U.: EEG detection
of early Alzheimer’s disease using psychophysical tasks. Clin. EEG Neurosci. 3, 141–150
(2005)
7. Rahman, Md., et al.: Overview and Current Status of Alzheimer’s Disease in Bangladesh,
pp. 27–42, 1 January 2017
8. Jack Jr, C.R., et al.: Magnetic resonance imaging in alzheimer’s disease neuroimaging
initiative 2. Alzheimer’s Dementia 11, 7 (2015)
9. Jongin, K., Lee, B.: Identification of Alzheimer’s disease and mild cognitive impairment
using multimodal sparse hierarchical extreme learning machine. Hum. Brain Mapp. 39(9),
3728–3741 (2018)
10. Veitch, D.P., et al.: Understanding disease progression and improving Alzheimer’s disease
clinical trials: Recent highlights from the Alzheimer’s disease neuroimaging initiative.
Alzheimer’s Dementia (2018)
11. Fujishima, M., Kawaguchi, A., Maikusa, N., Kuwano, R., Iwatsubo, T., Matsuda, H.:
Sample size estimation for Alzheimer‘s disease trials from Japanese ADNI serial magnetic
resonance imaging. J. Alzheimer‘s Dis. 56(1), 75–88 (2017)
12. Gallego-Jutglà, E., Solé-Casals, J., Vialatte, F.-B., Elgendi, M., Cichocki, A., Dauwels, J.: A
hybrid feature selection approach for the early diagnosis of Alzheimer‘s disease. J. Neural
Eng. 12(1), 016018 (2015)
13. Pellegrini, E., et al.: Machine learning of neuroimaging for assisted diagnosis of cognitive
impairment and dementia: a systematic review. Alzheimer’s Dement. Diagn. Assess. Dis.
Monit. 10(2018), 519–535 (2018)
14. Moradi, E., Pepe, A., Gaser, C., Huttunen, H., Tohka, J., Initiative, A.D.N.: Machine
learning framework for early MRI-based Alzheimer‘s conversion prediction in MCI subjects.
NeuroImage 104(2015), 398–412 (2015)
15. Albright, J.: Forecasting the progression of Alzheimer’s disease using neural networks and a
novel preprocessing algorithm. Alzheimer’s Dement. Transl. Res. Clin. Interv. 5, 483–491
(2019). ISSN 2352-8737
16. Tanveer, M., Richhariya, B., Khan, R.U., Rashid, A.H.: Machine learning techniques for the
diagnosis of alzheimer’s disease: a review, article. In: ACM Transactions on Multimedia
Computing, Communications and Applications, April 2020
17. Alonso, S.G., De La Torre-D´ıez, I., Hamrioui, S., L´opezCoronado, M., Barreno, D.C.,
Nozaleda, L.M., Franco, M.: Data mining algorithms and techniques in mental health: a
systematic review J. Med. Syst. 42(9), 161 (2018)
Alzheimer’s Disease Detection Using CNN 811
18. Ni, H., Yang, X., Fang, C., Guo, Y., Xu, M., He, Y.: Data mining-based study on sub-
mentally healthy state among residents in eight provinces and cities in china. J. Tradit. Chin.
Med. 34(4), 511–517 (2014)
19. Esmaeilzadeh, S., Belivanis, D.I., Pohl, K.M., Adeli, E.: End-to-end Alzheimer’s Disease
Diagnosis and Biomarker Identification. arXiv: 1810.00523 (2018)
20. Long, X., Chen, L., Jiang, C., Zhang, L.: Prediction and classification of Alzheimer disease
based on quantification of MRI deformation. PLoS One 12, 1–19 (2017)
21. Clark, D.G., McLaughlin, P.M., Woo, E., Hwang, K., Hurtz, S., Ramirez, L., Eastman, J.,
Dukes, R.M., Kapur, P., DeRamus, T.P., Apostolova, L.G.: Novel verbal fluency scores and
structural brain imaging for prediction of cognitive outcome in mild cognitive impairment.
Alzheimer’s Dement. (Amsterdam, The Netherlands) 2, 113–122 (2016)
22. Liu, M., Zhang, D., Shen, D.: Alzheimer’s disease neuroimaging initiative. Ensemble sparse
classification of Alzheimer’s disease. NeuroImage 60(2), 1106–1116 (2012). https://doi.org/
10.1016/j.neuroimage.2012.01.055
23. Davatzikos, C., Resnick, S.M., Wu, X., Parmpi, P., Clark, C.M.: Individual patient diagnosis
of AD and FTD via high-dimensional pattern classification of MRI. NeuroImage 41, 1220–
1227 (2008). [PubMed: 18474436]
24. Yoon, U., Lee, J.M., Im, K., Shin, Y.W., Cho, B.H., Kim, I.Y., Kwon, J.S., Kim, S.I.:
Pattern classification using principal components of cortical thickness and its discriminative
pattern in schizophrenia. NeuroImage 34, 1405–1415 (2007). [PubMed: 17188902]
25. Grimm, M.O., Rothhaar, T.L., Grösgen, S., Burg, V.K., Hundsdörfer, B., Haupenthal, V.J.,
Friess, P., Kins, S., Grimm, H.S., Hartmann, T.: Trans fatty acids enhance amyloidogenic
processing of the Alzheimer amyloid precursor protein (APP). J. Nutr. Biochem. 23, 1214–
1223 (2012)
26. Bengio, Y., et al.: Representation learning: a review and new perspectives. IEEE Trans.
Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013). arXiv:1206.5538. https://doi.org/10.
1109/tpami.2013.50. PMID 23787338. S2CID 393948
27. Bingham, E., Mannila, H.: Random projection in dimensionality reduction: Applications to
image and text data, 6 May 2014
28. Zhang, J., Wu, X., Hoi, S.C.H., Zhu, J.: Feature agglomeration networks for single stage face
detection. Neurocomputing 380, 180–189 (2020). ISSN 0925-2312. https://doi.org/10.1016/
j.neucom.2019.10.087
29. Joy, M.H., Hasan, M., Miah, A.S.M., Ahmed, A., Tohfa, S.A., Bhuaiyan, M.F.I., Rashid, M.
M., Zannat, A.: Multiclass MI-Task Classification using Logistic regression and Filter Bank
Common Spatial Patterns. Springer-Nature Singapore Pte Ltd., COMS2 2020, CCIS 1235,
pp. 160–170 (2020)
30. Teuwen, J., Moriakov, N.: Chapter 20 - Convolutional neural networks, Handbook of
Medical Image Computing and Computer Assisted Intervention, pp. 481–501. Academic
Press (2020). ISBN 9780128161760
An Analytical Intelligence Model
to Discontinue Products in a Transnational
Company
1 Introduction
A fundamental activity for the sales planning and operations in large companies with
value chains for the production, distribution, and marketing of consumer goods, is the
product discontinuation process. This product portfolio management process allows
optimizing the value chain. This process is generally driven by the commercial area of
the companies and is accompanied by the financial and supply chain area. These areas
make decisions about the portfolio empirically and based on the experience of the staff.
In the literature, there is evidence focused on solving the problem of discontinuing
products using techniques for classifying inventory types with a classic supply chain
approach. Other works are more focused on the application of machine learning
methods to categorize products based on consumer behavior. Transactional databases
are generally used, and rules are general for decision making. [1] show an application
to the retail sector on the type and quantity of products that they keep in their catalogs,
through the application of a hybrid model based on the k-means algorithm and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 812–822, 2021.
https://doi.org/10.1007/978-3-030-68154-8_70
An Analytical Intelligence Model to Discontinue Products 813
association rule mining. [2] shows a case of innovation and discontinuation but using
the firm as the unit of analysis.
[3] shows the application of text mining methodologies to assert the voice of the
customer by analyzing the case of a household and consumer products company. [4]
analyze the classification of products in an e-commerce framework using a deep multi-
modal architecture using text and images as inputs, showing the application of a deep
learning model. [5] develops an intelligent platform for the classification of products in
e-commerce based on natural language processing algorithms, vector space models, k-
nearest neighbor, Naive-Bayes classifier, and hierarchical classification. [6] address a
classification problem with big data. The goal is to properly classify millions of
products by type at Walmart-Labs. Applying large scales crowdsourcing, learning
rules, and in-house analysts.
Other studies such as the work of [7] use the Eigen-color feature with ensemble
machine learning combining artificial neural networks and support vector machines for
the classification of product images. The little identified on the application of product
classification does not address discontinuation and the vast majority focus on super-
vised models. However, the problem of this study lacks a dependent variable since
what is sought is to identify underlying information that allows grouping products for
later discontinuation based on a group of variables that address various areas of the
business beyond the physical characteristics of the product. For the above, this work
seeks to develop an analytical intelligence model to improve decision-making for the
discontinuation of products. The information used corresponds to commercial and
financial information for the entire product line in Mexico’s operation. The company
has 52 bottling plants and more than 300 distribution centers, its corporate headquarters
are located in Mexico City. The portfolio of products includes around 250 and up to
536 different product codes (SKUs). With the constant changes in consumer habits, the
proliferation of new product codes is a constant that different areas must deal with
when carrying out the quarterly sales and operations plan.
The application of an analytical intelligence model that allows integrating groups
susceptible of SKUs to be discontinued will allow having a standard methodology so
that this process is carried out in the most objective way possible based on technical
evidence. The work is organized as follows, in the first section, the theoretical
framework and the sources of information to be used are presented. The second section
presents the proposal for the model. The third section presents the main results iden-
tified. Finally, the conclusions and future lines of research.
The main variable that was historically taken into account to discontinue products
was the variable contribution, which is the result of subtracting from sales, variable
costs, and expenses, as well as the marketing expense of each product. It is the amount
that remains per product to cover costs and fixed expenses. The negative CV represents
a loss of economic value for the company. Presentations that correspond to specific
consumption occasions and that are important for the brand image were eliminated; as
well as products that according to their life cycle are in a phase of introduction to the
market, or growth.
It is important to mention that by transforming categorical variables into binary
variables, the amount of data is exponentially increased by generating a total of 761
variables, so the 16 original variables are transformed into 768 variables to be analyzed.
It is assumed that the discontinuation of any SKU will be for total channels and the
Customer Business Model variable will not be taken into account. As we can see, the
complexity of making a model with this database lies in the number of levels or factors
of each categorical variable. An exploratory analysis of the resulting database was
carried out and the levels of each categorical variable were converted into factors. The
extreme values of the base were eliminated using the criterion of the interquartile range
where the observations corresponding to the lower 5% and the upper 1% of the base
were discarded. 15 records were deleted. Subsequently, the analysis of Mixed Principal
Components was carried out, to be able to confirm if the data are grouped according to
the factors of the categorical variables and if any natural structure is identified in the
data.
An Analytical Intelligence Model to Discontinue Products 815
Figure 1 shows the eigenvectors of the first two principal components, the grouping
of the set of numerical variables with the first principal component, and the categorical
variables with the second component is observed. The first two mixed principal
components explain only 9% of the variance, so we cannot take the variables that are
most related to each one as a starting point for unsupervised modeling.
The graph of the observations concerning the first two mixed principal components
(Fig. 2) shows a natural grouping, the segmentation of 4 groups can be identified
according to the values of the first two principal components, the exploratory analysis
allows evaluating whether it is convenient for the construction of an unsupervised
model for the classification of products for their discontinuation.
The integration of mixed variables in the model implies the use of a specific distance
metric for the treatment of mixed data; in this case, the Gower distance matrix will be
used as the base input for the construction of the model.
816 G. Loy-García et al.
Where xi¼distðpi ;pj Þ is the distance for each observation from its closest neighbor and
yi¼distðqi ;qj Þ is the distance for each point from its closest neighbor in a data set simu-
lated using a uniform distribution. Since it is a database with mixed data, it is necessary
to establish an ad hoc metric, the Gower distance was used [11, 12].
Pn
i¼1 Wijk Sijk
GOWjk ¼ P n ð2Þ
i¼1 Wijk
An Analytical Intelligence Model to Discontinue Products 817
Where Wijk ¼ 0 if the objects j and k cannot be compared for variable i either because
xij or xik are not known. Additionally, for nominal variables:
Wijk = 1 if xij and xik are known; so
Sijk = 0 if xij 6¼ xik
Sijk = 1 if xij ¼ xik
To determine the number of partitions to be made, the Silhouette Coefficient was
used, which allows an optimization criterion to be used to determine the number of
partitions in a cluster using a measure of the quality of the cluster classification [13].
1X N
di s i
Silhouette ¼ ð3Þ
N i¼1 maxfdi si g
Where si is the mean distance to objects in the same cluster and di is the mean
distance to objects in the next nearest cluster. Observations with a large Silhouette
(1) are well grouped. A small Silhouette (0) means that the observation is between two
groups. Observations with a negative Silhouette are probably located in the wrong
group. A group of hard and soft cluster methodologies was selected to select the best
methodology to apply in the analyzed database. The selected methodologies were k-
means [14], k-medoids [13, 15], hierarchical agglomerative [16] and fuzzy clustering
[17, 18]. To select the best algorithm for the classification and profiling of the groups,
the validation of each model or each stage of the model is carried out by generating an
internal validation and the calculation of several stability means. To perform internal
validation of clusters, the Average Silhouette statistic is calculated. The grouping
quality will be better when the value of this statistic is close to 1. This statistic is
calculated for each selected algorithm by varying the number of clusters. In this way,
the algorithm and with the number of groups that the Silhouette statistic has closest to
1, should be chosen as the best model [13].
Taking into account the disaggregation of the categorical variables used in each
iteration, the increase in the dimension of the problem grows exponentially, due to all
the possible combinations of levels between the categorical variables. The proposed
model was coded on a free software platform (R) and later it was implemented on the
Microsoft Azure platform (Fig. 4).
The data processing power in the Data-Bricks module comes from a dedicated
processing cluster within the module with the following features: 140 GB of memory
and 20 cores. Additionally, it was configured with a minimum of 6 “Workers” or
parallel computing modules.
3 Results
In this case, the fuzzy cluster algorithm can no longer be calculated for K ¼ 3,
K ¼ 4; and K ¼ 5. We observe that the highest Silhouette index is obtained for the K-
Means and Hierarchical Agglomerative algorithms with K ¼ 3 and K ¼ 4 respectively
and it is verified that the optimal number of groups is 4. Given the complexity of the
problem, we seek to obtain partitions greater than two due to the desired segmentation
for discontinuation. Therefore, we selected the K-Means algorithm with 4 groups for
the classification of the base, graphically the clusters are represented on a principal
component map and dendrogram to observe the initial classification generated by the
selected model (Fig. 5). this is consistent with the segmentation identified in the
exploratory data analysis by applying mixed principal components.
Fig. 5. Cluster plot using the two principal components and dendrogram graph.
An Analytical Intelligence Model to Discontinue Products 819
The selection of the algorithm with four partitions allows integrating the groups of
SKUs, now it is necessary to build the profiles of each group to evaluate their char-
acteristics and define which group would be a candidate for discontinuation.
Definition of profiles within each group
The groups were summarized and characterized by the variable contribution that each
group generates per bottle and the percentage that it contributes to the total volume in
unit boxes to the business in Mexico. A total of 20, 218 records were excluded as they
had a negative variable contribution or a special reason, which is equivalent to 16% of
the volume in unit cases to the total of enterprise in Mexico. 82% of the remaining
volume is what was classified with this model. In this case, the outliers correspond to
the best products of the company.
$0.22 (
(a) (b
b)
$0.19
$
$0.17 $0.16
$0.13
993
34.6%
2
23.8%
18.1% 406
160
3.0% 2.3% 15 44
OU
UTLIERS 1 2 3 4 OUTLIE
ERS 1 2 3 4
Fig. 6. (a) Characterization and (b) dimension of the clusters generated in the first stage.
These are mostly products that do not make commercial sense due to the type of size
and presentation with which they are sold or because of their selling price that is too
onerous for states with a low socioeconomic status. The most representative sizes
belong to Personal Presentations and most correspond to Tin and Non-Returnable
Glass. These records are equivalent to discontinuing 195 SKUs, which do not represent
any profit for the company and only increase the complexity of the value chain as we
have previously pointed out.
4 Conclusions
References
1. Karki, D.: A hybrid approach for managing retail assortment by categorizing products based
on consumer behavior. Dublin, National College of Ireland. Ph.D. Thesis (2018)
2. Crowley, F.: Product and service innovation and discontinuation in manufacturing and
service firms in Europe. Eur. J. Innov. Manage. 20(2), 250–268 (2017)
3. Gonzalez, R.A., Rodriguez-Aguilar, R., Marmolejo-Saucedo, J.A.: Text mining and
statistical learning for the analysis of the voice of the customer. In: Hemanth, D., Kose,
U. (eds.) Artificial Intelligence and Applied Mathematics in Engineering Problems.
ICAIAME 2019. Lecture Notes on Data Engineering and Communications Technologies,
vol 43. Springer, Cham (2020)
4. Zahavy, T., Magnani, A., Krishnan, A., Mannor, S.: Is a picture worth a thousand words? A
deep multi-modal fusion architecture for product classification in e-commerce. In: The
Thirtieth Conference on Innovative Applications of Artificial Intelligence (IAAI) (2018)
5. Ding, Y., Korotkiy, M., Omelayenko, B., Kartseva, V., Zykov, V., Klein, M., Schulten, E.,
Fensel, D.: Golden bullet: automated classification of product data in e-commerce. In:
Proceedings of Business Information Systems Conference (BIS 2002), Poznan, Poland
(2002)
6. Sun, C., Rampalli, N., Yang, F., Doan, A.: Chimera: Large-scale classification using
machine learning, rules, and crowdsourcing. PVLDB 7(13), 1529–1540 (2014)
7. Oyewole, S.A., Olugbara, O.O.: Product image classification using Eigen Colour feature
with ensemble machine learning. Egypt. Inf. J. 19(2), 83–100 (2018)
8. Kassambara, A.: A practical guide to cluster analysis in R: unsupervised machine learning.
CreateSpace Independent Publishing Platform (2017)
9. Hopkins, B., Skellam, J.G.: A new method for determining the type of distribution of plant
individuals. Ann. Bot. Co. 18(2), 213–227 (1954)
10. Banerjee, A.: Validating clusters using the Hopkins statistic. In: IEEE International
Conference on Fuzzy Systems, pp. 149–153 (2004)
11. Gower, J.: A general coefficient of similarity and some of its properties. Biometrics 27, 857–
872 (1971)
12. Tuerhong, G., Kim, S.B.: Gower Distance-Based Multivariate Control Charts for a Mixture
of Continuous and Categorical Values. Elsevier, South Korea (2013)
13. Kaufman, L., Rousseeuw, P.J.: Finding Groups in Data: An Introduction to Cluster Analysis,
p. 1990. John Wiley & Sons Inc, Hoboken, NJ (1990)
822 G. Loy-García et al.
14. Lloyd, S.P.: Least squares quantization in PCM. IEEE Trans. Inf. Theory 28(1982), 129–137
(1982)
15. Park, H.-S., Jun, C.-H.: A simple and fast algorithm for K-medoids clustering. Expert Syst.
Appl. 36, 3336–3341 (2009)
16. Ward, J.H.: Hierarchical grouping to optimize an objective function. J. Am. Stat. Assoc. 58
(301), 236–244 (1963)
17. Dunn, J.C.: A fuzzy relative of the ISODATA process and its use in detecting compact well-
separated clusters. J. Cybern. 3(3), 32–57 (1973)
18. Bezdek, J.C.: Pattern Recognition with Fuzzy Objective Function Algorithms (1981)
19. Rodriguez-Aguilar, R.: Proposal for a comprehensive environmental key performance index
of the green supply chain. Mobile Netw. Appl. (2020)
Graph Neural Networks in Cheminformatics
1 Introduction
Past decades have witnessed the surge of data in every aspects of life and the devel-
opment of data science tools created to study them [1], especially in the field of
chemistry. In this revolution, machine learning stays the most important in solving the
problems of chemical data exploitation and understanding [2]. Machine learning is a
domain in artificial intelligence that focuses on learning from data with computational
models. In procedures which are dealing with a huge amount of data like drug discovery,
machine learning, particularly deep learning, is contemporary of great importance [3].
On the other side, the graph is important structural information of chemistry data,
hence it requires appropriate algorithms to learn and analyze them taking advantage of
their structures [4]. In the context of utilizing deep learning techniques for both han-
dling massive datasets and learning from graphs, novel algorithms specialized in graph
data are developed, which are currently known as graph neural networks (GNN).
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 823–837, 2021.
https://doi.org/10.1007/978-3-030-68154-8_71
824 H. N. T. Tran et al.
to the binding of a drug to a target location that results in a change in its behavior or
function [9]. This process is currently boosted up with the assistance of various
machine learning models.
proving that auto-encoder architecture can generate di-verse and novel molecules. To
improve the work of auto-encoder in molecule generation, [24, 25], and [26] adopted
RNN in their encoder or decoder, or both, and performed a high reconstruction rate.
[27] even took advantage of generative topo-graphic mapping (GTM) in their
RNN/auto-encoder model to visualize the auto-encoder latent space.
L¼DA ð3Þ
ov ¼ gðhv ; xv Þ ð8Þ
where f is a parametric function, called local transition function, which is shared among
all nodes and g is a parametric function, called local output function, used to produce
the output.
828 H. N. T. Tran et al.
Let H, O, X, and XN be the matrices constructed by stacking all the states, outputs,
features, and node features respectively. We have a compact form as following:
H ¼ FðH; XÞ ð9Þ
O ¼ GðH; XN Þ ð10Þ
where F is the global transition function and G is the global output function.
Let Ht be the tth iteration of H, the general form of the state is presented as:
H t þ 1 ¼ FðH t ; XÞ ð11Þ
With target information for a specific node denoted as tv and the number of
supervised nodes denoted as p, the loss of model will be:
X
p
loss ¼ ðti oi Þ ð12Þ
i¼1
Briefly, this vanilla model presents an effective method of modeling graphic data,
which is the first step towards integrating neural networks into the graph domain.
H ðl þ 1Þ ¼ rðD ~D
~ 12 A ~ 12 H ðlÞ W ðlÞ Þ ð13Þ
in which A ~ is the adjacency matrix of the undirected graph G with added self-
P
connections, computed as à = A + IN with IN is the identity matrix; D ~ ii ¼ ~
j Aij and
W is a layer-specific trainable weight matrix. r(.) represents an activation function.
(l)
H ðlÞ 2 Rnd is the matrix of activation in the lth layer (n is the number of nodes, d is the
dimension of the node feature vector), H(0) = X.
The definition of the GCN is demonstrated as:
Z¼D ~D
~ 2 A ~ 2 XH
1 1
ð14Þ
where H 2 RCF is a matrix of filter parameters with C input channels and F filters for
the feature.
Graph Neural Networks in Cheminformatics 829
X ht1
AGGMEAN ¼ u
ð16Þ
u2NðvÞ
jNðvÞj
• LSTM aggregator:
u ; 8u 2 pðNðvÞÞÞ
AGG ¼ LSTMð½ht1 ð17Þ
where h(0)
v = xv.
Based on GRNN, You et al. proposed a graph generative model called GraphRNN,
whose aim is to consider graphs under different node orderings as sequence repre-
sentations, and then to construct an auto-regressive generative model on these
830 H. N. T. Tran et al.
sequences [33]. The state transition function and the output function of the neural
network are presented as:
where hi 2 Rd is a vector representing the state of the generated graph, p is the node
ordering. An element Spi 2 {0,1}i−1, i 2 {1,…,n} denotes an adjacency vector repre-
senting the edges between node p(vi) and previous nodes p(vj), j 2 {1,…,i-1}. hi is the
distribution of the next node’s adjacency vector. ftrans and fout in general can be arbi-
trary neural networks.
Furthermore, the research work of [34] has extended GraphRNN of You et al.
(2018) to the model MolecularRNN in which they apply RNN for both state transition
function (NodeRNN) and output function (EdgeRNN).
Z ¼ GCNðX; AÞ ð22Þ
~ ¼ rðzT ; zu Þ
A ð23Þ
v
In the field of cheminformatics, GCN was early proposed by Duvenaud et al. who
studied a neural graph fingerprints architecture that employs CNN, and the model
training had matched or beaten the predictive performance of standard (fixed) finger-
prints on drug efficacy [36].
Differently, Kearns et al. employed ‘Weave modules’, which merge information
from all atoms and atom pairs to learn molecular features for molecular graph con-
volutional deep neural network models [37]. Compared to the random forest, logistic
regression, and a multitask neural network with two underlying layers, this graph
neural network framework demonstrated higher performance in all cases.
Graph Neural Networks in Cheminformatics 831
Table 1. Performance results of GNN and other machine learning-based methods in drug-target
interactions prediction
Article Evaluation Datasets Performance measurement
metrics Proposed Compared
method methods
[40] GNN-CNN k-NN RF SVM
AUC1 human 0.970 0.860 0.940 0.910
C.elegans 0.978 0.858 0.902 0.894
Precision human 0.923 0.798 0.861 0.966
C.elegans 0.938 0.801 0.821 0.785
[37] GCN MaxSim LR RF
AUC PubChem 0.908 0.754 0.838 0.804
MUV 0.858 0.638 0.736 0.655
Tox21 0.867 0.728 0.789 0.802
[21] GAT-GCN CNN SimBoost kernel-based
[10]
CI2 Davis 0.881 0.878 0.872 0.871
Kiba 0.891 0.863 0.836 0.782
MSE3 Davis 0.245 0.261 0.282 0.379
Kiba 0.139 0.194 0.222 0.411
[48] GAT-GNN docking CNN
AUC DUD-E 0.968 0.689 0.868
ChEMBL 0.633 0.572 –
MUV 0.536 0.533 0.518
PRAUC4 DUD-E 0.697 0.016 –
[49] GCN Smina neural AutoDock
fingerprints Vina
AUC DUD-E 0.567 0.642 0.704 0.633
MUV 0.474 0.593 0.575 0.503
1
Area under the receiving operating characteristic curve
2
Concordance index
3
Mean square error
4
Area under the precision-recall curve
5 Discussion
Most current models for drug-target interaction prediction are using GCN, mean-
while, graph autoencoders are preferably adopted in generative models, which are
meant to generate new compounds. There haven’t been many works on GRNN,
however, it is gaining more attention as in recent research by [34].
Various models of GNN have been being developed and extended in terms of
choosing methods for encoding and decoding graph data, the depth of networks and
selecting activation functions used in layers of neural networks, and the combination of
different learning methods in one model. From the literature, GNN has proved to be a
promising tool for studying cheminformatics with the ability of handling graph
structures, predicting from unlabeled data, and generating novel entities.
GNN is becoming state-of-the-art among machine learning techniques that are
created and practiced to solve cheminformatics problems. Although recent researches
have shown positive results, there is room for great improvement in this learning
method. The future of applying GNN in cheminformatics will be broad yet challenging
because research is expected to turn to discovering and designing a specific type of
drug, to reposition current drugs (to find new clinical uses of existing drugs), and to
enhance the performance of predictive models. Also importantly, there are still
demands of availability of various toolkits, programming libraries, and frameworks for
data mining and model building.
6 Conclusion
In this review, we have focused on recent applications of graph neural networks in drug
discovery. We first defined technical terms relevant to cheminformatics and indicated
that drug discovery is a key problem in cheminformatics. We provided the presentation
of the mathematical model of the initial graph neural network and its main variants:
graph convolutional neural networks, graph recurrent neural networks, and graph
autoencoders. Subsequently, we investigated recent researches employing graph neural
networks for drug-target interactions prediction and molecule generating. Along with
this, we discussed the performance of graph neural networks models and suggested
prospective topics in studying graph neural networks in cheminformatics.
Acknowledgement. This work is supported by the Fundamental Research Grant Scheme (FRGS)
of the Ministry of Higher Education Malaysia under the grant project number FRGS/1/2019/ICT02/
KDUPG/02/1.
References
1. Joshua Thomas, J., Pillai, N.: A deep learning framework on generation of image
descriptions with bidirectional recurrent neural networks. In: Advances in Intelligent
Systems and Computing, vol. 866. Springer International Publishing (2019). https://doi.org/
10.1007/978-3-030-00979-3_22
834 H. N. T. Tran et al.
2. Lipinski, C.F., Maltarollo, V.G., Oliveira, P.R., da Silva, A.B.F., Honorio, K.M.: Advances
and perspectives in applying deep learning for drug design and discovery. Front. Robot. AI 6
(November), 1–6 (2019). https://doi.org/10.3389/frobt.2019.00108
3. Chen, H., Engkvist, O., Wang, Y., Olivecrona, M., Blaschke, T.: The rise of deep learning in
drug discovery. Drug Discov. Today 23(6), 1241–1250 (2018). https://doi.org/10.1016/j.
drudis.2018.01.039
4. Li, J., Cai, D., He, X.: Learning Graph-Level Representation for Drug Discovery (2017).
http://arxiv.org/abs/1709.03741
5. Lo, Y.C., Rensi, S.E., Torng, W., Altman, R.B.: Machine learning in chemoinformatics and
drug discovery. Drug Discov. Today, 23(8), 1538–1546. Elsevier Ltd. (2018). https://doi.
org/10.1016/j.drudis.2018.05.010
6. Chan, H.C.S., Shan, H., Dahoun, T., Vogel, H., Yuan, S.: Advancing drug discovery via
artificial intelligence. Trends Pharmacol. Sci. 40(8), 592–604 (2019). https://doi.org/10.
1016/j.tips.2019.06.004
7. Rifaioglu, A.S., Atas, H., Martin, M.J., Cetin-Atalay, R., Atalay, V., Doǧan, T.: Recent
applications of deep learning and machine intelligence on in silico drug discovery: Methods,
tools and databases. Brief. Bioinform. 20(5), 1878–1912 (2019). https://doi.org/10.1093/bib/
bby061
8. Chen, R., Liu, X., Jin, S., Lin, J., Liu, J.: Machine learning for drug-target interaction
prediction. Molecules 23(9), 1–15 (2018). https://doi.org/10.3390/molecules23092208
9. Wang, H., Wang, J., Dong, C., Lian, Y., Liu, D., Yan, Z.: A novel approach for drug-target
interactions prediction based on multimodal deep autoencoder. Front. Pharmacol. 10, 1–19
(2020). https://doi.org/10.3389/fphar.2019.01592
10. Ding, H., Takigawa, I., Mamitsuka, H., Zhu, S.: Similarity-based machine learning methods
for predicting drug-target interactions: A brief review. Brief. Bioinform. 15(5), 734–747
(2013). https://doi.org/10.1093/bib/bbt056
11. Sachdev, K., Gupta, M.K.: A comprehensive review of feature-based methods for drug target
interaction prediction. J. Biomed. Inf. Elsevier (2019). https://doi.org/10.1016/j.jbi.2019.
103159
12. Thomas, J.J., Ali, A.M.: Dispositional learning analytics structure integrated with recurrent
neural networks in predicting students performance. In: Vasant, P., Zelinka, I., Weber, G.W.
(eds.) Intelligent Computing and Optimization. ICO 2019. Advances in Intelligent Systems
and Computing, vol 1072. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-
33585-4_44
13. Dahl, G.E., Jaitly, N., Salakhutdinov, R. Multi-task Neural Networks for QSAR Predictions,
pp. 1–21 (2014). http://arxiv.org/abs/1406.1231
14. Tian, K., Shao, M., Wang, Y., Guan, J., Zhou, S.: Boosting compound-protein interaction
prediction by deep learning. Methods 110, 64–72 (2016). https://doi.org/10.1016/j.ymeth.
2016.06.024
15. Segler, M.H.S., Kogej, T., Tyrchan, C., Waller, M.P.: Generating focused molecule libraries
for drug discovery with recurrent neural networks. ACS Central Sci. 4(1), 120–131 (2018).
https://doi.org/10.1021/acscentsci.7b00512
16. Gupta, A., Müller, A.T., Huisman, B.J.H., Fuchs, J.A., Schneider, P., Schneider, G.:
Generative recurrent networks for De Novo drug design. Mol. Inf. 37(1) (2018). https://doi.
org/10.1002/minf.201700111
Graph Neural Networks in Cheminformatics 835
17. Hirohara, M., Saito, Y., Koda, Y., Sato, K., Sakakibara, Y.: Convolutional neural network
based on SMILES representation of compounds for detecting chemical motif. BMC Bioinf.
19(Suppl 19), 83–94 (2018). https://doi.org/10.1186/s12859-018-2523-5
18. Lee, I., Keum, J., Nam, H.: DeepConv-DTI: prediction of drug-target interactions via deep
learning with convolution on protein sequences. PLoS Comput. Biol. 15(6), 1–21 (2019).
https://doi.org/10.1371/journal.pcbi.1007129
19. Öztürk, H., Özgür, A., Ozkirimli, E.: DeepDTA: Deep drug-target binding affinity
prediction. Bioinformatics 34(17), i821–i829 (2018). https://doi.org/10.1093/bioinform
atics/bty593
20. Trabelsi, A., Chaabane, M., Ben-Hur, A.: Comprehensive evaluation of deep learning
architectures for prediction of DNA/RNA sequence binding specificities. Bioinformatics 35
(14), i269–i277 (2019). https://doi.org/10.1093/bioinformatics/btz339
21. Karimi, M., Wu, D., Wang, Z., Shen, Y.: DeepAffinity: Interpretable deep learning of
compound-protein affinity through unified recurrent and convolutional neural networks.
Bioinformatics 35(18), 3329–3338 (2019). https://doi.org/10.1093/bioinformatics/btz111
22. Samanta, B., De, A., Jana, G., Chattaraj, P.K., Ganguly, N., Rodriguez, M.G.: NeVAE: a
deep generative model for molecular graphs. In: Proceedings of the AAAI Conference on
Artificial Intelligence, vol. 33, pp. 1110–1117 (2019). https://doi.org/10.1609/aaai.v33i01.
33011110
23. Lim, J., Hwang, S.Y., Moon, S., Kim, S., Kim, W.Y.: Scaffold-based molecular design with
a graph generative model. Chem. Sci. 11(4), 1153–1164 (2019). https://doi.org/10.1039/
c9sc04503a
24. Gómez-Bombarelli, R., Wei, J.N., Duvenaud, D., Hernández-Lobato, J.M., SánchezLen-
geling, B., Sheberla, D., Aguilera-Iparraguirre, J., Hirzel, T.D., Adams, R.P., Aspuru-Guzik,
A.: Automatic chemical design using a data-driven continuous representation of molecules.
ACS Central Sci. 4(2), 268–276 (2018). https://doi.org/10.1021/acscentsci.7b00572
25. Bjerrum, E.J., Sattarov, B.: Improving chemical autoencoder latent space and molecular de
novo generation diversity with heteroencoders. Biomolecules 8(4), 1–13 (2018). https://doi.
org/10.3390/biom8040131
26. Lim, J., Ryu, S., Kim, J.W., Kim, W.Y.: Molecular generative model based on conditional
variational autoencoder for de novo molecular design. J. Cheminform. 10(1), 1–9 (2018).
https://doi.org/10.1186/s13321-018-0286-7
27. Sattarov, B., Baskin, I.I., Horvath, D., Marcou, G., Bjerrum, E.J., Varnek, A.: De Novo
molecular design by combining deep autoencoder recurrent neural networks with generative
topographic mapping. J. Chem. Inf. Model. 59(3), 1182–1196. Research-Article (2019).
https://doi.org/10.1021/acs.jcim.8b00751
28. Scarselli, F., Gori, M., Tsoi, A.C., Hagenbuchner, M., Monfardini, G.: The graph neural
network model. IEEE Trans. Neural Networks 20(1), 61–80 (2009)
29. Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., Yu, P.S.: A Comprehensive Survey on
Graph Neural Networks, pp. 1–22 (2019). http://arxiv.org/abs/1901.00596
30. Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks.
In: 5th International Conference on Learning Representations, ICLR 2017 - Conference
Track Proceedings, pp. 1–14 (2017). https://arxiv.org/abs/1609.02907
31. Hamilton, W., Ying, Z., Leskovec, J.: Inductive representation learning on large graphs. In:
Advances in Neural Information Processing Systems, pp. 1024–1034 (2017)
32. Li, Y., Zemel, R., Brockschmidt, M., Tarlow, D.: Gated graph sequence neural networks. In:
4th International Conference on Learning Representations, ICLR 2016 - Conference Track
Proceedings, pp. 1–20 (2016)
836 H. N. T. Tran et al.
33. You, J., Ying, R., Ren, X., Hamilton, W.L., Leskovec, J.: GraphRNN: generating realistic
graphs with deep auto-regressive models. In: 35th International Conference on Machine
Learning, ICML 2018, vol. 13, pp. 9072–9081 (2018)
34. Popova, M., Shvets, M., Oliva, J., Isayev, O.: MolecularRNN: Generating realistic molecular
graphs with optimized properties (2019). https://arxiv.org/abs/1905.13372
35. Hajiramezanali, E., Hasanzadeh, A., Duffield, N., Narayanan, K.R., Zhou, M., Qian, X.:
Variational Graph Recurrent Neural Networks, pp. 1–12 (2019). http://arxiv.org/abs/1908.
09710
36. Duvenaud, D., Maclaurin, D., Aguilera-Iparraguirre, J., Gómez-Bombarelli, R., Hirzel, T.,
Aspuru-Guzik, A., Adams, R.P.: Convolutional networks on graphs for learning molecular
fingerprints. In: Advances in Neural Information Processing Systems, pp. 2224–2232 (2015)
37. Kearnes, S., McCloskey, K., Berndl, M., Pande, V., Riley, P.: Molecular graph
convolutions: moving beyond fingerprints. J. Comp. Aided Mol. Des. 30, 595–608
(2016). https://doi.org/10.1007/s10822-016-9938-8
38. Coley, C.W., Jin, W., Rogers, L., Jamison, T.F., Jaakkola, T.S., Green, W.H., Jensen, K.F.:
A graph-convolutional neural network model for the prediction of chemical reactivity.
Chem. Sci. 10(2), 370–377 (2019). https://doi.org/10.1039/c8sc04228d
39. Ryu, S., Lim, J., Hong, S.H., Kim, W.Y.: Deeply learning molecular structure-property
relationships using attention- and gate-augmented graph convolutional network (2018).
https://doi.org/10.1039/b000000x/been
40. Tsubaki, M., Tomii, K., Sese, J.: Compound-protein interaction prediction with end-to-end
learning of neural networks for graphs and sequences. Bioinformatics 35(2), 309–318
(2019). https://doi.org/10.1093/bioinformatics/bty535
41. Nguyen, T., Le, H., Quinn, T.P., Le, T., Venkatesh, S.: Predicting drug–target binding
affinity with graph neural networks. BioRxiv 12, 1–18 (2019). https://doi.org/10.1101/
684662
42. Thomas, J.J., Tran, H.N.T., Lechuga, G.P., Belaton, B.: Convolutional graph neural
networks: a review and applications of graph autoencoder in chemoinformatics. In: Thomas,
J.J., Karagoz, P., Ahamed, B.B., Vasant, P. (eds.) Deep Learning Techniques and
Optimization Strategies in Big Data Analytics, pp. 107–123. IGI Global (2020). http://doi.
org/10.4018/978-1-7998-1192-3.ch007
43. Niepert, M., Ahmad, M., Kutzkov, K.: Learning convolutional neural networks for graphs.
In: 33rd International Conference on Machine Learning, ICML 2016, vol. 4, pp. 2958–2967
(2016)
44. Kusner, M.J., Paige, B., Hemández-Lobato, J.M.: Grammar variational autoencoder. In: 34th
International Conference on Machine Learning, ICML 2017, vol. 4, pp. 3072–3084 (2017)
45. Simonovsky, M., Komodakis, N.: GraphVAE: towards generation of small graphs using
variational autoencoders. In: Lecture Notes in Computer Science (including subseries
Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). LNCS, vol.
11139, pp. 412–422 (2018). https://doi.org/10.1007/978-3-030-01418-6_41
46. De Cao, N., Kipf, T.: MolGAN: An implicit generative model for small molecular graphs
(2018). http://arxiv.org/abs/1805.11973
47. Bresson, X., Laurent, T.: A Two-Step Graph Convolutional Decoder for Molecule
Generation (2019). http://arxiv.org/abs/1906.03412
Graph Neural Networks in Cheminformatics 837
48. Lim, J., Ryu, S., Park, K., Choe, Y.J., Ham, J., Kim, W.Y.: Predicting drug-target interaction
using a novel graph neural network with 3D structure-embedded graph representation.
J. Chem. Inf. Model. 59(9), 3981–3988. Research-Article (2019). https://doi.org/10.1021/
acs.jcim.9b00387
49. Gonczarek, A., et al.: Interaction prediction in structure-based virtual screening using deep
learning. Comput. Biol. Med. (2017). https://doi.org/10.1016/j.compbiomed.2017.09.007
Academic and Uncertainty Attributes
in Predicting Student Performance
Abstract. This article investigates the use of academic and uncertainty attri-
butes in predicting student performance. The dataset used in this article consists
of 60 students who studied these subjects: Cloud Computing, Computing
Mathematics, Fundamentals of OOP, Object-Oriented SAD, and User Interface
Design. A Deep Learning model was developed that uses a Long Short-Term
Memory network (LSTM) and Bidirectional Long Short-Term memory network
(BLSTM) to predict the student grades. The results show combining different
types of attributes can be used in predicting the student results. The result is
further discussed in the article.
1 Introduction
The Use of ML technology in the education field has been increasingly popular in
recent years, with many types of research being conducted from analyzing student data,
making predictions, and helping prevent the student from dropping out of the
university.
Universities now offer a broader range of specialized degrees than ever before
(minors, double degrees, interdisciplinary and inter-university masters). [1] Therefore,
this results in generating a large amount of data from LMS used by students and
lecturers in the university.
The data stored in educational institutions repository plays an essential role in
extracting hidden and interesting patterns to assist every stakeholder in an educational
process.
Students’ performance is the primary concern of various stakeholders, including
educators, administrators, and corporations. For recruiting fresh graduates, academic
achievement is the main factor considered by the recruiting agencies. Therefore,
graduates have to work hard for excellent grades to rise to the expectations of recruiting
agencies [1].
University uses LMS tools to provide convenience for the student to interact with
the instructor, but what the university lack is the knowledge behind this student data.
University is not aware of the benefits that it may gain if it analyses this student data
generated from the LMS tool. Analyzing this student data will help the university
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 838–847, 2021.
https://doi.org/10.1007/978-3-030-68154-8_72
Academic and Uncertainty Attributes in Predicting Student Performance 839
understand why specific student(s) may not perform well and inform the instructor to
provide guidance and support to the student(s).
This paper proposes an DL model that can be applied to the student information,
analyze it, and predict their performance. Being able to analyze student data and make
an informed decision can have an impact on the university. University who put effort
into analyzing this data can help understand student performance by guiding students
when needed. The paper contributes to the universities trying to improve their students’
performance and the individual student who wants to know how likely it will progress
throughout the study. This paper shows how DL, together with a specific type of
attributes, can be applied to the student data to analyze it better and predict the student
performance results. The proposed model will assist the university in predicting student
performance.
The rest of this paper is organized as follows: In Sect. 1 will describe work that is
done by the other researchers in terms of the techniques, the methods of the work, and
the achieved results. Section 2 will be implementation, where it describes in detail the
proposed model, the type of attributes, the dataset used for the experiment, and a result
produced by the model. Section 3, which concludes, will highlight the paper contri-
bution and some limitations.
2 Related Works
Data prediction using machine learning is being used in different types of fields. Pre-
dicting student grades gained attention from universities who are seeking ways to
improve student experience with the university.
The research was done by this author [2], who wants to predict student performance
by developing an algorithm based on BLSTM called GritNet that achieves the
objective. The algorithm do not depend on feature selection because it learns from raw
input data, and it can also incorporate the student events based on the time-stamp. It
was used on The algorithm could be further improved by incorporating indirect data
such as student board activity and the interaction between the student and lecturer. This
kind of information will help GritNet to produce great performance prediction.
Utilizing those data gathered from LMS and applying them to an ML algorithm
may not be enough to predict student performance. There is another part where it
required attention to identify the features that will be used together with the algorithm
to perform the prediction. Selecting the right set of features plays an essential role in
producing a quality prediction.
This study by [3] presents an analysis of the performance feature selection to help
other researchers identify the best feature selection algorithm. During the experiment
various types of feature selection algorithms were used such as: CfsSubsetEval,
ChiSquaredAttributeEval, FilteredAttributeEval, GainRatioAttributeEval, Principal
Components, and ReliefAttributeEval. The result has shown that the principal com-
ponent produced better results. This could be further improved by using different
dataset sizes and perform parameter tuning of these features.
840 A. M. Ali et al.
This research by [4] wanted to predict student difficulties during a learning session
by using a tool called Digital Electronic Education and Design Suit (DEEDS). During
the experiment, the author has trained the following models: Artificial Neural Network
(ANN), Support Vector Machine (SVM), Logistic Regression (LR), Naïve Bayes (NB),
and Decision Tree (DT) on a dataset consist of 361 students. The result shows that
ANN & SVM have achieved a high accuracy result of 75% in predicting the student’s
difficulty for the next digital session. This also indicates that ANN & SVM can be
integrated into the DEEDS and Massive Open Online Course(MOOC) to help students
choose the appropriate session to study and work effectively. A future improvement
would be to use the K-means algorithm to study the student learning behavior during
the interaction with DEEDS so that the lecturer can provide further assistance to
students who need help.
Many types of research were conducted in the education field, where the research
goal is to predict student performance. Researchers used different ML algorithms and
various kinds of a dataset that achieved a certain prediction accuracy.
This paper proposes another work to improve on the weaknesses which were
identified by other researchers such as:
• The set of features used in other research was not effective as it focused entirely on
the student academic data.
• Lack of student’s previous grades in the dataset.
• The algorithms do not work with indirect data, such as student board activity and
behavior.
Academic and Uncertainty Attributes in Predicting Student Performance 841
This section has shown other research works where researches have implemented
various types of ML technique to predict the student performance. This paper proposes
using academic, behavior, and uncertainty data to predict student performance. The
next section will cover the implementation.
3 Methodology
In this section will cover the research implementation starting with the research
framework, and its components where will explain the overall structure in details.
The framework used in this research, as shown in Fig. (1), consist of 3 layers:
• Uncertainty attributes.
842 A. M. Ali et al.
3.4 Dataset
The dataset used for this experiment consist of 60 students who are studying 5 subjects
which are: Cloud Computing, Computing Mathematics, Fundamentals of OOP, Object
Oriented SAD, and User Interface Design.
These data are extracted from Canvas LMS and Oracle LMS. The following
Table 2 shows the types of attributes that were collected from Canvas and Oracle LMS.
Table (2) shows the set of attributes that will be used in this experiment. Table (3)
explains the attributes as shown in Table (2) in details:
Library activity refers to the total time the student spent in the library. This is
calculated based on the time the student enters and leaves the library. LMS activity
refers to the total time the student spent on the subject. This ranges from viewing
subject materials, taking an assessment, or attending an online lecture.
This dataset is collected from the LMS while the library activity is collected from
the library system, where it tracks student access/exit using the student card.
4 Implementation
This section will cover in detail the research implementation. Will show how the model
is developed by explaining the type of algorithms and the techniques used to produce
the result. LSTM and BLSTM were used to build the model and make the prediction.
These 2 networks were used because the dataset consists of student data, and it’s
sequential, and the model needs to predict the student progress throughout the course.
Using LSTM alone is not sufficient to make an accurate prediction because due to the
context of data, so will use BLSTM as well because it can move forward-backward in
analyzing the data to achieve a higher prediction accuracy. The data is split into an
80/20 ratio, where 80% of the total data was used for training, while the remaining 20%
was used for testing.
An activation function called Relu was used in the model together with a Loss
function called Mean Square Error and an Optimizer called Adam. The Dense layer
was set to 1 as the final value that the model expected to produce is a single value or
output. Executing the program has produced the result, as shown in Fig. (4).
The Blue color line indicates the actual student grades, while the Orange color line
indicates the predicted grades. This shows the model is learning from the data and
producing prediction grades.
846 A. M. Ali et al.
5 Conclusion
This paper has presented different types of attributes that can be applied to a model to
predict the student result, and it was done on a group of students who took the subjects.
A future improvement will be to apply this model to a larger number of data to monitor
how the model will produce the result and fine-tune it to make it better.
Acknowledgment. This research work is supported by UOW Malaysia KDU Penang University
College.
References
1. Rovira, S., Puertas, E., Igual, L.: Data-driven system to predict academic grades and dropout.
PLoS One 12(2), e0171207 (2017)
2. Kim, B.H., Vizitei, E., Ganapathi, V.: GritNet: Student performance prediction with deep
learning. arXiv preprint arXiv:1804.07405 (2018)
3. Zaffar, M., Hashmani, M.A., Savita, K.S.: Performance analysis of feature selection algorithm
for educational data mining. In: 2017 IEEE Conference on Big Data and Analytics (ICBDA),
pp. 7–12. IEEE, November 2017
4. Hussain, M., Zhu, W., Zhang, W., Abidi, S.M.R., Ali, S.: Using machine learning to predict
student difficulties from learning session data. Artif. Intell. Rev. 52(1), 381–407 (2019)
5. Mousa, A.E.D., Schuller, B.W.: Deep bidirectional long short-term memory recurrent neural
networks for grapheme-to-phoneme conversion utilizing complex many-to-many alignments.
In: Interspeech, pp. 2836–2840, September 2016
6. Xia, J., Pan, S., Zhu, M., Cai, G., Yan, M., Su, Q., Ning, G.: A long short-term memory
ensemble approach for improving the outcome prediction in intensive care unit. Comput.
Math. Methods Med. 2019, 1–11 (2019)
Academic and Uncertainty Attributes in Predicting Student Performance 847
7. Ameri, S., Fard, M.J., Chinnam, R.B., Reddy, C.K.: Survival analysis based framework for
early prediction of student dropouts. In: Proceedings of the 25th ACM International on
Conference on Information and Knowledge Management, pp. 903–912, October 2016
8. Thomas, J.J., Ali, A.M.: Dispositional learning analytics structure integrated with recurrent
neural networks in predicting students performance. In: International Conference on
Intelligent Computing and Optimization, pp. 446–456. Springer, Cham, October 2019
Captivating Profitable Applications
of Artificial Intelligence in Agriculture
Management
1 Introduction
There are about 14 crore farmers in India whose main means of livelihood is agri-
culture. With changes in ecological requirements [13], we have raised the question ‘as
to how we can use technology to replace some of the human activities and guarantee
efficiency in farming?’ Farmers need to reach the next level of profitability and effi-
ciency in farming. The situation is slowly changing right under our noses. Artificial
Intelligence is emerging as part of the solutions to improve agricultural productivity. AI
adoption in agriculture provides better farming systems and better ways to grow crops.
It also provides accurate and timely information on crop growth and on conditions to
detect and prevent crop disease, weed, and insect damage based on weather forecasts. It
helps in precision farming, crop management, crop yield estimation, c early warning
and mitigation of crop disaster, agricultural production planning, crop commodity
trading, and food security decision support [9]. While some of AI based agriculture
technology benefits include saving farmers time and money and boosting average crop
yields; others stem from more complex global concerns like weather uncertainties,
labor shortages, and rising demand for healthier, sustainably grown food [2].
[1]. In the present, irrigation and weeding has been mitigated by automated robotic
systems [15] Fig. 2.
AI is has been paired with drones to analyse aerial images and give agriculturists
data that can help them reduce planting costs, cut down water and fertilizer use, and
monitor crop health. Drones map out pre-planting patterns to help farmers maximize
the area to plant crops and implement the best planting pattern for individual crops.
This minimizes the time to map out the pattern on foot.
852 R. Sivarethinamohan et al.
Fig. 5. Fruit-fly
backgrounds, and different lighting. The pictures of two cows can be distinguished by
the physical features across their faces, including the muzzles and eyes, which ML
catches. This helps in classifying the cattle in terms of age, breed, and other categories.
For instance, the Mooofarm app is used for digitizing the breeding cycle of cattle., The
farmer uploads the basic inputs of his cattle to get real-time alerts such as the right time
to inseminate them and the right time to get a pregnancy diagnosis. The app also has an
inbuilt e-commerce platform that allows farmers to connect with input suppliers (food,
fodder seeds) and other service providers such as insurance and veterinarians. It also
allows making an entry for the medicines fed to the cattle, the kind of insemination
done, the amount of fodder given to the cattle, etc. This app helps farmer to reduce the
feed amount Fig. 7.
4.8 Chabots
A chat-bot will provide assistance to farmers using NLP. This bot answers their
questions related to agriculture practice and technology and provides advice and rec-
ommendations on specific farm problems [7].
4.9 Stethoscope
An excellent stethoscope provides the farmers with data and observations to make an
informed decision about their crops and thus takes the guess work out of farming.
That’s exactly done by fasal. The shoot and root borer insects in fruit and plantation
crops are detected through a stethoscope.
and AI to detect weeds in r fields and spray pesticides, fertilizer, and fungicides. These
applications are stored in the cloud and the farmer needs only a modem on the machine
to undertake heavier computing in the cloud. Using a combination of GPS, sensors, and
imaging, farmers can presumably have a better handle at how to deploy robotic
vehicles to till the soil [8].
5 Future of AI in Agriculture
The global Artificial Intelligence (AI) in the agriculture market is witnessing a CAGR
of 26.2% during the period 2019–2024 [16]. Digital farming and connected farm
services can impact 70 million Indian farmers in 2020, thereby adding $9 billion to
farmers incomes. In 2017, the Global AI in agriculture market was 240 million US$
and is expected to reach 1100 million US$ by the end of 2025 and a CAGR of 20.8%
during 2018–2025. Thus initiatives to increase digital literacy in the rural landscape can
be seen as a weapon to double farmers income in the near future The estimated number
Agricultural IoT devices are 75 million by 2020. 4.1 million data points farms are
estimated to be generated daily by 2050. The CAGR of AI in the agriculture industry
will be 22.68% during 2017–2021.
6 Conclusion
Artificial Intelligence is a program that can adapt itself to execute tasks in real-time
situations using cognitive processing similar to the human mind. AI can be appropriate
and efficacious in the agriculture sector as it optimises resource use and efficiency. It
solves the problem of the resource and labour scarcity to a great extent. Adoption of AI
is quite useful in agriculture. Artificial Intelligence can prove to be a technological
Revolution and a boon to agriculture to feed an increasing human population globally.
It will complement and challenge farmers to make the right decisions.
References
1. Harishankar, S., Kumar, R.S., Sudharsan, K., Vignesh, U., Viveknath, T.: Solar powered
smart irrigation system. Adv. Electr. Comput. Eng. 4, 341–346 (2014)
2. Ling, Y.: Application of artificial intelligence technology in agriculture. Comput. Knowl.
Technol. 202(29), 181–183 (2017)
3. Nagaraju, M., Chawla, P.: Systematic review of deep learning techniques in plant disease
detection. Int. J. Syst. Assur. Eng. Manag. 11, 547–560 (2020)
860 R. Sivarethinamohan et al.
4. Mekala, M.S., Viswanathan, P.: CLAY-MIST: IoT-cloud enabled CMM index for smart
agriculture monitoring system. Measurement 134, 236–244 (2019)
5. Murugesan, R., Sudarsanam, S.K., Malathi, G., Vijayakumar, V., Neelanarayanan, V.,
Venugopal, R., Rekha, D., Sumit, S., Rahul, B., Atishi, M., Malolan, V.: Artificial
Intelligence and Agriculture 5.0 8, 1870–1877 (2019)
6. Mishra, P., Polder, G., Vilfan, N.: Close range spectral imaging for disease detection in
plants using autonomous platforms: a review on recent studies. Curr. Robot. Rep. 1, 43–48
(2020)
7. Liu, S.Y.: Artificial intelligence (AI) in agriculture. In: IT Professional, vol. 22(3), pp. 14–
15, 1 May–June 2020. https://doi.org/10.1109/mitp.2020.2986121
8. Shi, Y.L.: Application of artificial intelligence technology in modern agricultural production.
South. Agric. Mach. 50(14), 73 (2019)
9. Sun, G., et al.: Application of artificial intelligence in intelligent agriculture. J. Jilin Normal
Univ. Eng. Technol. 35, 93–95 (2019)
10. Wang, N., Zhang, N., Wang, M.: Wireless sensors in agriculture and food industry- Recent
development and future perspective. Comput. Electron. Agric. 50, 1–14 (2006)
11. Hashimoto, Y., Murase, H., Morimoto, T., Torii, T.: Intelligent systems for agriculture in
Japan. IEEE Control Syst. Mag. 21(5), 71–85 (2001). https://doi.org/10.1109/37.954520
12. Yang, N., et al.: Tea Diseases Detection Based on Fast Infrared Thermal Image Processing
Technology, (wileyonlinelibrary.com) (2019). https://doi.org/10.1002/jsfa.9564
13. Senkevich, S., et al.: Optimization of the parameters of the elastic damping mechanism in
class 1, 4 tractor transmission for work in the main agricultural operations. In: Computing
and Optimization, Conference Proceedings ICO 2018, Springer, Cham (2018). ISBN 978–3-
030-00978-6
14. Kovalev, A., et al.: Optimization of the process of anaerobic bioconversion of liquid organic
wastes, intelligent computing and optimization. In: Proceedings of the 2nd International
Conference on Intelligent Computing and Optimization (ICO 2019), Springer International
Publishing (2019). ISBN 978-3-030-33585-4
15. Talaviya, T., Shah, D., Patel, N., Yagnik, H., Shah, M.: Implementation of artificial
intelligence in agriculture for optimisation of irrigation and application of pesticides and
herbicides. Artif. Intell. Agric. (2020). https://doi.org/10.1016/j.aiia.2020.04.002
16. https://www.businesswire.com/news/home/20200817005511/en/Artificial-Intelligence-in-
Agriculture-An-Industry-Overview-2020-2025-Featuring-Microsoft-IBM-and-Agribotix-
Among-Other-Major-Players—ResearchAndMarkets.com
17. Jia, L., Wang, J., Liu, Q., Yan, Q.: Application research of artificial intelligence technology
in intelligent agriculture. In: Liu, Q., Liu, X., Shen, T., Qiu, X. (eds) The 10th International
Conference on Computer Engineering and Networks. CENet 2020 Advances in Intelligent
Systems and Computing, vol 1274. Springer, Singapore (2021). https://doi.org/10.1007/978-
981-15-8462-6_25
18. Ahamed, B.B., Yuvaraj, D.: Framework for faction of data in social network using link
based mining process. In: International Conference on Intelligent Computing and
Optimization, pp. 300–309. Springer, Cham, October 2018
Captivating Profitable Applications of AI in Agriculture Management 861
19. Ahamed, B.B., Yuvaraj, D.: Dynamic secure power management system in mobile wireless
sensor network. In: International Conference on Intelligent Computing and Optimization,
pp. 549–558. Springer, Cham, October 2019
20. Yuvaraj, D., Sivaram, M., Ahamed, A.M.U., Nageswari, S.: An efficient lion optimization
based cluster formation and energy management in WSN based IoT. In: International
Conference on Intelligent Computing and Optimization, pp. 591–607. Springer, Cham,
October 2019
Holistic IoT, Deep Learning and
Information Technology
Mosquito Classification Using Convolutional
Neural Network with Data Augmentation
Abstract. Mosquitoes are responsible for the most number of deaths every year
throughout the world. Bangladesh is also a big sufferer of this problem. Dengue,
malaria, chikungunya, zika, yellow fever etc. are caused by dangerous mosquito
bites. The main three types of mosquitoes which are found in Bangladesh are
aedes, anopheles and culex. Their identification is crucial to take the necessary
steps to kill them in an area. Hence, a convolutional neural network
(CNN) model is developed so that the mosquitoes could be classified from their
images. We prepared a local dataset consisting of 442 images, collected from
various sources. An accuracy of 70% has been achieved by running the pro-
posed CNN model on the collected dataset. However, after augmentation of this
dataset which becomes 3,600 images, the accuracy increases to 93%. We also
showed the comparison of some methods with the CNN method which are
VGG-16, Random Forest, XGboost and SVM. Our proposed CNN method
outperforms these methods in terms of the classification accuracy of the mos-
quitoes. Thus, this research forms an example of humanitarian technology,
where data science can be used to support mosquito classification, enabling the
treatment of various mosquito borne diseases.
1 Introduction
Mosquitoes may seem to be tiny little creatures but they are one of the deadliest
animals of the world. They bring significant harm to humans since they are the main
reason behind various transmissible diseases like dengue, malaria, chikungunya, zika,
yellow fever etc. Their ability to carry and spread diseases to humans and animals
causes millions of killings each year. As stated by the World Health Organization,
deaths of millions of people every year are caused by mosquito bites [26]. With almost
2.5% case fatality, almost 500,000 people having severe dengue need to be hospitalized
each year annually [27]. Malaria is responsible for more than three hundred million
acute illnesses and it kills at least one million people every year. There are more than
3,000 species of mosquitoes but the most dangerous ones are aedes, anopheles and
culex because aedes cause dengue, yellow fever, chikungunya etc. Anopheles causes
malaria whereas culex causes zika, west nile virus etc.
Bangladesh, which is a densely populated country, and has a very low average
income, is one of the unhealthiest places to live in. Every year hundreds of people lose
their lives by mosquito bites and thousands get sick. They mainly get affected by
dengue, malaria and chikungunya. But in recent years, the problem has became acute.
In 2019, at least 18 people died of dengue and 16,223 got infected in Bangladesh by
August [3]. The number of malaria cases is also shocking. Approximately 150,000–
250,000 malaria instances are found in this country every year [20].
The goal of this research consists of developing a model that is capable of detecting
three different classes of mosquitoes: Aedes, anopheles and culex from a given input
image. The goal would be accomplished by training a model using machine learning
techniques on our dataset. Then we tried to improve the accuracy compared to other
systems and maintain roughly equal recognition rates for the classes.
The background and some related works on mosquito classification are presented in
the upcoming section. The methodology of this research is demonstrated in Sect. 3.
Then Sect. 4 gives an overview of the dataset constructed which is used in this research
along with data augmentation. Section 5 defines the implementation of the presented
system. After that, Sect. 6 shows the inspection of the result and finally, Sect. 7 gives
the conclusion of the paper by providing a brief description of the future works.
2 Related Work
There has been some work done on vision-based mosquito classification. The most
recent work on mosquito classification was done by Okayasu et al. [25]. They con-
structed their own dataset which consisted of 14,400 images of 3 types of mosquitoes.
They used three types of deep classification methods and showed a comparison of the
accuracy.
Motta et al. [23] presented a classification of adult mosquitoes. They trained CNN
to implement morphological classification of the mosquitoes. They used a dataset
consisting of 4,056 images of the mosquitoes. Using three neural network models:
LeNet, AlexNet and GoogleNet, they found the finest result (76%) with GoogleNet.
In 2018, Li-Pang Huang et al. [12] classified mosquitoes using edge computing
along with deep learning. They managed to have validation accuracy of 98%. On the
other hand, they achieved a test accuracy of 90.5%.
Fuchida et al. [8] showed the pattern and exploratory validation of a self-operating
factor for vision-based classification of mosquitoes. The factor could determine mos-
quitoes from other bugs, with the separation of the dialectal features. It has also used
the classification based on support vector machine (SVM). With a maximum recall of
98% having used different classification methods, they proved the efficiency and
validity of their proposed method. However, the classification of mosquito species was
not considered there.
Mosquito Classification Using Convolutional Neural Network 867
Ortiz et al. [28] presented a work based on convolutional neural network to classify
mosquito larvae. They used 1,118 images of larvae 8th segment. Using a pre-trained
model (VGG16), they achieved an accuracy of 97%.
MAM Fuad et al. [7] classified aedes aegypti larvae and float valve using transfer
learning and implementing Inception-V3. They performed the experiment on 534
images and used three different learning rates achieving an average accuracy of about
99%.
Minakshi et al. [21] classified seven species with sixty mosquitoes by using clas-
sifiers and indicating pixel merits. They implemented the random forest algorithm and
achieved a validation accuracy of 83.3%. But as their dataset was very small having
only sixty images, the classification was not very suitable.
Jeffrey Glick and Katarina Miller [9] classified insects using Hierarchical Deep
CNNs (HD-CNNs). They worked on 217,657 images of different insects. They
implemented their method on 277 distinctive classes.
Several research on mosquito classification using mosquito wingbeat have been
implemented too. Fanioudakis et al. [6] classified six species of mosquitoes using
observable recordings of their wingbeats. They achieved a classification accuracy of
96%. Their dataset consisted of 279,566 recordings of mosquitoes’. They used top-tier
deep learning approaches for implementing the classification.
Kiskin et al. [18] worked on wavelet transformations of audio recordings of
mosquitoes in 2017. They built a CNN model to detect the presence of mosquitoes by
their wingbeat recordings.
Kyukwang Kim et al. [16] built an automatic mosquito sensing and control system
using deep learning. They found an accuracy of 84% for Fully Convolutional Network
(FCN) and regression which was built on neural networks.
The key difference from the existing systems with our system is that they did not
use a custom CNN model for classifying the mosquitoes whereas our system has done
that. We know that CNN models perform better than other approaches in classifying
images as well as retrieval of images [24]. Therefore, custom CNN model has been
used as well as other pre-trained CNN models like VGG-16 to compare their perfor-
mances. As for the limitations are concerned, even though many of them used a dataset
consisting of a large amount of images, the results are not that great in comparison to
the dataset. In contrast, we tried to maintain a fair accuracy rate even after having a
dataset containing a small amount of images.
As there are not so many mosquito images found online, that restricts us to research
more on their classification. There is also no standard dataset for this purpose. So, the
dataset has to be built on our own to determine the species of the mosquito. We made
our own dataset by gathering the images of these mosquitoes and developed the
classification method in this research.
3 Methodology
The system was developed by Convolutional Neural Network(CNN) with data aug-
mentation in this research. Figure 1 illustrates the system flow chart of this system.
First of all, the model takes the images from the dataset. Then it starts preprocessing.
868 M. Akter et al.
After that, the images are augmented by using some augmentation functions. Finally,
the augmented dataset is run into the CNN model so that it can predict the class.
The preprocessing of the images has been done in a simple way. First of all, the
figures of the mosquitoes have been detected using OpenCV. Then, the body portions
of the mosquitoes have been cropped so that the unnecessary parts from the back-
ground are not available in the images. After that, the images are converted into gray-
scale so that the model could learn the images easily. Next, the images have been
normalized to a certain limit to be recognized in a better way. The feature extraction of
the images has been conducted by CNN itself. Images could be detected by using haars
features too [4].
Using a filter w(m, n), convolution over an image f(m, n) is defined in Eq. (1):
Xa Xb
wðm; nÞ f ðm; nÞ ¼ s¼a t¼b
wðs tÞf ðm s; n tÞ ð1Þ
ReLU activation function has been applied in the convolution layer. To make sure the
non-linearity of the model [10], ReLU is applied which is shown in Eq. (2):
The model has been provided with 442 images. It includes 4 convolutional layers:
1st Convolution Layer 2D, 2nd Convolution Layer 2D, 3rd Convolution Layer 2D and
4th Convolution Layer 2D. The 1st convolution layer has 16 filters of size 2 * 2
whereas the 2nd convolution layer has 32 filters. The 3rd convolution layer has 64
layers of size 2 * 2 and finally, the 4th convolution layer has 128 layers. The kernel size
is 2. The model consists of 2 * 2 pool size pooling layer after each convolution layer.
Max pooling layer has also been chosen after each convolution layer. ReLU activation
function has been administered in hidden layers like the convolution layer. After every
hidden layer, there is a dropout layer. The value of dropout was 0.5. The work of the
dropout layer is to deactivate 50% nodes from every hidden layer so that overfitting
[31] could be avoided. One Global Average Pooling layer has been added in the last
hidden layer which takes the average as it is suitable for feeding into our dense output
layer. The hyperparameters have been tuned by adding layers until the error did not
improve. Also, the value of the dropout layer has been chosen by experimenting with
multiple values. As the dropout value of 0.5 has helped to avoid overfitting more than
Mosquito Classification Using Convolutional Neural Network 869
other values, it has been selected as the final value of the dropout layer. The parameters
have been optimized to improve the result.
Finally, the output layer of the model includes three nodes because it has three
classes. As the activation function, SoftMax [33] has been used in the model.
ej
SoftmaxðmÞ ¼ P j ð3Þ
ie
As the model optimizer, we have used Adam [17] in our system and in case of the loss
function, we have used Categorical Crossentropy. ModelCheckpoint is also added as
callback function. The CNN model which has been composed for this study is illus-
trated in Table 1.
and 124 of culex species. Figure 2 displays some sample images of our dataset. We
took help from some sources like [19] to label the data correctly.
5 Implementation Process
The system’s code has been written and developed in a python programming language
using the Spyder IDE [30]. The libraries that we used in this research are Keras [11],
Tensorflow [1], NumPy [36], and Matplotlib [32]. Tensorflow has been selected for the
backend of the system and keras has been used in the system to provide built-in
functions like activation functions, optimizers, layers etc. Data augmentation has been
872 M. Akter et al.
performed by keras API. NumPy is used for numerical analysis. Sklearn is used for
generating confusion matrix, splitting train and test data, modelcheckpoint, callback
function etc. where matplotlib library has been used to make the graphical represen-
tations, such as confusion matrix, loss versus epochs graph, accuracy versus epochs
graph, etc. When an image is provided as an input in our system, it preprocesses the
image. The preprocessing is done in the exact way when the model is being trained.
Then it predicts the class based on the processes.
The intended model brings a strong output for the classification in spite of having a
dataset which does not contain many images. Although there is a small variation in the
classification rate for the three classes, it is quite decent. Our proposed model, the
convolutional neural network (CNN), could be able to gain the validation accuracy of
about 70% without data augmentation.
Figure 5 shows the confusion matrix of the system: with and without data
augmentation.
Here, the x-axis represents the predicted labels and the y-axis represents the true
levels. The validating accuracy of each class of the model (before and after augmen-
tation) is shown in Table 2. Before the augmentation, we can see that the individual
accuracy for anopheles is 70%. For aedes, the accuracy is 72% and finally, the accuracy
found for culex is 68%. But after the augmentation is done, the accuracy for anopheles
becomes 94% and the accuracy for both aedes and culex become 92%. With the help of
data augmentation, our proposed model could achieve an overall accuracy of 93%
which is way better than the accuracy without data augmentation.
Mosquito Classification Using Convolutional Neural Network 873
Figure 6 displays the training accuracy versus validation accuracy of the model before
augmentation. We can see that the training accuracy keeps increasing but the validating
accuracy increases up to 300 epochs. But it keeps decreasing in the last 200 epochs.
Figure 7 displays the training loss versus validation loss of the model before
augmentation. The training loss keeps decreasing until the end but the validation loss
decreases up to 100 epochs but keeps increasing after that.
874 M. Akter et al.
The training accuracy versus validation accuracy of the model after data aug-
mentation is shown in Fig. 8. We can see that both training accuracy and validation
accuracy keep increasing gradually until the end.
Figure 9 displays the training loss vs validation loss of the model after data aug-
mentation. Here both training and validation loss keeps decreasing until the end unlike
the validation loss without augmentation.
Mosquito Classification Using Convolutional Neural Network 875
Table 3 represents the comparison of model evaluation matrices before and after
data augmentation. The validation accuracy was 70%, the training accuracy was 86%,
the precision was 71%, the recall was 70% and finally, the F1-score was 69% without
augmentation. But after the data augmentation, the accuracy becomes 93% and the
training accuracy becomes 97%. Precision, recall and F1-score are 93%, 93% and 92%
respectively.
validation accuracy. On the other side, when it comes to applying the transfer learning
approach, VGG16 crosses the machine learning models in all performance evaluation
matrices as it gives an accuracy of 91%. But our proposed CNN method surpasses all
of them and gives the highest accuracy. In addition to that, an integration of data-driven
(CNN) and knowledge-driven (BRBES) approach can be proposed to portrait risk
assessment of a mosquito bite in the human body [14, 15, 22, 35].
This research focuses on determining the opportunities to improve and build a mos-
quito classifier, which could bring benefits to human beings. Therefore, the goal of the
research was to classify the mosquitoes when too many images of mosquitoes are not
available. The proposed Convolutional Neural Network model can give more efficiency
if it can be run on a good amount of data. The Convolutional Neural Network with data
augmentation should be more efficient and robust while comparing to other machine
learning methods in image processing. We will increase the data for each class to get
close to equal accuracy of each class since it is obvious that CNN is a procedure that is
data-driven. We can apply more augmentation functions to make the augmented dataset
bigger. We can gather a large amount of data so that we could make a standard dataset
for mosquitoes. We will also try to make the system capable of classifying the mos-
quitoes in real-time so that the system can be more efficient. It can be tried to build the
model more effectively to make a better mosquito classification system by using the
convolutional neural network (CNN) in the future.
References
1. Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S.,
Irving, G., Isard, M., et al.: Tensorflow: a system for large-scale machine learning. In: 12th
{USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16),
pp. 265–283 (2016)
2. Ahmed, T.U., Hossain, S., Hossain, M.S., Ul Islam, R., Andersson, K.: Facial expression
recognition using convolutional neural network with data augmentation. In: 2019 Joint 8th
International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd
International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 336–341.
IEEE (2019)
Mosquito Classification Using Convolutional Neural Network 877
3. Akram, A.: Alarming turn of dengue fever in Dhaka city in 2019. Bangladesh J. Infect. Dis.
6(1), 1–2 (2019)
4. Bong, C.W., Xian, P.Y., Thomas, J.: Face recognition and detection using haars features
with template matching algorithm. In: International Conference on Intelligent Computing &
Optimization, pp. 457–468. Springer (2019)
5. Chowdhury, R.R., Hossain, M.S., ul Islam, R., Andersson, K., Hossain, S.: Bangla
handwritten character recognition using convolutional neural network with data augmen-
tation. In: 2019 Joint 8th International Conference on Informatics, Electronics & Vision
(ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition
(icIVPR), pp. 318–323. IEEE (2019)
6. Fanioudakis, E., Geismar, M., Potamitis, I.: Mosquito wingbeat analysis and classification
using deep learning. In: 2018 26th European Signal Processing Conference (EUSIPCO),
pp. 2410–2414. IEEE (2018)
7. Fuad, M.A.M., Ghani, M.R.A., Ghazali, R., Izzuddin, T.A., Sulaima, M.F., Jano, Z.,
Sutikno, T.: Training of convolutional neural network using transfer learning for aedes
aegypti larvae. Telkomnika 16(4) (2018)
8. Fuchida, M., Pathmakumar, T., Mohan, R.E., Tan, N., Nakamura, A.: Vision-based
perception and classification of mosquitoes using support vector machine. Appl. Sci. 7(1), 51
(2017)
9. Glick, J., Miller, K.: Insect classification with heirarchical deep convolutional neural
networks. Convolutional Neural Netw. Vis. Recogn. (CS231N), Stanford Univ. Final
Rep. Team ID 283, 13 (2016)
10. Glorot, X., Bordes, A., Bengio, Y.: Deep sparse rectifier neural networks. In: Proceedings of
the Fourteenth International Conference on Artificial Intelligence and Statistics, pp. 315–323
(2011)
11. Gulli, A., Pal, S.: Deep Learning with Keras. Packt Publishing Ltd, Birmingham (2017)
12. Huang, L.P., Hong, M.H., Luo, C.H., Mahajan, S., Chen, L.J.: A vector mosquitoes
classification system based on edge computing and deep learning. In: 2018 Conference on
Technologies and Applications of Artificial Intelligence (TAAI), pp. 24–27. IEEE (2018)
13. Islam, M.Z., Hossain, M.S., ul Islam, R., Andersson, K.: Static hand gesture recognition
using convolutional neural network with data augmentation. In: 2019 Joint 8th International
Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International
Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 324–329. IEEE (2019)
14. Jamil, M.N., Hossain, M.S., Ul Islam, R., Andersson, K.: A belief rule based expert system
for evaluating technological innovation capability of high-tech firms under uncertainty. In:
2019 Joint 8th International Conference on Informatics, Electronics & Vision (ICIEV) and
2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR),
pp. 330–335. IEEE (2019)
15. Kabir, S., Islam, R.U., Hossain, M.S., Andersson, K.: An integrated approach of belief rule
base and deep learning to predict air pollution. Sensors 20(7), 1956 (2020)
16. Kim, K., Hyun, J., Kim, H., Lim, H., Myung, H.: A deep learning-based automatic mosquito
sensing and control system for urban mosquito habitats. Sensors 19(12), 2785 (2019)
17. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:
1412.6980 (2014)
18. Kiskin, I., Orozco, B.P., Windebank, T., Zilli, D., Sinka, M., Willis, K., Roberts, S.:
Mosquito detection with neural networks: the buzz of deep learning. arXiv preprint arXiv:
1705.05180 (2017)
19. Littig, K., Stojanovich, C.: Mosquitoes: Characteristics of anophelines and culicines (2005).
http://www.cdc.gov/nceh/ehs/docs/pictorial_key/mosquitoes.pdf. Accessed 06 Jan 2017
878 M. Akter et al.
20. Maude, R.J., Hasan, M.U., Hossain, M.A., Sayeed, A.A., Paul, S.K., Rahman, W., Maude,
R.R., Vaid, N., Ghose, A., Amin, R., et al.: Temporal trends in severe malaria in chittagong,
Bangladesh. Malaria J. 11(1), 323 (2012)
21. Minakshi, M., Bharti, P., Chellappan, S.: Identifying mosquito species using smart-phone
cameras. In: 2017 European Conference on Networks and Communications (EuCNC),
pp. 1–6. IEEE (2017)
22. Monrat, A.A., Islam, R.U., Hossain, M.S., Andersson, K.: A belief rule based flood risk
assessment expert system using real time sensor data streaming. In: 2018 IEEE 43rd
Conference on Local Computer Networks Workshops (LCN Workshops), pp. 38–45. IEEE
(2018)
23. Motta, D., Santos, A.Á.B., Winkler, I., Machado, B.A.S., Pereira, D.A.D.I., Cavalcanti, A.
M., Fonseca, E.O.L., Kirchner, F., Badaro, R.: Application of convolutional neural networks
for classification of adult mosquitoes in the field. PLoS ONE 14(1), e0210829 (2019)
24. Nandagopalan, S., Kumar, P.K.: Deep convolutional network based saliency prediction for
retrieval of natural images. In: International Conference on Intelligent Computing &
Optimization, pp. 487–496. Springer (2018)
25. Okayasu, K., Yoshida, K., Fuchida, M., Nakamura, A.: Vision-based classification of
mosquito species: comparison of conventional and deep learning methods. Appl. Sci. 9(18),
3935 (2019)
26. Omodior, O., Luetke, M.C., Nelson, E.J.: Mosquito-borne infectious disease, risk-
perceptions, and personal protective behavior among us international travelers. Prevent.
Med. Rep. 12, 336–342 (2018)
27. Organization, W.H., et al.: Dengue and severe dengue. Technical rep., World Health
Organization. Regional Office for the Eastern Mediterranean (2014)
28. Ortiz, A.S., Miyatake, M.N., Tünnermann, H., Teramoto, T., Shouno, H.: Mosquito larva
classification based on a convolution neural network. In: Proceedings of the International
Conference on Parallel and Distributed Processing Techniques and Applications (PDPTA),
pp. 320–325. The Steering Committee of The World Congress in Computer Science,
Computer … (2018)
29. Perez, L., Wang, J.: The effectiveness of data augmentation in image classification using
deep learning. arXiv preprint arXiv:1712.04621 (2017)
30. Raybaut, P.: Spyder-documentation. Available online at: pythonhosted. org (2009)
31. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a
simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–
1958 (2014)
32. Tosi, S.: Matplotlib for Python Developers. Packt Publishing Ltd, Birmingham (2009)
33. Tüske, Z., Tahir, M.A., Schlüter, R., Ney, H.: Integrating gaussian mixtures into deep neural
networks: softmax layer with hidden variables. In: 2015 IEEE International Conference on
Acoustics, Speech and Signal Processing (ICASSP), pp. 4285–4289. IEEE (2015)
34. Uddin Ahmed, T., Hossain, M.S., Alam, M., Andersson, K., et al.: An integrated cnn-rnn
framework to assess road crack. In: 2019 22nd International Conference on Computer and
Information Technology (ICCIT) (2019)
35. Ul Islam, R., Andersson, K., Hossain, M.S.: A web based belief rule based expert system to
predict flood. In: Proceedings of the 17th International conference on information integration
and web-based applications & services, pp. 1–8 (2015)
36. Walt, S.V.D., Colbert, S.C., Varoquaux, G.: The numpy array: a structure for efficient
numerical computation. Comput. Sci. Eng. 13(2), 22–30 (2011)
Mosquito Classification Using Convolutional Neural Network 879
37. Van der Walt, S., Schönberger, J.L., Nunez-Iglesias, J., Boulogne, F., Warner, J.D., Yager,
N., Gouillart, E., Yu, T.: scikit-image: image processing in python. PeerJ 2, e453 (2014)
38. Royalty free stock photos, illustrations, vector art, and video clips. in: Getty images. https://
www.gettyimages.com/. Accessed 17 Sept 2019
39. Stock images, royalty-free pictures, illustrations videos - istock. in: istockphoto.com. https://
www.istockphoto.com/. Accessed 20 Sept 2019
40. 1 million Stunning Free Images to Use Anywhere – Pixabay. https://pixabay.com/. Accessed
24 Sept 2019
41. Stock images, photos, vectors, video, and music. in: Shutterstock. https://www.shutterstock.
com/. Accessed 24 Sept 2020
Recommendation System for E-commerce
Using Alternating Least Squares
(ALS) on Apache Spark
1 Introduction
types of product in one place [1]. Pine claims that the design of one product is simply
no longer sufficient. Organizations must at least be able to develop multiple goods that
satisfy different customer needs. The e-commerce revolution helps companies to offer
more products to consumers. E-commerce is a process whereby goods, services and
information are purchased and sold or exchanged through computer networks,
including the Internet [2].
Nevertheless, this e-commerce generates vast amounts of information that cus-
tomers need to process before selecting which goods to purchase. The use of the
Recommendation System is the solution to this problem of overloading information.
The recommendation system is used to propose products for its customers through e-
commerce sites. The products may be recommended based on the online retailer, based
on the demographics of the buyer, or based on an analysis of the customers past buying
behavior in order to anticipate future behavior. Such strategies are usually a part of a
website’s personalization, since they allow any customer to follow a platform. Rec-
ommendation system automates web personalization, which enables individual cus-
tomization to this extent is one way of realizing the idea of pines on the web. So Pine is
presumably in agreement with JeffBezos, CEO of Amazon.com; “If I have two million
online customers, I should have two million web stores”.
The Recommendation System employs a range of technologies to filter the best
result and to provide users with information they need. Recommendation System is
divided into three broad categories: first one is collaborative filtering system, next one
is content based system and the last one is hybrid recommendation system [3]. Content-
based systems attempt to check the actions of the item as suggested. This operates by
learning new user’s behavior, based on their knowledge needed, which is provided in
objects by the user. It is a keyword-specific Recommendation System which uses
keywords to describe the items. Therefore, models work in a content-based Recom-
mendation System so that they can suggest comparable user’s products which have
been enjoyed or are currently browsing.
The collaborative filtering system is focused on similarities between customers’
need and the items. The suggested items for a new user are the ones that other similar
people in the past browsing history liked. In collaborative filtering, an item is com-
bined, similarities are recognized based on user ratings and new recommendation is
created based on multi-user comparisons. It faces however many difficulties as well as
limitations, such as data sparsity which is responsible for the evaluation of large items.
It is also difficult to predict on the basis of the nearest neighboring algorithm.
Scalability, when increasing the number of users and the number of items and the
last thing is cold starts, where bad relationships between people with the same mind.
Hybrid Recommendation System carries out their tasks by taking account of the
combination of content-based and collaborative filtering approaches so that a particular
item is suitable. The hybrid system is considered by many organizations to be the most
widely used Recommendation System because it is able to remove any defect that
could have arisen in the implementation of a Recommendation System.
The Main objective of this paper is to develop a recommendation system. Here for
developing a recommendation system a computationally effective algorithm for matrix
factorization ALS (Alternating Least Square) is proposed. ALS is one kind of col-
laborative filtering method is used to solve the overfitting issue in sparse data [4].
882 S. Gosh et al.
Apache Spark is used to implement the ALS algorithm. In this paper ALS algo-
rithm is also compared with other collaborative filtering method which is SGD
(Stochastic Gradient Descent) and SVD (Singular Value Decomposition).
In the first section of this research paper. In the second section of this research
paper, ALS Matrix Factorization Method is explained. In the third section of this
research paper, predictive accuracy metrics used to evaluate the system is defined. In
the following section, system prototype along with system workflow is described with a
diagram. In Sect. 6, system architecture of the recommendation system is illustrated.
Experimental Result is described in Sect. 7 and finally Conclusion and Future work of
the system is discussed.
2 Related Work
Many researchers have worked in the field of e-commerce utilizing techniques of big
data over with machine learning and data mining methods. A number of applications
have developed using those techniques. Recommendation system is used in large areas
such as e-commerce, social networks, education, government etc. The development of
recommendation system using machine learning methods of previous researches is
described in this section.
A Hybrid Recommendation model was designed by Wei et al. [5]. This study
solved the issue of Complete Cold start (CCS) and Incomplete Cold Start (ICS) prob-
lem. Those remaining problem was overcome by using combined feature of Collab-
orative filtering(CF) and deep learning neural network. A Specific Deep learning neural
network (SADE) used for extracting content features of items. CF model and time
SVD++ used to predict the rating by taking content features of cold start items. Netflix
large movie dataset used to perform their analysis. This research found that CCS can
able to do good recommendation than ICS. But their experiment done by investigating
less parameter and evaluated their performance by using RMSE rating prediction which
is not efficient for real recommendation system.
Using collaborative filtering algorithm on the basis of item a recommendation system
was developed by Kupisz and Unold [6]. They used hadoop and apache spark platform to
implement the system. MapReduce paradigm was used in this study to process large
datasets for reducing problem of parallel programming. The limitation of access time and
memory this prototype does not give better outcome for large amount of data.
Chen et al. [7] used the CCAM (co-clustering with augmented matrices) to develop
multiple methods like heuristic scoring, conventional classification and machine learning
to construct a recommendation system as well as the integration of content-based hybrid
recommendation systems in collaboration with collaborative filtering model.
Zhou et al. [8] developed a collaborative filtering based ALS Algorithm for the
Netflix Prize. ALS works to solve scalability issue of extensive datasets. This study
used ALS algorithm to develop a movie recommendation system for predicting rating
of users. Due to not refining the Restricted Boltzmann Machine (RBM) this system
cannot show moderately better result.
The item-based collaborative filtering strategies was used by Dianping et al. [9]. In
this system, the user item rating matrix is first examined and the associations between
Recommendation System for E-commerce 883
different items categorized and use these relationships are used to evaluate the user’s
recommendations.
Dev et al. [10] have proposed a scalable method for developing recommendation
systems based on similarity joins. To work with big data applications, MapReduce
framework was used to design the system. Using a method called extended prefix
filtering (REF), the process can significantly reduce the unnecessary overhead com-
putation such as redundant comparisons in the similarity computing step.
Zeng et al. [11] proposed PLGM to boost the accuracy, scalability and processing
power of large-scale data. Two matrix factorization algorithms, which are ALS and
SGD, were considered in their work. SGD based parallel matrix factorization was
implemented on spark and for its efficiency was compare with ALS and MLlib. Based
on test result, advantage and disadvantage of each model was analyzed. A number of
approaches to profile aggregation were studied and the model was adopted which gave
the best result in terms of efficiency and precision; such as PLGM and LGM have been
studied.
Jooa et al. [12] was implemented a recommendation system by using both col-
laborative filtering and association rules. Distance data from Global positing system
(GPS) was focused to recommend the items to customer for purchasing on the basis of
customer’s preference. The problem of information overloading solved by using Near
Field Communication (NFC). NFC used to select high preference of user by analyzing
time slot with association rules.
A combination of collaborative filtering and cluster techniques based recommen-
dation system was proposed by Panigrahi et al. [13]. Apache spark platform was used
to accelerate the running time of memory computation for performing recommenda-
tion. Spark native language scala used for speeding computational time. This study
focused to reduce the cold start problem which is limitation of traditional collaborative
filtering by correlating the customer to items on the base of features. But limitation of
programming feature and verbose code of scala programming language this model did
not evaluate better prediction on recommendation system.
An algorithm is proposed by Li et al. [14] for paralleling the frequent item set of
large dataset. This proposed algorithm discovered hidden pattern to help query rec-
ommendation. By calculating between cluster nodes this algorithm able to minimize
computation cost. They gathered better performance for distributed processing by
utilizing MapReduce framework concerning cluster computers. This study does not
analyzing the query logs for google search engine.
From the analysis of previous work, the drawback is shown for over fitting issue of
sparse data. Our proposed system overcomes this issue by applying matrix factorization
in ALS algorithm. The use of matrix factorization in our system also increases pre-
diction accuracy from previous researches.
The rating data can be represented as an n m matrix R where n users and m items.
The ðu; iÞth entry is rui in the matrix R which implies that ðu; iÞth ratings for ith item by
u user. The R matrix is sparse matrix as items do not receive ratings from many users.
Therefore, the R matrix has the most missing values.
Matrix factorization is the solution of this sparse matrix problem. There are two k
dimensional vectors which are referred to as “factors”.
1. xu is k dimensional vectors summarizing’s every user u.
2. yi is k dimensional vectors summarizing’s every item i.
Hence, rating for ith item by user u can be predicted by
Let,
xu ¼ x1 ; x2; . . .; xn; 2 Rk
yi ¼ y1 ; y2; . . .; yn; 2 Rk
Algorithm:
Alternating Least Squares (ALS)
Procedure ALS ðxu ; yi Þ
Initialization xu 0
Initialization matrix yi with random values
Repeat
Fix yi ; solve xu by minimizing’s the objective
function (the sum of squared errors)
Fix xu solve yi by minimizing the objective
function similarly
Until reaching the maximum iteration
Return xu ; yi
End procedure
Recommendation System for E-commerce 885
The recommendation system is evaluated using offline analysis. During offline analysis,
there is no actual users rather as large dataset is split into a training and test set. The
recommendation system is trained with the training dataset to predict the ratings given
by users in the test dataset and evaluation is carried out by comparing the resulting
prediction with the actual prediction in the test dataset. Popular predictive accuracy
metrics such as mean absolute error (MAE), root mean squared error (RMSE) and
mean user gain (MUG) are used to measure the accuracy of the prediction made by the
recommendation system.
1 X
MAE ¼ jr ðr Þ pi ðrk Þj
rk 2rui i k
ð3Þ
jrui j
sffiffiffiffiffiffiffiffi
1 X
RMSE ¼ ðjri ðrk Þ pi ðrk ÞjÞ2 ð4Þ
jrui j rk 2rui
ri ðrk Þ hi if pi ðrk Þ hi
UGðpi ðrk ÞÞ ¼ ð5Þ
hi ri ðrk Þ otherwise
1 X
MUG ¼ UG ðpi ðrk ÞÞ ð6Þ
jrui j rk 2rui
5 System Prototype
The proposed system helps online customers with swift selection of their desired
products among bigdata stored in a hadoop cluster. Customers, software bots/agents,
and robotic shopping assistants are performing data transactions in the online shopping
cart and e-commerce site frequently. Therefore, the size of information is growing up to
zattabytes and processing of this requires significant amount of computing resources
and time.
Apache spark runs on hadoop clusters. Apache spark can perform batch processing,
stream processing, graph processing, interactive processing and in-memory processing
of data. Applications built on Apache spark can process data in high speed as spark
stores the Resilient Distributed Dataset (RDD) in the RAM of clusters nodes. RDD is
886 S. Gosh et al.
immutable, logically partitioned, fault-tolerant records of the data as objects which are
computed on different cluster nodes [16–19]. The dataset are created by different types
of file such as json file, CSV file, text file etc. saved in the Hadoop file system. The text
file RDDs are created by sparkcontext’s textfile method.
In the system, the CSV file is loaded externally from HDFS in hadoop clusters. The
file in RDD is logically partitioned which are stored in multiple nodes of the cluster.
The file is partitioned in smaller blocks so that RDD data can be distributed equally
among threads. The full dataset is transformed coarse gained to take an input RDD
producing many output RDDs. The transformation is pipelined to improve the per-
formance. The transformation applies the computation like map(), filter(), reduceBy-
Key(), groupByKey() etc. Spark delivers final result of all intermediate transformations
and write out to the HDFS of hadoop clusters, this is referred to as action. Examples of
actions in spark are first(), take(), reduce(), collect(), count() etc.
The Recommendation system uses RDD based spark MLlib package. The spark
MLlib implements alternating least squares (ALS) algorithm to predict the user’s rat-
ings of different items. The algorithm uses Matrix Factorization technique to predict the
explicit preferences of users to items.
Fig. 1. Alternating least squares matrix factorization method on spark recommendation system
workflow
6 System Architecture
The recommendation system is designed for interactive e-commerce sites. In this case,
real time recommendation and retraining the recommendation system are two most
important requirements. Therefore, dealing with increasing volume of data and change
Recommendation System for E-commerce 887
of data are major challenges. The number of transactions on e-commerce sites varies
with time as new users arrive frequently. Predicting the new users recommendations is
important to keep users attracted and for good feedback.
The Movielens 1M data set has 71567 users ratings to 10681 movies [20]. The
users in this data set have performed rating s for minimum 20 movies. There are
10000054 ratings available in this data set. The Movielens data set is explicit feedback
data set here users give ratings for items.
The recommendation system architecture is depicted in the Fig. 2. The system is
running on Apache Spark 2.4.5, Apache Hadoop 3.2.1 and Aapche HBase. The
training dataset storage and retrieval operations along with Apache HBase data files are
managed by HDFS. The Apache Spark MLlib produces predictions and stores the
recommendations in HBase nosql database. The spark system consists of Master nodes,
cluster manager and worker nodes. The RDDs are created and processed in Worker
nodes and the cluster manager keeps records of RDD locations in the worker nodes.
The users visualize the recommendations in the e-commerce websites browsed by
smart phone clients and personal computer clients connected to the HTTP servers in the
cloud.
The system stores and manages master dataset in a Hadoop distributed file system
(HDFS) so that data can be accessed faster. The recommendations are pre-computed by
processing master dataset in batch using Apache Spark and are stored in HBase. Real
time views are computed on user data streaming by using Apache Spark streaming and
are also stored in HBase. The real time recommendations are calculated by merging the
batch view and real time view. Therefore, following the lambda architecture is a
sustainable option to develop the system.
888 S. Gosh et al.
The project undergoes a task of experimenting the difference of total execution time
required for calculating recommendations using Singular Value Decomposition train-
ing algorithm in non-distributed system and spark system. The experiment is conducted
on a desktop with AMD Sempron(tm) processor (2.8 GHz) and 4 GB RAM. The
experiment shows that the total execution time is reduced in spark system significantly
shown in the Table 1.
In this research, the recommender system is trained using Alternating Least Square
(ALS) algorithm. The Movielens performance metric RMSE value is dependent on
these parameters. The objective of the training is to decrease the value of RMSE and
the ALS model producing least minimum RMSE value is saved for future recom-
mendation. The system is implemented on Apache Spark and Apache Hadoop cluster.
Therefore, the hardware configuration of the cluster’s master node, data nodes, node
manager, resource manager, name node, secondary name node, and history servers
have impact on the performance.
The Fig. 3 shows the process of job submission in Spark. It is noticed that the
maximum time is required for building top recommendations for each user and each
movie. Because, during the job execution the datasets are divided into 20 RDDs
(Resilient Distributed Dataset) for user Factors and item Factors equally 10 each.
The results for these two factors are aggregated to produce recommendations.
Initial parameter for the system is shown in Table 2.
Recommendation System for E-commerce 889
Table 3 shows the top 5 recommendations of items for a particular user along with
the average rating and prediction values. The RMSE value resulted as 0.9 in this
experiment.
As shown in Table 4 the training process includes a RMSE threshold check step
where the lambda value can be changed to optimize the recommendation model. In
Fig. 4 by keeping the number of iteration fixed at 5 and by changing the lambda value
in the range between 0.005 to 1 the minimum RMSE value is found as 0.870 against
the lambda value 0.05. From Table 4 it has been shown that as when the lambda value
is between 0.005 and 0.006 the RMSE value is respectively 0.907 and 0.905. As the
lambda value is increased the value of RMSE is reduced than the previous value. After
that RMSE value is changed as the lambda value is increased and minimum RMSE
value is found when the lambda value is 0.05. The minimum RMSE value is 0.870 at
iteration number 5.
RMSE
1.1
RMSE 1
0.9
0.8
0.7 RMSE
Lamda
RMSE
1
0.95
RMSE
0.9
0.85 RMSE
0.8
ALS SVD++ SGD
Algorithm
Fig. 5. Graphical representation of RMSE comparison among ALS, SVD++ and SGD
Recommendation System for E-commerce 891
Apache spark is well suited for applications which require high speed query of data,
transformation and analytics results. Therefore, the recommendation system developed
in this research is implemented on Apache Spark. Also, the matrix factorization using
Alternating Least Squares (ALS) algorithm which is a type of collaborative filtering is
used to solve overfitting issues in sparse data and increases prediction accuracy. The
overfitting problem arises in the data as the user-item rating matrix is sparse. In this
research a recommendation system for e-commerce using alternating least squares
(ALS) matrix factorization method on Apache Spark MLlib is developed. The research
shows that the RMSE value is significantly reduced using ALS matrix factorization
method and the RMSE is 0.870. Consequently, it is shown that the ALS algorithm is
suitable for training explicit feedback data set where users provide ratings for items.
The proposed recommendation system architecture is designed following lambda
architecture so that the users can see the recommendations real time. This will in turn
bring enormous benefit to the customers during purchasing goods.
In future, the recommendation system will be integrated with content based fil-
tering, users click stream predictive analytics, deep collaborative filtering, and con-
volutional neural network based [21–24] recommendation. The integration might use
results produced by these different systems and correlate these results to show rec-
ommendations to the online users. Thus, the future research direction will be focusing
on hybrid models of recommendation systems [25–32].
References
1. Pine, B.J.: Mass Customization, vol. 17. Harvard Business School Press, Boston (1993)
2. Shahjee, R.: The impact of electronic commerce on business organization. Sch. Res.
J. Interdisc. Stud. 4(27), 3130–3140 (2016)
3. Verma, J.P., Patel, B., Patel, A.: Big data analysis: recommendation system with Hadoop
framework. In: 2015 IEEE International Conference on Computational Intelligence &
Communication Technology, pp. 92–97. IEEE, February 2015
4. Alzogbi, P., Koleva, G.: Lausen, towards distributed multi-model learning on apache spark
for model-based recommender. In: 2019 IEEE 35th International Conference on Data
Engineering Workshops (ICDEW), pp. 193–200. IEEE, April 2019
5. Wei, J., He, J., Chen, K., Zhou, Y., Tang, Z.: Collaborative filtering and deep learning based
hybrid recommendation for cold start problem. In: 2016 IEEE 14th International Conference
on Dependable, Autonomic and Secure Computing, 14th International Conference on
Pervasive Intelligence and Computing, 2nd International Conference on Big Data
Intelligence and Computing and Cyber Science and Technology Congress
(DASC/PiCom/DataCom/CyberSciTech), pp. 874–877. IEEE, August 2016
6. Kupisz, B., Unold, O.: Collaborative filtering recommendation algorithm based on Hadoop
and spark. In: 2015 IEEE International Conference on Industrial Technology (ICIT). IEEE
(2015)
7. Chen, Y.C.: User behavior analysis and commodity recommendation for point-earning apps.
In: 2016 Conference on Technologies and Applications of Artificial Intelligence (TAAI).
IEEE (2016)
892 S. Gosh et al.
8. Zhou, Y.H., Wilkinson, D., Schreiber, R.: Large scale parallel collaborative filtering for the
Netflix prize. In: Proceedings of 4th International Conference on Algorithmic Aspects in
Information and Management, Shanghai, pp. 337–348. Springer (2008)
9. Ponnam, L.T., et al.: Movie recommender system using item based collaborative filtering
technique. In: International Conference on Emerging Trends in Engineering, Technology,
and Science (ICETETS). IEEE (2016)
10. Dev, A.V., Mohan A.: Recommendation system for big data applications based on set
similarity of user preferences. In: International Conference on Next Generation Intelligent
Systems (ICNGIS). IEEE (2026)
11. Zeng, X., et al.: Parallelization of latent group model for group recommendation algorithm.
In: IEEE International Conference on Data Science in Cyberspace (DSC). IEEE (2016)
12. Jooa, J., Bangb, S., Parka, G.: Implementation of a recommendation system using
association rules and collaborative filtering. Proc. Comput. Sci. 91, 944–952 (2016)
13. Panigrahi, S., Lenka, R.K., Stitipragyan, A.: A hybrid distributed collaborative filtering
recommender engine using apache spark. In: ANT/SEIT, pp. 1000–1006, January 1026
14. Koren, Y., Bell, R., Volinsky, C.: Matrix factorization techniques for recommender systems.
Computer 42(8), 30–37 (2009)
15. Meng, X., Bradley, J., Yavuz, B., Sparks, E., Venkataraman, S., Liu, D., Freeman, J., Tsai,
D.B., Amde, M., Owen, S., Xin, D.: Mllib: machine learning in apache spark. J. Mach.
Learn. Res. 17(1), 1235–1241 (2016)
16. Li, H., Wang, Y., Zhang, D., Zhang, M., Chang, E.Y.: Pfp: parallel fp-growth for query
recommendation. In: Proceedings of the 2008 ACM Conference on Recommender Systems,
pp. 107–114, October 2008
17. Shanahan, J.G., Dai, L.: Large scale distributed data science using apache spark. In:
Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery
and Data Mining, pp. 2323–2324, August 2015
18. Sharma, S.: Dynamic hashtag interactions and recommendations: an implementation using
apache spark streaming and graphX. In: Data Management, Analytics and Innovation,
pp. 723–738. Springer, Singapore (2020)
19. Armbrust, M., Xin, R.S., Lian, C., Huai, Y., Liu, D., Bradley, J.K., Meng, X., Kaftan, T.,
Franklin, M.J., Ghodsi, A., Zaharia, M.M.: Spark sql: relational data processing in spark. In:
Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data,
pp. 1383–1394, May 2015
20. Harper, F.M., Konstan, J.A.: The movielens datasets: history and context. Acm. Trans.
Interact. Intell. Syst. (tiis) 5(4), 1–19 (2015)
21. Chowdhury, R.R., Hossain, M.S., ul Islam, R., Andersson, K., Hossain, S.: Bangla
handwritten character recognition using convolutional neural network with data augmen-
tation. In: 2019 Joint 8th International Conference on Informatics, Electronics & Vision
(ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition
(icIVPR), pp. 318–323. IEEE, May 2019
22. Ahmed, T.U., Hossain, M.S., Alam, M.J., Andersson, K.: An integrated CNN-RNN
framework to assess road crack. In: 2019 22nd International Conference on Computer and
Information Technology (ICCIT), pp. 1–6. IEEE, December 2019
23. Ahmed, T.U., Hossain, S., Hossain, M.S., ul Islam, R., Andersson, K.: Facial expression
recognition using convolutional neural network with data augmentation. In: 2019 Joint 8th
International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd
International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 336–341.
IEEE, May 2019
24. Islam, M.Z., Hossain, M.S., ul Islam, R., Andersson, K.: Static hand gesture recognition
using convolutional neural network with data augmentation, May 2019
Recommendation System for E-commerce 893
25. Sinitsyn, S., et al.: The concept of information modeling in interactive intelligence systems.
In: International Conference on Intelligent Computing & Optimization. Springer, Cham
(2019)
26. Alhendawi, K.M., Al-Janabi, A.A.: An intelligent expert system for management
information system failure diagnosis. In: International Conference on Intelligent Computing
& Optimization. Springer, Cham (2018)
27. Biswas, M., Chowdhury, S.U., Nahar, N., Hossain, M.S., Andersson, K.: A belief rule base
expert system for staging non-small cell lung cancer under uncertainty. In: 2019 IEEE
International Conference on Biomedical Engineering, Computer and Information Technol-
ogy for Health (BECITHCON), pp. 47–52. IEEE, November 2019
28. Kabir, S., Islam, R.U., Hossain, M.S., Andersson, K.: An integrated approach of belief rule
base and deep learning to predict air pollution. Sensors 20(7), 1956 (2020)
29. Monrat, A.A., Islam, R.U., Hossain, M.S., Andersson, K.: A belief rule based flood risk
assessment expert system using real time sensor data streaming. In: 2018 IEEE 43rd
Conference on Local Computer Networks Workshops (LCN Workshops), pp. 38–45. IEEE,
October 2018
30. Karim, R., Hossain, M.S., Khalid, M.S., Mustafa, R., Bhuiyan, T.A.: A belief rule-based
expert system to assess bronchiolitis suspicion from signs and symptoms under uncertainty.
In: Proceedings of SAI Intelligent Systems Conference, pp. 331–343. Springer, Cham,
September 2016
31. Hossain, M.S., Monrat, A.A., Hasan, M., Karim, R., Bhuiyan, T. A., Khalid, M.S.: A belief
rule-based expert system to assess mental disorder under uncertainty. In: 2016 5th
International Conference on Informatics, Electronics and Vision (ICIEV), pp. 1089–1094.
IEEE, May 2016
32. Hossain, M.S., Habib, I.B., Andersson, K.: A belief rule based expert system to diagnose
dengue fever under uncertainty. In:2 017 Computing conference, pp. 179–186. IEEE July
2017
An Interactive Computer System with Gesture-
Based Mouse and Keyboard
Abstract. Researchers around the world are now focused on to make our
devices more interactive and trying to make the devices operational with min-
imal physical contact. In this research, we propose an interactive computer
system which can operate without any physical keyboard and mouse. This
system can be beneficial to everyone, especially to the paralyzed people who
face difficulties to operate physical keyboard and mouse. We used computer
vision so that user can type on virtual keyboard using a yellow-colored cap on
his fingertip, and can also navigate to mouse controlling system. Once the user is
in mouse controlling mode, user can perform all the mouse operations only by
showing different number of fingers. We validated both module of our system by
a 52 years old paralyzed person and achieved around 80% accuracy on average.
1 Introduction
The uses of computers have become an integral part of our daily life and the human
computer interactions are becoming more convenient in everyday. While the majority
of the people take these facilities for granted, people with physical impairments face
many difficulties in properly using these devices. In particular, people with severe
movement disabilities may have physical impairments which significantly limit their
ability to control the fine motor. Therefore, they may not be able to type and com-
municate with a normal keyboard and mouse. In this situation, it is important to use
effective assisted technologies to ensure accessibility for such people.
2 Literature Review
There are traditional approaches for virtual keyboard and mouse systems which are
usually based on eye gestures. Our literature review focuses on the research works on
virtual keyboard and virtual mouse which were published in Elsevier, Springer, ACM
Digital Library, IEEE Digital Library etc. We discussed about few related works on
virtual keyboard and virtual mouse in the following two subsections.
In 2019, K. Hassan et al. presented a system to design and develop a hand gesture
based virtual mouse [20]. They captured different gestures via webcam and performed
mouse functions according to the gestures. This system achieved 78%–90% accuracy.
The system does not work efficiently in the complex or rough background.
As we can see from the reviewed literature, previous systems include either virtual
keyboard or virtual mouse. Those systems can’t fully eliminate the need of mouse and
keyboard completely. This work aims to build an interactive computer system which
can be operated without any physical mouse and keyboard.
3 Methodology
Figure 1 shows the overview of the process of interactive keyboard and mouse
controlling system. This work aims at creating a system that recognizes the colors and
hand gestures, and controls computer’s keyboard and mouse according to those ges-
tures using color detection technique.
Our system will use computer’s webcam and will display an onscreen keyboard
layout. Users will be able to type through the keyword using a yellow color cap on his
fingertip. User can also turn on mouse controlling system by pressing Mouse Control
Module button using that yellow color cap. After that, another live video frame will be
shown for tracking the hand movements to recognize mouse functions. Figure 2 rep-
resents the system architecture for virtual communication system.
898 D. Gupta et al.
Figure 3 displays a live demonstration of typing j using fingertip and Fig. 4 shows
how to navigate into mouse controlling system.
c2 ¼ a2 þ b2 2abcosðCÞ ð1Þ
The mask creates some specific region of the image according to certain rules.
Instead we draw contours from the mask. For object tracking, we loop through all the
contours. Convex hull of a set X of points in any space is defined as the smallest
convex set that contains X. Any deviation of the object from this convex hull can be
considered as convexity defect. The convex hull of a finite point set S can be defined as
the set of all convex combinations of its points.
To find the contours in the image, we have used cvFindContours() function of
OpenCV which uses an order finding method to detect edges. We are interested in
extracting the hand contour in the contour extraction process so that shape analysis can
be done to determine hand gestures. The hand contour convexity defects were
900 D. Gupta et al.
– if it detects two fingers, it will move the mouse cursor in the four directions (left,
right, up and down) according to the movement of the fingers, and
– if it detects four fingers and five fingers, then right button click and left button click
actions will be performed, respectively
Figures 5, 6 and 7 demonstrate mouse cursor movement, left button click and right
button click operations, respectively.
– Mouse functions: 6.
– Number of test for each function: 5.
– Total number of test: 30.
– Tested by: A 52 years old stroke patient who has very little control of his left side.
Figure 9 shows the number of times each of the six mouse functions works
accurately.
Figure 10 shows the confusion matrix for operations of virtual keyboard and
mouse. We performed each of our 24 tasks (eight words in the sentence, ten digits and
An Interactive Computer System with Gesture-Based Mouse and Keyboard 903
six mouse functions) five times. Our system successfully recognizes 95 operations out
of 120 operations.
In order to evaluate system performance the accuracy of the system was measured
using the Eq. (2) [23].
DF
Accuracy ¼ 100% ð2Þ
TF
Keyboard and mouse actually form an integral part of the computer system. Our system
architecture can facilitate the use of computer for the paralyzed people. We have
developed a virtual system where people can communicate with the computer without
using any physical keyboard and mouse. This could lead to a new age of Human
Computer Interaction in which physical contact with the computer would not be
necessary at all. The use of object detection and image processing in OpenCV for the
implementation of our work has proved to be practically successful and the task of
keyboard and mouse is achieved with good precision. This system can be beneficial to
certain people who have no control over their limbs.
Most of the applications require additional hardware which are often very expen-
sive. The motive of this work is to create this technology as cheaply as possible and to
create it under a standardized operating system as well. Though, our system can be
used as an alternative for physical keyboard and mouse, it still may perform less
accurately in a low light condition. This is a concern for further research. Moreover, the
work can be extended for a wide variety of environments and can be tested using the
sophisticated existing models [1, 3, 4, 7, 8, 11, 12, 15, 22].
References
1. Abedin, M.Z., Nath, A.C., Dhar, P., Deb, K., Hossain, M.S.: License plate recognition
system based on contour properties and deep learning model. In: 2017 IEEE Region 10
Humanitarian Technology Conference (R10-HTC), pp. 590–593. IEEE (2017)
2. Adajania, Y., Gosalia, J., Kanade, A., Mehta, H., Shekokar, N.: Virtual keyboard using
shadow analysis. In: 2010 3rd International Conference on Emerging Trends in Engineering
and Technology, pp. 163–165. IEEE (2010)
An Interactive Computer System with Gesture-Based Mouse and Keyboard 905
3. Ahmed, T.U., Hossain, S., Hossain, M.S., ul Islam, R., Andersson, K.: Facial ex-pression
recognition using convolutional neural network with data augmentation. In: 2019 Joint 8th
International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd
International Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 336–341.
IEEE (2019)
4. Asad, M.U., Mustafa, R., Hossain, M.S.: An efficient strategy for face clustering use in video
surveillance system. In: 2019 Joint 8th International Conference on Informatics, Electronics
& Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern
Recognition (icIVPR), pp. 12–17. IEEE (2019)
5. Bhuvana, S., Ashwin, E., Boopathi, R., Victor, A.D.: Virtual keyboard interaction with
system based on webcam (2017)
6. Cecotti, H.: A multimodal gaze-controlled virtual keyboard. IEEE Trans. Hum.-Mach. Syst.
46(4), 601–606 (2016)
7. Chowdhury, R.R., Hossain, M.S., ul Islam, R., Andersson, K., Hossain, S.: Bangla
handwritten character recognition using convolutional neural network with data augmen-
tation. In: 2019 Joint 8th International Conference on Informatics, Electronics & Vision
(ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition
(icIVPR), pp. 318–323. IEEE (2019)
8. Gupta, D., Hossain, E., Hossain, M.S., Andersson, K., Hossain, S.: A digital personal
assistant using bangla voice command recognition and face detection. In: 2019 IEEE
International Conference on Robotics, Automation, Artificial-intelligence and Internet-of-
Things (RAAICON), pp. 116–121. IEEE (2019)
9. Haria, A., Subramanian, A., Asokkumar, N., Poddar, S., Nayak, J.S.: Hand gesture
recognition for human computer interaction. Procedia Comput. Sci. 115, 367–374 (2017)
10. Hernanto, S., Suwardi, I.S.: Webcam virtual keyboard. In: Proceedings of the 2011
International Conference on Electrical Engineering and Informatics, pp. 1–5. IEEE (2011)
11. Islam, M.Z., Hossain, M.S., ul Islam, R., Andersson, K.: Static hand gesture recognition
using Convolutional neural network with data augmentation. In: 2019 Joint 8th International
Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International
Conference on Imaging, Vision & Pattern Recognition (icIVPR), pp. 324–329. IEEE (2019)
12. Islam, R.U., Hossain, M.S., Andersson, K.: A novel anomaly detection algorithm for sensor
data under uncertainty. Soft. Comput. 22(5), 1623–1639 (2018)
13. Jagannathan, M., Surya, M., BT, A.M., Poovaraghavan, R.: Finger recognition and gesture
based augmented keyboard (2018)
14. Keil, A., Albuquerque, G., Berger, K., Magnor, M.A.: Real-time gaze tracking with a
consumer-grade video camera (2010)
15. Noor, K., et al.: Performance analysis of a surveillance system to detect and track vehicles
using haar cascaded classifiers and optical flow method. In: 2017 12th IEEE Conference on
Industrial Electronics and Applications (ICIEA), pp. 258–263. IEEE (2017)
16. Patil, I.D., Lambhate, P.: Virtual keyboard interaction using eye gaze and eye blink. Int.
J. Recent Innov. Trends Comput. Commun. (IJRITCC) 3(7), 4849–4852 (2015)
17. Sahu, G., Mittal, S.: Controlling mouse pointer using web cam (2016)
18. Saraswati, V.I., Sigit, R., Harsono, T.: Eye gaze system to operate virtual keyboard. In: 2016
International Electronics Symposium (IES), pp. 175–179. IEEE (2016)
19. Shetty, S., Yadav, S., Upadhyay, R., Bodade, V.: Virtual mouse using colour detection
(2016)
20. Shibly, K.H., Dey, S.K., Islam, M.A., Showrav, S.I.: Design and development of hand
gesture based virtual mouse. In: 2019 1st International Conference on Advances in Science,
Engineering and Robotics Technology (ICASERT), pp. 1–5. IEEE (2019)
906 D. Gupta et al.
21. Shindhe, P.C., Goud, S.: Mouse free cursor control. Bonfring Int. J. Res. Commun. Eng.
(Spec. Issue Recent Advancements Electron. Commun. Eng. Eds. Dr. G.A. Bidkar, Dr.
C. Vijaya and Dr. S.B. Kulkarni) 6, 92–98 (2016)
22. Uddin Ahmed, T., Jamil, M.N., Hossain, M.S., Andersson, K., Hossain, M.S.: An integrated
real-time deep learning and belief rule base intelligent system to assess facial expression
under uncertainty. In: 9th International Conference on Informatics, Electronics & Vision
(ICIEV). IEEE Computer Society (2020)
23. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent computing and optimization. In: Conference
Proceedings on Intelligent Computing and Optimization, vol. 866. Springer, Heidelberg
(2018).https://doi.org/10.1007/978-3-030-00979-3
24. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent computing and optimization. In:
Proceedings of the 2nd International Conference on Intelligent Computing and Optimization,
vol. 1072. Springer, Heidelberg (2019). https://doi.org/10.1007/978-3-030-33585-4
25. Yousaf, M.H., Habib, H.A.: Virtual keyboard: real-time finger joints tracking for keystroke
detection and recognition. Arab. J. Sci. Eng. 39(2), 923–934 (2014)
Surface Water Quality Assessment
and Determination of Drinking Water Quality
Index by Adopting Multi Criteria Decision
Analysis Techniques
1 Introduction
Table 1. Water quality index (WQI) range, status, and possible usage of water sample (Brown
et al. 1972)
WQI Rating of water quality Probable use
value
0–25 Excellent water quality Drinking, irrigation, and industrial purpose
25–50 Good water quality Drinking, irrigation, and industrial purpose
50–75 Poor water quality Irrigation and industrial purpose
75–100 Very poor water quality For irrigation purpose
Above Unsuitable for drinking Proper treatment required for any kind of
100 purpose usage
Surface Water Quality Assessment and Determination 909
1.1 Motivation
Water Quality Index is a very effective way to tell us about the quality of the water.
Several standard methods have been prescribed for computing the WQI value. But
every technique has certain advantages and disadvantages. Moreover, most of the
methods have used the absolute approach in assigning weightage to the water quality
parameters. One of such popular process is WAWQI approach which is widely used
across the world.
Our present investigation has tried to enhance the accuracy of the method men-
tioned above by bringing some crucial modifications in its formula, which is broadly
discussed in Sect. 3. Moreover, the study has emphasized the relative approach in
assigning weightage to various parameters with the help of Multi-Criteria Decision
Analysis (MCDA) techniques. This approach will also tell us about the most significant
parameter in determining water quality based on the different criteria set by users across
the world.
with the conventional WAWQI method to check the new procedure’s effectiveness.
Finally, the study validates the efficiency of the MCDA based WQI approach in
assigning a weightage of priority to the water quality parameter with Sensitivity
Analysis of WQI.
2 Study Area
Jirania is a small town in India’s Tripura state situated on the banks of river Saidra with
longitude and latitude of 23.8132°N, 91.4362°E. Thirteen sampling stations (namely,
S1, S2… S13) are selected within Jirania, where two different water samples (one
sample consisting of tap water and the other comprising of treated drinking water) are
collected and tested from each of the locations. The tap water sample is termed as
‘INLET,’ and the purified water sample collected from the installed water cooler is
termed ‘OUTLET’. Figure 1 shows the Study area of Jirania which is obtained from
Google Earth Pro software.
3 Methodology
Collection of two water samples from each of the sampling stations of Jirania, Tripura.
One sample is the tap water (untreated) and the other sample is purified drinking water.
Performing different water quality tests on these water samples based on seven
parameters
Comparing the results of the tap water and purified water for each sampling station
to notice the change in the water quality
Determining the relative weighted significance value of the seven chosen water
quality parameters with the help of AHP -ANP MCDA Techniques
Comparison of the MCDA based WQI with the absolute conventional WAWQI
method
Performing sensitivity analysis of WQI to validate the most influential water quality
parameter in determining index calculation
The first work in this study was to carry out the collected water samples’ tests to assess
the water quality. Different instruments and procedures are available to analyze water
samples. To test various water quality parameters, multi-parameter water quality
checker instruments like HORIBA are of great use as values of different parameters can
912 D. Deb et al.
be obtained easily, which means it is not time-consuming. HORIBA has been used to
determine parameters like Total TDS, TSS, and pH for our study purpose. The EDTA
Titration process has selected the hardness of various water samples. Total Bacterial
count has been done using the Spread Plate technique. DO of all the collected samples
is measured using a DO meter.
The next aim of this study is to estimate the water quality index of the collected
water samples. Weighted Arithmetic WQI method was selected for this purpose. In
this study, two modifications in this standard formula are being computed:-
(i) The relative weights of the water quality parameters are determined using
MCDA Techniques such as Analytical Hierarchy Process or Analytic Network Process,
which will prioritize the impact of each of the parameters on the water quality.
(ii) The standard formula for Qi does not hold good for a few parameters such as
pH and DO. Generally, for consumption purposes, pH should be in the range of 6.5–8.5
and DO concentration stays typically between 5–18 mg/L. If the pH and DO values go
below these ranges, it is not acceptable for intake in both cases. If we go according to
the standard expression, even for lower costs of pH and DO, the quality rating scale
shows a negative value, which reduces the WQI value. Thus, we should consider only
the absolute value of the quality rating to compute the correct value of WQI, as shown
in Eq. 3. WQI of all the thirteen water samples is found using the traditional WAWQI
approach, and the weighted values are compared with our MCDA based method.
Validation of the efficiency of the new process is done using sensitivity analysis.
Estimation of WQI is done by the application of AHP and ANP methods.
The MCDA techniques are used to assign weightage and prioritize each of the seven
water quality parameters to determine the Drinking Water Quality Index. The AHP
based index is compared to the ANP based index to check the unit weights’ aptness.
Table 2 depicts the list of all the criteria and alternatives that are used in the process.
The first phase in the AHP process is to rank the Alternatives in the context of
specific criteria as per the study’s goal and decision maker’s judgment. The second step
is to determine the Equivalent number of Ranks (R) using the following equation:-
the row-wise average. Normalize the average to find the weight of the importance of
the alternatives.
In ANP, the six steps used in the AHP for computation of weighted significance to
the criteria will be carry forwarded further. Up next, the criteria are compared with each
other concerning each of the alternatives. As a result, each criterion will have another
weightage of significance. In the final step, multiply the weighted relevance of criteria
with the weighted significance of alternatives (which are computed in the sixth step of
AHP) and calculate the row-wise average. Normalize the average to find the updated
relative weights of alternatives.
Relative weighted significance to all the alternatives is computed using AHP and
ANP separately. And the results are compared and analyzed accordingly. If the values
obtained from both the MCDA techniques are almost exact, we will consider that value
as the alternative’s final relative weight. Still, if they show slightly different values,
then ensemble value can be regarded as the ultimate relative weighted significance to
the alternatives.
After computing updated weighted values, Water Quality Index can be found out
by Modified Weighted Arithmetic Water Quality Index Method formula. The
following steps have been worked out to determine the WQI value.
First Step - Relative Weightage value of all the selected water parameters is found
out with MCDA Techniques’ help. The summation of all the weights is 1.
Second Step - Assigning of a quality rating scale for each parameter, as below
ðVa ViÞ
Qi ¼ 100 ð2Þ
ðVs ViÞ
Wi is the relative unit weights obtained from MCDA methods for the ith parameter.
The first and the most crucial stage in computing AHP and ANP is to rank the water
quality parameters (Alternatives) with respect to the selected criteria. These rankings
play a pivotal role in determining the most significant parameter and take further
decisions. For example, considering the criterion ‘Cost,’ the alternative that involves
the least cost to perform the water sample test is ranked 1. The experimental test of the
parameter that requires maximum cost is given rank 7. Performing the Hardness test
914 D. Deb et al.
requires significantly less cost than other parameters, and consequently, it is ranked first
place under the ‘cost’ criterion. Similarly, Dissolved oxygen contributes to the
enhanced taste of drinking water. So, this parameter is rated 1 concerning criterion
‘Taste.’ Presence of bacteria influences the Potability of drinking water to a great
extent. Therefore, TPC is ranked 1 under the criterion’ ‘Potability Table 3.’
Table 3. Ranking the alternatives as per the criteria corresponding to the objective of the study.
Parameter Cost Potability Taste
TDS 4 7 4
TSS 2 3 6
DO 3 2 1
BOD 6 6 5
Hardness 1 4 7
TPC 7 1 3
pH 5 5 2
Table 4 compares the updated relative weights assigned to the alternatives derived
from the Analytic Hierarchy Process and Analytic Network Process, respectively.
Furthermore, the table shows the ensemble weights for each of the water quality
parameters, which are computed by taking the average of the weighted values by the
applied MCDA techniques. Since there is not much difference in the weighted sig-
nificance values of the alternatives by both the MCDA techniques; therefore, the
ensemble weights will be considered as the final updated values of the seven water
quality parameters.
Table 4. Relative weighing factor of the alternatives as per AHP and ANP
Water Weighted significance Weighted significance Ensemble
quality values as per the analytical values as per the analytic weighted
parameters hierarchy process network process significance
values
TDS 0.093 0.106 0.1
TSS 0.171 0.156 0.164
DO 0.207 0.213 0.21
BOD 0.078 0.082 0.08
Hardness 0.159 0.145 0.152
TPC 0.168 0.156 0.162
pH 0.126 0.142 0.134
From ANP and AHP approach, Dissolved Oxygen came out as the most influential
water quality parameter in deciding water quality. It can be seen that TSS and TPC are
also crucial water quality parameters carrying the higher weightage.
Surface Water Quality Assessment and Determination 915
Table 5 shows the test results of the water samples collected from all the sampling
stations. Two water samples are collected from each location. The first one is tap water,
which is represented as ‘INLET.’ The second one is the purified water collected from
the water cooler/water purifier, designated as ‘OUTLET.’ The drinking water quality
standards for the selected seven water quality parameters prescribed by the World
Health Organization (Third Edition) are also shown at the end of the table.
Table 5. Water quality analysis of the samples taken from thirteen sampling stations
TDS (mg/L) TSS (mg/L) Hardness DO (mg/L) BOD (mg/L) TPC pH
(mg/L) (colonies)
Inlet Outlet Inlet Outlet Inlet Outlet Inlet Outlet Inlet Outlet Inlet Outlet Inlet Outlet
S1 32 30 41 8 10 4 5.4 5.6 0.8 0.2 248 95 5.22 6.68
S2 32 29 10 5 6 10 3.9 3.3 1.7 1.3 254 190 5.95 6.36
S3 53 53 39 5 17 22 4.1 4.6 1.1 0.9 320 270 6.53 6.55
S4 58 56 28 14 12 12 3.3 3.8 1.2 0.3 165 50 6.37 6.83
S5 34 45 28 26 7 17 4 6.3 0.4 2.4 152 450 5.43 5.57
S6 94 93 32 24 19 11 5.7 6.9 1.6 1 260 20 5.7 6.02
S7 94 27 32 10 19 10 5.7 6.6 1.6 1.1 260 195 5.7 6.95
S8 94 80 32 3 19 12 5.7 6.5 1.6 1.2 260 65 5.7 6.89
S9 38 37 27 14 12 11 6.4 7 1.2 0.9 224 185 5.81 6.44
S10 38 36 24 9 10 8 7.4 7.4 1.9 0.8 224 77 5.25 6.02
S11 67 31 21 14 20 6 5.2 6.5 1.4 1.2 48 30 5.85 6.91
S12 67 51 21 2 20 16 5.2 6.3 1.4 0.8 48 21 5.85 6.42
S13 67 50 21 8 20 16 5.2 5.9 1.4 0.7 48 38 5.85 7.41
WHO 500 (mg/L) 20 (mg/L) 300 (mg/L) 5 (mg/L) 5 (mg/L) 300 8.5
Limits (colonies)
Table 6 depicts the water quality index value of all the tested water samples with
the help of a Modified Weighted Arithmetic WQI method in which MCDA techniques
have been used. Moreover, the level of water quality corresponding to the WQI value is
also enlisted to get a clear understanding of the quality of drinking water.
Figure 3 represents the comparison of the inlet and the outlet water samples of each
of the sampling stations through a line graph. It can be observed that apart from the
outlet/purified water sample of S5 (WQI value– 66.462), all other outlet samples are
found to be of good quality. Treated water of S8 and S12 is of Excellent quality as per
the WQI method. Except for S5, the WQI value of all the Outlet samples are lower than
their Inlet water samples.
916 D. Deb et al.
90
80
Water Quality Index (WQI)
70
60
50
40
30
20
10
0
S1 S2 S3 S4 S5 S6 S7 S8 S9 S10 S11 S12 S13
INLET 85.47 58.37 80.11 65.91 70.06 76.77 76.77 76.77 66.09 67.47 55.24 55.24 55.24
OUTLET 35.36 47.69 48.19 41.53 82.15 49.38 39.48 28.8 42.84 38.45 34.47 29.22 34.24
Table 7 shows the comparison between the absolute conventional WAQWI method
(Olayiwola et al. 2016) and the modified MCDA based WQI method. The unit weights
for every parameter are derived based on three criteria: Cost, Potability, and Taste. The
table shows that there is no significant difference in the WQI scores of the thirteen
collected water samples in most of the sites. However, the small fluctuations in the
value are due to the MCDA based relative unit weights, which prioritize the parameters
based on the factors mentioned above. Moreover, modification in the Qi formula has
also improved the accuracy of the MCDA based WQI score as the conventional
approach might not always work for water quality parameters such as pH and DO.
Surface Water Quality Assessment and Determination 917
Table 7. Comparison between WQI values derived from both the standard expression and
MCDA based approach
Sampling MCDA based WQI value of WQI value of the treated water samples
Stations the treated water samples computed from conventional Weighted
Arithmetic Water Quality Index approach
(Olayiwola et al. 2016)
S1 35.359 47.729
S2 47.693 47.579
S3 48.19 41.762
S4 41.532 46.154
S5 82.152 47.825
S6 49.376 37.719
S7 39.483 41.029
S8 28.799 38.362
S9 42.835 35.918
S10 38.448 28.089
S11 34.467 43.19
S12 29.222 32.085
S13 34.237 43.243
Table 8 depicts the sensitivity analysis of the MCDA based WQI of the water
samples collected from thirteen sampling sites to determine the critical parameter in
evaluating the water quality index. All seven variables are removed one at a time, and
the corresponding WQI score is computed. These WQI values are compared with the
original WQI value for that sampling site, and the changes in the scores are observed
accordingly. More the difference in the index value, superior is the parameter. From the
table, it can be observed that in 11 out of 13 sites, DO is the most influential parameter
as it can be seen that a significant decrease is happening in the WQI value after
eliminating DO. Thus, this analysis validates the efficiency of the MCDA based WQI,
which has also given maximum weightage to the DO parameter through its procedure.
Figure 4 shows the sensitivity analysis of MCDA based WQI at sampling sites S1,
S5, and S9, respectively. The Bar Graph indicates the dominance of the DO parameter
in the index calculation of water samples at the three locations as the WQI score shows
a significant reduction in the absence of the DO variable compared to the original WQI
value. However, apart from DO, TPC is also one of the critical parameters in the water
quality rating at site S5. Parameters have TDS and Hardness have hardly shown any
difference in the WQI score, making them least significant amongst the concerned
seven variables.
The competence to reduce the composite and varied information into a single value
to express the data in a simplified and consistent form makes the users’ water quality
index valuable. Based on previous literature, various methods have been used to cal-
culate WQI and each one of these methods has some merits and limitations. Lots of
researches are going on to improve this iterative process of evaluating WQI. Studies on
MCDA based WQI have been done earlier, but they have not considered the impact on
cost, Potability, and taste of the water quality parameter. These three parameters used in
the present study were not considered in any of the previous studies. That is why the
Surface Water Quality Assessment and Determination 919
Fig. 4. Sensitivity analysis of MCDA based WQI at sampling sites S1, S5, and S9
current investigation will not only approximate the WQI with AHP MCDA but will
also validate the same comparing with ANP MCDA incorporating the novel criteria
like Cost, Potability, and Taste. This MCDA approach will help improve the accuracy
of estimating WQI and present the most influential parameters in maintaining the water
quality. Thus, the users will focus on keeping the magnitude of that parameter within
the standard value.
6 Conclusion
WAWQI method is applied to determine WQI value. Two modifications in the standard
expression have been made; one uses MCDA techniques in assigning relative
weightage to the parameters to figure out the influence of the concerned parameters in
the quality of drinking water. The second modification uses the absolute term in the
Quality rating formula to get a more accurate and justified WQI value for the concerned
water. The results obtained from MCDA based water quality index show that the tap
water status collected from various sampling locations of Jirania, Agartala varies from
good to low quality. In comparison, most of the treated or purified water samples
collected from the thirteen sampling stations vary from Excellent to Good quality.
Sampling station 12 has the best water quality (Both Inlet and Outlet) than other
buildings, whereas sampling station 5 (Outlet purified water) shows unsatisfactory
results. The treated water samples of sampling stations 2 and 3 also show unsatisfactory
results for some of the water quality parameters. The application of the two MCDA
techniques in the present study has shown Dissolved Oxygen as the most influential
parameter in determining drinking water quality. It is observed that the unit weights
920 D. Deb et al.
obtained from these two methods are almost equal for each of the seven parameters.
Hence, ensemble weights from both the techniques have been considered, and the
calculation of WQI has been worked out accordingly. Sensitivity analysis of MCDA
based WQI is evaluated for validation of the new method, and the results have also
shown DO as the key parameter in the index calculation. While calculating relative
weighted significance with the help of MCDA processes, the order of preference for the
criteria is not standard or fixed. It can be inconsistent for a different section of users.
Furthermore, the selection of criteria may also vary depending upon the requirement of
the consumers. These factors might result in minor alterations in the value of relative
weights.
References
Bouslah, S., Djemili, L., Houichi, L.: Water quality index assessment of Koudiat Medouar
Reservoir, northeast Algeria using weighted arithmetic index method. J. Water Land Dev. 35,
221–228 (2017)
Brown, R.M., McClellan, N.I., Deininger, R.A., Tozer, R.G.: A water quality index—do we
dare? Water Sew Works 117, 339–343 (1972)
Chapelle, F.H., Bradley, P.M., McMahon, P.B., et al.: What does “water quality” mean? Ground
Water 47(6), 752–754 (2009). https://doi.org/10.1111/j.1745-6584.2009.00569.x
Chaurasia, A.K., Pandey, H.K., Tiwari, S.K., et al.: Groundwater quality assessment using water
quality index (WQI) in parts of Varanasi district, Uttar Pradesh, India. J. Geol. Soc. India 92
(1), 76–82 (2018)
Fondriest Environmental, Inc. Dissolved Oxygen Fundamentals of Environmental Measurements
[online] (2013). https://www.fondriest.com/environmental-measurements/parameters/water-
quality/dissolved-oxygen
Horton, R.K.: An index number system for rating water quality. J. Water Pollut. Control Fed. 37
(3), 300–306 (1965)
Intelligent Computing & Optimization, Conference proceedings ICO. Springer, Cham (2018).
ISBN 978-3-030-00978-6
Intelligent Computing and Optimization. In: Proceedings of the 2nd International Conference on
Intelligent Computing and Optimization 2019 (ICO 2019). Springer International Publishing.
ISBN 978-3-030-33585 -4
Karakuş, C.B.: Evaluation of groundwater quality in Sivas province (Turkey) using water quality
index and GIS-based analytic hierarchy process. Int. J. Environ. Health Res. 1–20 (2018). 10.
1080/09603123.2018.1551521
Lkr, A., Singh, M.R., Puro, N.: Assessment of water quality status of Doyang river, Nagaland,
India, using water quality index. Appl Water Sci. 10, 46 (2020)
Mokarram, M., Pourghasemi, H.R., Tiefenbacher, J.P.: Comparison analytic network and
analytical hierarchical process approaches with feature selection algorithm to predict
groundwater quality. Environ. Earth Sci. 78, 625 (2019)
Namugize, J.N., Jewitt, G.P.W.: Sensitivity analysis for water quality monitoring frequency in
the application of a water quality index for the uMngeni River and its tributaries, KwaZulu-
Natal. South Africa Water SA 44(4), 516 (2018)
Nandeesha, Khandagale, V.R., Mahesh, V.B., et al.: Assessment of groundwater quality by
integration of water quality index and GIS techniques. In: Proceedings of the 35th
International Conference of the Polymer Processing, PPS-35 (2020)
Surface Water Quality Assessment and Determination 921
Olayiwola, O., Fasakin, O.: The Use of water quality index method to determine the potability of
surface water and groundwater in the vicinity of a municipal solid waste dumpsite in Nigeria.
Am. J. Eng. Res. (AJER) 5(10), 96–101 (2016)
Sanders, E.R.: Aseptic laboratory techniques: plating methods. J. Vis. Exp. 63, e3064 (2012)
Sandle, T.: Microbiology laboratory techniques. Pharmaceutical Microbiology, Woodhead
Publishing, pp. 63–80 (2016). ISBN 9780081000229
Shah, K.A., Joshi, G.S.: Evaluation of water quality index for River Sabarmati, Gujarat. India.
Appl. Water Sci. 7(3), 1349–1358 (2015). https://doi.org/10.1007/s13201-015-0318-7
Tokatli, C.: Water quality assessment of yazir pond (Tekirdağ, Turkey): an application of water
quality index. Res. J. Biol. Sci. Biyoloji Bilimleri Araştırma Dergisi, E-, 12(1), 26–29 (2019).
ISSN: 1308-0261
Tuzkaya, G., Önüt, S., Tuzkaya, U.R., et al.: An analytic network process approach for locating
undesirable facilities: an example from Istanbul. Turkey. J. Environ. Manag. 88(4), 970–983
(2008)
Vaidya, O.S., Kumar, S.: Analytic hierarchy process: an overview of applications. Euro. J. Oper.
Res. 169(1), 1–29 (2006)
Vijaya Kumar, H., Patil, N.S., Prabhu, N.: Analysis of water quality parameters of groundwater
near Ranebennur industrial area, Haveri district, Karnataka, India. In: Proceedings of the 35th
International Conference of the Polymer Processing Society, PPS-35 (2020)
World Health Organization Guidelines for Drinking-water Quality. Third Edition, vol. 1 (2008).
ISBN 9789241547611
Zhang, Y., Wang, R., Huang, P., et al.: Risk evaluation of large-scale seawater desalination
projects based on an integrated fuzzy comprehensive evaluation and analytic hierarchy
process method. Desalination 478, 114286 (2020)
An Approach for Multi-human Pose
Recognition and Classification Using
Multiclass SVM
Abstract. The aim of this paper is to recognize human activities and classify
their activity using multi class support vector machine (SVM). There are many
ways to predict human activity either vision based or wearable sensor based. In
vision based, different types of sensors that can be used to address this task like a
Kinect sensor. In this paper, a vision-based algorithm is proposed to identify
multi-human activity so that we can predict the gesture Based on their action.
The model is based on segmentation and filtering the environment and identi-
fying human firstly, to identify frame based on semantic structure a dynamic
distance separability algorithm is leading to divide a shot into sub shots for
selecting appropriate key-frames in each sub shot by SVD decomposition. An
adaptive filter is used for filtering process. Then to compose a feature vector,
extraction of key poses is performed, and the meticulous Support vector
Machine performs classification, and activity recognition. The application of this
model is to recognize multi human and estimate their poses for classification.
The future research area of this paperwork will be predicted abnormality of
people’s health for medical purpose.
1 Introduction
The area of human activity recognition from unknown video sets there be limited
intelligent which can robustly and accurately recognize each class of human activities.
The difficulties in recognizing human activities come from several aspects. Firstly, the
feature vector data of human activities are high dimensional and secondly similar
activities by different subjects show large variations. Using depth sensors, like
Microsoft GenICam, Kinect or other similar devices, it is possible to design activity
recognition model exploiting depth maps, which are a better source of information
because they are not affected by environment like temperature and light variations, can
provide body shape, and simplify the problem of human activity detection and seg-
mentation [1] represents a smartphone inertial sensors-based approach for human
activity recognition. Efficient features are first extracted from raw data. The features
include mean, median, auto-regressive coefficients. Extracted from the depth frames the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 922–937, 2021.
https://doi.org/10.1007/978-3-030-68154-8_78
An Approach for Multi-human Pose Recognition 923
skeleton, joints allow impenetrable representation of the human body that can be used
in many applications [2].
As compared to the RGB images, depth image has several advantages for the aspect
of activity recognition. For example, depth images provide appropriate shape infor-
mation, which can be more preferential than the RGB images in many problems such
as classification, gesture detection, and activity recognition. Depth images also provide
additional body shape, and structure information, which have been accurately applied
to recover skeleton, joints from a single depth image-Sect. (3.6) [3]. The feature
extraction method based on light intensity and optical flow works well for the RGB
images. It is therefore, important to design a feature extractions method that is based on
specific characteristics of depth sequences (skeleton joint-reference- [4] equation no. 5,
features selection vector-reference [5], 3D depth structure) [4–6]. This extracted
information involved with the spatial properties such as legs, hands and bodies
movements difference during activity comparison and temporal relationship between
consecutive frames. Also, depth sensors are cost effective and feasible to use [7].
2 Related Work
In the last few years, many models have been proposed to detect human activities,
Devanne et al. [8] Proposed representing human actions by spatio-temporal motion
trajectories in a 60-dimensional space, they considered 20 joints, each of them with 3
coordinates. Then, an elastic metric which is a metric invariant to time and speed of the
action sequence, within a Riemannian shape space, is employed to represent the dis-
tance between two curves. The action recognition can be seen as a classification in the
Riemannian space, using a k-Nearest-Neighbor algorithm. Taha et al. [9] Exploit joints
spherical coordinates to represent the skeleton and a framework composed of a multi
class SVM and a discrete HMM to recognize activities constituted by many actions for
suspicious activity detection.
Ding et al. [6] develop a Spatio-temporal Feature, Chain (STFC) to represent the
human actions by the trajectories of joint action. Before using the STFC model, a graph
is used to wipe periodic sequences, making the solution more accurate to noise and
periodic sequence misalignment. Lorenzo s. [10] present a method to recognize human
actions based on the positions of joints. First, the body skeleton is decomposed in a set
of kinematic chains, and the position of each joint is expressed in a locally defined
reference system which make the coordinates invariant to body translations and rota-
tions. Wondering is a wearable sensor developed to detect and synthesis user’s
activities. The sensor used for old people to monitor their activities on daily bases, and
send it to their doctor or family member to know information about their health status
to help them to live safely. The system challenges to solve the issues of recognizing the
activity of the user, such as running, eating, walking, etc. Also, gaining useful infor-
mation from monitoring user activities and deliver these data to the healthcare Body.
User’s activities are stored in databases after that, the system filter these activities to get
useful data [11].
924 S. Md. R. H. Raj et al.
3 Model
The proposed algorithm for human activity prediction starts from a frame sequence
which are extracted from a live data set. These frames are then filtered through linear
filter, and then object selection is performed to determine a particular person which
activity we want to recognize after that compute skeleton joints and produce a vector of
features for each activity. A multi class machine learning (ML) algorithm, where each
class represents a different activity, is used for classification purpose.
3.1 Steps
Seven main steps constitute the whole algorithm which is discussed in the following: 1.
Frame segmentation. For segmentation, we take a live video as input and consider the
video frame based and determine the particular object like a human for activity
recognition. 2. Filtering. At the tracking stage, we find a correspondence between
hypothesized frame regions at current and previous frames, and find the true object
region among all the hypotheses for current frame. 3. Posture Features Extraction. To
evaluate the feature vectors which represent human postures the coordinates of the
skeletal joints are used.4. Postures Selection. Important postures for each activity
selected. 5. Activity Features Computation. A feature vector which represents the
whole activity is created and are used for classification. 6 Classification. 7. Activity
recognition. The classification is realized using a multi-class SVM.
The proposed model exploits spatial features computed from skeleton coordinates,
without including the time information features in the computation, to make the system
independent of the speed of movement. The feature extraction method has been
introduced in [11] and it is adopted with small differences. For each skeleton frame, a
posture feature vector is computed initially. Each joint is represented by Oi a three-
dimensional vector in the coordinate space of Kinect. The person may find at any place
within the coverage of Kinect and the coordinates of the same joint may assume
different values. It is necessary to recoup this effect, by using feature computation
algorithm. A straightforward method is to compensate to the position of the skeleton by
centering the coordinate area in one skeleton joint. Considering a skeleton composed of
J joints, O0 being the coordinates of the torso joint, and O2 being the coordinates of the
neck joint and the ith joint feature, di is as the distance vector between Oi and O0,
normalized to the distance between O2 and J0:
Oi oo
Di ¼ i ¼ 1; 2; . . .. . .J 1 ð1Þ
jjo2 Oojj
This feature vector is invariant to the position of the skeleton within the range of
Kinect; yet the invariance of the feature for building of the person is obtained by
considering the distance between the neck and torso joints. These features may be seen
as a set of distance matrix which connect each joint to the joint of the torso. A posture
feature vector g is created for each skeleton frame:
g ¼ ½ D1 ; D2 ; D3 ; . . .. . .. . .. . .. . .::DJ1 ð2Þ
P
k P
arg min ¼ jjgi bjjj
j¼1 gi2Aj ð3Þ
A
The N centroids are as the main postures, which are the most important feature
vectors. Where classical approaches based on key poses, proposed algorithm select the
most informative postures by considering all the sequences of each activity, in the
proposed algorithm the clustering is executed for each sequence. This will help to
avoids the application of a learning algorithm required to associate each frame to the
926 S. Md. R. H. Raj et al.
closest key pose and allows to have a more accurate representation of the activity.
The computation of a feature vector which models the whole activity is performed in
the third phase, starting from the N centroid vectors computed by the clustering
algorithm. In a word [b1, b2, b3,…, bK] vectors are sorted on the basis of the order in
which the cluster’s elements occur during the activity. The activity features vector is
definite of concatenating the sorted centroid vectors. For example considering an
activity featured by F = 12 and N = 5, after running the k-means algorithm, one of the
possible outputs could be the following sequence of cluster IDs: [2, 2, 2, 3, 3, 1, 1, 4, 4,
4, 5, 5], that means the first three posture vectors belong to cluster 2, the fourth and the
fifth are associated with cluster 3 and the sixth and seventh are in cluster 5 and so on. In
this case, the activity features vector is S = [B2,B3,B1,B4,B5]. A feature activity vector
has a dimension of 4N (J − 1) that can be handled without using dimensionality
reduction algorithms if K is small. The classification step target to associate each
feature activity vector to the correct activity. There are a number of machine learning
algorithms may be applied to fulfil this task, among them SVM is one. Considering a
number of d training vectors xi 2 Tn and a vector of labels y 2 Td, where yi 2 {−1,1},
a binary SVM can be formulated as follows (Fig. 1):
Pd
min ¼ 12 WT W þ C i1 ni
w; b; n
Is the optimal hyperplane that allows separation between groups in the feature
vector, C is a constant, and ni are nonnegative variables which are considered as
training error. The function / allows transforming between the features vector and a
reasonable higher dimensional space where the data are separable. Now we Consider
two training vectors xi and xj, the kernel function can be defined as
jjX1 X2
kðX1:X2Þ , expð Þ; r[ 0 ð8Þ
r2
An Approach for Multi-human Pose Recognition 927
Fig. 1. Activity segmentation and recognition (ASAR) model. The general scheme of the
activity recognition algorithm is composed of 7 steps. First segmentation then filtering then
object specification and then the posture feature vectors are computed for each skeleton frame
then the postures are selected and an activity features vector is created. Finally, a multiclass SVM
is exploited for activity classification.
It follows that C and r are the parameters that have to be estimated prior using the
SVM.
The idea we exploited is to use a multi class SVM, where each group or class
represents an activity of the dataset. Some approaches have been proposed in the
literature, to extend the role from a binary to a multi class classifier. In [2], the authors
compared many methods and found that the most suitable for practical use is the one
against- one method. It is implemented in LIBSVM [7] and it is approach used in this
work for classification. The one-against-one approach is based on the construction of
several binary SVM classifiers; in a broad sense, a number of N (N − 1)/2 binary
SVMs are necessary in an N-classes dataset. This happens because each SVM is trained
to distinguish between 2 classes, and the final decision is taken exploiting a voting
method among all the binary classifiers. During the training phase, the activity feature
vectors are given as inputs to the multi class Support Vector Machine (SVM), together
928 S. Md. R. H. Raj et al.
Fig. 2. Subsets of joints considered in the execution of proposed algorithm. The whole skeleton
is represented by 20 joints for representing a person, the selected ones are depicted as green
circles, while the unselected joints are represented by red squares. Subsets of (a) = 7, (b) = 11,
and (c) = 15 joints, and the whole set of (d) = 20 joints are selected.
with the label M of the action. In the test phase, the activity label is obtained from the
SVM classifier.
Experimental Results
The algorithm performance is evaluated on three publicly available data sets. The
reference test procedures have been considered for each dataset to perform an objective
comparison to previous works. The performance indicators are evaluated using four
different subsets of joints, shown in Fig. 3, started from a minimum of 7 up to a
maximum of 20 joints. Finally, to evaluate the performance in activity recognition
scenarios, some suitable activities related to HA are selected from the datasets, and
recognition accuracies are evaluated. After training, the test procedure can recognize
multi-human activity real-time.
Performance Analysis
The proposed model has been tested over the datasets detailed above, to ensure a fair
comparison to previous works which used the same data sets, following the recom-
mendations provided in each reference paper. Following this comparison, a more
activity recognition-oriented evaluation has been conducted, which consists of con-
sidering only suitable actions for each dataset, excluding the gestures. A subset of the
skeleton, joints that have to be included in the feature computation vector.
KARD Dataset
Gaglio et al. [11] collected the KARD dataset and proposed few evaluation experi-
ments on it. They considered two modalities of dataset splitting and three different
experiments. The experiments are as follows: (i) Experiment A: one-third of the data
set is considered for training, and the rest for testing.
An Approach for Multi-human Pose Recognition 929
(ii) Experiment B: two-thirds of the data set are considered for training, and the rest
for testing.
(iii) Experiment C: half of the data set is considered for training, and the rest for
testing. The activities constituting the data set are split in the following classes:
(i) Gestures and Actions.
(ii) Activity Set 1, Activity Set 2, and Activity Set 3, as listed in Table 1. Activity
Set 1, is the simplest one since it is composed of quite different activities while the
other two sets include more similar gestures and actions.
930 S. Md. R. H. Raj et al.
Table 1. Activity sets grouping different and similar activities from KARD dataset.
Activity set 1 Activity set 2 Activity set 3
Horizontal arm wave High arm wave Draw tick
Bend Catch cap Sit down
Two-hand wave side kick Drink
Forward kick Forward kick Toss paper
Phone call Draw tick Phone call
Stand up Hand clap Take umbrella
Walk Sit down Horizontal arm wave
Draw X Bend High throw
Each experiment has been repeated minimum ten times, randomly splitting training
and testing data. Finally, the “test-person” continuity is also performed that is a leave-
one-actor-out setting, consisting of training the model on nine of the ten the people of
the dataset, and testing on the tenth. In the “test-person” test, no hinds provided about
how to split the dataset, so, we assumed that whole data set of 18 activities is con-
sidered for activity classification. The only parameter which is set in the proposed
algorithm is the number of clusters N, and different subsets of skeleton joints are
considered. Considering J = 7, the maximum accuracy (98.2%) is shown in Fig. 5, and
it is obtained by N = 40. The minimum accuracy is 91.7% and it is obtained by N = 5.
The difference is quite high (6.6%) but this gap is reduced by considering more training
data. In fact, Experiment B shows a difference of 0.4% in Activity Set 1, 1.0% in
Activity Set 2, and 2.5% in Activity Set 3. Considering the number of selected joints,
the observation of Tables 2 and 3 gives us indication that not all the joints are nec-
essary to achieve good recognition accuracy. From Table 2, it can be noticed that the
Activity Set 1 and Activity Set 2, which are the simplest ones, has a good recognition
accuracy using a subset composed of J = 7 joints. The Activity Set 3, composed of
more similar activities, is better recognized with J = 11 joints. In this case it is not
necessary to consider all the skeleton, joint provided from KARD dataset for greater
accuracy. The results obtained with the “test-person” scenario is shown in Fig. 4. The
best result is obtained with N = 35.
Table 2. Accuracy (%) of the proposed algorithm compared to the other algorithm using KARD
dataset with different Activity Sets.
Activity Set 1 Activity Set 2 Activity Set 3
A B C A B C A B C
Gaglio et al. [27] 95.1 99.1 93.0 89.9 94.9 90.1 84.2 89.5 81.7
Enea Cippitelli [29] 98.2 98.4 98.1 99.8 100 99.7 91.6 95.8 93.3
Proposed (J = 7) 98.3 98.4 98.2 99.7 100 99.8 90.2 95.0 91.3
Proposed (J = 11) 98.0 99.0 97.7 99.9 100 99.6 91.8 96.8 95.3
Proposed (J = 15) 97.5 98.8 97.6 99.5 100 99.6 91.7 95.1 94.2
An Approach for Multi-human Pose Recognition 931
Fig. 4. Confusion matrix of the “test-person” test on the whole KARD dataset.
Using a number of J = 11 joints. The overall precision and recall are about 5%
higher than the previous work which uses the KARD dataset. In this situation, the
confusion matrix obtained is shown in Fig. 4. It can be observed that the actions are
distinguished very well, only toss paper and draw tick and drink show a recognition
accuracy equal or lower than 90%. The most critical activities are the draw X and draw
tick gestures, which are quite higher (Table 4).
From the point of view, only some activities are pertinent. Table 3 shows that the
eight actions which made the KARD dataset are recognized with an accuracy greater
than 99%, even if there are some similar actions such as stand up and sit down, or drink
and phone call. Considering only the Actions subset, the lower recognition accuracy is
95.1% and it is obtained with N = 5 and J = 7. It means that the algorithm is able to
reach a high recognition rate even if the feature vector is limited to 90 or less elements.
Table 3. Accuracy (%) of the proposed algorithm compared to the other using KARD dataset
where the dataset is splitted in Gestures and Actions for different experiment on the data.
Gestures Actions
A B C A B C
Gaglio et al. [29] 86.5 93.0 86.7 92.5 95.0 90.1
Enea Cippitelli [29] 89.9 95.9 93.7 99.1 99.6 99.4
Proposed (J = 7) 91.9 93.5 92.5 99.2 99.6 99.5
Proposed (J = 11) 92.9 96.9 94.7 99.0 99.9 99.1
Proposed (J = 15) 87.4 93.6 92.8 98.7 99.5 99.3
932 S. Md. R. H. Raj et al.
Table 4. Precision (%) and recall (%) of the proposed algorithm compared to the other
algorithms, using the whole KARD dataset and “test-person” activities.
Algorithm Precision Recall
Gaglio et al. [29] 84.8 84.5
Enea Cippitelli [29] 95.1 95.0
Proposed (J = 7) 95.3 95.7
Proposed (J = 11) 97.1 97.0
Proposed (J = 15) 96.0 95.8
separately and then merge the results at the time of computing average precision and
recall. In our approach simply considers two copies of the same action given as input to
the multiclass SVM and retrieves the classification results activities. The reduced
number of joints does not affect too much the average performance of the algorithm
that reaches a precision of 93.7%, and a recall of 92.5%, with J = 7. Using all the
available joints, on the other hand, brings to a more substantial reduction of the per-
formance, showing 88.9% and 87.7% for precision and recall, respectively, with
J = 15. The best results for the proposed algorithm were always obtained with a high
number of clusters (30 or 35). The reduction of this number affects the performance; for
example, considering the J = 11 subset, the worst performance is obtained with N = 5
and with a precision of 87.6% and a recall of 87.0%. From the ASAR point of view,
this dataset is composed only of actions, so the dataset does not have to be separated to
evaluate the performance in a scenario which is close to ASAR.
4.2.3. UTKinect Dataset. The UTKinect dataset is composed of 10 activities,
performed twice by 10 participants, and the evaluation method proposed in [16] is the
leave-one-outcross- validation (LOOCV), which means that the system is trained on all
the sequences of action except one and that one is used for testing. To reduce the
random effect of k-means each training or testing procedure is repeated 20 times. Since
the skeleton is captured using Microsoft SDK which provides 20 joints, for this
dataset all the different subsets of joints shown in Fig. 2 are considered. The considered
sequence of clusters is only N = [3, 4, 5], because the minimum number of frames
constituting an action sequence is 5. The results obtained compared with previous
works are shown in Table 7.
The best results for the proposed algorithm are obtained with J = 7, which is the
smallest set of joints considered. The result corresponds to a number of N = 4 clusters
but the difference with the other clusters is very low only 0.6% with N = 5 that
provided the worst result. The selection of different sets of joints from J = 7 to J = 15,
changes the accuracy only by a 2% from 93.1% to 95.1%. In this dataset, the main
drawback to the performance is given by the reduced number of frames that constitute
some action sequences and limits the number of clusters representing the actions
sequence (Table 8).
Vemulapalli et al. [32] reach the highest accuracy but the approach is much more
complex, the processing scheme includes Dynamic Time Warping after modeling
skeleton joints in a Lie group to perform temporal alignments and a specific modeling
called Fourier Temporal Pyramid before classification with one-versus-all multiclass
SVM. The UTKinect dataset to ASAR involves the consideration of a subset of
activities which includes only actions and discards gestures. Farther details, the fol-
lowing 5 activities have been selected: walk, sit down, pick up, carry and stand up. In
this condition, the highest accuracy (97.7%) is still given by J = 7, with 3 clusters, and
the lower one (95.1%) is represented by J = 15, again with 3 clusters. The confusion
matrix for these two configurations are shown in Fig. 5, where the main difference is
the reduced misclassification between the carry and activities walk that are very similar
to each other.
934 S. Md. R. H. Raj et al.
Table 5. Precision (%) and recall (%) of the proposed algorithm, in the different environments
of CAD-60, with J = 11 and N = 20.
Location Activity Test-person
Precision Recall
Living room Talking on phone 89.5 88.5
Talking on couch 93.9 100
Drinking water 89.5 88.5
Relaxing on couch 100 89.5
Average 93.23 91.6
Bathroom Brushing teeth 90.9 100
Wearing contact lens 100 85.2
Rinsing mouth 94.3 100
Average 95.06 95.01
Kitchen Cooking-chopping 89.7 100
Opening pill container 100 100
Cooking-stirring 100 82.1
Drinking water 97.0 100
Average 96.7 95.2
Bedroom Talking on phone 95.7 93.7
Opening pill container 98.0 100
Average 96.0 95.1
Office Talking on phone 100 89.5
Working on computer 100 100
Writing on whiteboard 100 96.8
Drinking water 89.7 100
Average 97.4 96.6
Global average 96.29 95.5
Table 6: Global precision (%) and recall (%) of the proposed algorithm for CAD-60 dataset and
“new-person” setting, with different subsets of joints, compared to other works.
Algorithm Precision Recall
Sung et al. [43] 67.9 55.5
Gaglio et al. [29] 77.3 76.7
Enea Cippitelli [29] 92.2 89.4
Proposed (P = 15) 91.9 89.7
Faria et al. [47] 91.1 91.9
Parisi et al. [48] 91.9 90.2
Proposed (P = 7) 96.7 93.5
Shan and Akella [46] 93.8 94.5
Proposed (P = 11) 95.9 94.5
An Approach for Multi-human Pose Recognition 935
Fig. 5. Confusion matrices of the UTKinect dataset with only ASAR related activities. (a) Best
accuracy confusion matrix, obtained with J = 7. (b) Worst accuracy confusion matrix, obtained
with J = 15.
Table 7. Global accuracy (%) of the proposed algorithm for UTKinect dataset and LOOCV
setting, with different subsets of joints, compared to other works.
Algorithm Accuracy
Jiang et al. [35] 91.9
Xia et al. [26] 90.9
Theodorakopoulos et al. [30] 90.95
Gan and Chen [24] 92.0
Ding et al. [31] 91.5
Zhu et al. [17] 91.9
Liu et al. [22] 96.0
Proposed (J = 15) 95.1
Proposed (J = 11) 97.2
Proposed (J = 20) 97.3
Proposed (J = 7) 97.1
Vemulapalli et al. [33] 97.1
Table 8. Global accuracy (%) of the proposed algorithm for Florence3D dataset and “test-
person” setting with different subsets of joints compared to other methods.
Algorithm Accuracy
Vemulapalli et al. [33] 90.9
Seidenari et al. [44] 82.0
Proposed (J = 7) 93.1
Proposed (J = 15) 92.7
Proposed (J = 11) 89.1
Anirudh et al. [34] 89.7
Taha et al. [28] 96.2
936 S. Md. R. H. Raj et al.
4 Discussion
Limiting the analysis to PSAR related activities only, the algorithm achieves surprising
results in all the datasets. In the considered method, the group of 8 activities defined as
Actions in the KARD dataset is recognized with an accuracy greater than 98.7%. The
CAD-60 dataset contains only actions for this all the activities, is considered within the
proper location and has been included in the evaluation, which provides precision and
recall of 94.9% and 95.5%, respectively. In the UTKinect dataset, the performance
improves considering only the PSAR activities. To improve recognition accuracy, not
only joints relative positions, but also, their velocity should be considered, for more
complex features. First, the proposed algorithm exploits a multi class SVM, which is a
good classifier, but it does not easy to make things understandable if an activity belongs
or not to any of the training classes. In a real scenario, it is possible to have a sequence
of frames for a particular action that does not represent any activity of the training set,
in this case, the SVM outputs the most likely class. The main problem is the seg-
mentation of actions. In many Data sets, the actions are represented as segmented
sequences of frames, but in the real cases, the algorithm has to handle a continuous
stream of frames and to segment actions by itself. Though some solutions for seg-
mentation has been proposed, most of them are based on thresholds on movements,
which can be highly data-dependent [11]. Also, this aspect has to be further investi-
gated, to have a system which is effectively applicable in a real PSAR scenario.
The main contribution of this work is classified different activity more accurately
through the multi-class SVM process and making model to identify and predict dif-
ferent activities of multiple people. For better accuracy Gaussian distribution is applied.
In this work, the activity recognition algorithm is based on skeleton data extracted from
an RGBD image and after segmentation it creates a feature vector representing the
whole activity. It is able to overcome state-of-the-art results in two, available data sets,
the KARD and CAD-60, outperforming more complex algorithms in many conditions
and case. The algorithm has been tested also, over other more challenging datasets, the
UTKinect, where it is outperformed only by the algorithms exploiting temporal
alignment techniques, or a combination of several machine learning methods. Future
works will concern the application of the activity recognition algorithm human health
abnormality prediction, considering the estimation of pose-based abnormality
classification.
An Approach for Multi-human Pose Recognition 937
References
1. Uddin, M.Md.Z., Almogren, M.A.: A robust human activity recognition system using
smartphone sensors and deep learning (2018)
2. Faria, D.R., Premebida, C., Nunes, U.: A probabilistic approach for human everyday
activities recognition using body motion from RGBD images. In: Proceedings of the 23rd
IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN
2014), Edinburgh, UK, pp. 732–737 (2014)
3. Parisi, G.I., Weber, C., Wermter, S.: Self-organizing neural integration of posemotion
features for human action recognition. Front. Neurorob. 9, 3 (2015)
4. Gasparrini, S., Cippitelli, E., Gambi, E., Spinsante, S., Fl´orez- Revuelta, F.: Performance
analysis of self-organising neural networks tracking algorithms for intake monitoring using
Kinect. In: Proceedings of the 1st IET International Conference on Technologies for Active
and Assisted Living (TechAAL 2015), Kingston, UK (2015)
5. Sung, J., Ponce, C., Selman, B., Saxena, A.: Unstructured human activity detection from
RGBD images. In: Proceedings of the IEEE International Conference on Robotics and
Automation (ICRA 2012), Saint Paul, Minn, USA, pp. 842–849 (2012)
6. Ding, W., Liu, K., Cheng, F., Zhang, J.: STFC: spatio-temporal feature chain for skeleton-
based human action recognition. J. Vis. Commun. Image Representation 26, 329–337 (2015)
7. Sung, J., Ponce, C., Selman, B.: Unstructured human activity detection from RGBD images.
In: Proceedings of the IEEE International Conference on Robotics and Automation (ICRA
2012), Saint Paul, Minn, USA, pp. 842–849 (2012)
8. Devanne, M., Wannous, H., Berretti, S., Pala, P., Daoudi, M., Del Bimbo, A.: Spacetime
pose representation for 3D human action recognition. In: New Trends in Image Analysis and
Processing—ICIAP 2013. Lecture Notes in Computer Science, vol. 8158, pp. 456–464.
Springer, Berlin (2013)
9. Taha, A., Zayed, H.H., Khalifa, M.E., El-Horbaty, E.-S.M.: Human activity recognition for
surveillance applications. In: Proceedings of the 7th International Conference on Information
Technology, pp. 577–586 (2015)
10. Seidenari, L., Varano, V., Berretti, S., Del Bimbo, A., Pala, P.: Recognizing actions from
depth cameras as weakly aligned multi-part Bag-of-Poses. In: Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition Workshops (CVPRW 2013),
Portland, Ore, USA, pp. 479–485 (2013)
11. Gaglio, S., Lo Re, G., Morana, M.: Human activity recognition process using 3-D posture
data. IEEE Trans. Hum.-Mach. Syst. 45(5), 586–597 (2015)
Privacy Violation Issues in Re-publication
of Modification Datasets
1 Introduction
Data utility and data privacy are both serious issues that must be considered when
datasets are released for public uses. To address these issues in released datasets, there
are various anonymization models to be proposed such as k-Anonymity [1], l-Diversity
[2], t-Closeness [3], and k-Likeness [4]. For privacy preservation with these
anonymization models, before datasets are released for public uses, all explicit iden-
tifier values of users are removed. Moreover, the unique quasi-identifier values of users
are suppressed or generalized by using their less specific values to be indistinguishable.
Unfortunately, with the best of our knowledge about these anonymization models, they
can be adequate to address privacy violation issues in static datasets. Thus, if the
specific dataset for public uses is dynamic and released on performing multiple times,
these anonymization models could be inadequate, i.e., the privacy data of users is
available in released datasets, it could be violated by using such an appropriate data
comparison technique. To rid this vulnerability of anonymization models, there are
various privacy preservation models to be proposed. A well-known privacy preser-
vation model is proposed to address privacy violation issues in dynamic datasets, it is
m-Invariance [5]. In addition, m-Invariance is generally proposed to address privacy
violation issues in dynamic datasets that are only allowed to change the data of them by
using insertion and deletion methods. The privacy preservation idea of m-Invariance is
that before datasets are released for public uses, all unique tuples are distorted by using
the additive counterfeit tuples until they are satisfied by the given privacy preservation
constraint. Aside from [5], in [6], the authors show that the incremental datasets also
have the concern of privacy violation issues that must be addressed. Aside from
dynamic datasets that are based on insertion and deletion methods, in [7] and [8], the
authors show that decremental datasets also have the concern of privacy violation
issues that must be addressed. To address these privacy violation issues, the authors
suggest that aside from released datasets are satisfied by the given privacy preservation
constraint, all possible comparison results between released datasets and their related
released dataset versions must also be satisfied by the given privacy preservation
constraint. Although these privacy preservation models can address privacy violation
issues in incremental and decremental datasets, they could be inadequate to address
privacy violation issues in dynamic datasets that allow to change the data of them by
using data updation methods. For this reason, a new appropriate privacy preservation
model for updation datasets (modification Datasets) is proposed in this work.
The organization of this work is as follows. In Sect. 2, the difference between
computer security and privacy preservation will be explained. In Sect. 3, an
anonymization model, l-Diversity, is presented. Then, in Sect. 4, we present privacy
violation scenarios in re-publication of modification datasets. Then, a privacy preser-
vation model that can use to address privacy violation issues in re-publication of
modification datasets to be proposed in Sect. 5. Subsequently, the experimental results
are based on the effectiveness and efficiency of the proposed model to be discussed in
Sect. 6. Finally, the conclusion of this work will be presented in Sect. 7.
3 Anonymization Model
One of the most well-known privacy preservation ideas has been proposed to address
privacy violation issues in released datasets, it is data anonymizations. The example of
privacy preservation models is based on data anonymizations to be k-Anonymity [1], l-
Diversity [2], t-Closeness [3], and k-Likeness [4]. For privacy preservation with these
anonymization models, before datasets are released for public uses, all explicit iden-
tifier values of users are removed, and the unique quasi-identifier values of users are
distorted by using data generalizations or data suppressions [12, 13] to be indistin-
guishable. To improve the readability of anonymization ideas, only l-Diversity is
focused on this work.
f l ðDj ; lÞ : Dj !l Dj
be the generalized function for transforming the raw dataset Dj to become the
released dataset Dj . That is, the unique values which are available in QIj j j
1 ; QI2 ; :::; QIq ,
j
of D are generalized by their less specific values to be indistinguishable such that each
group of the indistinguishable quasi-identifier tuples must relate to the sensitive values
in Sj to be at least l different values, it calls an equivalence class of Dj . For this reason,
Privacy Violation Issues in Re-publication of Modification Datasets 941
Dj can also be denoted in the form of the set of its equivalence classes, i.e.,
Dj ¼ fECj1 ; ECj2 ; :::; ECje1 ; ECje g. Without loss of generality, ECj1 [ ECj2 [ ::: [ ECje1
[ ECje ¼ Dj and ECj1 \ ECj2 \ ::: \ ECje1 \ ECje ¼ £. Also, let Dj ½QIj j j
1 ; QI2 ; :::; QIq
be the data projection over on the quasi-identifier attributes QIj j j j
1 ; QI2 ; :::; QIq of D ,
j j j j
and D ½S be the data projection over on the sensitive values S of D . Moreover,
ECjg ½QIj j j j j j
1 ; QI2 ; :::; QIq is the data projection over on QI1 ; QI2 ; :::; QIq of ECg , and
j
j j j
ECg ½S is the data projection over on the sensitive attribute S of ECg . j
With Table 2, we can conclude that after released datasets are satisfied by l-
Diversity constraints, they can guarantee that all possible query conditions of them
always have at least l different sensitive values to be satisfied. For this situation,
released datasets of l-Diversity seem to have not the concern of privacy violation
942 N. Riyana et al.
issues. Unfortunately, in this work, we show that although released datasets are sat-
isfied by l-Diversity constraints, they still have the concern of privacy violation issues
that must be addressed.
Let Dj1 and Dj be both specific released dataset versions of D such that they are
released at the timestamp j 1 and j respectively. Let ui be the target user of the
adversary. Let BGKj1 j
ui and BGKui be the adversary’s background knowledge about
j1
the target user ui at the timestamp j 1 and j respectively. Let ECj1
g ½QI1 ;
QIj1
2 ; :::; QIj1
q of Dj1 match to BGKj1 j1
ui , i.e., BGKui ðQI1
j1
[ QIj1
2 [ :::
[ QIj1
q Þ. Moreover, let ECjg ½QIj j j
1 ; QI2 ; :::; QIq be the related equivalence class of
j1
g ½QI1
ECj1 ; QIj1
2 ; :::; QIj1
q such that ECjg ½QIj j j
1 ; QI2 ; :::; QIq is available in D
j
j1
g ½QI1
and fully covered by ECj1 ; QIj1
2 ; :::; QIj1
q . Moreover, let ECjgg ½QIj
1;
QIj j
2 ; :::; QIq of D
j
match to BGKjui , i.e., BGKjui ðQIj j j
1 [ QI2 [ ::: [ QIq Þ. Let
j1
gg ½QI1
ECj1 ; QIj1
2 ; :::; QIj1
q be the related equivalence class of ECjgg ½QIj
1;
QIj j j1
2 ; :::; QIq such that ECgg ½QI1
j1
; QIj1
2 ; :::; QIj1
q is available in Dj1 and fully
covered by ECjgg ½QIj j j j1 j1
1 ; QI2 ; :::; QIq . For this situation, if ECg ½S ECjg ½Sj and
j j1 j1
ECgg ½S ECgg ½S do not satisfy the given value of l, the sensitive value of the
j
user ui is collected in Dj1 and Dj can be violated by using data inferring.
Example 2. Privacy Violation Issues in Released Datasets Based on the Difference
of Quasi-Identifier Values: Let Table 1 be the raw dataset at the timestamp j 1, and
its released dataset version is shown in Table 2. We suppose that after Table 2 is
released for public uses, Bob’s area, London, in Table 1 is updated to be Notting-
hamshire. Thus, the raw dataset version of Table 1 at the timestamp j is shown in
Table 3. Moreover, we suppose that the released dataset version of Table 3 that is
satisfied by 2-Diversity to be shown in Table 4. With Table 2 and 4, we can see that all
possible re-identification conditions to them always have at least 2 different salaries to
be satisfied. For this situation, Table 2 and 4 seem to have not the concern of privacy
Privacy Violation Issues in Re-publication of Modification Datasets 943
violation issues. Unfortunately, in this work, we show that Table 2 and 4 still have the
concern of privacy violation issues that must be addressed. That is, we suppose that
Bob is the target user of the adversary such that Bob’s salary is the target data.
Furthermore, we assume that adversary knows that Bob is a male who is 51 years old,
and he/she knows that after Table 2 is released, Bob’s area data, London, is updated to
be Nottinghamshire. For this situation, the adversary can be highly confident that such
a tuple is collected in EC1 of Table 2 must be Bob’s profile tuple at the timestamp
j 1. Moreover, the adversary can observe that EC1 of Table 2 fully covers EC1 of
Table 4. For this reason, the adversary can highly ensure that EC1 of Table 4 is the
related equivalence class of EC1 of Table 2. Therefore, the adversary can be highly
confident that $20000 is collected in EC1 of Table 2 must be Bob’s salary. In order to
ensure that $20000 is Bob’s salary information, Bob’s salary is collected in Table 4
must also be revealed. With Table 4, the adversary can see that its EC2 matches to
his/her background knowledge about Bob at the timestamp j, and he/she also see that
EC2 of this table fully covers EC2 of Table 2. Therefore, the adversary can ensure that
Bob’s salary is collected in Table 2 and 4 to be $20000.
Table 3. The raw dataset version of Table 1 is after Bob’s area information to be updated
Name Age Area Salary
Bob 51 Nottinghamshire $20000
David 52 Berkshire $23000
John 52 London $25000
Jennifer 51 Leicestershire $22000
Alice 51 Derbyshire $25000
From the example 2, it is so clear that although released datasets can guarantee that
all possible re-identification conditions to them always have at least l different sensitive
values to be satisfied, they still have the concern of privacy violation issues that must be
addressed.
944 N. Riyana et al.
Let Dj1 and Dj be both specific released dataset versions of D such that they are
released at the timestamp j 1 and j respectively. Let ui be the target user of the
adversary. Let BGKui be the adversary’s background knowledge aboutui . Let
j1
g ½QI1
ECj1 ; QIj1
2 ; :::; QIj1
q and ECjg ½QIj j j
1 ; QI2 ; :::; QIq match to BGKui . More-
j1
g ½QI1
over, ECj1 ; QIj1
2 ; :::; QIj1
q ¼ ECjg ½QIj j j
1 ; QI2 ; :::; QIq . For this situation, if
j1 j1 j j j j j1 j1
ECg ½S ECg ½S and ECg ½S ECg ½S does not satisfy the given value of
l, the sensitive values of ui is collected in Dj1 and Dj can be inferred that they are
j1
changed from ECj1 g ½S ECjg ½Sj to become ECjg ½Sj ECj1 g ½S
j1
.
Example 3. Privacy Violation Issues in Released Datasets Based on the Difference
of Sensitive Values: Let Table 1 be the raw dataset at the timestamp j 1, and its
released dataset version is shown in Table 2. We suppose that after Table 2 is released,
Alice’s salary is updated from $25000 to become $30000. Thus, the raw dataset version
of Table 1 at the timestamp j is shown in Table 5. Moreover, we suppose that Table 6 is
the released dataset version of Table 5 that it is satisfied by l-Diversity constraints. With
Table 2 and 6, we can also see that all possible re-identification conditions to them have
at least two different salaries to be satisfied. For this situation, Table 2 and 6 seem to
have not the concern of privacy violation issues. Unfortunately, in this work, we show
that Table 2 and 6 still have the concern of privacy violation issues that must be
addressed. Let Alice be the target user of the adversary such that Alice’s salary is the
target data of the adversary. We suppose that the adversary highly believes that Table 2
and 6 collect Alice’s tuple profile. Moreover, the adversary knows that after Table 2 is
released, Alice’s salary information is updated, and he/she further know that Alice is a
female who is 51 years old and lives in Derbyshire. For this situation, the adversary is
highly confident that EC2 of Table 2 and 6 must collect Alice’s profile tuple. After the
adversary compares the salaries which are available in both specific equivalence
classes, the adversary can be highly confident that Alice’s salary is changed from
$25000 to become $30000.
Table 5. The raw dataset version of Table 1 is after Alice’s salary information to be updated
Name Age Area Salary
Bob 51 London $20000
David 52 Berkshire $23000
John 52 London $25000
Jennifer 51 Leicestershire $22000
Alice 51 Derbyshire $30000
Privacy Violation Issues in Re-publication of Modification Datasets 945
From the example 2 and 3, they can be so clear that although released datasets are
satisfied by l-Diversity constraints, they still have the concern of privacy violation
issues that must be addressed. To rid this vulnerability of l-Diversity, a new privacy
preservation model for re-publication of modification datasets is proposed in this work,
it will be presented in Sect. 5.
In this section, a new privacy preservation model for modification datasets is proposed,
so called as b-Diversity. For privacy preservation with b-Diversity, aside from released
datasets are satisfied by the given privacy preservation constraint, all comparison
results between their related released dataset versions must also be satisfied by the
given privacy preservation constraint. In addition, it has the assumption that all ordered
pairs of released dataset versions which are constructed from the raw dataset D such
that they are released at the timestamp between 1 and j 1, they always are satisfied by
b-Diversity constraints. For this reason, only the released dataset version at the
timestamp j 1, Dj1 , is considered to construct the released dataset version Dj .
5.1 b-Diversity
Let Dj be the raw dataset at the timestamp j. Let Dj1 be the released dataset of Dj1 .
Let a positive integer b, where b 2, be the privacy preservation constraint. Let
equivalence class of Dj . Therefore, Dj can also be denoted by the set of its equiva-
lence classes, i.e., Dj ¼ fECj1 ; ECj2 ; :::; ECje1 ; ECje g.
j
X Xq h di QIjz
PLOSS ECjg ; W ¼ þ ðjWj qÞ ð1Þ
8dj
i
2ðECjg WÞ z¼1 H QIj
z
Pe
g¼1 PLOSS ðECg ; WÞ
j
j
DLOSS ðD ; WÞ ¼ j
ð2Þ
jD j q
Where,
– W is the set of the tuples which are holded or suppressed,
– hðdj j
i ½QIz Þ is the level of the generalized value which is available in the quasi-
j
identifier attribute QIj
z of the tuple di , and
– HðQIj j
z Þ is the height of generalization structure of QIz .
6 Experiment
In this section, the experimental results are based on the effectiveness and efficiency of
the proposed model to be discussed by comparing with its traditional model, l-Diversity
[2].
Moreover, all experiments are based on Adult Dataset [15]. With this dataset, it
contains about 48,843 tuples with six continuous attributes (i.e., age, fnlwgt, education-
num, capital-gain, capital-loss, and hours-per-week) and eight nominal attributes (i.e.,
workclass, education, marital-status, occupation, relationship, race, sex, and native-
country). In order to conduct the experiments effectively, only the education, age, sex,
native-country, race, and capital-loss attributes are available in the experimental dataset
such that the capital-loss attribute is set to be the sensitive attribute, and other attributes
are set to be the quasi-identifier attributes. Furthermore, all tuples that consist of “0”
and “?” to be removed. Thus, the experimental dataset only contains 2283 tuples, and
the characteristic of the experimental dataset is shown in Table 8. In addition, all
effective experimental results are evaluated by using the DLOSS metric which is pre-
sented in Sect. 5.2.
From the experimental results that are shown in Fig. 4, we can see that the number
of quasi-identifier attributes more influence to the data utility of released datasets. That
is, if the number of quasi-identifier attributes is increased, the data utility of released
datasets is decreased. The cause of decreasing the data utility of released datasets is
when the number of quasi-identifier attributes to be increased. It is size of equivalence
classes, i.e., more quasi-identifier attributes lead to be the larger size of equivalence
classes in released datasets. Unfortunately, we see that the number of quasi-identifier
attributes more influences to released datasets of the proposed model than released
datasets of l-Diversity. It is generally data privacy and data utility in released datasets to
be traded off.
In the third experiment, it is proposed to evaluate the effect of the value of l that
influences to the data utility of released datasets which are after satisfied by the pro-
posed model and l-Diversity. For experiments, the value of l is varied in the range
between 1 and 5. Also, in all experiments, 1600 tuples of experimental datasets are
random to update the sensitive value or the quasi-identifier values of them.
Privacy Violation Issues in Re-publication of Modification Datasets 951
From the experimental results that are shown in Fig. 5, we can see that the value of
l also influences to the data utility of released datasets. That is, if the value of l is
increased, the data utility of released datasets is decreased. Moreover, we can also see
that released datasets are satisfied by l-Diversity constraints, they often have the data
utility to be more than released datasets which are satisfied by the proposed model. The
cause of the less data utility in released datasets of the proposed model is the size of
equivalence classes that are available in the released datasets, i.e., they are often the
larger size than released datasets of l-Diversity.
6.2.2 Efficiency
In this section, both experiments for evaluating the efficiency of the proposed model are
presented such that they are based on the number of quasi-identifier attributes and the
value of l. By the effect of the number of quasi-identifier attributes, they are evaluated
by the limitations which are that the value of l is fixed to be 2 and the number of quasi-
identifier attributes is varied from 1 to 5 attributes. For evaluating the effect of the value
of l, all quasi-identifier attributes can be available in the experimental dataset and the
value of l is varied in the range between 1 and 5. Moreover, in all experiments, 1600
tuples of experimental datasets are random to update the sensitive value or the quasi-
identifier values of them. From the experimental results which are shown in Fig. 6 and
7, we observe that the greater number of quasi-identifier attributes and more value of l
are more using the execution time for transforming the experimental dataset to satisfy
the given privacy preservation constraints of the proposed model and l-Diversity.
Moreover, we can see that the proposed model uses the execution to be more than l-
Diversity. The cause of using more execution time in the proposed model is that aside
from the data of released datasets must be satisfied by the given privacy preservation
constraint, all compared results between released datasets and the related released
datasets of them are also considered to satisfy the given privacy preservation con-
straints of the proposed model.
952 N. Riyana et al.
7 Conclusion
References
1. Sweeney, L.: K-anonymity: a model for protecting privacy. Int. J. Uncertain. Fuzz. Knowl.
Based Syst. 10(5), 557–570 (2002)
2. Machanavajjhala, A., Kifer, D., Gehrke, J., Venkitasubramaniam, M.: L-diversity: Privacy
beyond k-anonymity. ACM Trans. Knowl. Discov. Data 1, 1 (2007)
3. Li, N., Li, T., Venkatasubramanian, S.: t-Closeness: privacy beyond k-anonymity and l-
diversity. In: 2007 IEEE 23rd International Conference on Data Engineering, Istanbul,
pp. 106–115 (2007)
4. Riyana, S., Natwichai, J.: Privacy preservation for recommendation databases. Serv.
Oriented Comput. Appl. 12(3–4), 259–273 (2018)
Privacy Violation Issues in Re-publication of Modification Datasets 953
5. Xiao, X., Tao, Y.: M-invariance: towards privacy preserving re-publication of dynamic
datasets. In: Proceedings of the 2007 ACM SIGMOD International Conference on
Management of Data (SIGMOD 2007), pp. 689–700. Association for Computing
Machinery, New York (2007)
6. Byun, J.W., Sohn, Y., Bertino, E., Li, N.: Secure anonymization for incremental datasets. In:
Jonker, W., Petković, M. (eds.) Secure Data Management. SDM 2006. Lecture Notes in
Computer Science, vol. 4165. Springer, Heidelberg (2006)
7. Riyana, S., Harnsamut, N., Sadjapong, U., Nanthachumphu, S., Riyana, N.: Privacy
preservation for continuous decremental data publishing. In: Chen, J.Z., Tavares, J., Shakya,
S., Iliyasu, A. (eds.) Image Processing and Capsule Networks. ICIPCN 2020. Advances in
Intelligent Systems and Computing, vol. 1200. Springer, Cham (2021)
8. Riyana, S., Riyana, N., Nanthachumphu, S.: An effective and efficient heuristic privacy
preservation algorithm for decremental anonymization datasets. In: Chen, J.Z., Tavares, J.,
Shakya, S., Iliyasu, A. (eds.) Image Processing and Capsule Networks. ICIPCN 2020.
Advances in Intelligent Systems and Computing, vol. 1200. Springer, Cham (2021)
9. Russell, D., Gangemi, G.T.: Computer Security Basics. O’Reilly & Associates Inc, USA
(1991)
10. Riyana, S., Riyana, N., Nanthachumphu, S.: Enhanced (k,e)-anonymous for categorical data.
In: Proceedings of the 6th International Conference on Software and Computer Applications
(ICSCA 2017). Association for Computing Machinery, New York, pp. 62–67 (2017)
11. Riyana, S., Harnsamut, N., Soontornphand, T., Natwichai, J.: (k, e)-Anonymous for ordinal
data. In: Proceedings of the 2015 18th International Conference on Network-Based
Information Systems (NBIS 2015), pp. 489–493. IEEE Computer Society, USA (2015)
12. Riyana, S., Nanthachumphu, S., Riyana, N.: Achieving privacy preservation constraints in
missing-value datasets. SN Comput. Sci. Appl. 12(3–4), 259–273 (2020)
13. Sweeney, L.: Achieving k-anonymity privacy protection using generalization and suppres-
sion. Int. J. Uncertain. Fuzz. Knowl. Based Syst. 10(5), 571–588 (2002)
14. Wang, G., Zhu, Z., Du, W., Teng, Z.: Inference analysis in privacy-preserving data re-
publishing. In: 2008 Eighth IEEE International Conference on Data Mining, Pisa, pp. 1079–
1084 (2008). https://doi.org/10.1109/ICDM.2008.118.
15. Kohavi, R.: Scaling up the accuracy of naive-bayes classifiers: a decision-tree hybrid. In:
Proceedings of the Second International Conference on Knowledge Discovery and Data
Mining (1996)
Using Non-straight Line Updates in Shuffled
Frog Leaping Algorithm
Abstract. This paper presents an improved version of the shuffled frog leaping
algorithm (SFLA), called the drunken SFLA. The updating process in the
original SFLA moves a solution toward a hopefully better solution in a straight-
line manner while the algorithm is trying to search for the optimal solution,
however, for some types of problems, moving in a straight-line manner might
not converge effectively or might not converge at all. Since the optimal solution
might not be on the straight line between solutions while the algorithm is
progressing. The idea is to add the angle to the updating process in order to
expand the search space in each step. The updating process of ‘Drunken SFLA’
is no longer in a straight line. This paper explores the possibility of modifying
the updating process of SFLA in a Non-Straight Line manner. The convergence
performance results of the proposed drunken SFLA and the original SFLA are
presented.
1 Introduction
The evolutionary algorithm (EA) is a heuristic-based approach that has been developed
to solving complicated problems in polynomial time and huge search space, classically
NP-Hard problems. For instance, genetic algorithms, particle swarms, ant colony,
SFLA, and anything else that would take too much time. The Shuffled Frog Leaping
Algorithm (SFLA) is a meta-heuristic which related to repeatedly update the frog
positions and segmenting the frogs in searching for food. This research compared the
principle and results of five algorithms based on evolutionary processes, including the
SFLA [1].
SFLA was presented by Eusuff and Lansey in 2003 [2]. This research proposes the
SFLA used to solve the discrete optimization problems. In this case, they applied SFLA
to determine the fitting size of the discontinuous pipe for guiding the new network
pipeline installation and expansion. Besides, it is utilized to find and try to access a
suitable global solution. SFLA is consists of global search and local exploration
step. The SFLA is a discrete search algorithm that imitates the search behavior of frogs
by grouping them to determine which group has the highest fitness function [3]. It has a
local exploration technique and global exploration steps. The local exploration step is
the process to divide all of the analyzed frogs into subgroups, called a memeplex.
Several researchers have modified SFLA to improve the efficiency of solving and have
been successfully implemented for their results [5, 6]. Accordingly, SFLA is broadly
used in various fields of science and industry.
In 2013, the fast-shuffled frog leaping algorithm (FSFLA) was proposed for two
reasons; First, each sub-group learns from groups using an update strategy. Second,
their algorithm speed is improved through all sorting and grouping of the individual at
regular intervals [7]. Their results show that the FSFLA is a high convergence accuracy
and fast speed. After that, in 2014, many researchers applied SFLA to improve an
optimization problem [8, 9]. The improved SFLA study the parameters. The experi-
mental results showed the improved SFLA solves the optimization problem more
efficiently in convergence accuracy compared to the original SFLA.
This paper is structured as follows: Sect. 1 as an introduction, Sect. 2, the materials
and the proposed method explained, a modified drunken SFLA using the Truncated
Gaussian random with a drift angle in updating process was described, and the impact
parameters used are presented. Section 3 explained the overview of the original SFLA
and the drunken SFLA. Section 4 shows the results of the graphs convergence between
the primary SFLA and three types of drunken SFLA and discussion. Section 5
explained the conclusion.
Pw ðnewÞ ¼ Pw þ D ð2Þ
In Eq. (1) vector D is the position updating, where RdðÞ is a random number with a
uniform distribution between 0 to 1, Pb has denoted the best frog position and Pw is
denoted the worst frog positions, respectively. Also, in Eq. (2), the new worst frog
position is denoted Pw ðnewÞ. If Pw ðnewÞ has better fitness than old Pw , then Pw ðnewÞ
956 K. Daoden and T. Thaiupathump
will be replaced to Pw , if not, the global frog (Pg ) was used instead of Pb in the
updating position process. If its fitness function value does not improve, the process
will generate randomly the new for Pw . This process is called ‘censorship’. For each
memeplex, a step is performed for a predetermined number of repetition iterations.
Then all the frogs in the entire population are shuffled for global exploration. Local
exploration and global search process will continue alternately until the predetermined
convergence criteria are satisfied.
In addition, the position updating of the modified drunken SFLA probably tend to
move the worst frog’s position to the area close to the middle area around the direct line
between the worst frogs’ and the best frogs’ positions. Thus, the drunken SFLA
replaces the uniformly distributed random number with a Truncated Gaussian distri-
bution. It is possible to fall in the middle of the vector D line, as shown in Fig. 1.
Using Non-straight Line Updates in Shuffled Frog Leaping Algorithm 957
This experiment added the directional movement of the frog’s position updating pro-
cess and the angle value is considering. The drift angle ðhÞ is an important parameter to
the study to investigates the use of the drift angle and compares the results of its which
range from 0 to 90. Further, this study using uniformly random number distribution
combined with the drift angle. Conclusively, this experiment combined the two types
of important parameters, using Truncated Gaussian distribution random number
combined with the drift angle parameter.
3.2 The Drift Angle ðhÞ is the First Parameter of the Proposed Method
The experiment defined the Truncated Gaussian distributed random number ðRTG Þ in
the range 0 to 1. The mean ðlÞ equals 0.5 used to substitute the uniform random
number in D, as shown in Eq. (3).
The drift angle ðhÞ is during in the range from 0 to 90. When the b is calculating in
Eq. (4), and h is the drift angle from the worst frog direct moving close to the best frog
in the 2-dimensional of the search area, calculated in Eq. (5). Dh is the parameter that
adjusts the deviation from the original direction and when i = 1 and 2. When the h is
the drift angle of deviation moving from the worst frog straightway to the best frog in
the 2D of the search space area. The process is calculating follow the Eq. (6) and (7),
which are parameters that improved from the original direction.
c0max
c ð6Þ
itermax
The shaded area (cmax ) from the Eq. (6) is a moving parameter that modifies the
process where the worst frogs randomly search around the best frogs through a local
exploration process. Moreover, the iter is the parameter number of algorithm iterations
while the cmax is the maximum initial deviation angle that starting from the original
direction. cmax is value between 0 and p/2. Therefore, a result of cmax is the highest
value at the initialization of the algorithm. It is decreasing as a better solution founded.
Since the h parameter in Eq. (7) is the angle between cmax and cmax , the variation of
the deviation angle will be the interval of [cmax , −cmax ], as described in this experiment.
This paper investigates studies and compares the value of drift angle (h), which is the
appropriate values as the explanation. In this experiment, start from angle equal 0, 10,
20, 30, 45, 60, and 90. When the h is smaller and larger, the convergence performance
results are poor. When the h is around 30 to 60, the convergence is better. In the
958 K. Daoden and T. Thaiupathump
simulation results, the best angle = 30, therefore, the reason to choose this value using
in the study. Four cases of these studies of original SFLA and Drunken SFLA
(DSFLA) are shown in Fig. 2.
Fig. 2. Four cases of the studies consist of the original SFLA and the Drunken SFLA (DSFLA).
Case 1. (a) Original SFLA model in locating with the frogs is a random number
where the Rd() is between 0 and 1. Case 2. (b) is DSFLA which distributed randomly
with a uniformly distributed random number where the RTG() is between 0 and 1. The r
is equal to 0.5. Case 3. (c) DSFLA are distributed randomly with a uniformly distributed
random number where the RTG() is between 0 and 1, the h is equal 30. And Case 4. (d) is
DSFLA with distributed randomly with a uniformly distributed random number where
the RTG() is between 0 and 1. The r is equal to 0.5. And the h is equal to 30.
Using Non-straight Line Updates in Shuffled Frog Leaping Algorithm 959
Fig. 3. The distributions of random numbers used in frogs’ position updating step in SFLA and
Drunken SFLA (DSFLA) with different r values.
Runtime (s.): The success rate was set up in global optimum at (−32, −32), with an
average of 10 runs.
In Fig. 4 shown the convergence rate of the drift angle ðhÞ with different values
comparing between h equal to 10, 30, 45, 60 and 90. The experiment achieved with the
same data set of the initial random population in solution space [−66,66]. When h is
equal to 10, the convergence rate is particularly slow. While the convergence rate for
the h equal to 20 caused is better than for h equal to 10, it still converges to the global
solution much slower. When the h equal to 30, the graph has shown a better conver-
gence performance than in the others, as shown by a solid line at the bottom of Fig. 4,
and when h equal to 45, 60 and 90 respectively. The convergence rates are slower than
the 30, for this reason, to choose h equal to 30 in this experiment.
Runtime (s.): The success rate was found in global optimum at (−32, −32) in 10
runs.
When r is greater than 1, the convergence efficiency can be expected to approach
the performance of the traditional SFLA.
1 X25
1
f ðxÞ ¼ þ ð8Þ
500 P
2
6
j¼1 jþ ðxi aij Þ
i¼1
962 K. Daoden and T. Thaiupathump
This result shows the convergence performance of four types. The comparison graph
between the traditional SFLA, the modified DSFLA using Truncated Gaussian distri-
bution with sigma ðrÞ equal to 0.5, the DSFLA using Truncated Gaussian distribution
with h equal to 30 and the DSFLA using Truncated Gaussian distribution with sigma
ðrÞ equal to 0.5 with the drift angle ðhÞ equal to 30. The mean ðlÞ value is equal to 0.5
in the range [0, 1].
Fig. 5. The Original SFLA comparison with DSFLA combine purposed method.
Figure 5 shows the convergence rates of the original SFLA and the modified
DSFLA. The original SFLA represents in a small black line. The DSFLA with the
truncated Gaussian distribution and the r equal to 0.5 uses the dashed line. The DSFLA
with the truncated Gaussian distribution and the h equal to 30 presents by the dashed
dot line, while the DSFLA with the drift angle h equal 30 is a solid black line,
respectively. The x-axis of the graph represents the number of iterations equal to 150,
while the y-axis shows the average population fitness level using in the log scale.
Using Non-straight Line Updates in Shuffled Frog Leaping Algorithm 963
6 Conclusion
References
1. Emad, E., Tarek, H., Donald, G.: Comparison among five evolutionary-based optimization
algorithms. Adv. Eng. Inform. 19, 43–53 (2005)
2. Muzaffar, E., Kevin, L.: Optimization of water distribution network design using the shuffled
frog leaping algorithm. J. Water Resour. Plan. Manag. 129, 210–225 (2003)
3. Muzaffar, E., Kevin, L., Fayzul, P.: Shuffled frog leaping algorithm: a memetic meta-
heuristic for discrete optimization. Eng. Optim. 38, 129–154 (2006)
4. Nagham, A., Ahamad, K.: De Jong’s sphere model test for a social-based genetic algorithm
(SBGA). Int. J. Comput. Sci. Netw. Secur. 8, 179–187 (2008)
5. Ziyang, Z.,Daobo, W.,Yuanyuan, L.: Improved shuffled frog leaping algorithm for
continuous optimization problem. In: IEEE Congress on Evolutionary Computation (2009)
6. Mohammad, A., Mohammd, A., Amin, S.: An efficient modified shuffled frog leaping
optimization algorithm. Int. J. Comput. Appl. 32, 0975–8887 (2011)
7. Lianguo, W.,Yaxing, G.: A fast shuffled frog leaping algorithm. In: IEEE International
Conference on Natural Computation (2013)
8. Guang-Yu, Z., Wei-Bo, Z.: An improved shuffled frog-leaping algorithm to optimize
component pick-and-place sequencing optimization problem. Expert Syst. Appl. 41, 6818–
6829 (2014)
9. Ehsan, B., Malihe, F.: An improved adaptive shuffled frog leaping algorithm to solve various
non-smooth economic dispatch problems in power systems. In: IEEE Intelligent Systems
(2014)
10. Kanchana, D., Trasapong, T.: A modified shuffled frog leaping algorithm using truncated
Gaussian distribution in Frog’s position updating process. In: Conference: Information
Science and Applications (ICISA), Ho Chi Minh City, Vietnam, vol. 376s (2016)
11. Zhen, W., Danhong, Z., Biao, W., Wenwen, C.: Research on improved strategy of shuffled
frog leaping algorithm. In: 34rd Youth Academic Annual Conference of Chinese
Association of Automation (YAC), Jinzhou, China (2019)
Efficient Advertisement Slogan Detection
and Classification Using a Hierarchical
BERT and BiLSTM-BERT
Ensemble Model
Md. Akib Zabed Khan1 , Saif Mahmud Parvez2(B) , Md. Mahbubur Rahman3 ,
and Md Musfique Anwar4
1
Bangladesh University of Business and Technology, Dhaka, Bangladesh
akibcseju21@gmail.com
2
Daffodil International University, Dhaka, Bangladesh
saif.mahmud.parvez@gmail.com
3
Crowd Realty, Tokyo, Japan
mahbuburrahman2111@gmail.com
4
Jahangirnagar University, Dhaka, Bangladesh
manwar@juniv.edu
people to identify their products from existing competitive brands. The slogan
must meet the criteria of a context. (e.g., food, automotive, business, etc.) [2] as
well as can be changed over time based on market demand.
Catchy phrase is usually used to catch the attention of the customers which
helps to distinguish a particular product from other similar products that are
available in the market. Many scholars analyzed and explained a number of lan-
guages and rhetorical devices that are common for advertising slogans. At the
realistic level, trademarks illustrate the utilization of full or fractional capitaliza-
tion as well as offbeat spelling, whereas, at the phonological level, the broad util-
ity of rhyme, similar sounding word usage, sound similarity, and onomatopoeia
can be seen for advertising slogans. Usage of pronouns, numerals, adjectives,
and verbs are common lexical features in the above methodology, whereas usage
of idioms and phrases, imperative sentences are usual notions at the syntactic
analysis level. Again, simile, puns, metaphor etc. seems to be a usual trade from
a semantic point of view of a slogan [3].
Perfect slogan finding for branding a particular product is not a simple task as
it usually takes a lot of brainstorming and time to generate good branding rhymes.
Natural language processing (NLP) is a technique that has already enabled a new
dimension for modeling human language, detecting the inner meaning of language,
and sentiment identification of a text. Sentence structure consolidates a number of
errands, such as dialect structure acknowledgment, lemmatization, morphological
division, discourse labeling, parsing, and stemming. Different coherent and fac-
tual computational strategies are utilized for the language structure and semantic
investigations [4]. In cooperation of deep neural networks gives it a boost to find
out the inner meaning in a more correct way [5].
In this work, we have used 10,023 renowned advertisement slogans and cre-
ated a model using BERT classifier to detect and classify the slogans into related
contexts. We also have evaluated if a slogan is a relevant slogan to the context
or not by incorporating text classification and deep learning approaches. Our
main objective has been to create a model that can detect whether a sentence
is a slogan or not and classify the respective context if the sentence seems to be
a slogan for representing a brand in advertising a product.
2 Literature Review
In English language, slogans have different stylistic features which differentiate
a slogan sentence from normal English sentences. Such features are, presence of
simile, usage of metaphors and puns, usage of different sound techniques etc.
Tatjana et al. proposed a method that focused on rhetorical devices such as
figurative language and sound techniques to classify 100 sampled slogans into
three groups [3]. However, the drawback of their method is that no automation
technique is used to generate or categorize the slogan.
Recent research works paid more attention for the automation of slogan gen-
eration. For example, nominal metaphors are generated on the basis of metaphor
interpretation models proposed in [6]. Using these metaphors and typical slogan
templates, the method is able to generate new slogans in a given context with its
966 M. A. Z. Khan et al.
data we have scraped through different websites1,2 and extracted the sentences.
This dataset is used for slogan detection. Later, we have scrapped the context
only for slogans along with slogan sentences for context classification such as food
& beverage, business, etc. Next, we have pre-processed the dataset to improve
the quality and performance of the dataset and the subsequent steps. We have
lower cased the sentences and removed the stopwords, numbers, unnecessary
characters, single word sentences to clean our data. We also applied stemming
and lemmatization to convert the words into corresponding root forms. Then,
we have shuffled our datasets so that slogan and non-slogan type sentences can
be distributed equally in all over the dataset.
sen1 = “W here is the beef ?” , and sent2 = ”Do not break anyone s heart”
pooling to detect the latent features from the sentences. The output of this layer
is first passed to a Dropout layer which passes its output to the Fully connected
layer. Finally, the output from Fully connected layer is passed to a Classifica-
tion/Output layer which classified all the sentences as slogan and non-slogan.
In the pre-trained BERT model, a bidirectional transformer encoder is used to
understand the latent syntactic and semantic features of the sentences. By imple-
menting the attention mechanism, the transformer model balances the relation-
ship of input and output. In the decoder portion, it calculates the relationship
of input sentence and output in parallel which improves the computational effi-
ciency and also removes the long distance information decay problem.
Slogan Text Features: There are many features exist in advertisement slogan
text which differentiate it from traditional text such as news, articles, etc. Some
of these main features are:
Context: The main feature of a slogan is that it indicates a particular context as
each slogan tries to get the attention of customers of a specific product category.
If the slogan fails to express the exact category according to the context, then
it doesn’t recognize as a good slogan. In reality, there are millions of brands
that have slogans but a majority of them are not good enough as well as not
expressed in right format. As a result, those slogans sometimes represent an
ambiguous context. Therefore, it is a challenging task to identify good slogans
with good context among all these unpopular advertising slogans.
Slogan Detection and Classification 969
Data Sparsity: Slogans usually consist of words ranging from only four to
seven. Due to this highly short number of words, it creates a massive problem
with sparse data. Context analysis becomes very difficult due to this sparse data
and if they are not processed properly they would cause serious problems in
sparse matrix representation.
Irregularity and Interactivity: As advertising slogans are really concise and
also brand focusing is the main concern, many common garbage words, unknown
nouns are seen in a common manner. Also, slogans are bound to be unique. As a
result, there is hardly a proper relationship among same types of slogans rather
than their specific contexts.
Feature Extraction: As the context is being the core feature of slogans, the
same context slogan tends to have the same type of words. Such as, a food
category tends to focus all the eating behaviour and food-related words whereas
an automotive category tends to focus on speed, energy and power focusing
words. Our dataset contains eight broad categories of contexts. Table 1 shows the
names of few of the contexts with an example. We have performed exploratory
data analysis on the overall dataset as well context wise slogan data. Common
words over the dataset are those words which are common to all the contexts.
As a result, these words create problems to distinguish the contexts that is
why they are discarded. Context wise common words indicate a context’s inner
nature. Both of these types of common words examples are shown in Table 2. To
fetch out these common words, first we have performed exploratory data analysis
(EDA) over the dataset. We got 30 of such common words over the dataset and
then discarded them. It helped the dataset to be clustered among the contexts.
Later, we performed topic modeling using LDA [14] to find the latent topics with
their relevant words. We also performed context wise EDA to cross check with
the LDA to get the top common words which are specific to a context.
After extracting and analyzing the features, the next step is to find the cor-
relation among the words. Firstly, we have classified the slogans from a pool
of sentences using the BERT classifier. Next, these slogans are processed based
on the common words. For context classification, we have used an ensemble of
BiLSTM and BERT to get the best result.
Ensemble Model: Both BiLSTM and BERT perform well to classify the con-
texts effectively. The three gate architecture of LSTM captures some temporary
information and generates candidates for selecting context traversing in both
ways. On the other hand, BERT uses transformer blocks which use self atten-
tion masks which tries to bring out the inner architecture of a slogan and predict
its context. In our method, we have used ensemble of BiLSTM and BERT and
used majority rules across these models to fetch the best probability score. The
final architecture of our model is shown in Fig. 2.
Our experiment has been conducted in two steps. In the first step, we have
experimented on a dataset which is a collection of 10,073 advertisement slogan
sentences and 10,223 regular single lined short sentences. After that, these 20,296
sentences are splitted into a training set and test set. Table 4 shows this distri-
bution, and Fig. 3 shows word cloud distribution of words into these two classes:
slogan data word cloud at left and non-slogan at right.
Slogan Non-slogan
Train 9062 (44.65%) 9204 (45.35%)
Test 1011 (4.98%) 1019 (5.02%)
972 M. A. Z. Khan et al.
Next, we have classified the slogan sentences from the dataset and split it into
train and test sets. Then, we have to again pre-process those identified slogans
for context classification. We discard some sentences containing garbage words
or of length of only one word. This processed dataset contains eight different
classes/categories of context as shown in Table 5.
We have used Accuracy, Precision, Recall and F1-score as stated in Eq. (11),
(12), (13) and (14), respectively for performance evaluation.
In the first part of our experiment, we classified slogan from non-slogan using
the BERT pre-trained model. Table 6. shows the output of the model.
Table 6. Accuracy and evaluation metrics for slogan detection using BERT
Model
BiLSTM BERT Ensemble Model
Accuracy 81.60% 82.21% 82.5%
974 M. A. Z. Khan et al.
5 Conclusion
References
1. Abdi, S., Irandoust, A.: The importance of advertising slogans and their proper
designing in brand equity. Int. J. Organ. Leadership 2(2), 62–69 (2013)
2. Kohli, C., Leuthesser, L., Suri, R.: Got slogan? Guidelines for creating effective
slogans. Bus. Horiz. 50(5), 415–422 (2007)
3. Dubovičienė, T., Skorupa, P.: The analysis of some stylistic features of English
advertising slogans. Žmogus ir žodis, t. 16(3), 61–75 (2014)
4. Rahim, A., Qiu, T., Ning, Z., Wang, J., Ullah, N., Tolba, A., Xia, F.: Social acquain-
tance based routing in vehicular social networks. Future Gener. Comput. Syst.
1(93), 751–60 (2019)
5. Dou, Z.Y., Wang, X., Shi, S., Tu, Z.: Exploiting deep representations for natural
language processing. Neurocomputing 21(386), 1–7 (2020)
Slogan Detection and Classification 975
6. Alnajjar, K., Kundi, H., Toivonen, H.: “Talent, Skill and Support.”: a method for
automatic creation of slogans. In: Proceedings of the Ninth International Confer-
ence on Computational Creativity, Salamanca, Spain, Association for Computa-
tional Creativity, June 25–29, (2018)
7. Iwama, K., Kano, Y.: Japanese advertising slogan generator using case frame and
word vector. In: 11th International Conference on Natural Language Generation,
pp. 197–198 (2018)
8. Tomašič, P., Papa, G., Žnidaršič, M.: Using a genetic algorithm to produce slogans.
Informatica, 39(2), (2019)
9. Yamane, H., Hagiwara, M.: Advertising slogan generation system reflecting user
preference on the web. IEEE ICSC 2015, 358–364 (2015)
10. Al-Makhadmeh, Z., Tolba, A.: Automatic hate speech detection using killer natu-
ral language processing optimizing ensemble deep learning approach. Computing
102(2), 501–22 (2020)
11. Jin, M., Luo, X., Zhu, H., Zhuo, H.H.: Combining deep learning and topic modeling
for review understanding in context-aware recommendation. In: NAACL: Human
Language Technologies, vol. 1, pp. 1605–1614 (2018)
12. Jeong, C.Y., Han, S.W., Nam, T.Y.: A hierarchical text rating system for objec-
tionable documents. JIPS 1(1), 22–26 (2005)
13. Sharif, O., Hoque, M.M.: Automatic detection of suspicious bangla text using logis-
tic regression. In: International Conference on Intelligent Computing & Optimiza-
tion, pp. 581-590. Springer, Cham (2019)
14. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. J. Mach. Learn.
Res. 3, 993–1022 (2003)
15. Jiang, S., Zhao, S., Hou, K., Liu, Y., Zhang, L.: A BERT-BiLSTM-CRF model
for chinese electronic medical records named entity recognition. In: ICICTA, pp.
166–169 (2019)
Chronic Kidney Disease (CKD) Prediction
Using Data Mining Techniques
Abstract. In the past decade, the rapid growth of digital data and global
accessibility through the modern internet has seen a massive rise in machine
learning research. In proportion to it, the medical data has also seen a massive
surge of expansion. With the availability of structured clinical data, researchers
have attracted scores to study clinical disease detection automation with
machine learning and data mining. Chronic Kidney Disease (CKD), also known
as the renal disorder, has been such a field of study for quite some time now.
Therefore, our research aims to study the automated detection of chronic kidney
disease using several machine learning classifiers with clinical data. The purpose
of this research work is to diagnose kidney disease using a number of machine
learning algorithms such as the Support Vector Machine (SVM) and the
Bayesian Network (BN) and to select the most effective one to assess the extent
of CKD patients. The amount of expertise in the medical field in relation to
CKD is limited. Many patients have to wait a long to get their test results. The
experience of medical staff is declining in value. Upon retirement, new
employees replace them. It helps professional doctors or medical staff in their
diagnosis of CKD. This paper’s primary purpose is to present a clear view of
Chronic Kidney Disease (CKD), its symptoms, and the process of early
detection that may help humanity be safe from this life-threatening disease.
1 Introduction
There have been many fatal diseases over past centuries—some of those formed
into an epidemic and affected millions. Deadly diseases like Smallpox, SARS, Mar-
liese, and polio has been cured with the vaccine in the past century. We are still fighting
to cure the likes of Cancer, HIV but the rVSV-ZEBOV vaccine is a single-dose vaccine
protocol that has been found to be safe and secure only against Ebolavirus strains of
Ebolavirus. This is the first FDA approval to vaccinate against Ebola.
One of the most deadly health hazards of our time is chronic kidney disease or in
abbreviation. CKD is defined as the gradual degradation of the normal kidney function,
and the consequence can be catastrophic. The kidney is essential for the filtering and
purification process of our blood. Without at least one functioning kidney, death is
imminent and inevitable within a few days. As it is a chronic disease and the symptoms
are mild and gradual, it often goes unnoticed for years until the very late stage.
So, the scope of this research is to build a model using data mining techniques to
predict if a patient does indeed have CKD by reviewing and analyzing symptoms and
various health parameters; Using Data Mining tools to classify those data and compare
the results acquired through different techniques.
In Bangladesh, a considerable portion of the population living below the poverty
line does not have sufficient access to the required medical attention. The public
medical sector, which is almost entirely run by the federal fund, does not have the
financial capability or the proper medical resources to incorporate this massive amount
of financially unprivileged people into the medical sector. As a result, a preeminent
health concern, such as chronic kidney disease. And renal failure is not diagnosed in
this population. As chronic kidney disorder often does not show any symptoms.
In most cases, patients develop ESRD (End-Stage Renal Disease) without knowing
they even had a chronic renal disorder in the first place. Chronic kidney disease- a
gradual loss of kidney function- when an advanced stage is reached. In end-stage
kidney disease, your kidneys can no longer function because they need to meet your
body’s needs. The lack of federal funding, Lengthy process in healthcare, persistent
lack of quality, and unsatisfactory service in public health institutes have forced the
general public to seek private healthcare. This consequently has flourished the private
sector but at the price of high medical costs. So a massive amount of people affected by
ESRD face dreadful consequences with the lack of proper treatment and dialysis, which
often leave the patient and the family in unfathomable sufferings and humanitarian
crises.
As the first step towards treatment in any medical condition is to getting diagnosed
first. The advancement of medical technology and the capacity to store medical data in
digital form have rejuvenated the idea of medical automation, and the data revolution
has all but made the possibility of artificially automated doctors more than just an
ambitious dream. An automated virtual system to classify CKD is still not entirely
convincing or decisive to the vast majority of doctors and medical personal, and rightly
so. But with more data, efficiency, and more accuracy, a future of automated artificial
medical assistant can become a reality. Followings are the list of things we tried to
accomplish in this project:
• Can the CKD disease be classified with clinical data?
• We are applying a probabilistic classifier to the dataset.
978 A. Pathak et al.
By analyzing the dataset, the CKD status of the patient should be classified. We
tried to find out the correlation of different attributes in the dataset in developing CKD.
The scope of this problem is to classify our dataset using machine learning algo-
rithms. We will attempt to explore the relationship between dataset symptoms to
determine interdependence in the development of chronic kidney disease.
In Bangladesh, an automated diagnosis system would reduce the lengthy process in
health care. With improved symptoms analyzing algorithms, the system can suggest
diagnostic tests to the users, reducing big hospitals’ time and cost.
2 Literature Review
Sujata Drall, Gurdeep Singh Drall, Sugandha Singh, and Bharat Bhushan Naib predict
chronic kidney disease and determine the accuracy of data mining algorithms such as
naïve Bayes classifier and k-nearest neighbour [18]. In this research, they worked with
more than 5 attributes. The classification algorithm’s performance has been compared
based on accuracy, precision, and total execution time for the prediction of Chronic
Kidney disease.
S Dilli Arasu, Dr. R. Thirumalaiselvi has reviewed several algorithms, such as
clustering and classification algorithms [17]. In this research, they dealt with the
algorithm used in a different research paper. They tried to find out the better one
classification algorithm, which has been used to identify chronic kidney disease and its
stages.
Dr. S. Vijayarani, S. Dhayanand has predicted chronic kidney disease and find out
better performance accuracy in less time. They used Naïve Bayes and Support Vactor
Machine algorithms [2]. In this research, They worked with 5 attributes. The execution
performance of the Naïve Bayes classifier was 1.29, which was better than the Support
Vector Machine with a performance of 3.22 to predict CKD.
Tabassum S, Mamatha Bai BG, Jharna Majumdar has developed a system for
predicting chronic kidney disease and find out the better performance accuracy of used
data mining algorithms. In this research, they used Expectation-Maximization, Artifi-
cial Neural Network, C4.5 algorithm, where EM got 70%, ANN got 75%, and C4.5 got
96.75% accuracy [19].
Pushpa M. Patil has predicted chronic kidney disease with data mining algorithms
[15]. This research dealt with those algorithms used in a different research paper and
tried to determine the better one.
3 Methodology
Data Mining is a technique where a large volume of preexisting raw data in the
database is processed or altered to needs and analyzed to reveal useful patterns and new
relations among attributes for achieving various goals. Data mining is also called
knowledge discovery in databases, also known as KDD (Fig. 1).
Chronic Kidney Disease (CKD) Prediction 979
Data mining is a powerful and new field having various techniques to analyses
recent real-world problems. It converts the raw data into useful information in various
research fields and finds the patterns to decide the future medical field [9]. There are
robust data mining techniques developed and used in data mining projects recently for
knowledge discovery from the database.
In this project, we will use the SVM and NB algorithms that will perform classi-
fication and predictability in the database to extract information and classify patients
into two categories: chronic kidney disease (ckd) and not a chronic disease (notckd). In
this section, we propose a method of extracting action rules for CKD datasets showing
in Fig. 2.
980 A. Pathak et al.
Classification is done for the data mining assigned to the objects in a group to the
target categories. Classification is when predicting the final results will be based on the
given input dataset [12]. The set of rules attempts to discover relationships among the
attributes that might make it feasible to expect the outcome. The classification tech-
nique aims to appropriately be expecting the target class for every case within the
records. Example: Whether the affected person is affected by Chronic Kidney Disease
or not.
A. Prediction of CKD
A Naive Bayes classifier is general classifiers with probabilities based on Bayes theory.
It has strong (innocent) beliefs of independence in facilities. SMO is an algorithm that
solves a quadratic programming problem that occurs when training support vector
machines. IBk is one example-based learner who uses the k training examples class,
close to the test example class. The parameter k is the number of immediate neighbours
to be used for the estimate [26].
B. Data Preprocessing
Before implementing any taxonomy algorithm on data, the data must be cleaned in
advance and is called the pre-processing step. During this pre-processing stage, many
processes occur, such as estimating missing values, deleting noisy data such as outliers,
normalizing, and balancing unbalanced data. Real-world data usually do not contain
values. One way to deal with missing values is to leave the entire record containing the
missing value, called case deletes.
C. Estimate the Steps by Calculating the GFR
Different steps are calculated by calculating the GFR from the given properties using
the MDRD formula. GFR is traditionally measured as the renal clearance of an ideal
filter marker, such as insulin from plasma. The calculated GFR is considered the
standard for everyday clinical use but not practical due to the complexity of the
measurement process. Estimating GFR based on the filter marker (usually serum cre-
atinine) is now widely accepted as a basic test.
MDRD formula:
GFR (mL/min/1.73m2) = 175 (Scr)-1.154 (Age)-0.203 (0.742 if female)
(1.212 if African American)
used to measure GFR. But inulin is not an internal constitution of the body. So an
external infusion of inulin is required for the detection of inulin clearance. In most
practical cases, creatinine clearance is used for this process. Creatinine (a direct result
of protein metabolism) is an endogenous substance used for GFR measurement pretty
accurately. The estimation of GFR is used to classify kidney impairment [11]. A low
GFR rate is indicative of a disruptive kidney function. Table 1 illustrates the various
stages of Kidney disease and its severity.
The above two interfaces represent the processed and non-processed data used in this
research [7]. The first one is for non-processed data, and the other is for pro-cessed data
with the probability of CKD or non-CKD. The data which are used in the data mining
technique cannot be used without processing. That’s why data must be processed
before using in data mining technique [8] (Fig. 3).
To test our classification and evaluate its performance, we apply a 10-fold cross-
certification test, which is a technology that divides the original set to train the model
for training, and has a test set to evaluate it. After applying the pre-processing and
Chronic Kidney Disease (CKD) Prediction 985
manufacturing methods, we try to analyze the data and find out the distribution of
values in terms of model performance and accuracy (Table 7).
450
400
350
300
250
200
150
100
50
0
Correctly Incorrectly
Time to build
classified classified Accuracy (%)
model (s)
instances instance
SVM 0.43 392 8 61.25
NB 0.04 382 18 56.5
In conclusion, the use of data mining methods for data risk analysis is very important in
the health sector because it first gives the power to fight diseases and therefore saves
people’s lives by reversing treatment. In this work, we used a number of learning
algorithms to assess patients with chronic renal failure (ckd) and patients with this
disorder (notckd). Simulation results have shown that SVM classification has proven its
performance in predicting the best results in terms of accuracy and minimum execution
time. The result of the testing does not achieve 100%. It may be due to calculation and
weight. To get a more accurate result, the calculation of the similarity has to modify
with the weight. The experts’ opinion and the view will determine the most critical
features (weight), and it will be used in similarity computation by weighted average. By
doing that, the most accurate calculation will be determined. In the future, data col-
lection using the accelerometer sensor will be collected for older people with non-
neurodegenerative diseases. An in-depth Artificial Neural Network can be done before
getting better performance. Many details can be used to train the learning model.
Finally, bring out some of the features that will be most helpful in training the learning
model. In the future, the information-driven approach may be used to remove uncer-
tainty as a legal system based on expertise.
References
1. Andrew, K., Bradley, D., Shital, S.: Predicting survival time for kidney dialysis patients: a
data mining approach. Comput. Biol. Med. 35, 311–327 (2005)
2. Vijayarani1, S., Dhayanand, S.: Data mining classification algorithms for kidney disease
prediction. In: International Journal on Cybernetics & Informatics (IJCI), vol. 4 (2015)
3. Dulhare, U.N., Ayesha, M.: Extraction of action rules for chronic kidney disease using Naïve
bayes classifier. In: 2016 IEEE International Conference on Computational Intelligence and
Computing Research (ICCIC) pp. 1–5, IEEE (2016)
4. Firman, G.: Definition and Stages of Chronic Kidney Disease (2009) http://www.
medicalcriteria.com/site/index.php?option=com_content&view=article&id=142%
3Anefckd&catid=63%3Anephrology&Itemid=80&lang=en
5. Jiawei, H., Micheline, K.: Data Mining, Concepts and Techniques, Second (Eds.), Elsevier
Publication (2003)
6. Jinn-Yi, Y., Tai-Hsi, W., et al.: Using data mining techniques to predict hospitalization oh
hemodialysis patients, Elsevier, vol. 50(2), pp. 439–448 (2011)
7. Kirubha, V., Manju Priya, S.: Survey on data mining algorithms in disease prediction. Int.
J. Comput. Trends Technol. 38(3), 24–128 (2016)
8. Jena, L., Narendra, K.K.: Distributed data mining classification algorithms for prediction of
chronic- kidney-disease. Int. J. Emerg. Res. Manage. Technol. 4(11), 110–118 (2015)
9. Kumar, M.: Prediction of chronic kidney disease using ran-dom forest machine learning
algorithm. Int. J. Comput. Sci. Mobile Comput. 5(2), 24–33 (2016)
10. Dunham, M.H., Sridhar, S.: Data Mining: Introductory and Advanced Topics. Pearson
Education, Dorling Kindersley (India) Pvt Ltd (2006)
11. Naganna, C., Kunwar, S.V., Sithu, D.S.: Role of at-tributes selection in the classification of
chronic kidney disease patients. In: International Conference on Computing, Communication
and Security (ICCCS), pp. 1–6.4-5 Dec 2015
988 A. Pathak et al.
12. National Kidney Foundation (NKF): Clinical practice guidelines for chronic kidney disease:
evaluation, classification and stratification. Am. J. Kidney Disease 39, 1–266 (2002)
13. Pavithra, N., Shanmugavadivu, R.: Survey on data mining techniques used in kidney related
diseases. Int. J. Modern Comput. Sci. 4(4), 178–182 (2016)
14. Pushpa, M.P.: Review on prediction of chronic kidney disease using data mining techniques.
Int. J. Comput. Sci. Mobile Comput. 5(5), 135–141 (2016)
15. Pushpa, M.P.: Review on prediction of chronic kidney disease using data mining techniques.
Int. J. Comput. Sci. Mobile Comput. 5(5), 135–141 (2016)
16. Ruey Kei, C., Yu-Jing, R.: Constructing models for chronic kidney disease detection and
risk estimation. In: Proceedings of 22nd IEEE International Symposium on Intelligent
Control, Singapore, pp. 166–171.International Conference on Internet of Things and
Applications (IOTA), p. 5. Pune, IEEE (2011)
17. Dilli, A., Thirumalaiselvi, R.: Review of chronic kidney disease based on data mining
techniques. Int. J. Appl. Eng. Res. vol. 12(23), 13498–13505 (2017)
18. Sujata, D., Gurdeep, S.D., Sugandha, S., Bharat, B.: Chronic kidney disease prediction using
machine learning: a new approach. Int. J. Manage. Technol. Eng
19. Tabassum, S., Mamatha Bai, B.G., Jharna, M.: Analysis and prediction of chronic kidney
disease using data mining techniques. Int. J. Eng. Res. Comput. Sci. Eng. (IJERCSE) 4(9),
25–32 (2017)
20. Uma, N.D., Ayesha, M.: A review on prediction of chronic kidney disease using
classification techniques. In: 4th International Conference on Innovations in Computer
Science & Engineering (2016)
21. Wu, X., Kumar, V., Quinlan, J.R., Ghosh, J., Yang, Q., Motoda, H., McLachlan, GJ., Ng,
A., Liu, B., Philip, S.Y., Zhou, Z.H.: Top 10 algorithms in data mining. Knowl. Inf. Syst. 14
(1):1–37 (2008)
22. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent Computing & Optimization. Springer,
Berlin (2018)
23. Intelligent Computing and Optimization. In: Proceedings of the 2nd International
Conference on Intelligent Computing and Optimization 2019 (ICO 2019), Springer
International Publishing, ISBN 978–3-030-33585 -4 (2019)
24. Lominadze, D., Schuschke, D.A., Joshua, I.G., Dean, W.L.: Increased ability of erythrocytes
to aggregate in spontaneously hypertensive rats. Clin. Exp. Hy-pertains. 24(5), 397–406
(2002) https://doi.org/10.1081/ceh-120005376
25. Dulhare, U.N., Ayesha, M.: Extraction of action rules for chronic kidney disease using Naïve
bayes classifier. In: 2016 IEEE International Conference on Computational Intelligence and
Computing Research (ICCIC), pp. 1–5, (2016) 10.1109/ICCIC.2016.7919649
26. Naganna, C., Kunwar, S.V., Sithu, D.S.: Role of attributes selection in classification of
Chronic Kidney Disease patients. In: International Conference on Computing, Communi-
cation and Security (ICCCS), pp 1–6. 4-5 Dec 2015
27. Dua, D., Graff, C.: UCI Machine Learning Repository [http://archive.ics.uci.edu/ml]. Irvine,
CA: University of California, School of Information and Computer Science (2019)
28. UCI Machine Learning Repository: Kidney failure Data Set. https://archive.ics.uci.edu/ml/
datasets/Chronic_Kidney_Disease
29. Overview of the KDD Process. http://www2.cs.uregina.ca/*dbd/cs831/notes/kdd/1_kdd.
html
30. http://www.datasciencecentral.com/profiles/blogs/python-resources-for-top-data-mining-
algorithms
31. Basma, B., Hajar, M., Abdelkrim, H.: Performance of data mining techniques to predict in
healthcare case study: chronic kidney failure disease. In: International Journal of Database
Management Systems (IJDMS), vol. 8 (2016)
Multi-classification of Brain Tumor Images
Based on Hybrid Feature Extraction Method
1 Introduction
unsupervised learning method using the CNN algorithm for brain tumor image clas-
sification into different classes. The performance of the classifier in terms of accuracy
was found approximately 98.7% to categories the brain tumors into different classes.
Moreover, gumaei et al. [8] employed a supervised learning method named regularized
extreme learning to classify MRI images as glioma, meningioma, and pituitary
respectively. The classification method provides an accuracy of approximately
94.233% because of using the PCA-based normalized GIST as a hybrid feature
extraction technique.
In another study, Kaplan et al. [9] designed a modified feature extraction technique
named local binary pattern (LBP) based on angle and distance to classify the brain
tumors into three different classes. Then different classification methods were applied
by using KNN, ANN, RF, and linear discriminant analysis (LDA). Among these, the
KNN showed the highest classification accuracy of 95.56% with a distance-based LBP
feature extraction method. Ismael et al. [5] developed a classification method using a
multilayer perceptron for MRI brain tumor classification. The hybrid feature extraction
process including DWT and Gabor filter was employed to produce input statistical data
for the classification network. The classification network was trained based on the
back-propagation algorithm to provide the highest accuracy of 91.9% for brain disease
classification into multiple classes.
The major contributions of this work are summarized as (i) develop an automatic
classification method for brain tumors classification into three classes. (ii) Implement a
simple and effective hybrid feature extraction technique that uses SWT and GLCM for
developing a total of 13 statistical features that contains the most significant infor-
mation of the tumor images. (iii) Enhance the classification accuracy with the aid of
these extracted statistical features. (iv) Analyze the efficacy of the designed method by
comparing it with existing classification methods, where the same datasets are used.
The resting part of the paper is arranged as follows: Sect. 2 gives an explanation of
the methodology of the proposed system with an extensive explanation of each of the
steps. Section 3 presents the results with analysis and at the end of the section, a
comparative study is also conducted. Finally, Sect. 4 shows concluding remarks.
Figure 1 illustrates the flow diagram of the proposed multi-class classification method,
in which the system starts with the preprocessing of the tumor images. Then without
any segmentation process, the hybrid feature extraction method is utilized with the
composition of SWT and GLCM. The hybrid technique produces a total of 13 statis-
tical features of the brain tumor images. Finally, the ANN classifies the brain tumor
images according to the statistical feature.
Multi-classification of Brain Tumor Images 991
Image Preprocessing
Min-Max Normalization
SWT GLCM
Statistical Features
Image Classification
ANN
Meningioma
Glioma Pituitary
f ðx; yÞ Vmin
f ðx; yÞ ¼ ð1Þ
Vmax Vmin
Where f ðx; yÞ represents each pixel value of the brain image. Moreover, Vmin and Vmax
represent the minimum and maximum value of the image. This normalization rule
excludes all the negative and large values of the input image that range from 0 to 255 as
displayed in Fig. 2(a). After preprocessing, the images are converted into an intensity-
based contrast enhanced image of values from 0 to 1 as illustrated in Fig. 2(b).
Fig. 2. Preprocessing steps; (a) The input image, and (b) The contrast enhanced output image.
approximation (LL), diagonal (HH), vertical (HL), and horizontal (LH) images. Then
the sub-band LL is utilized to evaluate the next 2-level of decomposition. Finally, in the
3-level decomposition, the tumor image produces eight sub-bands images represents as
LLL, LLH, LHL, LHH, HLL, HLH, HHL, and HHH. The 3-level of SWT image
decomposition, in which the approximation sub-band, LLL is applied for further
processing and also for finding the nine (mean, variance, standard deviation, entropy,
root means square (RMS), inverse difference moment (IDM), smoothness, kurtosis,
skewness) statistical features shown in Table 1.
3 Results Analysis
The performance of the classifier in terms of training, testing, and validation is shown
in Fig. 4. It is shown that the mean squared error (MSE) is diminished at epoch 33 and
the performance of the classifier is found superior at this epoch with a minimum valued
error of approximately 0.04.
Figure 5 presents the testing performance of the modeled classifier based on the
confusion matrix (CM) for classifying the brain tumor images into three classes. The
predicted classes (system output) are represented along the x-axis while the actual
classes (ground truth) are represented along the y-axis. From this matrix, the Accuracy,
Precision, and Sensitivity have been calculated using the following equations.
TP þ TN
Accuracy ¼ 100% ð2Þ
TP þ TN þ FP þ FN
TP
Precision ¼ 100% ð3Þ
TP þ FP
TP
Sensitivity ¼ 100% ð4Þ
TP þ FN
Where
• TP (true positive) = Positive predicted classes are correctly identified as positive.
• TN (true Negative) = Negative predicted classes are correctly identified as negative.
996 K. A. Sathi and Md. Saiful Islam
The accuracy metrics developed from the confusion matrices are shown in Table 3.
The highest performance based on accuracy, precision, and sensitivity is found for the
pituitary class. The accuracy is achieved 96.96% for glioma tumor, 96.70% for
meningioma tumor, and 98.69% for a pituitary tumor.
Figure 6 presents the receiver operating characteristic (ROC) curve of the proposed
classifier with the variation of different threshold values to shows the capability of the
classifier for identifying the brain tumor types correctly. The region under the ROC
curve for each type of tumor is 0.97, 0.96, and 0.98 for Glioma, Meningioma, and
Pituitary tumors respectively.
Multi-classification of Brain Tumor Images 997
Fig. 7. Variation of modeled classifier accuracy with different feature extraction techniques.
4 Conclusion
This paper introduced a hybrid feature extraction based techniques with the compo-
sition of ANN for classifying the brain tumor MRI images into three classes as glioma,
meningioma, and pituitary respectively. The performance of the ANN classifier in
terms of accuracy is quite improved than the modern classifier because of employing
the SWT and GLCM based hybrid feature extraction method that extracts the statistical
information of the preprocessed tumor images. Moreover, the proposed ANN classifier
gives a satisfactory classification accuracy of 96.2% that also improves the discrimi-
nation capability among three types of brain tumor images. Therefore, the modeled
classifier can provide an enormous impact on the clinical diagnosis and treatment of
brain cancer disease. In the future, the proposed approach can be applied to solve
another biomedical classification problem.
References
1. Rashid, M.H,, Mamun, M.A., Hossain M.A., Uddin, M.P.: Brain tumor detection using
anisotropic filtering, svm classifier and morphological operation from mr images. In: 2018
International Conference on Computer, Communication, Chemical, Material and Electronic
Engineering (IC4ME2), pp. 1–4. IEEE (2018)
2. Raju, A.R., Pabboju, S., Rao, R.R.: Brain image classification using dual-tree m-band
wavelet transform and naïve bayes classifier. In: Intelligent Computing in Engineering 2020,
pp. 635–642. Springer, Singapore (2020)
3. Zhang, L., Han, Z., Islem, R., Yaozong, G., Qian, W., Dinggang, S.: Malignant brain tumor
classification using the random forest method. In: Joint IAPR International Workshops on
Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern
Recognition (SSPR), pp. 14–21. Springer, Cham (2018)
Multi-classification of Brain Tumor Images 999
4. Siddiqui, M.F., Mujtaba, G., Reza, A.W., Shuib, L.: Multi-class disease classification in
brain MRIs using a computer-aided diagnostic system. Symmetry 9(3), 37 (2017)
5. Ismael, M.R., Abdel-Qader, I.: Brain tumor classification via statistical features and back-
propagation neural network. In: 2018 IEEE International Conference on Electro/Information
Technology (EIT), pp. 0252, IEEE (2018)
6. Ramdlon, R.H., Entin, M.K., Tita., K.: Brain tumor classification using mri images with k-
nearest neighbor method. In: 2019 International Electronics Symposium (IES), pp. 660–667.
IEEE (2019)
7. Sultan, H.H., Salem, N.M., Al-Atabany, W.: Multi-classification of brain tumor images
using deep neural network. IEEE Access 27(7), 69215–69225 (2019)
8. Gumaei, A., Hassan, M.M., Hassan, M.R., Alelaiwi, A., Fortino, G.: A hybrid feature
extraction method with regularized extreme learning machine for brain tumor classification.
IEEE Access 11(7), 36266–36273 (2019)
9. Kaplan, K., Kaya, Y., Kuncan, M., Ertunç, H.M.: Brain tumor classification using modified
local binary patterns (LBP) feature extraction methods. Med. Hypo. 25, 109696 (2020)
10. Cheng, J.: Brain tumor dataset (version 5) (2017) https://doi.org/10.6084/m9.figshare.
1512427.v5
11. Kumar, R., Gupta, A., Arora, H.S., Pandian, G.N., Raman, B.: CGHF: a computational
decision support system for glioma classification using hybrid radiomics-and stationary
wavelet-based features. IEEE Access 21(8), 79440–79458 (2020)
12. Vasant, P., Ivan, Z., Gerhard-Wilhelm, W.: Intelligent computing & optimization (ICO
2018). Vol. 866. Springer (2018)
13. Daoud, M., Mayo, M.: A survey of neural network-based cancer prediction models from
microarray data. Artif. Intell. Med. 1(97), 204–214 (2019)
14. Badwan, J.: Predicting the quality of mis characteristics and end-users’ perceptions using
artificial intelligence tools: expert systems and neural network. In: Intelligent Computing and
Optimization: Proceedings of the 2nd International Conference on Intelligent Computing and
Optimization 2019 (ICO 2019). Vol. 1072. Springer Nature (2019)
15. Mathew, A., Reema, P., Babu, A., Thara, N.K.: Brain tumor segmentation and classification
using DWT, Gabour wavelet and GLCM. In: 2017 International Conference on Intelligent
Computing, Instrumentation and Control Technologies (ICICICT), pp. 1744–1750. IEEE
(2017)
An Evolutionary Population Census
Application Through Mobile Crowdsourcing
1 Introduction
The census is the most serious and complex process which requires mapping the entire
country, mobilizing and training a large number of enumerators, conducting a massive
public campaign, canvassing all households, collecting individual information, com-
piling vast amounts of data in paper or electronic form, and analyzing and dissemi-
nating the data. In most developing countries, censuses are the primary source of data
and it helps to make the decision for doing the task which are required for country
development [1, 2].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 1000–1010, 2021.
https://doi.org/10.1007/978-3-030-68154-8_84
An Evolutionary Population Census Application 1001
In this section we present studies on the terminologies related to the system which are
important to understand Mobile Crowdsourcing process. This section also contains
brief discussion on related previous works.
There are many examples which are implemented using mobile crowdsourcing
concept. Some of example or related work is describe bellow:
Quora is the most popular website which is an example of crowdsourcing. The
crowd can answer the dropped questions simultaneously [18].
Uber provides a private car while on the go. This on-demand service is already
available in over forty countries and saves the waiting time in the taxi line. The app
allows riders to pay via PayPal or credit card, compare the fare rates and be picked up
from their current location in a few minutes [17].
Mobile Crowdsourcing System (MCS) consists of three main part namely MCS
Service Provider (SP), end user and MCS worker. SP accepts service requests from
MCS end users and deals with the requests. It selects proper MCS workers and assigns
relevant tasks to them. MCS’s end users are the requester of MCS services who can
make a request to the worker through SP. To perform the assigned task a mobile user
may act as a worker of MCS system.
The tools which are required for designing census application describe in below:
Java is an Object Oriented Programming (OOP) Language like PHP and C++. To
develop an android application using java, we need a development environment.
Android Studio is a very popular and powerful IDE development environment for java.
XML stands for Extensible Mark-up Language. XML is a design tool for android.
XML defines a set of rules for making a document to human-readable and machine-
readable format [8, 9].
To develop an application for Android platform android SDK (software develop-
ment kit) is used which have a set of developing tools.
Firebase is a mobile and web app development platform which provides many
services like phone and email authentication system, storage, real time database,
information security.
The Firebase Realtime Database is a cloud-hosted NoSQL database that stores and
synchronizes between users in Realtime. Firebase Authentication provides backend
services, easy-to-use SDKs and ready-made UI libraries to authenticate users to app.
3 Methodology
In this work we focus on developing of a crowdsource based census system that is able
to perform census process. This section mainly focuses on the overall system archi-
tecture of the proposed system and the procedure to achieve this in details.
system is replaced by digital census process using smart phone to reduce cost and time.
Figure 1 shows a general structure of census process that is implemented by the mobile
crowdsourcing system and machine learning.
4 Implementation Details
4.1 Implementation of Application Front-End
4.1.1 Phone Verification Module
A worker inputs his phone number in the required field and a verification code is sent to
this number within a few second. If the phone number is already active on this phone, it
automatically read the verification code.
Fig. 7. All modules captured from App front end in a single image
In the above Fig. 7, at first the worker should complete the registration process with
his mobile number and should confirm the cell number by submitting a verification
code upon receiving it from server. If someone haven’t any account yet he should first
create one and then have to log in with this user name and password. After logging in
he can add a voter, can see the voter list according to region. He can add the infor-
mation of a new voter by submitting the necessary data of that voter in the required
fields. This data includes his/her (voter) name, age, district, thana, sex, educational
qualification etc. then he can submit this information. At any time, a worker can see the
voter-list as well as the details information about these voters according to region.
In this chapter we have discussed about what and how the information is measured in
this system and the performance of our proposed application to survey the used
feedbacks. This survey was performed among 50 peoples. The experimental and survey
results are given in table and chart in the subsequent sections.
1008 I. H. Mukul et al.
Fig. 8. Pi-Chart for survey result ‘Yes’ (Left image) and ‘Need to develop’ (Right image)
An Evolutionary Population Census Application 1009
6 Conclusion
Now a days Mobile Crowdsourcing System (MCS) is the most important and very
much popular technology. Many applications are designed based on MCS. In com-
munication system or census process or business purpose, this technique is used
increasingly. MCS provides an easy and comfortable life to person or an organization.
Population Census process is more important issue for a country. Based on Census
result measure the change of country’s condition. In census time, it takes more time and
huge cost. MCS based census application provides an easy way to complete census
process of a country. To design this application main contribution of this giant is
android platform. Worldwide thousands of the developers are dedicated to develop
world class application in this platform. Besides android and, google establishes
Firebase, a mobile and web application platform in support of developing world class
applications. Firebase provides a high quality and secured database for the developer.
On the other hand, the most efficient developed application can be done by using the
Java library. As android is a derivative of Java, integration is quite flexible. Again,
using the Firebase platforms in the background makes the application secured.
References
1. Feng, W., Yan, Z., Zhang, H., Zeng, K., Xiao, Y., Hou, Y.T.: A survey on security, privacy,
and trust in mobile crowdsourcing. IEEE Internet Things J. (2018). https://doi.org/10.1109/
JIOT.2017.2765699
2. Feng, W., Yan, Z.: MCS-chain: decentralized and trustworthy mobile crowdsourcing based
on blockchain. Future Gen. Comput. Syst. (2019). https://doi.org/10.1016/j.future.2019.01.
036
3. Ma, Y., Sun, Y., Lei, Y., Qin, N., Lu, J.: A survey of blockchain technology on security,
privacy, and trust in crowdsourcing services. World Wide Web (2020). https://doi.org/10.
1007/s11280-019-00735-4
4. Wang, Y., Huang, Y., Louis, C.: Respecting user privacy in mobile crowdsourcing. Science
2(2), 50 (2013)
1010 I. H. Mukul et al.
5. Yuen, M.C., King, I., Leung, K.S.: A survey of crowdsourcing systems. In: Proceedings -
2011 IEEE International Conference on Privacy, Security, Risk and Trust and IEEE
International Conference on Social Computing, PASSAT/SocialCom 2011 (2011). https://
doi.org/10.1109/PASSAT/SocialCom.2011.36
6. Wang, Y., Huang, Y., Louis, C.: Towards a framework for privacy-aware mobile
crowdsourcing. In: Proceedings - SocialCom/PASSAT/BigData/EconCom/BioMedCom
2013 (2013). https://doi.org/10.1109/SocialCom.2013.71
7. Yuen, M.C., King, I., Leung, K.S.: A survey of crowdsourcing systems. In: Proceedings -
2011 IEEE International Conference on Privacy, Security, Risk and Trust and IEEE
International Conference on Social Computing, PASSAT/SocialCom 2011 (2011). https://
doi.org/10.1109/PASSAT/SocialCom.2011.36
8. Fang, C., Yao, H., Wang, Z., Wu, W., Jin, X., Yu, F.R.: A survey of mobile information-
centric networking: research issues and challenges. IEEE Commun. Surv. Tutorials (2018).
https://doi.org/10.1109/COMST.2018.2809670
9. Phuttharak, J., Loke, S.W.: Logiccrowd: A declarative programming platform for mobile
crowdsourcing. In: Proceedings - 12th IEEE International Conference on Trust, Security and
Privacy in Computing and Communications, TrustCom 2013 (2013). https://doi.org/10.
1109/TrustCom.2013.158
10. Liza, F.Y.: Factors influencing the adoption of mobile banking: perspective Bangladesh.
Glob. Discl. Econ. Bus (2014). https://doi.org/10.18034/gdeb.v3i2.164
11. Jonathon Rendina, H., Mustanski, B.: Privacy, trust, and data sharing in web-based and
mobile research: participant perspectives in a large nationwide sample of men who have sex
with men in the United States. J. Med. Internet Res. (2018). https://doi.org/10.2196/jmir.
9019
12. Pan, Y., de la Puente, M.: Census Bureau guideline for the translation of data collection
instruments and supporting materials: documentation on how the guideline was developed.
Surv. Meth. (2005)
13. U.S. Census Bureau. Population and Housing Unit Estimates. Annual Estimates of the
Resident Population for the United States, Regions, States, and Puerto Rico (2018)
14. Australian Bureau of Statistics. (2017). 2071.0 - Census of Population and Housing:
Reflecting Australia - Stories from the Census, 2016. Census of Population and Housing:
Reflecting Australia - Stories from the Census, (2016). https://doi.org/10.4018/978-1-59904-
298-5
15. United Nations Secretariat Dept. of Economic and social Affairs Statistics Division Census
Data Capture Methodology, New York, September 2009
16. https://digital.gov/2013/12/03/census-mobile-app-showcases-localstatistics/
17. https://www.maketecheasier.com/crowdsourcing-mobile-apps/
18. https://www.quora.com
IoT-Enabled Lifelogging Architecture
Model to Leverage Healthcare Systems
1 Introduction
The number of hospitals in urban areas is increasing day by day. Many hospitals
contain several hospital cabins. It is difficult to monitor the cabin individually
to avoid unwanted environmental conditions like fire occurrence, excessive noise,
improper room temperature, increasing CO2 , over-attendance of guests, smoking
inside the cabin, etc. Over-attendance of guests create noise and enhances the
room temperature and ratio of CO2 , which consequently disturbs the patients
unwanted situation occurs, the hospital authority can take proper steps. Here
we implement a wireless communication system, which is also cost-effective.
for monitoring air quality by using various gas sensors. Here the sensors data are
stored in an SD card and send the data through a wireless medium to the inter-
net(database) and the web server where the received data will calculate to draw
a final air quality result. They said that the data would transmit to the SD card
database only when a proper internet connection is available, but maintaining
a decent internet connection available for 24 hrs is not possible for developing
countries.
The authors in [25] proposed to develop a system for air quality monitoring
by using a wireless system. Here they wanted to design a system by using two gas
sensor and one temperature sensor. A microcontroller will receive the sensors’
data and then send it to a PC to calculate the values via a wireless media. If
the calculated result is not satisfactory, the MC will generate an alarm. In this
paper, they said to use wireless media, but which wireless media will use are not
defined. The wireless media they used for this system can only send 64 bytes
every five seconds. For that reason, they got a lower bit rate than they expected.
Sneha et al. [26] designed an embedded system for monitoring air quality.
Here, they proposed to develop a system by using air quality sensors and tem-
perature sensors. A microcontroller will receive the sensor data and then send
it to the processor to calculate the data and compare it with the given patient
sustainable value. If the calculated result is not satisfactory, it will send a mes-
sage over the phone to the authority to take the proper air purifier or humid-
ifier/dehumidifier. Here they want to develop the system by using an electrical
circuit. However, nowadays, wireless media are cheaper, easy install system, and
instinctive to use. For monitoring environmental issues and patient health con-
ditions, several works have been done. But most of the system designed either
for air quality monitoring or patient monitoring. Besides, most of the previous
work is based on a wired system and not that much instinctive and efficient
compared to our system. By making it completely wireless, it may become naive
for the related authority. Sometimes the environmental condition of a hospi-
tal cabin became worse, and the authority cannot take proper steps to control
the unwanted situation for not getting the information at the right time. Also,
a patient’s health condition may become down while the patient is alone in
the cabin. In our proposed system, we develop a system for the hospital cabin
to monitor the cabin environmental conditions and patient health condition at
some point. We want to implement the system using a wireless microcontroller
WeMos D1 and microcomputer Raspberry Pi 3, to wirelessly process the whole
work. Using a microcomputer like Raspberry Pi 3 will ensure efficacy, low cost,
scalability, low power usage, and overall proficiency.
3 System Description
3.1 System Architecture
Our system’s main purpose is to monitor the hospital cabins’ environment and
patient physical conditions centrally so that the hospital authority can take
proper steps to control the situation. In this proposed system, we used WeMos
Lifelogging Architecture for Healthcare 1015
D1 to take analog data from the sensors and convert them into digital data.
In each cabin of a hospital, there will be a WeMos D1 to take the sensor data
which will exist in each cabin. Then, WeMos D1 will continuously broadcast the
received data over IP address to the router. A Raspberry Pi 3 will perform the
central work. An IP address list exists in both the router and the Raspberry Pi
3. The Raspberry Pi 3 takes an IP address from the IP list and establishes a
connection with the WeMos D1, which holds that IP address. The Raspberry Pi
3 asks for data from the Wemos D1 through a router. The router then matches
the given IP address by the Raspberry Pi 3 with the IP address of the related
WeMos D1 from the IP address list, which it contains. Then the WeMos D1
sends the digital data to the Raspberry Pi 3 through the router. Raspberry Pi
3 checks the received data and compares it to the hospital authority’s given
standard values. If Raspberry Pi 3 founds any unusual data, it will inform the
admin. Based on that, the admin will take proper steps to control the unwanted
situation. Then, the Raspberry Pi 3 will move to the next WeMos D1 from the IP
address list. The overall system architecture and block diagram of our proposed
method are presented in Fig. 1 and Fig. 2, respectively.
The logic diagram if our system is depicted in Fig. 3. At first, each WeMos D1
retrieves the sensor values from five individual sensors. After that, the Raspberry
Pi scans the IP of each WeMos D1 from a list of IP address. By using the IP
of WeMos D1, Raspberry Pi will request for data. If the reading data is true,
Raspberry Pi will receive the sensor data where A1 is for the temperature sensor,
A2 is for the amount of CO2 sensor, A3 is for the movement sensor, A4 is for
the pulse sensor, and A5 is for noise/sound sensor. For each sensor, a threshold
1016 S. Zaman et al.
value will be given to the Raspberry Pi. If the receive data of A1 is greater
than 33◦ Celsius or A5 is greater than 10 dB, then set the alarm to the buzzer
and inform the admin computer; otherwise, the data of the next sensor will be
examined. If the receive data of A2 is greater than 4000 kPa, then the alarm will
set and inform to admin; otherwise, it will analyze the following sensor data.
Lifelogging Architecture for Healthcare 1017
If the data of A3 greater than 4.5◦ , it will inform the admin computer and set
the alarm; otherwise, it will proceed to the next sensor to read the data. If the
receiving value of A4 is greater than 16000bp, it will inform the admin computer
and set the alarm to the buzzer. If every sensor receives data is equal, or below
the threshold value, then the receiving data by the WeMos D1 is accepted. Then
the Raspberry Pi will change the IP of WeMos D1 to read the next WeMos D1
data, which is given to the IP list. The total system will proceed in the same
way.
where, T = 9.31 mV
Noise Sensor: For measuring noise value, we use two equations. For converting
the analog data into digital, we use the following equation:
Here, the resolution of ADC is 1024 as we use 10 bit ADC and system voltage
is 5 V. After measuring the converted voltage, we convert it into decibel as the
sound parameter is measuring in decibel. We use the following equation for
converting the data into decibel:
1018 S. Zaman et al.
Gas and Pulse Sensor: For gas and pulse sensor, the direct analog to digital
data is used.
4 System Implementation
In our proposed system, we used two controller boards. One is Raspberry Pi,
which can be regarded as a Nano computer, and the other one is a WeMos
D1 R2, which is a wireless communication-enabled micro-controller board. The
code related to WeMos D1 R2 imported in every WeMos D1 R2 with specific IP
addresses and IP sources. The analog reading of the sensors is passed to their
respective WeMos D1 R2 with the ADS1115 analog to digital converter module
for communicating with the analog sensors. Then the converted digital data is
processed for further data transmission to the Raspberry pi. PHP programming
language has been used to receive the data from the server, and Python program-
ming language to reprocess the data and execute the commands at Raspberry Pi
retrieved from the sensor reading through WeMos D1 R2 and ADS1115 module.
By analyzing these values, Raspberry Pi will notify the admins. The Wemos D1
R2 is the micro-controller that will be acting as a client in the proposed system.
The Raspberry Pi is the central controller of our system that acts as a server. In
the proposed system, the Raspberry Pi fits perfectly because it has 1 Ghz 64-bit
single-core BCM2835 with 512 MB ram, one micro USB port, a mini HDMI
port, and HAT-compatible 40 GPIO pins which can read and write digital data
which is more than enough for the system. Unlike Raspberry Pi 3, Raspberry Pi
also has built-in a wireless module with 802.11 b/g/n wireless LAN and Blue-
tooth 4.1 with BLE feature by which we can transmit the data faster with less
energy consumption. For implementing the system, we have used WeMos D1
R2, which is both Arduino and NodeMCU compatible board, including some
extra features. Unlike Arduino Uno, it has 11 digital input/output pins, and all
pins have interrupt/pwm /I2C /one-wire supported (except for D0). It has one
analog input (3.2 V max input) and has a micro USB connection port for both
communication and power source of 5v. It has a Power jack that can handle up to
9 − 24 V power input. However, the most useful feature is that it uses an ESP-
8266EX micro-controller, enabling wireless communication and providing 4MB
of flash memory. It can be operated in both 80 mhz and 160 mhz clock speed.
Unfortunately, WeMos D1 R2 has only one analog pin built-in, in which it lacks
from Arduino Uno. But, WeMos D1 R2 has a built-in I2C port, which makes a
remedy to the problem. I2C is a multi-master-slave serial computer bus that is
used for attaching lower-speed peripheral ICs to the micro-controller. Using I2C
port, ADS1115, a 16bit, four programmable channel analog to digital converter,
the other sensors have been connected. ADS1115 has ten connectors. They are
SDA, SCL, Address pin, Alert pin, and the other four connectors are analog pins
Lifelogging Architecture for Healthcare 1019
denoted with A0-A3. The ADS1115 uses SDA and SCL pin of WeMos D1 R2 to
communicate with it. The address pin is for configuring the bit rate. We will be
using 16bit from making the threshold more precise. The four-analog pin is used
for taking data for the analog sensor.
The system uses a total of five sensors for collecting the data of the hospital
cabin. They are LM35 temperature sensor, KY−038 microphone sound sensor,
MG811 CO2 gas sensor module, ADXL345 Tilt/movement sensor, and pulse
sensor SEN−11574. The LM35 temperature sensor has been used that acts for
change in surroundings temperature and drops 10 mV for per 1◦ C change is the
temperature; this is what helps to measure the environment’s temperature. It has
three output pins, and from the plain side of the sensor first one is VCC, which
can handle from 4v to 30v. The second one is the signal pin, which gets connected
with the A0 port of the WeMos D1 R2 board, and the last one the ground pin.
The KY-038 microphone, sound sensor module, has been used to detect the
sound level in a cabin. It consists of four pins, which are VCC, GND, A0, and
D0. The VCC and GND pins are for power input, which should be constant 5v.
This can provide both analog and digital output. But as we measure the sound
level, we need to use the A0 pin, which would be connected to the A0 pin of
the ADS1115 module. The value returned from ADS1115 will be threshold by
WeMos D1 R2. MG811 gas sensor module has an MG811 gas sensor embedded
with the L293 interface module. The MG811 sensor is sensitive to CO2 gas and
operated with the voltage of 5–6 V. The MG811 gas sensor has six pins, but the
L293 interface turns it into four. These four pins are the same as the KY-038
sensor module. The ADXL345 Tilt sensor is a small, thin, low powered 3-axis
MEMS sensor, also known as an accelerometer sensor with a high resolution
of 13bit measurement at up to (+−) 16g. This sensor can be communicated
via both SPI and I2C digital interface. We will be using the I2C interface for
communicating with the sensor. The sensor can be operated from 2.0−3.6 V with
a 40 uA current supply. The sensor can sense the static acceleration of gravity
and dynamic acceleration resulting from motion or shock. The sensor is set under
the cabin bed of the patient to sense the patient’s inappropriate behavior. The
pulse sensor SEN-11574 will be the only sensor that will be physically connected
with the patient to measure the patient’s pulse rate. The sensor is very straight
forward and has three output pins: VCC, GND, and signal pin. As the pulse
sensor is analog, a graph can be directly generated from its sensor value via
Arduino IDE. The sensor will be connected through the ADS1115 module, and
the data will be processed for providing precise human-readable output.
should be connected to other three digital pins according to the software code
declaration.
Step 7 (Connecting Pulse Sensor): The signal pin of the pulse sensor is
connected to the A2 pin of the ADS1115 module. The sensor operates in 2.0–3.6
V, so the VCC should be connected with the 3.3 V port of WeMos.
Step 8 (Connecting Buzzer with Raspberry Pi): The buzzer operates with
5v. So, the buzzer’s positive pin should be connected with the Pin 23 of the Pi
Zero, and the other pin should be grounded.
Step 9 (Powering up the devices): Both WeMos D1 R2 and Raspberry
Pi operate with a 5v power supply. But two devices should be powered with
separate power supply as the system is wireless. For the perfect operation of the
system, the router should also be powered up.
In Fig. 4 and Fig. 5, we presented the connection of WeMos with the sensors
and circuit diagram of raspberry pi and buzzer.
Fig. 8. Buzzer turns ON when movement sensor value (= 5.26) exceeds threshold (=
4.5)
Fig. 9. Buzzer turns ON when the sensors value greater than threshold and turns OFF
while the sensors value less than threshold.
6 Conclusion
This paper implements an autonomous wireless communication system to
observe the hospital cabins’ surrounding environment and patients’ health con-
dition. In this system, each cabins sensor data can report to the admin if the
received data is less than or greater than the pre-set threshold values. The advan-
tage of our system is that a hospital’s overall situation can be monitored centrally
with little cost and effective setup. This approach would reduce hospitals’ man-
ual activity, and the hospital authority would be acknowledged immediately if
any unusual condition occurs. This system is easy to operate so any user can
easily handle it. This system will help the hospital authority to monitor all the
cabins centrally for 24 h and take proper steps at the right time if any unwanted
situation occurs. Our proposed implementation scheme can be adapted to other
sectors of IoT applications, e.g., smart transportation, smart city, smart liv-
ing, and so on. In future, we have a plan to discover pattern, identify behavior
and patient status using the extracted data. We also have a plan to optimize
1024 S. Zaman et al.
the energy consumption of the IoT devices and perform local training without
sharing the sensitive information with any external entity.
References
1. Lee, I., Lee, K.: The Internet of Things (IoT): applications, investments, and chal-
lenges for enterprises. Bus. Horiz. 58(4), 431–440 (2015)
2. Madakam, S., Lake, V., Lake, V., Lake, V.: Internet of Things (IoT): a literature
review. J. Comput. Commun. 3(05), 164 (2015)
3. Farooq, M.U., Waseem, M., Mazhar, S., Khairi, A., Kamal, T.: A review on internet
of things (IoT). Int. J. Comput. Appl. 113(1), 1–7 (2015)
4. Singh, A., Payal, A., Bharti, S.: A walkthrough of the emerging IoT paradigm:
visualizing inside functionalities, key features, and open issues. J. Netw. Comput.
Appl. 143, 111–151 (2019)
5. Gubbi, J., Buyya, R., Marusic, S., Palaniswami, M.: Internet of Things (IoT): a
vision, architectural elements, and future directions. Future Generation Comput.
Syst. 29(7), 1645–1660 (2013)
6. Marjani, M., Nasaruddin, F., Gani, A., Karim, A., Hashem, I.A.T., Siddiqa, A.,
Yaqoob, I.: Big IoT data analytics: architecture, opportunities, and open research
challenges. IEEE Access 5, 5247–5261 (2017)
7. Imteaj, A., Rahman, T., Hossain, M.K., Alam, M.S., Rahat, S.A.: An IoT based fire
alarming and authentication system for workhouse using Raspberry Pi 3. In: 2017
International Conference on Electrical, Computer and Communication Engineering
(ECCE), pp. 899–904. IEEE (2017)
8. Imteaj, A., Rahman, T., Hossain, M.K., Zaman, S.: IoT based autonomous percip-
ient irrigation system using raspberry Pi. In: 2016 19th International Conference
on Computer and Information Technology (ICCIT), pp. 563–568. IEEE (2016)
9. Zhu, Z.-T., Ming-Hua, Yu., Riezebos, P.: A research framework of smart education.
Smart Learn. Environ. 3(1), 4 (2016)
10. Imteaj, A., Amini, M.H.: Distributed sensing using smart end-user devices: path-
way to federated learning for autonomous IoT. In: 2019 International Conference
on Computational Science and Computational Intelligence (CSCI), pp. 1156–1161.
IEEE (2019)
11. Kalaycı, B., Özmen, A., Weber, G.W.: Mutual relevance of investor sentiment and
finance by modeling coupled stochastic systems with MARS. Ann. Oper. Res. 1–24
(2020)
12. Imteaj, A., Rahman, T., Begum, H.A., Alam., M.S.: IoT based energy and gas
economic home automation system using Raspberry Pi 3. In: 2017 4th International
Conference on Advances in Electrical Engineering (ICAEE), pp. 647–652. IEEE
(2017)
13. Reyna, A., Martı́n, C., Chen, J., Soler, E., Dı́az, M.: On blockchain and its inte-
gration with IoT. challenges and opportunities. Future Generation Comput. Syst.
88, 173–190 (2018)
14. Chen, S., Hui, X., Liu, D., Bo, H., Wang, H.: A vision of IoT: applications, chal-
lenges, and opportunities with china perspective. IEEE Internet of Things J. 1(4),
349–359 (2014)
15. Intelligent Computing and Optimization, Proceedings of the 2nd International
Conference on Intelligent Computing and Optimization 2019 (ICO 2019). Springer
International Publishing, ISBN 978-3-030-33585 -4
Lifelogging Architecture for Healthcare 1025
Abstract. Frequent Pattern Mining (FPM) has been playing an essential role in
data mining research. In literature, many algorithms have proposed to discover
interesting association patterns. However, frequent pattern mining in a large-
scale dataset is a complex task because of a prohibitively large number of the
smaller pattern need to be generated first to identify the extended pattern. The
complexity in terms of computational time is a big concern. In this paper, we
have proposed an updated novel FPM algorithm that uses a boolean matrix and
then decomposed the matrix vertically. Here we have reduced the computational
time by reducing a few steps and added itemset if and only if satisfy the
condition instead of pruning after added into the candidate itemset. Our pro-
posed algorithm has reduced the computational time at least two times on the
generation of the boolean matrix (i.e. Q), frequent-1 matrix (i.e., F1) and
frequent-1 itemset (i.e., L1), which has improved the efficiency of the algorithm.
1 Introduction
The frequent pattern mining (FPM) is using to identify the association rules among the
items which are present in the dataset. This technique is also known as Association
Rules Mining (ARM). In FPM, by applying some statistical methods able to generate
compelling and unrecognized patterns, itemset, substructures from the dataset. One of
the essential techniques is FPM, which enables to discover important feature from the
given itemset. In many data mining techniques, the FPM used as a base technique to
analyze, like association, relationship, and correlation.
Many algorithms have been proposed to make more intense the techniques in FPM
by different researchers. Though, improvements are still required to increase the per-
formance by reducing the computational time of the existing FPM algorithms. Some of
them are the well-reputed algorithms of FPM includes Apriori [1–3], FP Growth [1–4],
Partitioning [1], Dynamic Itemset Counting (DIC) [6], etc. One of the base algorithms
is Apriori to discover the frequent pattern. The significant challenges of this algorithm
are lengthy computational time and massive consumption of memory due to the
multiple time scan of the transactional dataset. To reduce computational time and
memory consumption a boolean load-matrix based frequent pattern mining has been
developed [15]. Hence, improvements are still required to increase the performance of
the algorithm in terms of time complexity.
In our research paper, we have proposed an improved boolean load-matrix based
novel algorithm to reduce computational time.
This paper is structured as follows. In Sect. 2, we have presented a few previous
researchers that have been done in the literature. In Sect. 3, we have proposed an
improved Boolean load-matrix based novel algorithm. In Sect. 4, we have discussed on
the obtained result with example. In Sect. 5, performance evaluation and complexity
analysis between the existing and proposed algorithm. Finally, in Sect. 6, we have
concluded our research paper.
2 Related Work
Frequent pattern mining is a significant area of data science. Using the FPM huge
amount of are mined to explore some mysterious pattern or subsequence for generating
the association rules. Several algorithms have been proposed by different researchers to
make more intense the techniques in FPM. Still, there are some scopes to improve the
algorithm efficiency in terms of execution time means by reducing the execution time
while finding the frequent pattern. The work mentioned in [1, 3] proposed an efficient
algorithm to mine the association rules from the transaction dataset. In [1] proposed an
improved algorithm to generate a new frequent itemset but still needed to scan the
previous frequent itemset to computed support that also a costly process. This proposed
algorithm avoids the pruning technique instead of using the pruning technique we have
included in the candidate itemset only while satisfying the criteria. Many hybrid
Apriori algorithms have developed and proposed by different researchers [3, 7–9]. In
[2], Author proposed eight performance metrics, such as execution time, memory
consumption, completeness, and interestingness for association rules mining. In [3],
AprioriTid has used to present a hybrid algorithm to generate the association rules for
discovering the large, frequent itemset. From paper [3], we have influenced and start
work to create large frequent-itemset. In [4], Author proposed a frequent pattern tree
(FP-tree) structure, an extended prefix tree structure to store compressed, crucial
information about the frequent patterns. In [5], Athors reviewed and presented a
comparison of different algorithms for Frequent Pattern Mining (FPM) to develop a
more efficient FPM algorithm. In [6], They proposed an algorithm for finding large
itemsets which use fewer passes over the data than classic algorithms and uses fewer
candidate itemsets than methods-based sampling. In [7], Based on confidence and
classification, another Apriori algorithm has been suggested. In [8], a new FPM
algorithm has been proposed, which has given time consumption, CPU utilization, and
efficient memory utilization during knowledge discovery from datasets by combining
the Weighted Apriori algorithm and Apriori Hash T algorithm. Weighted Apriori,
Itemset combination will be generated frequently, and it will increase the candidate
itemset. The weight computation for each transaction takes more time to execute, with
1028 S. Roy et al.
less accuracy. Apriori Hash T, high computational requirement, high memory uti-
lization. In [9] proposed an Apriori-Growth algorithm based on the Apriori algorithm
and FP-tree structure to mine the frequent pattern.
The work reported in [10], by analyzing the Apriori algorithm, researchers pro-
posed a DC_Apriori algorithm, which has increased the efficiency of the algorithm.
Here they avoided the generation of repeated candidate itemset. They also compared
their proposed algorithm with the MC_Apriori algorithm [11] based on the compressed
matrix.
In the research [12], researchers proposed an improved algorithm, which includes
the following improvement. To generate the candidate items by decreasing the number
of scans the dataset, to pruning frequent itemset, and candidate itemset generation to
improve joining efficiency and also used overlap strategy to count support to increase
the efficiency of the algorithm. In [13], An improved Apriori algorithm has been
proposed, where dataset scan performed only once, and all the transactions have
transformed into the components of a two-dimensional array with a weight value. An
improved algorithm proposed in [14], In this paper count-based method used for
pruning the candidate itemset to reduce dataset scans.
In the research [15], researchers proposed an algorithm to discover the frequent
pattern using a boolean matrix base where researchers reduced the dataset scan time. In
this paper, they transformed the transactional dataset into the Boolean matrix, which
has is using for further process. Researchers also used the pruning process to delete the
itemset from the candidate itemset if each subset of an itemset is not present in the
previous frequent itemset and finally combined all frequent itemset to discover the final
frequent itemset.
In the work reported in [16], researchers compared Apriori and Eclat algorithms for
association rules mining by applying them on a real-world dataset. This paper high-
lighted how it becomes more computationally expensive for Apriori compared to Eclat
to mine frequent patterns at very low support threshold levels. In [17], A new algorithm
has been proposed that combines ECLAT algorithm and a method that uses special
candidates to find frequent patterns from the database.
By studying the above research papers, we have gathered some knowledge and
found a tremendous amount of research completed by many researchers on frequent
pattern mining (FPM). Day by day, the data volume is increasing. From that massive
amount of data, find out the frequent pattern with minimum computational time is still
challenging. For the sake of that, in this paper, we have proposed an improved boolean
load-matrix based novel algorithm to reduce the computational time that also usable in
a distributed system to share the data loads among the available nodes.
3 Proposed Algorithm
Input: D, min_sup_count
Output: L
1: Scan D
2: Construct Qm×n and set sup_count for each item set
for each i, i ≤ m set sup_counti = 0;
for each j, j ≤ n
if ITj Ii then Q[i][j] = 1 and sup_counti += 1;
else Q[i][j] = 0; end if
end for
if (sup_counti ≥ min_sup_count) then set
sup_count[i] = sup_counti; add the tuple i to F1
matrix; add the item to L1; end if
end for
3: Vertically split Qm×n into multiple matrices such as
Load1m×x, Load2m×x, ..., Loadtm×x, where x > 0.
4: for each i, i ≤ m
for each j, (i + 1) ≤ j ≤ m
if q1[i] ≠ q2[j], q1, q2 Lk−1 then
q = q1[i] q2[i]; include_in_ck = true;
for (k-1) subset p, p q
if each p Lk-1 then include_in_ck = false;
break; end if
end for
if include_in_ck = true then Ck = Ck q; end if
end if
end for
end for
5: for each r, r Ck
compute sup_itemset = Load1.r + Load2.r + ...+
Loadt.r.
if sup_itemset ≥ min_sup_count then add it to Lk; end
if and end for
6: Repeat steps 4 to 5 till Lk−1 ≠ .
7: L = Li.
In this algorithm, the m and n represent total no of items and transactions in the
dataset D, respectively. The min_sup_count is the predefined minimum support count
threshold, and Q is a Boolean matrix with size m n. The items are stored in the rows,
and transactions of an item are stored in the columns. Lk is the frequent k itemset that
has generated from Ck, where Ck is the candidate itemset that has developed from Lk-1
itemset, whereas k 2. After scanning dataset D, the Q matrix has constructed. In the
1030 S. Roy et al.
matrix Q, if the item is present in a specific transaction, then put decimal value one
(i.e.,1) otherwise zero (i.e., 0), at the same time calculating support count and stored in
sup_count array.
Here, sup_count is a two-dimensional array where the row contains itemset, and the
column contains the support count value of the respective itemset. After scanning each
tuple, compare with min_sup_count count value to determine whether it goes to fre-
quent k itemset or not. Further, based on the number of nodes, decomposed the fre-
quent one itemset into multiple loads (i.e., Load1, Load2, …., Loadt). To calculate the
support count of frequent k itemset, we have applied the bitwise AND operator (i.e., &)
on the decomposed matrix.
Algorithm 1 shows a pseudo-code of improved boolean load-matrix-based FPM
algorithm and the improvement part of the algorithm is highlighted (step 2). Initially,
dataset D scanned then a boolean matrix (i.e. Qmn) is constructed and calculated
support count for each row i, which is stored in sup_count[i] and compared with
min_sup_count. If sup_count[i] min_sup_count, then enter rows is added to the k-
1 matrix (i.e. F1) and listed in k-1 itemset (i.e. L1). After that, vertically split the
Q matrix into multiple matrices (i.e., Load1, Load2, Load3, …., Loadt) where the size
of each matrix size is m x where x is a positive number. Here the matrix is
decomposed based on the available nodes to use in a distributed environment and each
load is assigned to individual nodes to execute.
Now, join operation is performed between two sets of items from Lk-1 items (i.e., q)
and joined such a way that q1[i] \ q2[i] = ⌀. Then the k-1 subsets are generated from
q (i.e., p) and if each p is present in Lk-1, then stored in Ck. Then, calculate the support
count of each itemset r 2 Ck (i.e., sup_itemset) from the simulation of Load.r values.
Then, Load.r is computed by applying bitwise AND operation (i.e., &) on the Load
matrices, which is stored binary digits. Finally, combining all the frequent itemset (i.e.,
{L1 [ L2 [ L3 [ …… Lk-1 [ Lk}) generated the final frequent itemset L.
In this segment of the paper, we have discussed the proposed algorithm. Let assume,
the transactional dataset D presented in Table 1, which has contained the set of items
I = {I0, I1, I2, I3, I4, I5, I6, I7, I8} and set of transactions T = {IT0, IT1, IT2, IT3, IT4, IT5,
IT6, IT7, IT8, IT9}. The min_sup_count is a predefined threshold for minimum support
count. Ck is a candidate k itemset that has generated from k-1 itemset where k 2.
The main intention of this researches is to find out the large, frequent itemset with less
time consumption to draw the association rules. In this discussion, according to our
proposed algorithm, we have presented as a transactional dataset in Table 1, which will
convert as the boolean matrix. In this matrix, we have shown the items as rows and the
transactions as columns.
An Improved Boolean Load Matrix-Based Frequent Pattern Mining 1031
We have constructed the boolean matrix Q by scanning each itemset Ii. If specific
transaction ITi present, then we have inserted 1 for the position Q[Ii][ITi], if not present,
then inserted 0. For example, if IT1 2 I1, then we have inserted value one (i.e.,1) at the
position [I1] [IT1] of boolean matrix Q (i.e., Q[I1] [IT1] = 1). Again IT2 2 I1, so we
have inserted value one (i.e.,1) at the position [I1] [IT2] of boolean matrix Q (i.e., Q[I1]
[IT1] = 1). But IT5 62 I1, so we have inserted zero (i.e.,0) at the position [I1] [IT5] of
boolean matrix Q (i.e., Q[I1] [IT5] = 0). In the same way, we have prepared the entire
boolean matrix represented in Table 2. At the same time, we have calculated support
count (i.e., sup_count) for each row and stored it into the sup_count array with the
same row indicator. In this analysis, we have considered value four (i.e., 4) as the
minimum support count (i.e., min_sup_count = 4). After calculating the support count
of each row, we have compared with the minimum support count (i.e., min_sup_count)
to determine whether it goes to frequent-1 itemset or not.
Table 2. Matrix 1 with all items; some of them will not include in the frequent matrix but just
shown to explain.
IT0 IT1 IT2 IT3 IT4 IT5 IT6 IT7 IT8 IT9 sup_count
I0 1 1 0 1 1 0 1 0 0 0 5
I1 0 1 1 1 1 0 1 0 1 0 6
I2 0 0 0 0 0 1 1 0 0 1 3
I3 1 0 0 1 0 1 1 1 0 1 6
I4 0 1 1 0 0 1 1 0 1 0 5
I5 1 1 1 0 1 0 1 0 1 1 7
I6 0 0 0 0 1 1 0 1 0 0 3
I7 1 0 0 0 0 0 0 1 1 1 4
I8 0 1 0 0 1 0 1 1 0 1 5
Table 3 represents all items with boolean values and respective row sup_count.
After comparing the minimum support count value with support count value, we
1032 S. Roy et al.
figured out that row no 3 and 7 will not include in the frequent-1 matrix, which is F1.
All items of frequent matrix-1 are listed as frequent itemset, which is L1.
After that, we have split the dataset into multiple loads based on the number of
available nodes. In this example, we have considered-2 nodes available. So, we have
divided the dataset into the two loads, which are Load1 and Load2 that represented in
Table 4.
Then we have to perform the join operation on frequent-1 itemset (i.e., L1) to
generate candidate-2 itemset, which is C2. After completing the join operation, we have
to check that each subset of a joined itemset is present in the previous itemset (i.e., Lk-1)
or not. If each subset of a joined itemset is present in the Lk-1 itemset, then include the
joined itemset into the candidate-2 itemsets; otherwise, the joined itemset will not
involve in the candidate-2 itemsets. After performing these steps, we will get C2, which
is shown below.
An Improved Boolean Load Matrix-Based Frequent Pattern Mining 1033
fI0 ; I1 g; fI0 ; I3 g; fI0 ; I4 g; fI0 ; I5 g; fI0 ; I7 g; fI0 ; I8 g; fI1 ; I3 g; fI1 ; I4 g; fI1 ; I5 g; fI1 ; I7 g; fI1 ; I8 g;
C2 ¼
fI3 ; I4 g; fI3 ; I5 g; fI3 ; I7 g; fI3 ; I8 g; fI4 ; I5 g; fI4 ; I7 g; fI4 ; I8 g; fI5 ; I7 g; fI5 ; I8 g; fI7 ; I8 g
All subset of C2 present in L1 (i.e., C2 2 L1). Now we have to calculate the support
count of all itemset to obtain frequent itemset 2 (i.e., L2). Support count calculated from
the candidate-2 itemsets by performing bitwise AND operator (i.e., &) on each row
values for individual load matrices. In this example, to calculate the support count for
{I0, I1}, we have to perform bitwise AND (i.e., &) operation between the row values of
I0 and I1 from load1 and load2 matrix. For load1 calculated value is 3 (i.e. {1, 1, 0, 1,
1} & {0, 1, 1, 1, 1}) and for load2 value is 1 (i.e. {1, 1, 0, 1, 1} & {0, 1, 1, 1, 1}). To
calculate the support count of {I0, I1}, we have to add all load matrices result, which is
4 (i.e., 3 + 1 = 4). In the same way, we have to calculate the support count value of
remaining itemset for candidate-2 itemset (i.e., C2). The support count calculation
process of each itemset for C2 is exhibit below.
Now we have to compare each itemset support count value with the minimum
support count value to determine whether it will go to the frequent itemset-2 or not. If
the support count of an itemset is greater than the minimum support count value (i.e.,
min_sup_coun), then added to the frequent itemset-2 (i.e., L2). L2 is shown below.
Similarly, we have generated candidate itemset 3 with all possible itemset (i.e., C3),
which has generated from L2 by performing join operation with L2 itemset itself.
1034 S. Roy et al.
We did not include some of the itemsets that are not present in L2. We have included if
each subset is present in L2 (i.e., C3 2 L2). For example, one possible itemset was {I0,
I1, I4}. Possible subset of {I0, I1, I4} with max k-1 items (i.e. 2) are {I0, I1}, {I0, I4}, {I1,
I4} as {I0, I4} 62 L2 that is why we did not include {I0, I1, I4} itemset into the C3. The
candidate itemset 3 (i.e., C3) demonstrates below.
After getting the final candidate itemset-3, we have to calculate the support count of
each itemset, which demonstrates bellow.
Since we have found both itemsets support count value if greater than or equal to
minimum support count (i.e., 4), so they have added into the frequent itemset-3 (i.e.,
L3), the L3 demonstrates below.
L3 ¼ fI1 ; I4 ; I5 g
Now, we have to find out the candidate itemset-4 by performing the join operation
on the L3 itemset itself. But we have found that L3 contains only one itemset so we
cannot complete the join operation. So, we will not get any other candidate set.
Finally, we have to generate the final frequent itemset by combining all the frequent
itemset. Which is L = {L1 [ L2 [ L3 [ L4 [ …. Lk-1 [ Lk}. The last large, frequent
itemset (i.e., L) is presented below from the given dataset.
fI0 g; fI1 g; fI3 gfI4 g; fI5 g; fI7 g; fI8 g; fI0 I1 g; fI0 I5 g
L¼
fI1 I4 g; fI1 I5 g; fI4 I5 g; fI5 I8 g; fI1 ; I4 ; I5 g
The performance evaluation is made by comparing the proposed method with the
existing method [15]. The experimental data were generated from a machine with the
configurations 2.90 GHz, Intel(R), Core (TM)-i7 7820HQ CPU, 16 GB main memory.
Python Language is used to implement the proposed design. Both existing method [15]
and Proposed method are coded and execute in the same platform. In this performance
evaluation, the one Load is considered. Apriori and FP growth Algorithm dataset from
Kaggle [18] is used for the performance evaluation. The dataset contains 12 items and
12526 transactions. In this experiment the minimum support count (min_sup_count) is
An Improved Boolean Load Matrix-Based Frequent Pattern Mining 1035
set 20 for mining the dataset. Noted that the minimum support count value can be
varied depends on needs.
In this experiment, we have found that our proposed algorithm has given at least
two times better performance than the existing boolean load-matrix based algorithm in
terms of computational time on the generation of the boolean matrix (i.e., Q), frequent-
1 matrix (i.e., F1) and frequent-1 itemset (i.e., L1). The frequent-1 matrix (i.e., F1) and
frequent-1 itemset (i.e., L1) generation execution time for existing and proposed
algorithms.
In Eq. (1) shows the time complexity to construct the boolean matrix (i.e., Q),
frequent-1 matrix (i.e., F1), and frequent-1 itemset (i.e., L1) of the existing algorithm.
In Eq. (2) shows the time complexity to generate the boolean matrix (i.e., Q), frequent-
1 matrix (i.e., F1), and frequent-1 itemset (i.e., L1) of the proposed algorithm. Equa-
tion (2) is for step-2 of the proposed algorithm. Here T is the time complexity.
Fig. 1. Comparison result between existing and proposed boolean-load matrix FPM algorithms.
6 Conclusion
References
1. Fang, X.: An improved apriori algorithm on the frequent itemset. In: International
Conference on Education Technology and Information System (ICETIS 2013). IEE (2013)
2. Ghafari, S.M., Tjortjis, C.: A survey on association rules mining using heuristics. In:
Interdisciplinary Reviews: Data Mining and Knowledge Discovery, vol. 9, no. 4, p. e1307.
Wiley (2019)
An Improved Boolean Load Matrix-Based Frequent Pattern Mining 1037
3. Nath, B., Bhattacharyya, D.K., Ghosh, A.: Incremental association rule mining: a survey. In:
Interdisciplinary Reviews: Data Mining and Knowledge Discovery, vol. 3, no. 3, pp. 157–
169. Wiley (2013)
4. Han, J., Pei, J., Yin, Y., Mao, R.: Mining frequent patterns without candidate generation: a
frequent-pattern tree approach. In: Data Mining and Knowledge Discovery, vol. 8, no. 1,
pp. 53–87. Springer, Netherlands (2004)
5. Chee, C.-H., Jaafar, J., Aziz, I.A., Hasan, M.H., Yeoh, W.: Algorithms for frequent itemset
mining: a literature review. In: Springer's Artificial Intelligence Review, pp. 1–19 (2018)
6. Brin, S., Motwani, R., Ullman, J.D., Tsur, S.: Dynamic itemset counting and implication
rules for market basket data. In: ACM SIGMOID Record, pp. 255–264 (1997)
7. Chun-sheng, Z., Yan, L.: Extension of local association rules mining algorithm based on
apriori algorithm. In: International Conference on Software Engineering and Service
Science, pp. 340–343. IEEE, China (2014)
8. Thakre, K.R., Shende, R.: Implementation on an approach for Mining of Datasets using
APRIORI Hybrid Algorithm. In: International Conference on Trends in Electronics and
Informatics, pp. 939–943. IEEE, India (2017)
9. Wu, B., Zhang, D., Lan, Q., Zheng, J.: An efficient frequent patterns mining algorithm based
on apriori algorithm and the FP-tree structure. In: International Conference on Convergence
and Hybrid Information Technology, vol. 1, pp. 1099–1102. IEEE, South Korea (2008)
10. Du, J., Zhang, X., Zhang, H., Chen, L.: Research and improvement of apriori algorithm. In:
International Conference on Information Science and Technology, pp. 117–121. IEEE,
China (2016)
11. Yu, N., Yu, X., Shen, L., Yao, C.: Using the improved apriori algorithm based on
compressed matrix to analyze the characteristics of suspects. Int. Express Lett. Part B Appl.
Int. J. Res. Surv. 6(9), 2469–2475 (2015)
12. Yuan, X.: An improved apriori algorithm for mining association rules. In: Conference
Proceedings, vol. 1820, no. 1. AIP (2017)
13. Liao, B.: An improved algorithm of apriori. In: International Symposium on Intelligence
Computation and Applications (ISICA), pp. 427–432. Springer, Berlin (2009)
14. Wu, H., Lu, Z., Pan, L., Xu, R., Jiang, W.: An improved aprioribased algorithm for
association rules mining. In: International Conference on Fuzzy Systems and Knowledge
Discovery, vol. 2, pp. 51–55, IEEE, China (2009).
15. Sahoo, A., Senapati, R.: A Boolean load-matrix based frequent pattern mining algorithm. In:
International Conference on Artificial Intelligence and Signal Processing (AISP), IEEE,
India (2020). ISSN:2572-1259
16. Robu, V., Santos, V.D.: Mining frequent patterns in data using apriori and eclat: a
comparison of the algorithm performance and association rule generation. In: 6th
International Conference on Systems and Informatics (ICSAI). IEEE, China (2019).
ISBN:978-1-7281-5257-8
17. Jin, K.: A new algorithm for discovering association rules. In: International Conference on
Logistics Systems and Intelligent Management (ICLSIM), ISBN 978-1-4244-7331-1, IEEE,
China (2010)
18. (Dataset for Apriori and FP growth Algorithm). https://www.kaggle.com/newshuntkannada/
dataset-for-apriori-and-fp-growth-algorithm
Exploring CTC Based End-To-End
Techniques for Myanmar Speech Recognition
1 Introduction
2 Related Work
Most of the early works in Myanmar speech recognition are mostly Hidden Markov
Model (HMM) based systems. Hidden Markov Model is used together with Gaussian
Mixture Model (GMM) and Subspace Gaussian Mixture Model (SGMM) and the
performance of the model increases with the increased use of training data [9].
HMM-GMM models are also used in spontaneous ASR systems and tuning
acoustic features, number of senones and Gaussian densities tends to affect the per-
formance of the model [10].
Since the Deep Neural Networks (DNNs) have gained success in many areas, they
are also used in automatic speech recognition systems usually in the combination with
HMMs [11–14]. Research has shown that the DNN, CNN, and Time Delay Network
models outperformed the HMM-GMM models [14].
However, many of these works needed several components where we have diffi-
culties to tune each individual component. End-to-end ASR models simplified these
components into a single pipeline and achieved state-of-the-art results in many sce-
narios. To the best of our knowledge, no research has been conducted on end-to-end
training in Myanmar ASR systems. In this paper, we introduce an end-to-end Myan-
mar ASR model with the CTC based approach.
1040 K. M. M. Chit and L. L. Lin
This section introduces the CTC based end-to-end model architecture. We explore the
architectures with the initial layers of VGGNet [15] and up to 6 layers of BLSTM.
Batch Normalization is applied after every CNN and BLSLM layer to stabilize the
learning process of the model. A fully connected layer with the softmax activation
function is followed after the BLSTM layers. The CTC loss function is used to train the
model. Figure 1 shows the model architecture, which includes a deep CNN encoder and
a deep bidirectional LSTM network.
Label
CTC Loss
Fully Connected
BLSTM
Spectrogram
Fig. 1. The architecture of CTC based end-to-end ASR model. The different architectures are
explored by varying the number of BLSTM layers from 3 to 6 and by removing and adding deep
convolutional layers.
Fig. 2. The architecture of convolutional layers (VGGNet). Each convolutional block consists of
two convolutional layers with Batch Normalization and a max pooling layer. Each convolutional
layer uses the ReLU activation function.
A stack of BLSTM layers is followed after the CNN layers. The models with 3 to
6 layers of BLSTM are experimented using 512 hidden units in each layer and
direction. Since the Batch Normalization is shown to improve the generalization error
and accelerate training [16], Batch Normalization is added after every BLSTM layer.
4 Experimental Setup
4.1 Description for Myanmar Speech Corpus
Due to the scarcity of public speech corpora for the Myanmar language, we build a corpus
of read Myanmar speech. The dataset contains 9908 short audio clips each with a tran-
scription. The content of the corpus is derived from the weather news of the “Department
of Meteorology and Hydrology (Myanmar)” [18] and “DVB TVnews” [19].
The dataset contains 3 female and 1 male speakers. The audio clips are recorded in
a quiet place with a minimum of background noise. The lengths of audio clips vary
from 1 to 20 s and the total length of all audio clips is nearly 26 h. Audio clips are in
single-channel 16-bit WAV format and are sampled at 22.05 kHz.
Most of the transcriptions are collected from the above-mentioned data sources and
some are hand-transcribed. Since the transcriptions are mixed with both Zawgyi and
Unicode font encodings, all the texts are firstly normalized into Unicode encoding.
Myanmar Tools [20] is used to detect the Zawgyi strings and ICU Transliterator [21] is
used to convert Zawgyi to Unicode. The numbers are also expanded into full words and
the punctuations and the texts written in other languages are dropped. The audio clips
and transcriptions are segmented and aligned manually.
For the experimental purpose, we randomly split 70% ( 18 h) of the total data as
the training set, 10% ( 2 h) as the development set and 20% ( 5 h) as the test set.
We make sure to contain the audio clips of all four speakers in each dataset (Table 1).
The learning rate is reduced by the factor of 0.2 if the validation loss stops improving
for certain epochs. The batch size of 8 is used in all experiments. The early stopping
call back is also used to prevent overfitting. The different output label encoding
methods are compared: character-level, syllable-level and sub-word level encodings.
We use a set of regular expression rules [22] to segment the syllable-level tokens and
use Byte Pair Encoding (BPE) to create the sub-word tokens. Since the experiments are
performed with different label encoding methods, the results are evaluated using both
CER and SER metrics. All the experiments are carried out on NVIDIA Tesla
P100 GPU with 16GB GPU memory and 25 GB RAM (Google Colab). We do not use
the external language model in this work. We will further investigate the effect of
combining the decoding process with the external language model in the future.
5 Experimental Results
Table 2. Comparison of the effect of using convolutional layers and different label encoding
units on both development and test sets. Each model contains the same number of convolutional
layers, 5 layers of LSTM layers and a fully connected layer. The input features are not
downsampled for the models without convolutional layers. The experiments are evaluated with
both character error rate (CER) and syllable error rate (SER) metrics.
Unit With CNN Without CNN
Dev Test Dev Test
CER SER CER SER CER SER CER SER
Char 4.69 11.86 4.72 12.38 5.29 12.12 5.67 12.61
Syllable 21.95 20.67 22.68 22.34 22.09 23.46 23.41 24.83
BPE 100 16.44 18.80 13.72 16.58 19.51 26.69 18.72 25.27
BPE 300 9.61 19.12 10.38 20.35 9.65 20.88 9.98 21.83
BPE 500 10.77 24.73 11.58 27.34 22.08 34.61 22.67 36.03
Table 3. Comparison of character-level ASR models with different numbers of BLSTM layers.
Convolutional layers are used in every experiment. A fully connected layer is followed after the
BLSTM layers. The hidden unit size of BLSTM layers is set to 512 units.
Number of BLSTM Layers Dev Test
CER SER CER SER
3 4.65 11.85 5.03 13.24
4 4.58 12.54 4.93 13.75
5 4.69 11.86 4.72 12.38
6 5.26 14.45 5.45 15.46
6 Conclusion
In this paper, we explore the CTC based end-to-end architectures on the Myanmar
language. We empirically compare the various label encoding methods, different
depths of BLSTM layers and the use of convolutional layers on the low-resource
Myanmar speech corpus. This work shows that a well-tuned end-to-end system can
achieve state-of-the-art results in a closed domain ASR even for low-resource
Exploring CTC Based End-To-End Techniques for Myanmar Speech Recognition 1045
languages. As future work, we will further investigate the integration of the language
model to our end-to-end ASR system and will explore the other end-to-end multi-
tasking techniques.
Acknowledgments. The authors are grateful to the advisors from the University of Information
Technology who gave us helpful comments and suggestions throughout this project. The authors
also thank Ye Yint Htoon and May Sabal Myo for helping us with the dataset preparation and for
technical assistance.
References
1. Amodei, D., Anubhai, R., Battenberg, E., Case, C., Casper, J., Catanzaro, B., Chen, J.,
Chrzanowski, M., Coates, A., Diamos, G., Elsen, E., Engel, J., Fan, L., Fougner, C., Han, T.,
Hannun, A., Jun, B., LeGresley, P., Lin, L., Narang, S., Ng, A., Ozair, S., Prenger, R.,
Raiman, J., Satheesh, S., Seetapun, D., Sengupta, S., Wang, Y., Wang, Z., Wang, C., Xiao,
B., Yogatama, D., Zhan, J., Zhu, Z.: Deep Speech 2: end-to-end speech recognition in
English and Mandarin. arXiv:1512.02595 [cs.CL] (2015)
2. Graves, A., Fernández, S., Gomez, F., Schmidhuber, J.: Connectionist temporal classifica-
tion: labelling unsegmented sequence data with recurrent neural networks. In: ICML 2006:
Proceedings of the 23rd International Conference on Machine Learning. ACM Press, New
York (2006)
3. Chan, W., Jaitly, N., Le, Q., Vinyals, O.: Listen, attend and spell: a neural network for large
vocabulary conversational speech recognition. In: 2016 IEEE International Conference on
Acoustics, Speech and Signal Processing (ICASSP), pp. 4960–4964. IEEE (2016)
4. Graves, A.: Sequence transduction with recurrent neural networks, arXiv:1211.3711 [cs.NE]
(2012)
5. Zweig, G., Yu, C., Droppo, J., Stolcke, A.: Advances in all-neural speech recognition. In:
2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),
pp. 4805–4809. IEEE (2017)
6. Hannun, A., Case, C., Casper, J., Catanzaro, B., Diamos, G., Elsen, E., Prenger, R.,
Satheesh, S., Sengupta, S., Coates, A., Ng, A.Y.: Deep speech: scaling up end-to-end speech
recognition, arXiv:1412.5567 [cs.CL] (2014)
7. Shan, C., Zhang, J., Wang, Y., Xie, L.: Attention-based end-to-end speech recognition on
voice search. In: 2018 IEEE International Conference on Acoustics, Speech and Signal
Processing (ICASSP), pp. 4764–4768. IEEE (2018)
8. Li, J., Zhao, R., Hu, H., Gong, Y.: Improving RNN transducer modeling for end-to-end
speech recognition. In: 2019 IEEE Automatic Speech Recognition and Understanding
Workshop (ASRU), pp. 114–121. IEEE (2019)
9. Mon, A.N., Pa, W.P., Thu, Y.K.: Building HMM-SGMM continuous automatic speech
recognition on Myanmar Web news. In: International Conference on Computer Applications
(ICCA2017), pp. 446–453 (2017)
10. Naing, H.M.S., Pa, W.P.: Automatic speech recognition on spontaneous interview speech.
In: Sixteenth International Conferences on Computer Applications (ICCA 2018), Yangon,
Myanmar, pp. 203–208 (2018)
11. Nwe, T., Myint, T.: Myanmar language speech recognition with hybrid artificial neural
network and hidden Markov model. In: Proceedings of 2015 International Conference on
Future Computational Technologies (ICFCT 2015), pp. 116–122 (2015)
1046 K. M. M. Chit and L. L. Lin
12. Naing, H.M.S., Hlaing, A.M., Pa, W.P., Hu, X., Thu, Y.K., Hori, C., Kawai, H.: A Myanmar
large vocabulary continuous speech recognition system. In: 2015 Asia-Pacific Signal and
Information Processing Association Annual Summit and Conference (APSIPA), pp. 320–
327. IEEE (2015)
13. Mon, A.N., Pa Pa, W., Thu, Y.K.: Improving Myanmar automatic speech recognition with
optimization of convolutional neural network parameters. Int. J. Nat. Lang. Comput.
(IJNLC) 7, 1–10 (2018)
14. Aung, M.A.A., Pa, W.P.: Time delay neural network for Myanmar automatic speech
recognition. In: 2020 IEEE Conference on Computer Applications (ICCA), pp. 1–4. IEEE
(2020)
15. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image
recognition. In: International Conference on Learning Representations (ICRL) (2015)
16. Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing
internal covariate shift, arXiv:1502.03167 [cs.LG] (2015)
17. Hannun: Sequence Modeling with CTC. Distill. 2 (2017)
18. Department of Meteorology and Hydrology. https://www.moezala.gov.mm/
19. DVB TVnews. https://www.youtube.com/channel/UCuaRmKJLYaVMDHrnjhWUcHw
20. Google: Myanmar Tools. https://github.com/google/myanmar-tools
21. ICU - International Components for Unicode. https://site.icu-project.org/
22. Thu, Y.K.: sylbreak. https://github.com/ye-kyaw-thu/sylbreak
IoT Based Bidirectional Speed Control
and Monitoring of Single Phase Induction
Motors
1 Introduction
In this modern age, induction motors (IM) are widely used in industry, automotive,
aerospace, military, medical, and domestic equipment and appliances because of their
low costs, robustness, high power, and low maintenance [1]. The bulk of the electric
power that is being consumed globally, is by single-phase induction motors (SPIMs)
driving either a fan, pump, or compressor type of load, including their applications in
heating, ventilating, and air conditioning [2]. Traditionally, motor-driven systems run
at a nearly constant speed and are usually designed to provide a load margin of 20–30%
over the full load value for a long duration. A significant portion of the input energy is
wasted for controlling their outputs to meet the variable load demands where the power
drawn from the utility remains essentially the same as at full load. This energy loss can
be avoided by driving the motors at a speed that results in the desired regulated outputs.
The input power decreases as the speed is decreased. This decrease can be estimated by
recognizing that in induction motor,
have been for monitoring and speed control of induction motors. While bidirectional
rotation of motors has not been widely investigated.
This paper addresses this gap by providing a simple web-based communication
solution for bidirectional speed control and monitoring of multiple single-phase
induction motors. The speed control of the induction motors has been performed by the
stator voltage control method using the pulse width modulation (PWM) technique [20].
The direction of rotation of the motor has been changed by reversing the stator mag-
netic field by swapping the connection of auxiliary stator winding with simple, low-
cost relay switching. The parameters of the prototype design are observed in LCD by a
microcontroller and shared to remote computer/mobile using IoT through GPRS which
makes the system flexible and user friendly.
Figure 1 shows the proposed system's architecture in the block diagram form. It mainly
consists of a web application developed on a remote computer or smartphone, a
GSM/GPRS module for interfacing with the server system, a Microcontroller, an LCD,
IGBTs with gate drive circuits, Bridge rectifiers, IR sensors, a relay module, and two
single-phase induction motors (SPIM).
Fig. 1. Block diagram of IoT based bidirectional speed control and monitoring of SPIM.
The IoT based bidirectional speed and status observation of the motors is accom-
plished by the web application. The GSM/GPRS module acts as an interface between
the web server and the microcontroller unit to send and receive command and output
signals. Following the specific command input from the field operator, the programmed
1050 A. Rahman et al.
microcontroller generates PWM signals. These signals are sent through the isolation
circuit to the IGBT gate driver circuits to operate the IGBT switches. By the PWM AC
chopping technique, input AC voltages to the stator windings of the induction motors
are varied. As a result, the motor speeds are controlled. The IR sensors are used for
measuring the RPM of the motors and the results are compared with fine-tuned digital
tachometer readings. The relay module with four sets of relays is employed for the
motor on/off and direction control. The overall motor status is observed in the LCD and
the web application developed in the remote computer or smartphone.
The circuit diagram of the proposed scheme for the controlling and monitoring of
induction motors with IoT is depicted in Fig. 2. Functionally, the system can be
divided into four major functional units described as follows:
harmonic content compared to other methods. Here the supply voltage is chopped and
modulated with a high switching frequency IGBT gate drive signal generated from the
microcontroller unit. The change in the duty cycle of the IGBT switch changes the
effective value of the load voltage and current. The chopped voltage can be expressed
by multiplying the sinusoidal line voltage with the switching signal. The detailed
analysis is given in [20]. The switching function can be written by the Fourier series
expansion of the pulse for one period as:
X1
d ð t Þ ¼ a0 þ n¼1
ðan cosnxs t þ bn sinnxs tÞ ð3Þ
ton ton
a0 ¼ ¼ ¼D ð4Þ
T ton þ toff
1
an ¼ sinð2pnDÞ ð5Þ
np
1
bn ¼ ½1 þ cosð2pnDÞ ð6Þ
np
The high-frequency terms in Eq. (8) are filtered by the motor’s inner inductance
that results in the load voltage to be expressed as:
D:V m
V Lrms ðtÞ ¼ p ð10Þ
2
As the input power required by the induction motor varies as the speed cubed,
decreasing the motor speed to half of the rated speed by voltage control method to meet
the variable load demands the input power drawn from the utility will drop to one-
eighth of the rated power input. So there is a significant reduction in the power that is
being consumed.
1052 A. Rahman et al.
Fig. 6. Web application based panel for monitoring the motor status
The laboratory implementation of multiple SPIM speed and direction control and
monitor using IoT technology is shown in Fig. 8. As described earlier, an operator from
a remote location can observe the system and provide command inputs for the desired
operation of motors.
IoT Based Bidirectional Speed Control and Monitoring of SPIMs 1055
Fig. 8. A laboratory prototype of IoT based bidirectional speed of single-phase induction motor
This proposed system offers ON/OFF control, clockwise (CW) and counter-
clockwise (CCW) rotation control, speed control as a percentage of rated speed through
the specially designed website over the internet. The motor parameters, measured by
using the IR sensors, are processed by the microcontroller and shared to remote
computer/mobile by using GPRS and also displayed in LCD for field operators.
Concerning the variation of the PWM duty cycle, the SPIM speed variation has been
observed in the web server and LCD and also compared with field measurement data
obtained from a fine-tuned digital tachometer. It has been found that the deviation of
field-measured tachometer-data and IR sensor-based remote data is negligible. This
comparison is presented in Fig. 9 by using a comparative bar diagram.
Fig. 9. Digital Tachometer and IR sensor-based speed measurement comparative bar diagram.
1056 A. Rahman et al.
Concerning the PWM duty cycle, the practically measured RPM values are plotted
against percent duty cycle variation as shown in Fig. 10 and a second-order polynomial
Eq. (11) has been derived.
Here, N = Speed values in RPM; D = Duty cycle in percentage (%). This equation
can be used to get motor speed empirically based on the corresponding percent of duty
ratio. For various desired speed values of single-phase induction motor, the required
percent of PWM duty cycles obtained from the above equation are summarized in
Table 1.
Table 1. Percentage of the duty cycle for different motor speed values
Speed, N (RPM) Duty cycle, D (%) Speed, N (RPM) Duty cycle, D (%)
1700 55.87 2100 65.75
1800 58.03 2200 69.01
1900 60.36 2300 72.96
2000 62.91 2400 78.40
IoT Based Bidirectional Speed Control and Monitoring of SPIMs 1057
5 Conclusion
Multiple SPIM speed and direction control and monitoring using IoT technology have
been presented in this paper. The speed control of the induction motors has been
performed by the stator voltage control method using the PWM technique. The
ON/OFF and the bidirectional rotation (CW/CCW) of the motors have been controlled
by the cheap and simple operation of relays. The motor operational parameters have
been monitored remotely in a computer/smartphone using IoT through GPRS. For a
variation of the PWM duty cycle, the SPIM speed variation has been observed and was
compared with field measurement data obtained from a fine-tuned digital tachometer. It
has been found that the deviation of direct field measurement using the tachometer and
IR sensor-based remote measurement is little. Also from the analysis, it has been shown
possible to calculate the desired rpm of the motor beforehand following the percentage
of the duty cycle from the second-order polynomial equation derived. In industrial,
manufacturing, agricultural, and household applications, the proposed IoT based smart
control system can be introduced with more sensors to continuously track multiple
control variables of machines, plants, etc. It is also possible to develop an android
based application to make the system more user-friendly.
References
1. Sarb, D., Bogdan, R.: Wireless motor control in automotive industry. In: 2016 24th
Telecommunications Forum (TELFOR), pp. 1–4. Belgrade (2016). https://doi.org/10.1109/
TELFOR.2016.7818790
2. Mohan, N.: Power Electronics: A First Course, 1st edn. Wiley, USA (2012)
3. Internet World Stats. https://www.internetworldstats.com/stats.htm. Accessed 03 June 2019
4. Doshi, N.: Analysis of attribute-based secure data sharing with hidden policies in smart grid
of IoT. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.) Intelligent Computing & Optimization.
ICO 2018. Advances in Intelligent Systems and Computing, vol. 866. Springer, Cham
(2019). https://doi.org/10.1007/978-3-030-00979-3_54
5. Khan, A.A., Mouftah, H.T.: Energy optimization and energy management of home via web
services in smart grid. In: IEEE Electrical Power and Energy Conference, pp. 14–19, London
(2012). https://doi.org/10.1109/EPEC.2012.6474940
6. Cheah, P.H., Zhang, R., Gooi, H.B., Yu, H., Foo, M.K.: Consumer energy portal and home
energy management system for smart grid applications. In: 10th International Power &
Energy Conference (IPEC), pp. 407–411. Ho Chi Minh City (2012). https://doi.org/10.1109/
ASSCC.2012.6523302
7. Kuzlu, M., Pipattanasomporn, M., Rahman, S.: Communication network requirements for
major smart grid applications in HAN, NAN, and WAN. Comput. Netw. 67, 74–88 (2014).
https://doi.org/10.1016/j.comnet.2014.03.029
8. Tsai, W., Shih, Y., Tsai, T.: IoT-type electric fan: remote-controlled by smart-phone. In:
Third International Conference on Computing Measurement Control and Sensor Network
(CMCSN), pp. 12–15. Matsue (2016). https://doi.org/10.1109/CMCSN.2016.17
9. Kuzlu, M., Rahman, M.M., Pipattanasomporn, M., Rahman, S.: Internet-based communi-
cation platform for residential DR programmes. IET Networks 6(2), 25–31 (2017). https://
doi.org/10.1049/iet-net.2016.0040
1058 A. Rahman et al.
10. Kamble, I., Patil, Y.M.: A review of parameters monitoring and controlling system for
industrial motors using wireless communication. Int. J. Res. Appl. Sci. Eng. Technol. 7(1),
47–49 (2019). https://doi.org/10.22214/ijraset.2019.1010
11. Potturi, S., Mandi, R.P.: Critical survey on IoT based monitoring and control of induction
motor. In: IEEE Student Conference on Research and Development (SCOReD), pp. 1–6.
Selangor, Malaysia (2018). https://doi.org/10.1109/SCORED.2018.8711222.
12. Patil, R.R., Date, T.N., Kushare, B.E.: ZigBee based parameters monitoring system for
induction motor. In: IEEE Students' Conference on Electrical, Electronics and Computer
Science, Bhopal, pp. 1–6 (2014). https://doi.org/10.1109/SCEECS.2014.6804469
13. Khairnar, V.C., Sandeep, K.: Induction motor parameter monitoring system using zig bee
protocol & MATLAB GUI: Automated Monitoring System. In: Fourth International
Conference on Advances in Electrical, Electronics, Information, Communication, and Bio-
Informatics (AEEICB), Chennai, pp. 1–6 (2018). https://doi.org/10.1109/AEEICB.2018.
8480992
14. Joshi, P.S., Jain, A.M.: Wireless speed control of an induction motor using PWM technique
with GSM. IOSR J. Electr. Electron. Eng. 6(2), 1–5 (2013). https://doi.org/10.9790/1676-
0620105
15. Rekha, V.S.D., Ravi, K.S.: Induction motor condition monitoring and controlling based on
IoT. Int. J. Electron. Electr. Comput. Syst. 6(9), 74–89 (2015)
16. Şen, M., Kul, B.: IoT-based wireless induction motor monitoring. In: XXVI International
Scientific Conference Electronics (ET), pp. 1–5. Sozopol (2017). https://doi.org/10.1109/ET.
2017.8124386.
17. Venkatesan, L., Kanagavalli, S., Aarthi, P.R., Yamuna, K.S.: PLC SCADA based fault
identification and protection for three-phase induction motor. TELKOMNIKA Indone-
sian J. Electr. Eng. 12(8), 5766–5773 (2014)
18. Kathiresan, S., Janarthanan, M.: Design and implementation of industrial automation using
IOT and Bluetooth LE. Int. J. Adv. Res. Trends Eng. Technol. (IJARTET) 3(19), 335–338
(2016). https://doi.org/10.20247/IJARTET.2016.S19040062
19. Çakır, A., Çalış, H., Turan, G.: Remote controlling and monitoring of induction motors
through internet. TELKOMNIKA Indonesian J. Electr. Eng. 12(12), 8051–8059 (2014).
https://doi.org/10.11591/telkomnika.v12i12.6719
20. Yildirim, D., Bilgic, M.: PWM AC chopper control of single-phase induction motor for
variable-speed fan application. In: 2008 34th Annual Conference of IEEE Industrial
Electronics, Orlando, FL, pp. 1337–1342 (2008). https://doi.org/10.1109/IECON.2008.
4758148.
Missing Image Data Reconstruction Based
on Least-Squares Approach
with Randomized SVD
1 Introduction
Missing data problems have been important issues in many research fields such as
biology, medical, engineering, image processing, remote sensing etc. In image pro-
cessing, pixels in some images may be corrupted or simply missing. In some cases,
pixels may be randomly deleted in order to save network bandwidth during data
transmission. The missing data reconstruction methods are extensively developed to
restore those corrupted or missing pixels.
In 2008, Emmanuel J. C. et al. [1] considered matrix completion in the recovery of
a data matrix from a sampling of its entries which led to the notions of missing data
reconstruction algorithms. For example, Jin Z. et al. [2] and Feilong C. et al. [3]
demonstrated the reconstruction algorithm by working on sparse matrix completion.
approach is shown to use least execution time, for sufficiently small dimensions of the
POD basis, while providing the same level of accuracy when compared to the other
approach. Finally, some concluding remarks are given in Sect. 5.
2 Background Knowledge
X ¼ URV T ð1Þ
where U 2 Rmm and V 2 Rnn are matrices with orthonormal columns. The column
vectors of U and V are left and right singular vectors, respectively, denoted as ui and vi ,
R ¼ diagðr1 ; . . .; rr Þ 2 Rmn , r ¼ minfm; ng where r1 r2 . . . rr [ 0, are
known as singular values. Using the subscript T to denote the transpose of the matrix.
The left singular vectors of X are eigenvectors of XX T and the right singular vectors of
X are eigenvectors of X T X.
Truncated SVD is obtained by computing full SVD and then truncating it by
selecting the top k dominant singular values and their corresponding singular vectors
such that
X
k
Xk ¼ Uk Rk VkT ¼ ri ui vTi ð2Þ
i¼1
Where k\r is the numerical rank, Uk 2 Rmk and Vk 2 Rnk are matrices with
orthonormal columns Rk ¼ diagðr1 ; . . .; rk Þ 2 Rkk where r1 r2 . . . rk [ 0.
Here, Xk ¼ ½~x1 ; ~x2 ; . . .; ~xk is the best rank k approximation to the matrix X by the
low-rank matrix approximation error measured by 2-norm:
X
n X
r
kX Xk k22 ¼ r2k þ 1 ; or kX Xk k2F ¼ kxi ~xi k22 ¼ r2l ð3Þ
i¼1 l¼k þ 1
^k ¼ U
X ^kV
^kR ^kT ; ð4Þ
where k\r is the numerical rank, U ^ k 2 Rmk and V ^k 2 Rnk are matrices with
orthonormal columns R ^k 2 R kk
is a diagonal matrix with the singular values
^1 r
r ^2 . . . r
^k [ 0: Details are given as follows.
Define the random projection of a matrix as,
Y ¼ XX ð5Þ
Let ^sc :¼ C T ^s 2 Rnc and ^sg :¼ GT ^s 2 Rng . Then, the known components and the
unknown components are given in the vectors ^sc and ^sg , respectively. The overview of
the steps described in this section for approximating an incomplete image is shown in
Fig. 2.
1064 S. Intawichai and S. Chaturantabut
y y
The solution of the above problem is given by a ¼ Uc ^sc where Uc ¼ ðUcT Uc Þ1 UcT is
the Moore-Penrose inverse. That is,
ng
with known entries in sc nc
and unknown entries in sg , where n nc ng
OUTPUT : Approximation of sg
n ns
Step 1. Create snapshot matrix : S [ s1 , s2 , ..., sns ] and let r rank (S )
Step 2. Construct basis U of rank k r for S
2
Step 3. Find coefficient vector a from sc using Least-Squares problem: mink sc U c a 2
a
Next, we consider the optimal basis which is obtained from the singular value
decomposition (SVD) because it is optimal in the least-squares sense. The basis defined
above can be obtained from the left singular vector of the matrix S.
Recall that S ¼ ½s1 ; s2 ; . . .; sns 2 Rnns and k\r ¼ rank (SÞ. The SVD of S is
S ¼ URV T , where U ¼ ½u1 ; . . .; ur 2 Rnr and V ¼ ½v1 ; . . .; vr 2 Rns r are matrices
with orthonormal columns and R ¼ diagðr1 ; . . .; rr Þ 2 Rrr with r1 r2 . . .
rr [ 0.
The optimal solution of least-squares problem minSk kS Sk k2F , rank (Sk Þ ¼ k is
2 2 Pr
Sk ¼ Uk Rk VkT with minimum error S Sk 2 ¼ r2k þ 1 or S Sk F ¼ r2l . Then
l¼k þ 1
optimal orthonormal basis of rank k (the POD basis) of dimension k is the matrix
formed by the first k left singular vectors, i.e. Uk ¼ ½u1 ; . . .; uk 2 Rnk , k r.
However, it could be computational intensive to obtain the standard SVD. We will
use the rSVD methods in Sect. 2.2 to find the optimal orthonormal basis, which reduce
the computational complexity of the standard approach for computing SVD.
4 Numerical Experiments
In this Section, we show the numerical experiments of the rSVD and SVD methods for
constructing the projection basis and compare the CPU times together with the
reconstruction errors. Here, we use the relative error in 2-norm.
We use a standard test image, called Lena picture. This image is considered in
gray scale color with size 512 512 shown in Fig. 3. We consider this image with
2:75% missing pixels in a block of missing pixels with size 15 20 pixels spreading
over the test image. We consider two cases with different patch sizes for image
reconstructing:
Case I. Clipping the image to 256 patches with size 32 32 pixels or the
dimension of the matrix representing this image is 1024 256.
Case II. Clipping the image to 1024 patches with size 16 16 pixels or the
dimension of the matrix representing this image is 256 1024.
1066 S. Intawichai and S. Chaturantabut
Fig. 3. The standard test image, called Lena picture (a) The original gray-scale image
(b) Incomplete image with missing pixels.
Note that, the numbers of the known patches must be enough for computing the
basis. Also, the available pixels in the corrupted patch are sufficiently used to recon-
struct the missing pixels. The performance of both cases are shown in Fig. 4 and Fig. 5.
Fig. 4. The reconstructed images (when clipped to 256 patches with size 32 32 pixels) by
using the rSVD method. The images in (a)-(e) show the reconstruction using basis of rank
k = 10, 15, 20, 25, 30, respectively.
The relative error of the reconstruction results and the computational time for
constructing basis set are shown in Fig. 6 and Fig. 7, respectively. In addition, the
efficacy of the proposed reconstruction strategy is measured by the peak signal-to-noise
ratio (PSNR), as shown in Table 1 and Fig. 8.
Fig. 5. The reconstructed images of case I, (1a)–(3a) and (1b)–(3b) show the images when using
the rSVD method and SVD method, respectively with rank k = 10, 20, 30. Similarly, the
reconstructed images of case II, (1c)–(3c) and (1d)–(3d) show the images when using the rSVD
method and SVD method, respectively with rank k = 10, 20, 30.
1068 S. Intawichai and S. Chaturantabut
Fig. 6. Relative errors of the reconstruction results of the Lenna picture when 2.75% of pixels
are missing: (a) Case I and (b) Case II.
Fig. 7. Computational time for computing basis used in the reconstruction of the Lenna picture
when 2.75% of pixels are missing: (a) Case I and (b) Case II.
Table 1. Reconstruction performances of images under using rSVD and SVD method with
different rank k.
PSNR Case I. Case II.
(dB) rSVD SVD rSVD SVD
k = 10 32.3814 32.3476 32.2426 31.3460
k = 15 32.4651 32.3421 32.2796 31.0623
k = 20 32.0866 32.0097 31.7313 30.9716
k = 25 32.2250 32.0922 31.9056 30.6867
k = 30 32.1546 31.8980 31.7345 30.5484
From Table 1, both cases show that using the rSVD method is a little more accurate
than SVD method. Although the rSVD method gives slightly more PSNR values in
some test cases, it uses significantly less computation time.
Missing Image Data Reconstruction 1069
Fig. 8. Reconstruction performances in term of PSNR: (a) Case I and (b) Case II.
5 Conclusions
This work presented an image reconstruction approach. A given incomplete image was
first divided into many patches, the only corrupted patches are reconstructed while the
known patches are used to compute the POD basis. The available image pixels in
patches were used to form a low-dimensional subspace and approximate the missing
pixels by applying the least-squares method. By using the available pixels in the
corrupted patches around the missing pixels, the incomplete image is shown to be
approximated efficiently and accurately. Instead of using a traditional approach based
on the standard SVD for computing this low-dimension basis, this work efficiently used
rSVD, which was shown in the numerical tests to give substantially less reconstruction
time with the same order of accuracy.
Acknowledgments. The authors gratefully acknowledge the financial support provided by this
study was supported by Thammasat University Research Fund, Contract No. TUGR 2/12/2562
and Royal Thai Government Scholarship in the Area of Science and Technology (Ministry of
Science and Technology).
References
1. Emmanuel, J.C., Benjamin, R.: Exact matrix completion via convex optimization. Found.
Comput. Math. 9, 717–772 (2008)
2. Jin, Z., Chiman, K., Bulent, A.: A high performance missing pixel reconstruction algorithm
for hyperspectral images. In: Proceedings of 2nd International Conference on Applied and
Theoretical Information Systems Research (2012)
3. Feilong, C., Miaaomiao, C., Yuanpeng, T.: Image interpolation via low-rank matrix
completion and recovery. IEEE Trans. Circuits Syst. Video Technol. 25(8), 1261–1270
(2015)
4. Pete, G.W.S.: On the early history of the singular value decomposition. Siam Rev. 35, 551–
566 (1993) http://www.jstor.org/stable/2132388
5. Dan, K.: A singularly valuable decomposition: the svd of a matrix. The College Math. J. 27
(1), 2–23 (1996)
1070 S. Intawichai and S. Chaturantabut
6. Dimitris, A., Frank, M.: Fast computation of low-rank matrix approximations, J. Assoc.
Comput. Mach., 54 (2007)
7. Gilbert, S.: Linear Algebra and its Applications, 3rd edn. Harcourt Brace Jovanovich, San
Diego, USA (1988)
8. Gene, H.G., Charies, F.V.L.: Matrix Computations, 3rd ed., Johns Hopkins University Press,
Baltimore, Maryland (1996)
9. Harry, C.A., Claude, L.P.: Singular value decompositions and digital image processing,
24, 26–53 (1976)
10. Samruddh, K., Reena, R.: Image compression using singular value decomposition. Int.
J. Adv. Res. Technol. 2(8), 244–248 (2013)
11. Lijie, C.: Singular Value Decomposition Applied To Digital Image Processing, Division of
Computing Studies. Arizona State University Polytechnic Campus, Mesa, Arizona, 1–15
12. Sven, O.A., John, H.H., Patrick, W.: A critique of SVD-based image coding systems. In:
IEEE International Symposium on Circuits and Systems (ISCAS), vol. 4, pp. 13–16 (1999)
13. Vitali, V.S., Roger, L.: Fast PET image reconstruction based on svd decomposition of the
system matrix. IEEE Trans. Nuclear Sci. 48(3), 761–767 (2001)
14. Rowayda, A.S.: SVD based image processing applications: state of the art, contributions and
research challenges. Int. J. Adv. Comput. Sci. and Appl. 3(7), 26–34 (2012)
15. Davi, M.L., João, P.C.L.C., João, L.A.C.: Improved MRI reconstruction and denoising using
SVD-based low-rank approximation. In: Workshop on Engineering Applications, pp. 1–6
(2012)
16. Sarlos, T.: Improved approximation algorithms for large matrices via random projections. In:
47th Annual IEEE Symposium Foundations of Computer Science, pp. 143–152(2006)
17. Liberty, E., Woolfe, F., Martinsson, P.G., Rokhlin, V., Tygert, M.: Randomized algorithms
for the low-rank approximation of matrices. Proc. Natl. Acad. Sci. 104(51), 20167–20172
(2007)
18. Halko, N., Martinsson, P.G., Tropp, J.A.: Finding structure with randomness: probabilistic
algorithms for constructing approximate matrix decompositions. Siam Rev. 53(2), 217–288
(2011)
19. Siriwan, I., Saifon, C.: An application of randomized singular value decomposition on image
reconstruction using least-squares approach, Preprint in Thai Journal of Mathematics (2018)
20. Siriwan, I., Saifon, C.: A Numerical Study of Efficient Sampling Strategies for Randomized
Singular Value Decomposition, Thai Journal of Mathematics, pp. 351–370 (2019)
An Automated Candidate Selection
System Using Bangla Language
Processing
1 Introduction
Data mining is a logical process used to find relevant data from a large data set.
It is the process of analyzing data from different aspects and summarizes them
into useful information [11]. Data mining helps us extract this information from
an extensive dataset by finding patterns in the given data set. Patterns that are
certain enough according to the user’s measures are called knowledge [18]. Data
mining is a sub-process of knowledge discovery in which the various available data
sources are analyzed using different data mining algorithms. Like data mining,
text mining is used in extracting essential pieces of information from the text
[3]. Since all data on the web and social media are available in a fuzzy and
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 1071–1081, 2021.
https://doi.org/10.1007/978-3-030-68154-8_90
1072 M. M. Islam et al.
random manner, it is sometimes hard for humans to understand and process the
data effectively. In that type of case, text mining tools help to create a bridge
and relationship among the texts. This process is also known as Knowledge
Discovery from Texts (KDT) [15]. There are several text mining applications,
such as speech recognition, social media data analysis, content enrichment, etc.
[1,2].
Recruitment, or selecting the right candidates from a vast pool of candidates,
has always been a fundamental issue as far as employers are concerned. Generally,
the recruitment process follows like this for a company. Candidates send CV
to the company. Recruiter selects some candidates from those CVs followed by
conducting personality and other technical eligibility evaluation tests, interviews,
and group discussions. The human Resource (HR) department is one of the most
critical departments of a company. The HR department plays a vital role in
selecting an expert candidate for a particular post. The first task of the HR
department is to shortlist the CVs of various candidates who applied for the
specific post. High level of uncertainty when the HR department checks all the
CVs, ranks them and selects a candidate for a particular job position because of
many applicants. This uncertainty occurs due to the various occupation domain
expert’s different opinions and preferences in the decision-making process. This
evaluation process involves excessive time consumption and monotonous work
procedure. To reduce the load of the HR department in the recruiting process,
an automation tool like CV ranker is very helpful.
Recently, some studies have focused on automated candidate selection system
in [9,13,16,17,20]. An elaborate discussion on these methodologies is given in
Sect. 2. The common phenomenon of the most existing techniques is done for a
CV that is written in Bengali. In Bangladesh, as Bengali is our native language,
and for many government posts, there is a need for a CV written in Bengali.
However, any automated candidate selection system for Bengali CV has not been
done yet.
In this study, we have developed an automated candidate selection system
that takes the CVs (written in Bengali) of candidates and the employer’s require-
ments as input. After that, our system extracts information of the candidates
using Bangla Language Processing (BLP) from the CVs, generates an average
Cosine Similarity Score for each CV. Finally, our system ranks the candidates
according to these average Cosine Similarity Scores and returns the dominant
candidates’ list.
The rest of this paper is organized as follows: We present the related works
on the automated candidate selection system in Sect. 2. Section 3 provides the
details of the proposed methodology of the automated candidate selection sys-
tem. Our experimental results are shown in Sect. 4. Finally, we conclude the
paper in Sect. 5 by providing the future directions of this research.
2 Related Work
Few contributions are regarding the automated candidate selection system. Such
as, Evanthia et al. [7] implement automated candidate ranking, based on objec-
An Automated Candidate Selection System 1073
tive criteria that can be extracted from the applicant’s LinkedIn profile. The
e-recruitment system was deployed in a real-world recruitment scenario, and
expert recruiters validated its output. Menon et al. [14] represent an approach
to evaluate and rank candidates in a recruitment process by estimating emo-
tional intelligence through social media data. The candidates have to apply for a
job opening by filling an online resume and accessing their Twitter handle. This
system estimates the candidate’s emotion by analyzing their tweets, and the
professional eligibility is verified through the entries given in the online resume.
Faliagka et al. [6] present an approach for evaluating job applicants in online
recruitment systems, using machine learning algorithms to solve the candidate
ranking problem, and performing semantic matching techniques. The system
needs access to the candidate’s full profile and the recruiters’ selection criteria,
including assigning weights to a set of candidate selection criteria. The pro-
posed scheme can output consistent candidate rankings compared to the ones
assigned by expert recruiters. De Meo et al. [5] propose an Extensible Markup
Language (XML)-based multi-agent recommender system for supporting online
recruitment services. XML is a standard language for representing and exchang-
ing information. It embodies both representation capabilities, which are typical
of Hypertext Markup Language, and data management features, which are char-
acteristic of the Database Management System. Kessler et al. [12] present the
E-Gen system: Automatic Job Offer Processing system for Human Resources
implementing two tasks - analysis and categorization of job postings. Fazel
et al. [8] present an approach to matchmaking job seekers and job advertise-
ments. When searching for jobs or applicants, a job seeker or recruiter can ask
for all the job advertisements or applications that match his/her application in
addition to expressing his/her desired job-related descriptions. Getoor et al. [10]
represent a survey on Link Mining using the Top-k query processing technique.
Link mining refers to data mining techniques that explicitly consider these links
when building predictive or descriptive models of the linked data.
Most of the CV ranking research is done on the CV, written in the English
language. In this study, we have developed an automated candidate selection
system that takes the CVs (written in Bengali) of candidates and the employer’s
requirements as input. Then, it extracts information of the candidates using
Bangla Language Processing (BLP) from the CVs, generates an average Cosine
Similarity Score for each CV. Finally, it ranks the candidates according to these
average Cosine Similarity Scores and returns the dominant candidates’ list.
3 Methodology
also learn some new traits about words in a vocabulary. Besides, it can also
capture the context of a word in a document, semantic and syntactic similarity,
relation with other words, etc. Word2Vec is a popular method to construct such
word embedding. This method takes input as a large corpus of text and produces
a vector space as output. Each vector consists of many hundred dimensions or
more than a hundred dimensions, with each unique word from the large corpus
of text being assigned a corresponding vector in the space. Word vectors are
positioned in the vector space so that words that share common contexts in
the corpus are located close to one another in the space. Considering the fol-
lowing sentence: “Have a good day” with exhaustive vocabulary V = Have, a,
good, day. One-hot encoding of this vocabulary list is done by putting one for
the element at the index representing the corresponding word in the vocabulary
and zero for other words. One-hot encoded vector representation of these words
are: Have = [1,0,0,0]; a=[0,1,0,0]; good=[0,0,1,0]; day=[0,0,0,1]. After creating
a one-hot encoded vector for each of these words in V, the length of our one-
hot encoded vector would be equal to the size of V (=4). Visualization of these
encodings represented by 4-dimensional space, where each word occupies one of
the dimensions and has nothing to do with the rest (no projection along the
other dimensions).
In this study, the Common Bag of Words (CBOW) (based on Neural Net-
work) method is used to obtain Word2Vec. CBOW method takes each word’s
context as the input and tries to predict another word similar to the context
[20]. For the sentence “Have a great day,” if the input to the Neural Network is
the word (“great”), the model will try to predict a target word (“day”) using
a single context input word (“great”). More specifically, the one-hot encoding
of the input word is used to measure the output error compared to the one-hot
encoding of the target word (“day”).
In Fig. 3, a one hot encoded vector of size V goes as input into the hidden
layer of N neurons, and the output is again a V length vector with the elements
being the softmax values. Wvn(V*N dimensional matrix) is the weight matrix
that maps the input x to the hidden layer. W’nv (N*V dimensional matrix)is
1076 M. M. Islam et al.
Fig. 3. A simple CBOW model with only one word in the context
the weight matrix that maps the hidden layer outputs to the final output layer.
The hidden layer neurons copy the weighted sum of inputs to the next layer. The
only non-linearity is the softmax calculations in the output layer. But, the above
CBOW model in Fig. 3 used a single context word to predict the target. This
CBOW model takes C context words. When Wvn is used to calculate hidden
layer inputs, we take an average over all these C context word inputs.
Cosine similarity returns a list of scores for each skill. After that final score
for a CV is calculated by taking the mean of all the scores. Thus, all the CV
goes through the model and generates scores, which is then stored in a database
to rank them. Finally, Recruiters can see all the CVs in sorted order.
For implementing our study, we have collected 50 CVs (written in Bengali), and
all the experiments are done on these CVs. All the data are extracted from these
CV documents using UTF-8 encoding, and the output of the extracted data is
shown in Fig. 4.
Here, all the data in Fig. 4, is not needed for calculating cosine similarity.
So, important keywords like “Skills,” “CGPA,” etc. are extracted to match the
company requirements and calculate average cosine similarity. All the important
keywords extracted from CV is shown in Fig. 5a.
Fig. 5. Snapshot of the system which indicates important keywords and company
requirements required
1078 M. M. Islam et al.
Fig. 6. Bar plot of all the CV’s average cosine similarity scores
5 Conclusion
In this paper, we have narrated an automated candidate selection system that
ranks all the CVs (written in Bengali) by extracting information and calculating
an average Cosine Similarity Score. Automating the complete candidate selection
task may help the HR agencies save time, cost, and effort to search and screen
the pioneer applicants from vast applications. There are many automated can-
didate ranking systems available online in which all the CV format is in English.
But we have developed an automated candidate selection system suitable with
the help of Bangla Language Processing (BLP) for a CV that is written in Ben-
gali. In the system performance evaluation, we have used 50 CVs of technical
background testing of the system and found that the system works efficiently to
return the best candidates by matching the given requirements with the candi-
date’s qualifications. Altogether, the system performs better in filtering the CV
documents written in Bengali and the candidates based on the CV document’s
information. But, Our current method can not capture temporal features. That
is why it will fail to find sequential similarity between sentences. Suppose the
recruiter wants to prioritize a candidate’s recruitment qualities via giving quality
tags sequentially in a sentence. In that case, the current method will fail to rank
CVs according to the priorities.
Our future work includes a system that can directly take a job description
from the recruiters and then evaluate CVs more dynamically. We are planning to
start with a bi-directional LSTM encoder with the cosine similarity computing
layer to give our model the capability of understanding both the semantic and
syntactic meaning of a sentence.
References
1. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent computing & optimization.
Springer, Berlin (2018)
1080 M. M. Islam et al.
20. Yasmin, F., Nur, M.I., Arefin, M.S.: Potential candidate selection using informa-
tion extraction and skyline queries. In: International Conference on Computer Net-
works, Big data and IoT, pp. 511–522. Springer (2019)
21. Zhang, Y., Jatowt, A., Tanaka, K.: Towards understanding word embeddings:
Automatically explaining similarity of terms. In: 2016 IEEE International Con-
ference on Big Data (Big Data), pp. 823–832. IEEE (2016)
AutoMove: An End-to-End Deep Learning
System for Self-driving Vehicles
Abstract. End to End learning is a deep learning approach that has been used
to solve complex problems that would usually be carried out by humans with
great effect. A deep structure was designed within this study to simulate
humans’ steering patterns in highway driving situations. The architecture of the
network was based on image processing algorithm which is integrated with deep
learning convolutional neural network (CNN). There are five aspects in this
work, which enables the vehicle to detect the lanes, detect the speed of the
vehicle, detect the angle of the road, recognize the objects on the road and
predict the steering angle of the vehicle. A self-derived mathematical model is
used to calculate the road angles for the prediction of vehicle’s steering angles.
The model is trained on 2937 video frame samples and validated on 1259
samples with 30 epochs. The video of the local road was set as the output which
will show the difference between actual steering angles and predicted steering
angle. The experiments have been carried out in a newly built industrial park
with suitable industry 4.0 standard design of urban smart development.
1 Introduction
left and if the angle is more than five degree the steering angle will turn right The rest
of the article are organized such as, literature review in Sect. 2. Section 3 discussed the
methodology, and navigation of the self-driving vehicle. The implementation of the
five stages are described in Sect. 4. Section 5 has covered the integration of the image
processing algorithm with convolutional neural network (CNN) for the automatic
navigation. Section 6 concludes the work done.
2 Literature Review
connected layer. The input image is separated into YUV planes and send to the network
(Bojarski et al. 2016, 3-6).
Next is the training. Selecting the frames to use is the first step to train a neural
network. The collected data is labelled with weather condition, road type, and driver’s
activity (staying in lane, turning and switching lanes). The data that driver was staying
in a lane was selected and others were discarded. Then the video sampled at 10 FPS
(Bojarski et al. 2016, 3-6). After that is data augmentation. The data will be augmented
after selecting the final set of frames by adding artificial rotations and shifts to teach the
network how to restore from a poor orientation or position. The final stage is simu-
lation. Pre-recorded videos will be taken by the simulator from a forward-facing on-
board camera and images that relatively what would appear if the CNN were, instead,
steering the vehicle will be generated. In conclusion, during training, the system learns
to detect the outline of a road without the explicit labels. Thomas, J. J et al., has referred
multiple works in deep learning which we have used for the literature review.
2.3.2 YOLO
YOLO stands for “You only look once”. According to Redmon (2016), it is an object
detection algorithm that runs quicker than R-CNN because of its simpler architecture.
Classification and bounding box regression will be done at the same time. Figure 2
shows how YOLO detection system works.
A single convolutional network predicts several bounding boxes and class proba-
bilities for those boxes simultaneously. YOLO trains on full images and the detection
performance will be optimized directly. Thomas, J. J., Tran, H. N has worked on graph
neural network application which has been referred in this work for literature.
3 Methodology
4 Implementation
4.1 Lane Detection
Lane detection is implemented by colour thresholding. The first step is to convert the
image to grayscale identifying the region of interest. The pixels with a higher bright-
ness value will be highlighted in this step. Figure 4 shows the code snippet to convert
the image to grayscale and Fig. 5 shows the image that is converted to grayscale.
1086 S. Ramasamy and J. Joshua Thomas
#read image
img = cv2.imread('road png.PNG')
#convert image to grayscale
gray = cv2.cvtColor(img,
cv2.COLOR_BGR2GRAY)
After removing all the noises, higher order polynomial is needed to incorporate to
detect curved roads. For that, the algorithm must have an awareness of previous
detection and a value of confidence. In order to fit a higher order polynomial, the image
is sliced in many horizontal strips. On each strip, straight line algorithm was applied,
and pixels were identified corresponding to lane detections. Then, the pixels will be
combined all the stripes and a new polynomial function will be created that fits the best
(Bong et al. 2019). Finally, a blue mask is attached on the lane to see the difference
between the lanes and the other images on the road. Figure 7 and 8 shows the lane
detection on two different sets of roads.
Fig. 7. Lane detection on curved road Fig. 8. Lane Detection on straight road
In this work, white line markers are used to measure the speed of the vehicle. When
a line marker is present, number of frames will be tracked until the next lane marker
appears in the same window. Then, the number of frame rate of the video is also taken
which will be extracted from the video to determine the speed. Figure 9 shows the
architecture to calculate the linear speed of the vehicle.
Based on the Fig. 9, on a straight road, the far white lines will be ignored because
the distance between the dash cam and the white line is very far from the video.
Therefore, nearest white straight line which can be seen fully from the dash cam will be
taken. Starting and endpoint of the white dashed line will be taken to measure the
speed. When it goes from one frame to another, the position of the white dashed lines
will be measured (mcCormick 2019). Then, when the white dashed lines reach the end
point of it, the number of frames took to reach the point will be calculated. Equation 1
is the calculation to calculate the speed.
distance
Speed ¼ ; where time is number of frames
time
AutoMove: An End-to-End Deep Learning System 1089
To get the position of the white dashed lines, a cv2.rectangle is used. A rectangle is
drawn in between of two lines. Then, using matplotlib, the output is plotted to see the
position of rectangle. It is plotted to the nearest white line. Figure 10 shows the output
of the drawn rectangle.
Based on the output, if the number of frame counts is 14, then the speed of the
vehicle will be 10.7 km/h. This is calculated using the Eq. 1. Figure 11 shows the
output using matplotlib and cv2 and Fig. 12 shows the speed of the vehicle on a
straight road.
Fig. 11. Matplotlib speed detection Fig. 12. Speed detection actual road
To measure the speed on the curve, the same method is used. However, the position
of rectangle will be changed according to the position of the white lines. Figure 13
shows the speed of the vehicle on a curve road.
1090 S. Ramasamy and J. Joshua Thomas
y ¼ ao þ a1 x þ a2 x 2 þ e
In this formula, a and b parameters represent the constants and usually set to a = 1
and b = 100. Figure 14 shows the code snippet for Rosenbrock function.
def curve ( x, t, y ):
return x[0] * t * t + x[1] * t + x[2]
Fig. 15. Road angle calculation Fig. 16. Curved road angle average
Based on the Fig. 15, the difference angle for each frame is taken by subtracting top
angle with bottom angle of the rectangle which is being placed on the white-dashed
lines. However, the average angle is calculated using a formula called exponential
moving average, (EMA) (Fig. 16).
Fig. 17. Top angle of straight road Fig. 18. Bottom angle of straight road
Figure 19 and 20 shows the top and bottom angle of curved road which are used to
determine the difference angle (Fig. 21).
In this work, YOLO algorithm is used for object detection as it is faster compared
to other classification algorithms. In real time, YOLO algorithm process 45 frames per
second (Redmon and Farhadi 2016). Furthermore, even though it makes localization
error, less false positives will be predicted in the background. To make use of YOLO
algorithm, an image is divided as grids of 3 3 matrixes.
1092 S. Ramasamy and J. Joshua Thomas
Fig. 19. Top angle of curved road Fig. 20. Bottom angle of curved road
4.4 Dataset
To develop this work, a new dataset was created. A One Plus 6 model smartphone was
placed on a car’s dashboard. A video of the driving was taken at Jalan Barat Cassia 2,
Batu Kawan, Penang which consist of straight and curve roads. The video size is
1920 1080 and the frame rate is 30fps per second.
To determine the accuracy of the steering angle, the difference between actual and
predicted steering angle were measured. At the last frame of the video, the accuracy for
the whole model is 91.3784%. Table 1 shows the actual and predicted angle for straight
road.
Fig. 23. Straight road: AutoMove Fig. 24. Curved road: AutoMove
Figure 23, 24 shows the output of autonomous navigation straight road and curved
road. Based on the output, the steering of the vehicle is quite straight as the steering
wheel does not turn on a straight road. This is because, the steering angle is in between
zero to five degrees which was set to drive straight.
However, when the road is curved, the accuracy reduces to 85% which can be seen
in Table 2. Furthermore, the steering wheel is set according to few parameters, for
instance, if the steering angle is less than zero degrees, then the steering wheel will turn
left whereas if the steering angle is more than five degrees, the steering wheel will shift
to right. Table 2 shows the actual and predicted angle for curved road.
1094 S. Ramasamy and J. Joshua Thomas
5 Conclusion
This work has developed mainly to reduce accidents in Malaysia. According to Oth-
man O. Khalifa (2008), the United Nations has categorised Malaysia as 30th in the
ranking of highest number of fatal accidents with an average of 4.5 causalities per
10,000 vehicles. Furthermore, this system was also developed to help elderly people
who are too old to drive safely. This is because as driver age increases, changes in
flexibility, visual acuity, reaction time, memory and strength will affect the ability to
drive (Lutin et al. 2013).
Acknowledgement. The authors would like to thank the UOW Malaysia KDU Penang
University College on selecting this work for the launch of the new Batu Kawan campus,
mainland Penang, Malaysia. There is an urgent need for low-cost but high throughput novel road
analysis operations for self-driving vehicle in Batu Kawan UOW Malaysia Campus, hence this
simulation might provide to be one of the ways forward.
References
Thomas, J.J., Karagoz, P., Ahamed, B.B., Vasant, P.: Deep learning techniques and optimization
strategies in big data analytics. IGI Global 10, 978 (2020). https://doi.org/10.4018/978-1-
7998-1192-3
Thomas, J.J., Tran, H.N.T., Lechuga, G.P., Belaton, B.: Convolutional graph neural networks: a
review and applications of graph autoencoder in chemoinformatics. In: Thomas, J.J., Karagoz,
P., Ahamed, B. B., Vasant, P., (Ed.), Deep Learning Techniques and Optimization Strategies
in Big Data Analytics (pp. 107–123). IGI Global (2020) https://doi.org/10.4018/978-1-7998-
1192-3.ch007
Assidiq, A., Khalifa, O., Islam, M., Khan, S.: Real time lane detection for autonomous vehicles.
In: 2008 International Conference on Computer and Communication Engineering (2008)
AutoMove: An End-to-End Deep Learning System 1095
Bojarski, M., Del Testa, D., Dworakowski, D.: End to End Learning for Self-Driving Cars
(2016). https://arxiv.org/abs/1604.07316 Accessed 10 March 2019
Croucher, M.: Minimizing the Rosenbrock Function (2011). https://demonstrations.wolfram.
com/MinimizingTheRosenbrockFunction/ Accessed 11 October 2019
Curve Fitting and Interpolation (2019) [ebook] (2019). http://www.engineering.uco.edu/
*aaitmoussa/Courses/ENGR3703/Chapter5/ch5.pdf Accessed 6 October 2019
García Cuenca, L., Sanchez-Soriano, J., Puertas, E., Fernandez Andrés, J., Aliane, N.: Machine
learning techniques for undertaking roundabouts in autonomous driving. Sensors 19(10),
2386 (2019)
Geethapriya, S., Duraimurugan, N., Chokkalingam, S.: Real-Time Object Detection with
Yolo. Int. J. Eng. Adv. Technol. (IJEAT), 8(38) 578–581 (2019) https://www.ijeat.org/wp-
content/uploads/papers/v8i3S/C11240283S19.pdf Accessed 13 Nov. 2019
Huang, R., Pedoeem, J., Chen, C.: YOLO-LITE: A Real-Time Object Detection Algorithm
Optimized for Non-GPU Computers (2018). https://arxiv.org/pdf/1811.05588.pdf Accessed
13 Nov. 2019
Jalled, F.: Face Recognition Machine Vision System Using Eigenfaces (2017) https://arxiv.org/
abs/1705.02782 Accessed 3 March 2019
Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural
networks. Commun. ACM 60(6), 84–90 (2017)
Laguna, R., Barrientos, R., Blázquez, L., Miguel, L.: Traffic sign recognition application based
on image processing techniques. IFAC Proc. Vol. 47(3), 104–109 (2014)
Lane detection for self driving vehicles (2018). https://mc.ai/lane-detection-for-self-driving-
vehicles/ Accessed 8 Oct. 2019
Lutin, J., Kornhauser, L.A., Lerner-Lam, E.: The Revolutionary Development of Self-Driving
Vehicles and Implications for the Transportation Engineering Profession. Ite Journal (2013).
https://www.researchgate.net/publication/292622907_The_Revolutionary_Development_of_
SelfDriving_Vehicles_and_Implications_for_the_Transportation_Engineering_Profession
Accessed 23 Nov. 2019
Majaski, C.: Comparing Simple Moving Average and Exponential Moving Average (2019).
https://www.investopedia.com/ask/answers/difference-between-simple-exponential-moving-
average/ Accessed 9 Nov. 2019
McCormick, C.: CarND Advanced Lane Lines (2017). https://github.com/colinmccormick/
CarND-Advanced-Lane-Lines Accessed 12 Oct. 2019
Pant, A.: Introduction to Linear Regression and Polynomial Regression (2019). https://
towardsdatascience.com/introduction-to-linear-regression-and-polynomial-regression-
f8adc96f31cb Accessed 1 Nov. 2019
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You Only Look Once: Unified, Real-Time
Object Detection pp. 1–4 (2016)
Redmon, J., Farhadi, A.: YOLO9000: Better, Faster, Stronger (2016). https://pjreddie.com/media/
files/papers/YOLO9000.pdf Accessed 11 Nov. 2019
Sermanet, P., Eigen, D., Zhang, X., Mathieu, M.: OverFeat: Integrated Recognition, Localization
and Detection using Convolutional Networks (2013). https://arxiv.org/abs/1312.6229
Accessed 7 Nov. 2019
tf.keras.Model| TensorFlow Core r2.0 (2019) https://www.tensorflow.org/api_docs/python/tf/
keras/Model Accessed 23 Nov. 2019
Trubia, S., Canale, A., Giuffrè, T., Severino, A.: Automated Vehicle: a Review of Road Safety
Implications as Driver of Change (2017)
Vasant, P., Zelinka, I., Weber, G.: Intelligent Computing & Optimization. Springer (2018)
Vasant, P., Zelinka, I., Weber, G.: Intelligent Computing and Optimization. Springer
International Publishing (2019)
1096 S. Ramasamy and J. Joshua Thomas
Venketas, W.: Exponential Moving Average (EMA) Defined and Explained (2019). https://www.
dailyfx.com/forex/education/trading_tips/daily_trading_lesson/2019/07/29/exponential-
moving-average.html Accessed 12 Nov. 2019
Vitelli, M., Nayebi, A.: CARMA: A Deep Reinforcement Learning Approach to Autonomous
Driving (2016). https://www.semanticscholar.org/paper/CARMA-%3A-A-DeepReinforce
ment-Learning-Approach-to-Vitelli-Nayebi/b694e83a07535a21c1ee0920d47950b4800b08bc
Accessed 16 Nov. 2019
Wagh, P., Thakare, R., Chaudhari, J., Patil, S.: Attendance system based on face recognition
using eigen face and PCA algorithms. In: 2015 International Conference on Green
Computing and Internet of Things (ICGCIoT) (2015)
Joshua Thomas, J., Pillai, N.: A deep learning framework on generation of image descriptions
with bidirectional recurrent neural networks. In: Vasant, P., Zelinka, I., Weber, G.W.
(eds) Intelligent Computing & Optimization. ICO 2018. Advances in Intelligent Systems and
Computing, vol. 866. Springer, Cham (2019) https://doi.org/10.1007/978-3-030-00979-3_22
Joshua Thomas, J., Belaton, B., Khader, A.T.: Visual analytics solution for scheduling
processing phases. In: Vasant, P., Zelinka, I., Weber, G.W., (eds) Intelligent Computing &
Optimization. ICO 2018. Advances in Intelligent Systems and Computing, vol. 866. Springer,
Cham (2019) https://doi.org/10.1007/978-3-030-00979-3_42
Thomas, J. J., Ali, A. M.: Dispositional learning analytics structure integrated with recurrent
neural networks in predicting students performance. In: International Conference on
Intelligent Computing & Optimization. pp. 446–456. Springer, Cham (2019)
Bong, C.W., Xian, P.Y., Thomas, J.: Face recognition and detection using haars features with
template matching algorithm. In: International Conference on Intelligent Computing &
Optimization. pp. 457–468. Springer, Cham (2019)
Thomas, J.J., Fiore, U., Lechuga, G. P., Kharchenko, V., Vasant, P.: Handbook of Research on
Smart Technology Models for Business and Industry. IGI Global (2020) https://doi.org/10.
4018/978-1-7998-3645-2
Thomas, J.J., Wei, L.T., Jinila, Y.B., Subhashini, R.: Smart computerized essay scoring using
deep neural networks for universities and institutions. In: Thomas, J.J., Fiore, U., Lechuga, G.
P., Kharchenko, V., Vasant, P. (Ed.), Handbook of Research on Smart Technology Models
for Business and Industry. pp. 125–152. IGI Global (2020) https://doi.org/10.4018/978-1-
7998-3645-2.ch006
An Efficient Machine Learning-Based
Decision-Level Fusion Model to Predict
Cardiovascular Disease
1 Introduction
Coronary heart failure is one of the primary causes of mortality globally. It is
commonly regarded as a primary disease in the old and middle ages. In world-
wide, this Disease (CAD) particularly has the highest rate of mortality. In prac-
tice, the disease is treated by physicians, but there are very few cardiac experts
in contrast to the patient’s number. Particularly, the traditional method of diag-
nosis of the disease requires time, and it is expensive. Furthermore, at the initial
stage, the symptoms are mild, so people usually ignore it until it gets serious.
False diagnosis, expensive tests are the main reasons people can not depend so
much on doctors. Money also plays a crucial role in this issue. Researchers are
trying their best to develop a more effective intelligent system for the diagnosis
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 1097–1110, 2021.
https://doi.org/10.1007/978-3-030-68154-8_92
1098 H. B. Kibria and A. Matin
of heart disease, and also a number of smart models have been developed over
time. The main reason behind our motivation for designing a combined model
is that it would improve the health care sector and reduce patients’ struggle. It
will also save a lot of time for both physicians and the patients, and again, the
patients will save from having an extra expense.
A medical diagnosis system to predict cardiovascular disease based on
machine learning gives more accurate results than traditionally, and the treat-
ment cost reduces [5]. Medical diagnosis is a test where a physician tries to
identify the disease by experimenting with the symptoms and the lab values
obtained from the test. Various algorithms are used in that task [7].
In this paper, an automated medical diagnosis system is used to classify
coronary disease based on a machine-learning algorithm. The system is based
on a backpropagation algorithm that uses an Artificial Neural Network learning
method. Also, the KNN algorithm is used for disease classification, and the per-
formance of the two algorithms is observed. Finally, the two algorithms’ decision
is fused by summation to make a single decision that gave a better result than
each individual model [8]. So, the problem we are solving in this paper is to
detect patients with or without heart disease by analyzing the medical data. If
the patients do not have heart disease, then there is nothing to worry about, but
if the result is positive (the patient has heart disease), he can undergo further
treatment. As machine learning can process huge datasets, it can give results at
very low cost and within a short time, improving the quality of health care and
saving the physicians from a lot of pressure.
This paper is presented as follows. Cardiovascular disease and the importance
of early identification of disease are introduced in the first section. The second
section discusses previous studies of medical diagnosis based on machine learning.
The third section describes the method that is used to develop the model. In
segment four, the paper analyzes the experimental findings of our suggested
decision level fusion model. Then in section five, our fusion model’s performance
has been compared with the related work done by so far. Finally, in section six,
we end with a conclusion and possible future development of the work.
2 Related Study
SVM linear-kernel and got an accuracy of 86.8% using 80% data in training.
They have experimented with different train-test ratios in their work, but the
accuracy was not satisfactory. 86.66% accuracy was achieved using 75% data in
training. The best result among other algorithms was the SVM linear kernel.
The main weakness of this work is its poor accuracy.
In [5], researchers proposed a system that was designed with a multilayer
perceptron neural network. Backpropagation was used for training and their pro-
posed system. They have calculated different accuracy by adjusting the number
of sizes in the hidden layer. With eight nodes in the hidden layer, the maximum
accuracy of 95% was reached with PCA. Other performance parameters were
observed with respect to different sizes of the hidden layer.
Another study was done [12], where two supervised data mining algorithms
were applied to classify a patient’s heart disease using two classification models,
Naı̈ve Bayes and Decision tree classifier. The decision tree classifier predicted
better with an accuracy of 91%, where Naı̈ve Bayes showed 87% accuracy. In
[1], researchers analyzed machine learning methods for different types of dis-
ease prediction. Logistic regression (LR), XGBoost (XGB), random forest (RF)
and LSTM, which is a special kind of recurrent neural network, were used for
prediction. Here XGBoost performed better than LSTM.
In a recent study [13], researchers used four different algorithms for heart
disease prediction. The highest accuracy was achieved from KNN, which was
90.789% using test-rain split. Naı̈ve Bayes and random forest both gave 88.15%
and 86.84% accuracy, respectively, using the Cleveland dataset. No complex or
combinational model for higher accuracy was introduced in this study, which is
a drawback.
There is scope for improving these work accuracy by implementing a fuzzy
method or combining separate algorithms to make a new one. So we have pre-
ferred to use the fusion model to classify disease, which can provide greater
accuracy.
3 Proposed Solution
This research aims to design a decision-level fused model that can improve accu-
racy and identify heart disease efficiently. This fusion model classifies patients
with and without heart disease by merging two machine learning-based algo-
rithms. The proposed solution is displayed in Fig. 1:
In this proposed architecture, there are several steps. First of all, the data
were preprocessed. Using the chi-square test, only the critical features were
selected that contribute most to the prediction. For the training with KNN,
the top 10 important features were chosen for prediction. And for the artificial
neural network, all 13 features were taken. After training, we fused the individ-
ual models’ decision by summation to get a more accurate outcome. The phases
are described in detail below.
The dataset UCI for heart disease was used in heart disease classification. The
Cardiovascular disease data set has been taken from the UCI machine learning
repository [2]. It has 303 instances with 13 features with some missing values.
This repository holds a huge and varied range of data sets that are of interest
to various sectors. The machine learning group uses these data to contribute to
developments in different domains. The repository had been developed by David
Aha [2]. The dataset is labeled into those two classes. The target value contains
two categories: 0 and 1. 0 means no heart disease, and 1 represents having heart
disease, which is a binary classification. The explanation of the dataset has been
shown in Table 1 [2].
Features Description
Age Lifetime in years (29–77)
Gendr Instance of gender (0 = Female, 1 = Male)
ChstPainTyp The type of chest pain (1 = typical angina,
2 = atypical angina, 3 = non-anginal pain,
4: asymptomatic)
RestBlodPresure Resting blood pressure in mm Hg [94, 200]
SermCholstrl Serum cholesterol in mg/dl [126–564]
FstingBlodSugr Fasting blood sugar >120 mg/dl (0 = False, 1= True)
ResElctrcardigrphic Results of Resting ECG (0: normal, 1: ST-T wave
abnormality, 2: LV hypertrophy)
MaxHartRte Maximum heart rate achieved [71, 202]
ExrcseIndcd Exercise induced angina (0: No, 1: Yes)
Oldpek ST depression induced by exercise relativeto rest
[0.0, 62.0]
Slp Slope of the peak exercise ST segment
(1 = up-sloping, 2 = flat, 3 = down-sloping)
MajrVesels Number of major vessels coloured by fluoroscopy
(values 0–3)
Thl Defect types: value 3 = normal, 6 = fixed defect,
7 = irreversible defect
Decision-Level Fusion Model to Predict Cardiovascular Disease 1101
Fig. 2. (continued)
Figure 2 shows the data distribution of all 13 attributes in terms of the target
class. The target class distribution has been displayed here as well.
Feature scaling was applied to limit the range of variables. We have used min-
max normalization for feature scaling.
Qv − min(Qv )
Qv = (1)
max(Qv ) − min(Qv )
where Qv is an original value, Qv is the normalized value. And then, the dataset
has been splitted into two parts: testing and training. 75% of the data were used
for training and 25% for testing. 10% data was taken from training data as a
validation set. The validation set prevents training data from over-fitting.
There are 13 features in our dataset, and only ten of them were used for the
training of KNN. The top significant features were selected using the chi-square
test. It selects the features that have the best connection with output. Sex,
fasting blood sugar, and resting ECG were excluded with the help of the chi-
square test. So ten features out of 13 were used in KNN. Figure 3 has represented
the ten most essential features according to their score.
Decision-Level Fusion Model to Predict Cardiovascular Disease 1103
test data; thus, the fusion model’s accuracy increases. Usually, both separate
algorithms gave accurate decisions for the same test data; this situation occurs
only with one or two test data, and that’s made the fusion model more reliable
and efficient.
4 Performance Evaluation
A decision-level fusion model has been constructed by using an artificial neural
network and k-nearest neighbor algorithm. The two algorithms were applied
individually on the same dataset, and then the decision score of each algorithm
was combined to improve the accuracy. And we got a better result after fusion.
(a) Train and validation accuracy (b) Train and validation loss
Fig. 5. Train and validation accuracy and loss of ANN in terms of training number
Fig. 6. Variation of the accuracy of train and test regarding the KNN neighbor number
Decision-Level Fusion Model to Predict Cardiovascular Disease 1107
Fig. 7. Confusion matrix and ROC curve of decision level fusion model
Figure 5 displays the training and validation accuracy and loss with respect
to the epoch. Fifty epochs have been used. In the graph, validation is mentioned
as a test.
The relationship between the test and training accuracy regarding the vari-
ance of KNN neighbour numbers has been shown in Fig. 6. From this graph, we
selected the neighbor number k = 12 and achieved an accuracy of 88.16% for
test data.
Table 3 shows the various performance parameters, which increases when
two algorithms’ decisions were combined. Thus the decision score level fusion
became more efficient for predicting disease. Also, performance parameters have
been displayed. Here the fusion’s model accuracy has been represented only with
testing data, as fusion model was not trained.
Approach Tp Fp Accuracy (train) Accuracy (test) Precision Recall F1-score Roc-auc score
Fn Tn
ANN 40 2 82.35 92.10 92 92 92 91.73
4 30
KNN 41 1 81.15 88.16 89 88 88 87.04
8 26
ANN+KNN 41 5 – 93.4 94 93 93 93.94
1 29
In Table 3, the accuracy for test and train data has been measured. And for
the fusion model, only the accuracy for test data is displayed as decision-level
fusion has been implemented after training two algorithms. The fusion model’s
accuracy has risen 1% comparing ANN and 5% comparing KNN. All the other
1108 H. B. Kibria and A. Matin
parameters also improved. The confusion matrix is generated from the training
dataset.
From the values in Table 3, it can be said that ANN performed well com-
pared to KNN, and the performance of the fusion model increased after merging.
The decision-level fusion model performed much better than the two individual
algorithms. Here each model is used independently for prediction, and then the
decision score from each model has been combined to improve accuracy.
Figure 7(a) shows the confusion matrix of our fusion model. The miss classi-
fication rate is 6.57%. And in Fig. 7(b), the ROC curve has been displayed.
The comparison of the performance of the three algorithms (ANN, KNN, and
ANN+KNN) was displayed in Figs. 8 to show to improvement of fusion model.
6 Conclusion
This work aims to develop a decision score level fusion model that can provide a
better result than one individual algorithm. This model creates a final decision
using the decision score from two other models. Artificial neural network and
k-nearest neighbor both has given good results individually for the prediction of
cardiovascular disease. By merging the decision score of two algorithms, a signif-
icant improvement is noticed. If one algorithm somehow gives the wrong result
for a particular data and other predicts right about it, so there is a possibility
that the correct result will be obtained in the fusion model. That’s why in the
medical diagnosis fused model is dominating. Our model gave an accuracy of
93.42%, and the separate model’s accuracy was 92.10% and 88.16% for ANN
and KNN, respectively. In this paper, only the cardiovascular disease dataset
has been used. This decision level fusion model can also be used for other med-
ical datasets. In the future, this fused model can be enhanced by using different
algorithms. Furthermore, more than two classification algorithms can be used to
build a new model to gain more accurate result.
References
1. Christensen, T., Frandsen, A., Glazier, S., Humpherys, J., Kartchner, D.: Machine
learning methods for disease prediction with claims data. In: 2018 IEEE Interna-
tional Conference on Healthcare Informatics (ICHI), pp. 467–4674. IEEE (2018)
2. Dua, D., Graff, C.: UCI machine learning repository (2017). http://archive.ics.uci.
edu/ml
3. Harimoorthy, K., Thangavelu, M.: Multi-disease prediction model using improved
svm-radial bias technique in healthcare monitoring system. J. Ambient Intell. Hum.
Comput. 1–9 (2020)
1110 H. B. Kibria and A. Matin
4. Hasan, T.T., Jasim, M.H., Hashim, I.A.: Heart disease diagnosis system based on
multi-layer perceptron neural network and support vector machine. Int. J. Curr.
Eng. Technol. 77(55), 2277–4106 (2017)
5. Karayılan, T., Kılıç, Ö.: Prediction of heart disease using neural network. In: 2017
International Conference on Computer Science and Engineering (UBMK), pp. 719–
723. IEEE (2017)
6. Kotsiantis, S., Kanellopoulos, D., Pintelas, P.: Data preprocessing for supervised
leaning. Int. J. Comput. Sci. 1(2), 111–117 (2006)
7. Louridi, N., Amar, M., El Ouahidi, B.: Identification of cardiovascular diseases
using machine learning. In: 2019 7th Mediterranean Congress of Telecommunica-
tions (CMT), pp. 1–6. IEEE (2019)
8. Matin, A., Mahmud, F., Ahmed, T., Ejaz, M.S.: Weighted score level fusion of iris
and face to identify an individual. In: 2017 International Conference on Electrical,
Computer and Communication Engineering (ECCE), pp. 1–4. IEEE (2017)
9. Mohan, S., Thirumalai, C., Srivastava, G.: Effective heart disease prediction using
hybrid machine learning techniques. IEEE Access 7, 81542–81554 (2019)
10. Nourmohammadi-Khiarak, J., Feizi-Derakhshi, M.R., Behrouzi, K., Mazaheri, S.,
Zamani-Harghalani, Y., Tayebi, R.M.: New hybrid method for heart disease diag-
nosis utilizing optimization algorithm in feature selection. Health Tech. 1–12 (2019)
11. Ramya Perumal, K.A.: Early prediction of coronary heart disease from cleveland
dataset using machine learning techniques. Int. J. Adv. Sci. Technol. 29(06), 4225 –
4234, May 2020. http://sersc.org/journals/index.php/IJAST/article/view/16428
12. Krishnan, S., Geetha, J.S.: Prediction of heart disease using machine learning algo-
rithms. In: 1st International Conference on Innovations in Information and Com-
munication Technology (ICIICT). IEEE (2019)
13. Shah, D., Patel, S., Bharti, S.K.: Heart disease prediction using machine learning
techniques. SN Comput. Sci. 1(6), 1–6 (2020)
14. Telnoni, P.A., Budiawan, R., Qana’a, M.: Comparison of machine learning classifi-
cation method on text-based case in twitter. In: 2019 International Conference on
ICT for Smart Society (ICISS). vol. 7, pp. 1–5. IEEE (2019)
15. Thomas, J., Princy, R.T.: Human heart disease prediction system using data min-
ing techniques. In: 2016 International Conference on Circuit, Power and Computing
Technologies (ICCPCT), pp. 1–5. IEEE (2016)
16. Vasant, P., Zelinka, I., Weber, G.W.: Intelligent Computing & Optimization, vol.
866. Springer, Berlin (2018)
Towards POS Tagging Methods
for Bengali Language: A Comparative
Analysis
Fatima Jahara, Adrita Barua, MD. Asif Iqbal, Avishek Das, Omar Sharif ,
Mohammed Moshiul Hoque(B) , and Iqbal H. Sarker
1 Introduction
POS tagging has significant importance in many NLP applications such as pars-
ing, information retrieval, speech analysis, and corpus development. Moreover,
it is used as a pivotal component to build a knowledge base for natural language
analyzer. It makes the syntactic parser effective as it resolves the problem of
input sentence ambiguity. Tagging of words is significantly useful since they are
used as the input in various applications where it provides the linguistic signal on
how a word is being used within the scope of a phrase, sentence, or document.
POS tagging directly affects the performance of any subsequent text process-
ing steps as it makes the processing easier when the grammatical information
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 1111–1123, 2021.
https://doi.org/10.1007/978-3-030-68154-8_93
1112 F. Jahara et al.
about the word is known. Usually, supervised and unsupervised approaches are
employed in POS tagging, which are further divided into rule-based, stochas-
tic based, and transformation based methods. The rule-based POS tagging uses
a dictionary or lexicon for taking the possible tags of a word. The stochastic
method considers the highest frequency or probability value to assign a POS tag.
Few stochastic tagging methods such as N-grams, CRFs, and HMMs have been
implemented for Bengali, English and other languages [1–3]. The transformation-
based method combines rule-based and stochastic techniques such as Brill
Tagger.
Designing a POS tagger is a very challenging task for a resource poor lan-
guage like Bengali. POS tagging of the Bengali sentence is complicated due
to its complex morphological structure, the dependency of the subject on verb
infections, person-verb-tense-aspect agreement and the scarcity of pre-tagged
resources [4,5]. Moreover, the ambiguity of a word with multiple POS tags and
the lack of availability of language experts in Bengali language posses other
obstacles that need to overcome. Most of the previous works on POS tagging
in Bengali neither highlighted the tagging effectiveness nor investigated their
appropriateness. Thus, to address this issue, this paper empirically investigates
the performance of 16 POS tagging methods using a supervised approach on
a corpus containing 7390 Bengali sentences. Comparative analysis in terms of
execution time and accuracy are reported, which helps to decide the suitable use
of POS tagging technique for various language processing tasks in Bengali.
2 Related Work
Different approaches have been explored on POS tagging in Bengali and other
languages. Stochastic and transformation based methods are the most widely
used techniques where a large dataset is prerequisite to achieve good perfor-
mance. Hasan et al. [6] showed a comparative analysis of n-grams, HMM and
Brill transformation-based POS tagging for south Asian languages. A tagset of
26 tags used for Bengali, Hindi and Telegu languages which gained 70% accu-
racy for Bengali using Brill tagger. Another work implemented the trigram and
HMM tagging methods [7] for the Marathi language. A comparison between the
stochastic (HMM, Unigram) and transformation based (Brill) methods is pre-
sented by Hasan et al. [8]. This work used a small training set of 4048 tokens in
Bengali and experimented with two different tagsets (12-tag and 41-tag). The
results revealed that Brill tagger performed better than the other stochastic
methods. A stochastic based approach proposed by Ekbal et al. [9] concluded
that a maximum entropy-based method outperforms the HMM-based POS tag-
ging method for Bengali. Ekbal et al. [10] developed a POS tagger for Bengali
sentences using CRF in the name entity recognition task. PVS et al. [11] showed
that CRF, along with transformation-based learning, achieved 76.08% accuracy
for Bengali POS tagging.
The supervised tagging methods demanded a large amount of tagged data
to achieve high accuracy. Dandapat et al. [12] used a semi-supervised method of
Towards POS Tagging Methods for Bengali Language: 1113
POS tagging with HMM and Maximum Entropy (ME). Hossain et al. [13] devel-
oped a method that checked whether the construction of the Bengali sentence
is valid syntactically, semantically or pragmatically. They designed a rule-based
algorithm using context-free grammars to identify all POS meticulously. Roy
et al. [14] developed a POS tagger that identifies 8 POS tags in Bengali using
grammar and suffix based rules. However, they only considered word-level tag
accuracy, which failed to identify the correct tag sequence in sentences. Sakiba
et al. [15] discussed a POS tagging tool where they used a predefined list of POS
tags and applied rules to detect POS tags from Bengali texts. Their method used
a very small data set containing 2000 sentences, and it faced difficulties due to
the limited rules. Chakrabarti et al. [16] proposed a POS tagging method using
the layered approach. Rathod et al. [17] surveyed different POS tagging tech-
niques such as rule-based, stochastic, and hybrid for Indian regional languages
where the Hybrid method performed better. Most of the previous approaches
of POS tagging in Bengali experimented on a smaller dataset which limits to
investigate their effectiveness on evaluation concerning diverse tagsets. In this
work, a bit larger dataset consisting of 7390 sentences are used.
3 Methodology
The key objective of our work is to investigate the performance of different types
of POS tagging techniques under supervised approach. To serve our purpose, we
used a tagged corpus in Bengali developed by Linguistic Data Consortium [18].
The overall process of the work consists of five significant steps: tokenization,
training/testing corpus creation, POS tagger model generation, tagged sentence
generation, and evaluation. Figure 1 illustrates the abstract representation of
the overall process to evaluate POS taggers.
3.2 Tokenization
Tokenization is the task of slicing a sequence of character into pieces, called
tokens [20]. A token is a string of contiguous characters grouped as a semantic
unit and delimited by space and punctuation marks. These tokens are often
1114 F. Jahara et al.
loosely referred to as terms or words. In our corpus, 7390 tagged sentences were
tokenized into a total of 115 K tokens with 22,330 unique tokens. A sample
tagged sentence and its corresponding tokens are shown in the following.
The tokenized dataset is divided into two sets: train corpus and test corpus. The
training corpus consists of 98 K tagged tokens, while the test corpus contained
17 K tokens. The data in the test corpus is untagged to use in the testing phase
for evaluation. A data sample in the training corpus (with 11-tagset and 30-
tagset) and the testing corpus is illustrated in the following.
Towards POS Tagging Methods for Bengali Language: 1115
The training set is used to train different POS tagging methods. Each tagging
method is trained on training corpus tagged with 11 and 30 tagsets–each of
16 POS tagging methods used in a unique way to generate their corresponding
training models. N-gram, HMM, TnT, and CRF tagger models generate feature
matrices which is used in calculating the probability of the tags. The Brill tagger
generates rules used to estimate tags, and the HMM model makes a transition
probability matrix called Markov Matrix. An N-gram tagger follows the Bag of
Words approach while the CRF tagger uses a statistical approach.
C(τi−(n−1) , ..., τi )
P (τi |τi−(n−1) , ..., τi−1 ) = (2)
C(τi−(n−1) , ..., τi−1 )
HMM Tagger. In HMM, the hidden states are the POS tags (τ1 , τ2 , ....., τn )
and the observations are the words themselves (ω1 , ω2 , ......, ωn ). Both transition
and emission probabilities are calculated for determining the most appropriate
tag for a given word in a sequence.
The overall probability of a tag τi given a word ωi is,
P (τi |ωi ) = P (ωi |τi ).P (τi |τi−1 ).P (τi+1 |τi ) (3)
Here, P(ωi —τi ) is the probability of the word ωi given the current tag τi ,
P(τi —τi−1 ) is the probability of the current tag τi given the previous tag τi−1
and P(τi+1 —τi ) is the probability of the future tag τi+1 given the current
tag τi .
Towards POS Tagging Methods for Bengali Language: 1117
TnT Tagger. In TnT, the most appropriate sequence is selected based on the
probabilities of each possible tag. For a given sequence of words of length n, a
sequence of tags is calculated by the Eq. 4.
n
arg max [P (τi |τi−1 , τi−2 ).P (ωi |τi )]P (τn |τn−1 ) (4)
τ1 ...τn
i=1
where, P(τi —τi−1 , τi−1 ) denotes the probability of the current tag τi given the
two previous tags τi−1 , and τi−2 . P(ωi —τi ) indicates the probability of the word
ωi given the current tag ωi , and P(τn —τ(n−1) ) denotes the probability of the
tag τn given the previous tag τ(n−1) .
Here, fk (τi ,τi−1 ,ωi ) represents a feature function whose weight is θk , and Z(w)
is a normalization factor which is the sum of all possible tag sequences.
τ 1 → τ2 (6)
Stochastic taggers can be used as a back-off tagger with the Brill Tagger. In
our work, to investigate the effect of back-off tagger on tagging performance, we
used uni-gram, bi-gram, tri-gram, uni-gram+bi-gram, uni-gram+bi-gram+tri-
gram, HMM, TNT, and CRF tagging methods with Brill tagger.
The generated POS tagging model predicts the highest probability of a tag
against the token and labels it with the appropriate POS tag. This process reads
the untagged tokens and calculates the probabilities of different tags based on
the trained tagger model. Stochastic tagger models (such as N-gram, HMM,
TnT, and CRF) use the feature matrices to calculate the probability of the tags.
The transformation-based (i.e., Brill tagger) model use the generated rules to
estimate the probability of the tags. After POS tagging of each token, individual
1118 F. Jahara et al.
lists of tagged tokens are created for each sentence. Algorithm 1 describes the
process of converting the tagged tokens lists into the tagged sentences.
Algorithm 1: Tagged tokens to tagged sentence generation
T ←List of Tagged Tokens
tagged sentence ←[] List initialization
for t ∈ T do
S ←”” Tagged sentence initialization
for token ∈ t do
S ←S + token[word] + ”\” + token[tag] + ””
end
tagged sentence.append(S);
end
To examine the effect of the various corpus size on the performance of POS
taggers, the taggers were trained with different amounts of tokens from the
same corpus. The tagged corpus is partitioned into different sizes as train sets
such as 10 K, 20 K, 30 K, 40 K, 50 K, 60 K, 70 K, 80 K, 90 K, 100 K, 115 K. Each
model is trained with each partitioned data sets individually and tested over the
17 K untagged testing dataset. Figure 2 shows the performance of the different
POS tagging methods for various sizes of training corpus using a 30-tagset.
From the Figure, it observed that Brill+CRF tagger has the highest accuracy
even when the corpus size is small. Again, both CRF and Brill+CRF tagger
reached almost 75% (for 30-tagset) and 85% (for 11-tagset) of accuracy with a
10 K tagged set. The accuracy of each method increased sharply with the rise of
the data set and becomes almost steady at 100 K.
The performance of the tagger also depends on the execution time. The faster
the execution time, the better the performance. We have computed the execution
time of taggers into 11-tagset and 30-tagset. Table 3 shows the performance
comparison concerning execution time among 16 tagging techniques.
The amount of time required to train a model using the train set decides the
training time of the tagger. From Table 3, it is observed that the HMM tagger
requires the least training time (0.25 s) whereas, Brill+TnT requires the highest
training time (333.38 s) for 30-tagset. For the 11-tagset, HMM consumed 0.25 s
and Brill+TnT required 164.56 s of training time. In the case, if testing time,
1120 F. Jahara et al.
30 Tagset 11 Tagset
POS tagger Training Testing Execution Training Testing Execution
time (s) time (s) time (s) time (s) time (s) time (s)
Unigram 0.43 0.02 0.45 0.34 0.02 0.37
Bigram 0.6 0.03 0.63 0.59 0.03 0.62
Trigram 0.76 0.04 0.8 0.67 0.03 0.7
Uni-gram+bi-gram 1.0 0.04 1.03 1.04 0.04 1.07
Uni-gram+bi- 1.79 0.05 1.84 1.66 0.05 1.71
gram+tri-gram
HMM 0.25 6.88 7.13 0.25 3.8 4.05
TnT 0.53 44.87 45.39 0.5 19.01 19.51
CRF 49.94 0.14 50.08 15.97 0.11 16.09
Brill+uni-gram 21.54 0.16 21.7 16.3 0.2 16.51
Brill+bi-gram 67.08 0.79 67.87 42.97 0.77 43.74
Brill+tri-gram 74.24 0.94 75.17 55.75 0.85 56.6
Brill+uni- 12.36 0.2 12.56 12.79 0.21 13.0
gram+bi-gram
Brill+uni- 9.64 71.67 81.31 10.62 75.45 86.07
gram+bi-
gram+tri-gram
Brill+HMM 39.84 5.76 45.6 25.24 3.5 28.73
Brill+TnT 333.38 44.81 378.19 164.56 22.56 187.12
Brill+CRF 88.56 0.48 89.04 36.83 0.5 37.33
Towards POS Tagging Methods for Bengali Language: 1121
uni-gram tagger utilized the lowest tagging time (0.2 s) in both tagsets, whereas
Brill+unigram+bigram+trigram required the highest tagging time about 71.67 s
(for 30-tagset) and 75.45 s (for 11-tagset) respectively.
The execution time determines the tagging speed of the POS tagging tech-
niques. Figure 3 shows the execution time required for each tagging methods on
our dataset. Results indicate that the Brill+TnT demand more execution time
compared to other POS tagging methods.
From the results, it is investigated that Brill taggers, along with other back-
off taggers achieved the higher accuracy but they lag in execution time. The
Brill+CRF obtained the highest accuracy of 91.83% (for 11 tagset), but it
requires a higher execution time (37.33 s). On the other hand, the CRF method
achieved 90.84% of accuracy and consumes 16.09 s for execution. Thus, there is a
trade-off between accuracy and execution time. Taking into consideration both
the accuracy and execution time, it revealed that the CRF method provided
better POS tagging performance compared to other techniques.
5 Conclusion
In this work, we have illustrated and investigated the different POS tagging
techniques for the Bengali language. A comparative analysis of 16 POS tagging
techniques (eight stochastic-based and eight transformations-based) on a tagged
corpus consisting of 1,15,000 tokens have been reported. The comparative analy-
sis revealed that Brill with CRF technique achieved the highest accuracy among
1122 F. Jahara et al.
other POS tagging techniques, but it requires more execution time. CRF can
be maintained as a good trade-off between accuracy and execution time, and
this method can be used as a better POS tagging technique. Tagging methods
that include both statistical and linguistic knowledge may produce a better per-
formance. The performance of the other tagging techniques such as TAGGIT,
CLAWS, Xerox, and Hybrid can be investigated further on the larger tagged
corpus with more unique words. These issues will be addressed in the future.
References
1. Dandapat, S., Sarkar, S.: Part of speech tagging for Bengali with Hidden Markov
Model. In: Proceedings of NLPAI Machine Learning Competition (2006)
2. Diesner, J.: Part of speech tagging for English text data. School of Computer
Science Carneige Mellon University Pittsburgh, PA 15213
3. Manju, K., Soumya, S., Idicula, S.M.: Development of a POS tagger for Malayalam-
an experience. In: International Conference on ARTCom, pp. 709–713. IEEE (2009)
4. Haque, M., Hasan, M.: Preprocessing the Bengali text: an analysis of appropriate
verbs (2018)
5. Bhattacharya, S., Choudhury, M., Sarkar, S., Basu, A.: Inflectional morphology
synthesis for Bengali noun, pronoun and verb systems. Proc. NCCPB 8, 34–43
(2005)
6. Hasan, M.F., UzZaman, N., Khan, M.: Comparison of Unigram. HMM and Brill’s
POS tagging approaches for some South Asian languages, Bigram (2007)
7. Kumawat, D., Jain, V.: POS tagging approaches: a comparison. Int. J. of Com.
App. 118 (6) (2015)
8. Hasan, F.M., UzZaman, N., Khan, M.: Comparison of different POS tagging tech-
niques (N-Gram, HMM and Brill’s tagger) for Bangla. Advances and Innovations
in Systems. Computing Sciences and Software Engineering, pp. 121–126. Springer,
Berliin (2007)
9. Ekbal, A., Haque, R., Bandyopadhyay, S.: Maximum entropy based Bengali part
of speech tagging. J. Res. Comput. Sci. 33, 67–78 (2008)
10. Ekbal, A., Haque, R., Bandyopadhyay, S.: Bengali part of speech tagging using
conditional random field. In: International Conference on SNLP2007, pp. 131–136
(2007)
11. PVS, A., Karthik, G.: Part-of-speech tagging and chunking using conditional ran-
dom fields and transformation based learning. Shallow Parsing South Asian Lang.
21, 21–24 (2007)
12. Dandapat, S., Sarkar, S., Basu, A.: Automatic part-of-speech tagging for Bengali:
an approach for morphologically rich languages in a poor resource scenario. In:
Proceedings of 45th Annual Meeting of ACL Companion, pp. 221–224 (2007)
13. Hossain, N., Huda, M.N.: A comprehensive parts of speech tagger for automatically
checked valid Bengali sentences. In: International Conference ICCIT, pp. 1–5. IEEE
(2018)
14. Roy, M.K., Paull, P.K., Noori, S.R.H., Mahmud, H.: Suffix based automated parts
of speech tagging for Bangla language. In: International Conference on ECCE, pp.
1–5. IEEE (2019)
15. Sakiba, S.N., Shuvo, M.U.: A memory efficient tool for Bengali parts of speech
tagging. In: Artificial Intelligence Techniques for Advanced Computer Application,
pp. 67–78. Springer (2020)
Towards POS Tagging Methods for Bengali Language: 1123
16. Chakrabarti, D., CDAC, P.: Layered parts of speech tagging for Bangla. Language
in India: Problems of Parsing in Indian Languages (2011)
17. Rathod, S., Govilkar, S.: Survey of various POS tagging techniques for Indian
regional languages. Int. J. Comput. Sci. Inf. Technol. 6(3), 2525–2529 (2015)
18. Bali, Kalika, M.C., Biswas, P.: Indian language part-of-speech tagset: Bengali.
Philadelphia: Linguistic Data Consortium (2010)
19. Sankaran, B., Bali, K., Choudhury: A common parts-of-speech tagset framework
for Indian Languages (01 2008)
20. Rai, A., Borah, S.: Study of various methods for tokenization. In: Application of
IoT, pp. 193–200. Springer (2021)
BEmoD: Development of Bengali
Emotion Dataset for Classifying
Expressions of Emotion in Texts
1 Introduction
Recognizing emotion in expression involves the task of attributing an emotion
label to an expression preferred from a set of pre-specific labels of emotion. There
are several application areas where there is a need to understand and interpret
emotions in text expressions such as business, politics, education, sports, and
entertainment. Recognizing emotion in text expression is one of the critical tasks
in NLP, which demands an understanding of natural language. The hindrance
commences at the sentence level in which an emotion is stated through the
semantics of words and their connections; as the level enhances, the problem’s
difficulty grows. Nevertheless, not all opinions are stated explicitly; there are
metaphors, mockery, and irony.
c The Author(s), under exclusive license to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 1124–1136, 2021.
https://doi.org/10.1007/978-3-030-68154-8_94
BEmoD: Development of Bengali Emotion Dataset 1125
2 Related Work
Different corpora have been developed to detect emotions in text for various
languages. Alm et al. [4] developed a dataset consisting of 185 tales, in which
labeling is done at the sentence level with sadness, happy, neutral, anger-disgust,
fear, positive surprise, and negative surprise classes. A dataset consisting of blog
posts representing eight emotion labels, including Ekman’s six basic emotions
[5]. ISEAR2 [6] is a corpus with joy, fear, anger, sadness, disgust, shame, and
guilt classes. Semantic Evaluation (SemEval) introduced a complete dataset in
English for emotion analysis [7]. To perform sentiment analysis, corpus of several
languages, such as Arabic [8], Czech [9], and French [10], were created. Another
corpus proposed by SemEval [11] consisting of Arabic, English, and Spanish
tweets, and these are labeled with eleven emotions. A recently developed corpus
consists of textual dialogues where each discussion is either labeled as anger,
happy, sad, or others [12]. A tweeter corpus containing Hindi-English code mixed
1126 A. Das et al.
text was annotated with Ekman’s six basic emotions by two languages proficient
and the quality validated using Cohen’s Kappa coefficient in [13].
Although Bengali is a resource-poor language, few works have been con-
ducted on sentiment analysis by classifying Bengali words or texts into positive,
negative, and neutral classes. Das et al. [14] approached with semi-automated
word-level annotation using the Bangla Emotion word list, which was translated
from English WordNet Affect lists [15]. The authors focused heavily on words
rather than the context of the sentences and built an emotion dataset of 1300
sentences with six classes. Prasad et al. [16] tagged sentiments of tweet data
based on the emoticon and hashtags. They build a sentiment dataset having
999 Bengali tweets. Nafis et al. [17] developed a Bangla emotion corpus con-
taining 2000 Youtube comments, which were annotated with Ekman’s six basic
emotions. They measured the majority vote for final labeling and came up tag-
ging with four emotion labels as some of them are ambiguous. Rahman et al.
[18] used Bengali cricket Dataset for sentiment analysis of Bengali text in which
2900 different online comments were labeled into positive, negative, and neu-
tral sentiment. A couple of work [19] and [20] developed datasets for sentiment
analysis from Bengali text. The first work considered 1000 data of restaurant
reviews, whereas the second considered 1200 data. Most of the previous work
Bengali dataset developed to analyze sentiment into positive, negative, or neu-
tral. However, BEmoD intends to use classifying emotion in Bengali text into
six basic emotion classes.
4.2 Preprocessing
Pre-processing performed in two phases: manual and automatic. In the manual
phase, “typo” errors eliminate from the collected data. We took Bangla academy
BEmoD: Development of Bengali Emotion Dataset 1129
supported accessible dictionary (AD) database [22] to find the appropriate form
of a word. If a word existed in input texts but not in AD, then this word was
considered to be a typo word. The appropriate word searched in AD and the
typo word was replaced with this corrected word. For example, the text,
The whole corpus was labeled manually, followed by majority voting to assign a
suitable label. The labeling or annotation tasks were performed by two separate
groups (G1 and G2). G1 consists of 5 postgraduate students having a Computer
Engineering background and working on NLP. An expert group (G2) consists of
three academicians and working on NLP for several years. They performed label
verification by selecting an appropriate label. The majority voting mechanism
uses to decide the final emotion class of emotion expression. The unique ultimate
label of expression chosen by pursuing the process described in Algorithm 1.
The majority voting by the G1 annotators has decided the original label of data.
The original label was considered the ultimate if this label matched the expert
label (G2). When the label of G1 and G2 was mismatched, then it was sent to the
groups for further discussion. Both groups accepted a total of 4950 data labels
among 5700 data. The remaining 750 data was sent for discussion. Both groups
are agreed about 250 data label after discussion and added to BEmoD. About
500 data have been precluded due to the disagreement between groups. This
exclusion may happen due to the texts with neutral emotion, implicit emotion,
and ill-formatted. Holding verifying 5200 data, including their labels saved in
*.xlsx format.
5 Analysis of BEmoD
6 Evaluation of BEmoD
We investigate how much the annotators were agreed in assigning emotion classes
by using Cohen’s kappa (κ) [24]. We also measure the density of emotion words,
high-frequency emotion words, and distribution of emotion words with Zipf’s
law, respectively [25].
Category Total emotion data Total sentences Total words Unique words
Anger 723 1126 18977 7160
Fear 859 1494 19704 6659
Surprise 649 1213 16806 6963
Sadness 1049 1864 26986 8833
Joy 993 1766 25011 8798
Disgust 927 1651 22992 8090
Fig. 3. Distribution of word frequencies:(a) Zipf curve (b) Zipf curve (log-log) scale
We investigated the available datasets on Bengali text emotions and find out
their characteristics in terms of the number of data, number of emotion cat-
egories, and types of emotions. Table 6 summarizes the properties of several
available datasets along with developed BEmoD. The summarizing indicates
that the developed dataset is larger than the available datasets.
Classification of implicit emotion is the most critical problem because of
its unapparent nature within the expression, and thus, its solution demands to
interpret of the context. Emotions are complicated; humans often face problems
to express and understand emotions. Classifying or detecting emotions in text
enhances the complicity of interpreting emotion because of the lack of apparent
facial expressions, non-verbal gestures, and voice [25]. Recognition of emotion
automatically is a complicated task. A machine should handle the difficulty of
linguistics phenomena and the context of written expression.
1134 A. Das et al.
7 Conclusion
Emotion recognition and classification are still developing areas of research, and
challenges for low-resource languages are daunting. The scarcity of the bench-
mark dataset is one of the vital challenges to perform the emotion classification
task in the Bengali language. Thus, in this work, we presented a new corpus
(called BEmoD) for emotion classification in Bengali texts and explained its
development processes in details. Although few datasets are available for emo-
tion classification (mostly consider positive, negative, and neutral classes) in
Bengali, this work adopted basic six emotions, including joy, fear, anger, sad-
ness, surprise, and disgust. This work revealed several features of emotion texts,
especially concerning each different class, exploring what kinds of emotion words
humans use to express a particular emotion. The evaluation of BEmoD shows
that the developed dataset followed the distribution of Zipf’s law and main-
tained an agreement among annotators with an excellent κ score. However, the
current version of BEmoD has 5200 emotion texts. This amount is not sufficient
to apply in deep learning algorithms. Therefore, more data samples should be
considered with implicit and neutral emotion categories. BEmoD can be consid-
ered to annotate emotion expressions in terms of various domains. These are left
for future research.
References
1. Liu, B.: Sentiment analysis and subjectivity, 1–38 (2010)
2. Garg, K., Lobiyal, D.K.: Hindi emotionnet: a scalable emotion lexicon for sentiment
classification of hindi text. ACM Trans. Asian Low-Resour. Lang. Inf. Process.
19(4), 1–35 (2020)
3. Eckman, P.: Universal and cultural differences in facial expression of emotion. In:
Nebraska Symposium on Motivation, vol. 19, pp. 207–284 (1972)
4. Alm, O.C., Roth, D., Richard, S.: Emotions from text: machine learning for text-
based emotion prediction. In: Proceeding in HLT-EMNLP, pp. 579–586. ACL, Van-
couver, British Columbia, Canada (2005)
5. Aman, S., Szpakowicz, S.: Identifying expressions of emotion in text. In: Inter-
national Conference on Text, Speech and Dialogue, pp. 196–205. Springer, Berlin
(2007)
BEmoD: Development of Bengali Emotion Dataset 1135
6. Scherer, K.R., Wallbott, H.G.: Evidence for universality and cultural variation of
differential emotion response patterning. J Per. Soc. Psy. 66(2), 310–328 (1994)
7. Pontiki, M., Galanis, D., Pavlopoulos, J., Papageorgiou, H., Androutsopoulos,
I., Manandharet, S.: Semeval-2014 task 4: aspect based sentiment analysis. In:
International Workshop on Semantic Evaluation, pp. 27–35. ACL, Dublin, Ireland
(2014)
8. Al-Smadi, M., Qawasmeh, O., Talafha, B., Quwaider, M.: Human annotated Ara-
bic dataset of book reviews for aspect based sentiment analysis. In: International
Conference on Future Internet of Things and Cloud, pp. 726–730. IEEE, Rome,
Italy (2015)
9. Ales, T., Ondrej, F., Katerina, V.: Czech aspect-based sentiment analysis: a new
dataset and preliminary results. In: ITAT, pp. 95–99 (2015)
10. Apidianaki, M., Tannier, X., Richart, C.: Datasets for aspect-based sentiment anal-
ysis in French. In: International Conference on Lan. Res. & Evaluation, pp. 1122–
1126. ELRA, Portorož, Slovenia (2016)
11. Mohammad, S., Bravo-Marquez, F., Salameh, M., Kiritchenko, S.: Semeval-2018
task 1: affect in tweets. In: International Workshop on Semantic Evaluation, pp.
1–17. ACL, New Orleans, Louisiana (2018)
12. Chatterjee, A., Narahari, K.N., Joshi, M., Agrawal, P.: Semeval-2019 task 3:
emocontext: contextual emotion detection in text. In: International Workshop on
Semantic Evaluation, pp. 39–48. ACL, Minneapolis, Minnesota, USA (2019)
13. Vijay, D., Bohra, A., Singh, V., Akhtar, S.S., Shrivastava, M.: Corpus creation and
emotion prediction for hindi-english code-mixed social media text. In: Proceedings
of the 2018 Conference of the North American Chapter of the Association for
Computational Linguistics: Student Research Workshop, pp. 128–135 (2018)
14. Das, D., Bandyopadhyay, S.: Word to sentence level emotion tagging for Bengali
blogs. In: Proceedings of the ACL-IJCNLP 2009 Conference Short Papers, pp.
149–152 (2009)
15. Strapparava, C., Valitutti, A., et al.: Wordnet affect: an affective extension of
wordnet. In: Lrec, vol. 4, p. 40. Citeseer (2004)
16. Prasad, S.S., Kumar, J., Prabhakar, D.K., Tripathi, S.: Sentiment mining: an app-
roach for Bengali and Tamil tweets. In: 2016 Ninth International Conference on
Contemporary Computing (IC3), pp. 1–4. IEEE (2016)
17. Tripto, N.I., Ali, M.E.: Detecting multilabel sentiment and emotions from Bangla
youtube comments. In: 2018 International Conference on Bangla Speech and Lan-
guage Processing (ICBSLP), pp. 1–6. IEEE (2018)
18. Rahman, A., Dey, E.K.: Datasets for aspect-based sentiment analysis in Bangla
and its baseline evaluation. Data 3(2), 15 (2018)
19. Sharif, O., Hoque, M.M., Hossain, E.: Sentiment analysis of Bengali texts on online
restaurant reviews using multinomial naıve bayes. In: International Conference on
Advance in Science, Engineering & Robotics Technology, pp. 1–6. IEEE, Dhaka,
Bangladesh (2019)
20. Ruposh, H.A., Hoque, M.M.: A computational approach of recognizing emotion
from Bengali texts. In: International Conference on Advances in Electrical Engi-
neering (ICAEE), pp. 570–574. IEEE, Dhaka, Bangladesh (2019)
21. Dash, N.S., Ramamoorthy, L.: Utility and Application of Language Corpora.
Springer (2019)
22. Accessible dictionary. https://accessibledictionary.gov.bd/. Accessed 2 Jan 2020
23. Full emoji list. https://unicode.org/emoji/charts/full-emoji-list.html. Accessed 7
Feb 2020
1136 A. Das et al.
24. Cohen, J.: A coefficient of agreement for nominal scales. Educ. Psychol. Meas.
20(1), 37–46 (1960)
25. Alswaidan, N., Menai, M.B.: A survey of state-of-the-art approaches for emotion
recognition in text. Knowl. Inf. Syst. 62, 2937–2987 (2020)
26. Landis, J.R., Koch, G.G.: The measurement of observer agreement for categorical
data. Biometrics, 159–174 (1977)
Advances in Engineering
and Technology
Study of the Distribution Uniformity
Coefficient of Microwave Field of 6 Sources
in the Area of Microwave-Convective Impact
1 Introduction
Many agricultural processing processes, such as drying, are characterized by their high
energy intensity. In this connection, the development of energy-efficient equipment
remains relevant. Nowadays the development of energy-efficient equipment is con-
necting with accumulating and handling the mount of information. Obtained data this
way can be oriented at optimized algorithms of the managing equipment and also
education deeply for building smart control systems. Improving the post-harvest pro-
cessing of grain does not lose its relevance. At the same time, the improvement of these
processes using electrophysical effects is becoming increasingly popular [3, 4, 8, 10,
11, 15]. Many researchers consider exposure to ultrasound, ozone-air mixtures, aero-
ions, infrared radiation, microwave field, etc. At the same time, almost all of these
factors are associated with a high unevenness of impact on the volume of the processed
material. Thus, there is a need to assess the uniformity of the studied factor over the
layer of the processed material [1, 2, 6, 7, 12]. This article will consider an experi-
mental study of the propagation of the microwave field in a dense layer of wheat
subjected to microwave convective drying.
2 Main Part
2.1 Research Method
In order to take into account the influence of uneven distribution of the field by volume,
enter the coefficient of uniform distribution of the layer Кun. This coefficient is the ratio
of the average tension in the chamber volume to the maximum one. Similarly, this
coefficient can be calculated as the square root of the ratio of the average power
absorbed by the grain material in the volume of the product pipeline to the maximum
one:
rffiffiffiffiffiffiffiffiffiffi sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
Emid Qmid 5:56 1011 Emid 2 f e0 tgd
Kun ¼ ¼ ¼ ; ð1Þ
Emax Qmax 5:56 1011 Emax 2 f e0 tgd
where Emid – the average value of the amplitude of the electric field in the volume of
the processed material, V/m; Emax – the maximum value of the amplitude of the electric
field in the volume of the microwave convection processing zone, V/m; Qmid, Qmax –
average and maximum power dissipated in dielectric material, W/m3.
If we take into account the fact that the depth of penetration into the material is the
distance at which the amplitude of the incident wave decreases by e times, then the
minimum uniformity coefficient Кun min, which should be taken into account can be
defined as: Кun min = 1/e = 0.368.
Further, the uniformity coefficient of propagation of the electromagnetic field can
be used to assess the applicability of the design of the microwave convective treatment
zone together with the applied.
To solve the problem of propagation of electromagnetic waves in a waveguide with
a dielectric, it is necessary to solve two dimensional Helmholtz equations in various
sections of the waveguide [4, 5, 9, 12], in particular, in an empty waveguide (product
line) and in a filled with dielectric, also write the boundary conditions connecting these
equations.
2.2 Modelling
This problem can be solved numerically. Currently, there is a large number of spe-
cialized software that uses various numerical methods [5].
CST Microwave Studio was chosen as the framework for conducting the numerical
experiment. The most flexible calculation method, implemented in Microwave Studio
as a transient solver, can calculate the device over a wide frequency range after cal-
culating a single transient characteristic (as opposed to the frequency method, which
Study of the Distribution Uniformity Coefficient of Microwave Field 1141
requires analysis at many frequency points). The results obtained by this product are
based on the Finite Integration Technique (FIT), which is a sequential scheme for
discretizing Maxwell's equations in integral form.
The desired coefficient cannot only be calculated based on the presented simulation,
but also based on experimental measurements on control points. The Table 1 presents
the calculated coefficient for the given dimensions of the product pipeline and three
types of applied waveguides for the propagation of the microwave field in a dense layer
of wheat with a humidity of 16%.
It was found that the results of modeling the distribution of the electromagnetic
field in the zone of microwave convective influence of the installation containing
sources of microwave power for processing the grain layer indicate a high level of its
unevenness in the volume of the product pipeline.
The obtained data on the dependence of the uniformity coefficient of the electro-
magnetic field propagation in the zone of microwave convective exposure on the
humidity of wheat can be approximated by a third-degree polynomial of the form:
Kun ¼ b0 þ b1 W þ b2 W 2 þ b3 W 3 ; ð2Þ
Table 2. Calculated values of the proportionality coefficients for calculating the uniformity
coefficient.
Product section, mm mm b0 b1 b2 b3 R2
200 200 2.2753 −0.2525 0.0114 −0.0002 0.9992
200 300 0.1887 0.0336 −0.0023 4⋅10–5 0.9972
(a) (b)
Fig. 1. Product line model 200 200 600 mm in cross section: a) three-dimensional model;
b) simulation results.
The experiment was conducted for wheat of three humidity levels: 14, 20, 26%.
Taking into account the fact that even one experiment in this case will include
325 measurement points and the symmetry of the design, the number of control points
can be reduced by 4 times. Moreover, in order to exclude the mutual influence of the
sensors on each other, they can be distributed throughout the volume.
Table 3. The uniformity coefficient of the electromagnetic field propagation calculated from
experimental data.
W, % Kun1 Kun2 Kun3 Kun4 Kun5 Kun mid
16 0.38128 0.390358 0.402989 0.426276 0.357204 0.391621
20 0.28125 0.340313 0.330938 0.300313 0.33 0.316563
24 0.252511 0.250549 0.271037 0.247713 0.256229 0.257601
0.45
1
0.4
2
Uniformity coefficient
0.35
0.3
0.2
15 17 19 21 23 25
Moisture, W , %
Differences in experimental data from the calculated data obtained on the basis of
electrodynamic modeling are 15–20% and may be due to inaccuracy of information
about the dielectric properties of the material, its blockage, measurement errors, etc.
3 Conclusions
References
1. Agrawal, S., Raigar, R.K., Mishra, H.N.: Effect of combined microwave, hot air, and
vacuum treatments on cooking characteristics of rice. J. Food Process Eng. e13038 (2019).
https://doi.org/10.1111/jfpe.13038
2. Ames, N., Storsley, J., Thandapilly, S.J.: Functionality of beta-glucan from oat and barley
and its relation with human health. In: Beta, T., Camire, M.E. (eds.) Cereal Grain-Based
Functional Foods, pp. 141–166. Royal Society of Chemistry, Cambridge (2019)
3. Bansal, N., Dhaliwal, A.S., Mann, K.S.: Dielectric characterization of rapeseed (Brassica
napus L.) from 10 to 3000 MHz. Biosyst. Eng. 143, 1–8 (2016). https://doi.org/10.1016/j.
biosystemseng.2015.12.014
4. Basak, T., Bhattacharya, M., Panda, S.: A generalized approach on microwave processing
for the lateral and radial irradiations of various Groups of food materials. Innov. Food Sci.
Emerg. Technol. 33, 333–347 (2016)
5. Budnikov, D.A., Vasilev, A.N., Ospanov, A.B., Karmanov, D.K., Dautkanova, D.R.:
Changing parameters of the microwave field in the grain layer. J. Eng. Appl. Sci. N11
(Special Issue 1), 2915–2919 (2016)
6. Dueck, C., Cenkowski, S., Izydorczyk, M.S.: Effects of drying methods (hot air, microwave,
and superheated steam) on physicochemical and nutritional properties of bulgur prepared
from high-amylose and waxy hull-less barley. Cereal Chem. 00, 1–3 (2020). https://doi.org/
10.1002/cche.10263
7. Izydorczyk, M.S.: Dietary arabinoxylans in grains and grain products. In: Beta, T., Camire,
M.E. (eds.) Cereal Grain-Based Functional Foods, pp. 167–203. Royal Society of
Chemistry, Cambridge (2019)
8. Nelson, S.O.: Dielectric Properties of Agricultural Materials and Their Applications.
Academic Press, Cambridge (2015). 229 p.
Study of the Distribution Uniformity Coefficient of Microwave Field 1145
9. Pallai-Varsányi, E., Neményi, M., Kovács, A.J., Szijjártó, E.: Selective heating of different
grain parts of wheat by microwave energy. In: Advances in Microwave and Radio Frequency
Processing, pp. 312–320 (2007)
10. Ranjbaran, M., Zare, D.: Simulation of energetic- and exergetic performance of microwave-
assisted fluidized bed drying of soybeans. Energy 59, 484–493 (2013). https://doi.org/10.
1016/j.energy.2013.06.057
11. Smith, D.L., Atungulu, G.G., Sadaka, S., Rogers, S.: Implications of microwave drying
using 915 MHz frequency on rice physicochemical properties. Cereal Chem. 95, 211–225
(2018). https://doi.org/10.1002/cche.10012
12. Vasilev, A.N., Budnikov, D.A., Ospanov, A.B., Karmanov, D.K., Karmanova, G.K.,
Shalginbayev, D.B., Vasilev, A.A.: Controlling reactions of biological objects of agricultural
production with the use of electrotechnology. Int. J. Pharm. Technol. (IJPT) 8(N4), 26855–
26869 (2016)
13. Vasiliev, A.N., Ospanov, A.B., Budnikov, D.A., et al.: Improvement of grain drying and
disinfection process in the microwave field. Monography. Almaty: Nur-Print (2017).155
p. ISBN 978-601-7869-72-4
14. Vasiliev, A.N., Goryachkina, V.P., Budnikov, D.: Research methodology for microwave-
convective processing of grain. Int. J. Energy Optim. Eng. (IJEOE) 9(2), 11 (2020). Article:
1. https://doi.org/10.4018/IJEOE.2020040101.
15. Yang, L., Zhou, Y., Wu, Y., Meng, X., Jiang, Y., Zhang, H., Wang, H.: Preparation and
physicochemical properties of three types of modified glutinous rice starches. Carbohyd.
Polym. 137, 305–313 (2016). https://doi.org/10.1016/j.carbpol.2015.10.065
Floor-Mounted Heating of Piglets with the Use
of Thermoelectricity
Abstract. The article discusses the problem of energy saving when providing
comfortable conditions for keeping young animals. The question of the use of
thermoelectric modules as a source of thermal energy in installations for local
heating of piglets is considered. A functional technological scheme of floor-
mounted heating of suckling piglets is proposed. In the scheme the energy of the
hot circuit of thermoelectric modules is used to heat the panel. The energy of the
cold circuit of the thermoelectric module is used to assimilate the heat from the
removed ventilation air. The heat exchange of the heated panel with the envi-
ronment is considered. Much attention is paid to the study of the prototype of
thermoelectric installation and its laboratory tests. The energy efficiency of using
thermoelectric modules as energy converters in thermal technological processes
of local heating of young animals has been substantiated.
1 Introduction
Various types of equipment have been developed for heating young animals [3, 4].
However, the most widespread in practice are infrared electric heaters and heating
panels [5]. The unevenness of the heat flow in the area where animals are located and
the incomplete correspondence of the radiation spectrum to the absorption spectrum of
heat waves by animals are the main disadvantages of IR heaters. In addition, a common
disadvantage of local heaters is rather high energy consumption.
Heat-cold supply by means of heat pumps including using thermoelectricity
belongs to the field of energy-saving environmentally friendly technologies and
becomes more widespread in the world [6].
The analysis of the use of Peltier thermoelectric modules (thermoelectric assem-
blies) shows that thermoelectric assemblies are compact heat pumps allowing for the
creation of energy-efficient plants for use in various technological processes of agri-
cultural production [7–10].
The purpose of the research is to substantiate the parameters and develop an energy
saving installation for local heating of piglets using Peltier thermoelectric elements.
A sow is located in the area of the stall, the piglets are mainly located in the area fenced
off from the sow. In this area there is a heated panel where the piglets rest. Suckling
piglets freely pass through the passage in the separation grate to the sow for feeding.
Based on our analysis of various schemes of thermoelectric assemblies [11, 12] it
was found that the “liquid-air” heat exchange scheme should be adopted for energy and
structural parameters for creating a local heating system for piglets using a thermo-
electric heat pump. The technological scheme of the installation of local floor-mounted
heating of piglets is shown in Fig. 1.
The installation consists of a thermoelectric assembly, which in turn includes the
estimated number of Peltier elements 3, the air cooler of cold circuit 1 with an exhaust
fan 2 installed in the bypass of a pigsty ventilation system, the water cooler of hot
junction 4 connected by direct and return pipes 8 through a circulation pump 9 with a
heated panel 6. Thermal power of the thermoelectric assembly is regulated by changing
the current strength flowing through the thermocouples. Power, control and manage-
ment of the circuit is carried out by the block 5.
To assess the performance of a thermoelectric water heater circulating through a
heating panel, consider the heat balance equations for hot and cold junctions of a
thermoelectric element [13], which have the form:
Since QH and QC represent the amount of heat per unit time, the work of electric
forces (power consumption) can be determined by Eq. (5)
W ¼ QH QC ð5Þ
Taking into account Eqs. (1) and (2) as well as relations (3) and (4), the Eq. (5) can
be rewritten in the following form:
W ¼ eI ðTH TC Þ þ I 2 R; ð6Þ
case, the thermocouple works like a heat pump, transferring heat from the environment
to the heated object.
To analyze the energy efficiency of heat pumps, turn to Eq. (5), which can be
rewritten in the following form:
1 ¼ QH =W QC =W ¼ kh kc ; ð7Þ
Qpc þ 0; 5QR QT
kh ¼ : ð8Þ
eIðTH TC Þ þ I 2 R
The smaller the temperature difference between the junctions TH – TC is, the higher
the efficiency of the thermocouple in heating mode is.
From the point of view of electric energy consumption, the most economical mode
of operation is the mode of a heat pump, in which the heating coefficient tends to the
maximum. When operating in heating mode, there is no extreme dependence of the
heating coefficient on current.
The thermoelectric assembly is a compact heat pump that absorbs thermal energy
from the cold side of the module and dissipates it on the hot side [14]. The use of
thermoelectric assemblies makes it possible to simultaneously heat and cool objects in
different technological processes whose temperature is not adequate to the ambient
temperature.
The power of a heating device for a heated panel should compensate for heat loss of
the site heated to the required temperature on its surface [15]. From the conditions of a
quasi-stationary regime, the energy balance of a heated area free of animals has the
form:
P ¼ P1 þ P2 þ P3 ; ð9Þ
where: P1 is loss of heat transfer through the surface of a site to a room, W; P2 is heat
transfer from side surfaces of a site, W; P3 is loss of heat transfer through a base to a
floor array, W.
The magnitude of the convective component of heat transfer is determined by the
well-known Eq. (10):
where ta is the air temperature in a room, °C; tb is the required temperature on the
surface of a heated panel, °C; ac is heat transfer coefficient from the panel to the
environment, W/m2K; F is surface area of the heated panel, m2.
These air parameters should be taken in accordance with the norms of technological
design of agricultural livestock facilities. The sizes of sites F vary depending on the
purpose of a farm. For piglets it is from 0.5 to 1.5 m2.
1150 D. Tikhomirov et al.
Fig. 2. Calculation of the power of a heating panel. 1 is base; 2 is heat carrier; 3 is thermal
insulation.
Re ¼ va l=t; ð11Þ
where va is the relative air mobility in the room, m/s; l is length of a heated panel -
characteristic size, m; t is kinematic viscosity of air, m2/s.
Heat transfer coefficient in a laminar boundary layer is determined by Eq. (12):
where C0 = 5.76 W/m2K is the emissivity of a completely black body; eb is the integral
degree of blackness of the surface of a heated panel; F is surface area of the heated
panel, m2; Tb and Tbe are surface temperatures of a heated panel and building envelope
indoors, K.
General losses from the surface of a heated panel as a result of convective and
radiant heat transfer are:
Floor-Mounted Heating of Piglets with the Use of Thermoelectricity 1151
P1 ¼ Pc þ Pr : ð14Þ
tt ¼ tb þ P1 db =ðFkb Þ; ð15Þ
where tb is the surface temperature of a heated panel (base), °C; db is the thickness of a
base, m; kb is the coefficient of thermal conductivity of a base material, W/mK.
Heat losses from the side surface P2 of the heated panel can be neglected due to its
small area.
A preliminary calculation of the heat loss of the heated panel through the floor P3 is
carried out taking into account a number of assumptions:
• The values of floor temperature in winter should be higher than negative values and
dew points;
• The heated panel does not affect the thermal regime of a floor at its location.
• The value of floor temperature is taken below the calculated values in the range of
5–10 °C to ensure system stability.
P ¼ P1 þ P2 þ P ð17Þ
Table 1 presents the calculation of the power of the heated panel made for the
maximum allowable temperature and air velocity in the room for suckling piglets.
Table 1. (continued)
Parameter Designation Unit measure Value
Radiant component of heat loss Pr W 150.2
General losses from a heated panel surface P1 W 162.7
Base thickness db m 0.003
Thermal conductivity of a base material kb W/mK 0.4
Coolant temperature tt °C 36.2
Conductivity coefficient of thermal insulation kins W/mK 0.03
Thermal insulation layer thickness dins m 0.04
Floor temperature tf °C 5
Heat loss through a floor P3 W 44.3
Power of a heated panel (max) P W 207.0
The estimated maximum power of the panel was about 207 W (Table 1). The
minimum rated power of the panel for its surface temperature of 32 °C will be 174 W
in accordance with the presented method.
Based on the performed calculations, a prototype of the installation for local heating of
piglets was made (Fig. 3) using a heated floor panel with thermoelectric modules
(TEC1–12706). Water is used as a heat carrier that circulates in a closed loop.
The thermal energy and structural calculation of the thermoelectric assembly, cold
and hot circuit radiators was performed according to the methodology [16] and taking
Floor-Mounted Heating of Piglets with the Use of Thermoelectricity 1153
into account the results of [17]. The main parameters of the local floor-mounted
installation for heating piglets are given in Table 2.
waste heat of removed ventilation air to supply a cold junction, it is possible to reduce
the power consumed from an electrical network to heat the panel itself. This heat pump
generates 1.3 kWh of thermal energy per 1 kWh of consumed electrical energy in the
set mode.
It is possible to use infrared heaters located above piglets if the combined heating of
piglets is necessary [18]. They might be the irradiators [19] with uniform heat flux and
high reliability.
4 Conclusions
Analysis of the experimental results of the prototype and theoretical studies suggest the
possibility of using Peltier thermoelectric modules in local heating systems for young
animals as an energy-saving source of thermal energy.
The use of local heating systems, based on Peltier elements as heat pumps, can
significantly reduce the energy consumption (up to 30%) for heating young animals
compared to existing direct heating plants (electric irradiators, heated floor mats,
brooders, etc.). The heat output of the panel with an area of 1,0–1.4 m2 should be 170–
200 W, and the total electric power of the thermoelements – 120–140 W.
The estimated payback period for local heaters of young animals, based on Peltier
thermoelectric elements, is about 3.5 years.
Further research will be aimed at determining the conditions for the optimal
operation of thermoelectric systems to obtain the maximum increase in power taken by
the hot circuit compared to the power consumed from an electrical network. The
justification of heat, power and structural parameters of local installations for heating
piglets will be also carried out.
The heat power of the panel is determined for static modes of heat exchange with
the environment. The study of heat exchange dynamics, taking into account the degree
of filling of the panel with piglets, will allow further development of an adaptive
control system to maintain a set temperature on its surface.
References
1. RD-APK 1.10.02.01–13. Metodicheskie rekomendacii po tekhnologicheskomu proek-
tirovaniyu svinovodcheskih ferm i kompleksov [Guidelines for the technological design of
pig farms and complexes]. Ministry of Agriculture of the Russian Federation, Moscow,
Russia (2012)
2. Trunov, S.S., Rastimeshin, S.A.: Trebovaniya k teplovomu rezhimu zhivotnovodcheskih
pomeshchenij s molodnyakom i predposylki primeneniya lokal’nogo obogreva [Require-
ments for the thermal regime of livestock buildings with young animals and the prerequisites
for the application of local heating]. Vestnik VIESKH 2(27), 76–82 (2017)
3. Valtorta, S.: Development of microclimate modification patterns in animal husbandry. In:
Stigter, K. (eds.) Applied Agrometeorology. Springer, Heidelberg (2010). https://doi.org/10.
1007/978-3-540-74698-0_92
Floor-Mounted Heating of Piglets with the Use of Thermoelectricity 1155
4. Tikhomirov, D., Vasilyev, A.N., Budnikov, D., Vasilyev, A.A.: Energy-saving automated
system for microclimate in agricultural premises with utilization of ventilation air. Wireless
Netw. 26(7), 4921–4928 (2020). Special Issue: SI. https://doi.org/10.1007/s11276-019-
01946-3
5. Samarin, G.N., Vasilyev, A.N., Zhukov, A.A., Soloviev, S.V. Optimization of microclimate
parameters inside livestock buildings. In: Vasant, P., Zelinka, I., Weber, G.W. (eds.)
Intelligent Computing & Optimization. ICO 2018. Advances in Intelligent Systems and
Computing, vol 866. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-00979-3_35
6. SHostakovskij, P.G.: Teplovoj kontrol’ ob”ektov na baze termoelektricheskih sborok
[Thermal control of objects based on thermoelectric assemblies]. Komponenty i tekhnologii
9, 142–150 (2011)
7. Tihomirov, D.A., Trunov, S.S., Ershova, I.G., Kosolapova, E.V.: Vozdushno-teplovaya
zavesa s ispol’zovaniem termoelektricheskih modulej v doil’nom bloke fermy KRS [Air-
thermal curtain using thermoelectric modules in the milking unit of the cattle farm].
Vestnik NGIEI 1, 47–57 (2020)
8. Amirgaliyev, Y., Wojcik, W., Kunelbayev, M.: Theoretical prerequisites of electric water
heating in solar collector-accumulator. News Nat. Acad. Sci. Repub. Kaz. Ser. Geol. Tech.
Sci. 6, 54–63 (2019)
9. Trunov, S.S., Tikhomirov, D.A.: Termoelektricheskoe osushenie vozduha v sel’skohozya-
jstvennyh pomeshcheniyah [Thermoelectric air drainage in agricultural premises]. Nauka v
central’noj Rossii 2(32), 51–59 (2018)
10. Kirsanov, V.V., Kravchenko, V.N., Filonov, R.F.: Primenenie termoelektricheskih modulej
v pasterizacionno-ohladitel’nyh ustanovkah dlya obrabotki zhidkih pishchevyh produktov
[The use of thermoelectric modules in pasteurization-cooling plants for processing liquid
food products]. FGOU VPO MGAU, Moscow, Russia (2011)
11. Chen, J., Zhang, X.: Investigations of electrical and thermal properties in semiconductor
device based on a thermoelectrical model. J. Mater. Sci. 54(3), 2392–2405 (2019)
12. Reddy, B.V.K., Barry, M., Li, J.: Mathematical modeling and numerical characterization of
composite thermoelectric devices. Int. J. Therm. Sci. 67, 53–63 (2013)
13. Stary, Z.: Temperature thermal conditions and the geometry of Peltier elements. Energy
Convers. Manage. 33(4), 251–256 (1992)
14. Ismailov, T.A., Mirzemagomedova, M.M.: Issledovanie stacionarnyh rezhimov rabo-ty
termoelektricheskih teploobmennyh ustrojstv [Study of stationary modes of operation of
thermoelectric heat exchange devices]. Vestnik Dagestanskogo gosudarstvennogo tekhnich-
eskogo universiteta. Tekhnicheskie nauki 40(1), 23–30 (2016)
15. Wheeler, E.F.; Vasdal, G., Flo, A., et al.: Static space requirements for piglet creep area as
influenced by radiant temperature. Trans. ASABE 51(1),2008) 278–271 )
16. Pokornyj, E.G., SHCHerbina, A.G.: Raschet poluprovodnikovyh ohlazhdayushchih ustrojstv
[Calculation of semiconductor cooling devices]. Nauka, Lipetsk, Russia (1969)
17. Trunov, S.S., Tihomirov, D.A., Lamonov, N.G.: Metodika rascheta termoelektricheskoj
ustanovki dlya osusheniya vozduha [Calculation Method for Thermoelectric Installation for
Air Drying]. Innovacii v sel’skom hozyajstve 3(32), 261–271 (2019)
18. Ziemelis, I, Iljins, U, Skele, A.: Combined local heating for piglets. In: International
Conference on Trends in Agricultural Engineerin, Prague, Czech Republic, pp. 441–445
(1999)
19. Kuz’michyov, A.V., Lyamcov, A.K., Tihomirov, D.A.: Teploenergeticheskie pokazateli IK
obluchatelej dlya molodnyaka zhivotnyh [Thermal energy indicators of IR irradiators for
young animals]. Svetotekhnika 3, 57–58 (2015)
The Rationale for Using Improved Flame
Cultivator for Weed Control
Abstract. The work substantiates the problem of weed control and the need to
improve control methods in the production of agricultural products, the applied
control methods are considered. The design of an improved flame cultivator for
burning weeds in the near-stem strips of orchards and vineyards (rows and rows
of perennial plantings) by the thermal method, as well as the technological
process of its operation and conclusions, is given.
1 Introduction
It is necessary to understand the possible soil infestation with weed seeds in order
to analyze their composition and monitor their growth and the difference in weed
infestation depends on the weed seeds concentration in the upper layer of the soil [4].
It is established that the decrease productivity of cultivated crops reaches 35% or
higher when their soreness. Weed vegetation not only reduces soil fertility, it also helps
to generate pests and diseases. Weed vegetation rapidly reduces the harvest and its
existence complicates the harvesting process, besides the quality of the grown product
is reduced.
The temperature of soils full of weeds decreases from 2 to 4 °C, this helps to reduce
the vital activity of bacteria in them, restraining the development of decay of organic
matter by this and reducing the number of useful components [5]. This explains the
interest of agricultural producers in development and integration of modern technical
means for weed control, those are allowing to reduce the cost of growing crops.
Agricultural production does consist not in absolute elimination of weeds, but in
keeping them in volume, does not negatively affect the yield of cultivated crops in
present conditions. The main factors of significant infestation are high productivity,
viability and resistance of weed vegetation to the protection measures used.
Classic measures of protection from weeds are chemical and preventive. These
measures make it possible to reduce the number of weeds to a cost-effective threshold
value of harmfulness in their collective use. They also have a number of other well-
known disadvantages. A certain proportion of herbicide solution particles carried away
by uncontrolled wind movement has a negative effect on the surrounding atmosphere.
Modification of weeds occurs and the impact on most of their chemical methods
varieties of weed destruction is reduced [6]. Researchers from many countries are
searching for other ways to destroy weed vegetation for this reason. So, fire methods of
weed control have General scientific and practical significance as well as classical
methods [7].
2 Main Part
The use of flame cultivator in vegetable growing is also effective. Good results can
be seen when using this method on areas sown with onions, beets, carrots and others,
The Rationale for Using Improved Flame Cultivator 1159
The amount of weed vegetation in the areas where crops are cultivated in-creases
their harmful impact. It accompanies a decrease in crop productivity. This circumstance
contributed to the establishment of a mathematical interpretation of the quantitative
1160 M. Abdulgalimov et al.
relationship between the weed vegetation excess and the particular crop yield, which is
determined by an exponential regression equation quite reliably:
Y ¼ aebx þ c; ð1Þ
where Y is crop productivity of cultivated crops on a particular area that has a weed
infestation %, g/m2, ton/Ha;
x is excess weed vegetation, %, g/m2, ton/Ha;
e = 2,718 is base of natural logarithms;
a is the value that determines the decrease crop productivity with the maximum
weed infestation of the cultivation area;
b is the value that determines the expression of the decrease in crop productivity
with weed infestation of the cultivation area;
c is the value that determines the volume of saving crop productivity with maximum
weed infestation of the cultivation area.
The following levels of weed infestation harmfulness limits have been established
subject to the impact of cultivated crops weed vegetation: phytocenotic (an excess
weed vegetation does not have a significant impact on cultivated crops), critical (the
excess of weed vegetation contributes to a decrease in the productivity of cultivated
crops from 3.0 to 6.0%), economic (the increase crop productivity reaches from 5.0 to
7.0%, that is provided by reducing to a minimum the amount of weed vegetation).
The economic limit of harmfulness is determined by the formula:
Ca ¼ Ya P; ð2Þ
S ¼ B l = 10000; Ha ð5Þ
Q ¼ Vg = S; L=Ha ð6Þ
The value of the destructive flame temperatures (°K) dependent on the exposure
time is determined by the formula:
Results
The flame cultivator must comply with the following main conditions based on the
technological neutralization weeds scheme:
– to provide a significant level of weed control;
– to not harm cultivated plants within the limits of the safe lane width;
– provide copying of the terrain movement in the transverse-longitudinal-vertical
planes;
– have protective devices against damage to the working elements in contact with
obstacles.
Weeds destruction in the ground between the rows does not imply significant
problems with the support current agricultural tools. However, the issue of weed
control in rows and protection zones of cultivated crops is still not fully resolved,
despite the large number of technical means and methods available in the current
agricultural sector [18].
Agricultural machinery and equipment manufacturers are mainly focused on the
creation agricultural aggregates for the destruction weed vegetation without the use of
toxic drugs, this is relevant due to the demand for ecological agricultural production at
present stage. Gas-powered units that use the gas-burning flame energy to control
weeds are considered the result of these foundations formation [19, 20]. They combine
in itself high productivity with low construction and maintenance costs and the ability
to effectively perform joint work on crop care and weed control.
The control of weeds and their seeds should be effective and carried out by high-
performance technical means at minimal cost [21, 22]. Implementation of weed vege-
tation neutralization process on the soil surface is allowed using a flame cultivator [23].
The design of the proposed advanced flame cultivator (Figs. 4, 5 and 6) includes
gas sections-burners 1 consisting of removable paired pipe sections (burners) 5 with
The Rationale for Using Improved Flame Cultivator 1163
deaf flanges 20 at the ends and holes at the bottom for screwing gas-air nozzles 6
connected by mixer pipes 19 having holes on top for connecting the air pipe 3 and the
gas pipe 2 with the nozzle 4; reflector shields 7; rails-skates (bearing) 8 with vertical
plates-knives (vertical) 9 to maintain the straight movement of the unit; gas-air nozzles
(terminal) 10; damper 11 welded to the probe 12 and the sleeves 13; brackets with
stoppers 15; rod 16 connected to each other by return springs 17 [23].
Discussion
The use of the proposed flame cultivator for weed control reduces the number of soil
processing and the cost of caring for cultivated crops, and the fire method of destroying
weeds is environmentally safe.
It is established that the destruction weeds occurs due to pruning at considerable
cost and only at different stages of their formation in a significant part of the available
technical means and aggregates for weed control.
The device novelty is the improved flame cultivator construction developed for the
burning weeds technological process [23].
3 Conclusion
Negative results of the use methods for weed control indicate the need for their
improvement.
The improved technology and design of the proposed flame cultivator ensure the
effectiveness of the destroying weed vegetation process using the proposed flame
cultivator.
Economic feasibility is substantiated by a decrease number of soil processing while
destroying weed vegetation relative to other technologies.
It is necessary to continue the modernization of both existing and developed flame
cultivator for the destruction weeds constructions.
References
1. Perederiyeva, V.M., Vlasova, O.I., SHutko, A.P.: Allelopaticheskiye svoystva sornykh
rasteniy i ikh rastitel’nykh ostatkov v protsesse mineralizatsii [Allelopathic properties of
weeds and their plant residues in the process of mineralization]. KubGAU, Krasnodar 09
(73), 111–121 (2011). https://ej.kubagro.ru/2011/09/pdf/11. (in Russian)
2. Hatcher, P.E., Melander, B.: Combining physical, cultural and biological methods: prospects
for integrated non-chemical weed management strategies. Weed Res. 43(5), 303–322 (2003).
https://doi.org/10.1046/j.1365-3180.2003.00352.x
3. Abouziena, H.F., Hagaag, W.M.: Weed control in clean agriculture: a review. Planta
Daninha. 34(2), 377–392 (2016). https://doi.org/10.1590/S0100-83582016340200019
4. Dorozhko, G.R., Vlasova, O.I., Perederiyeva, V.M.: Sposob obrabotki - faktor reg-
ulirovaniya fitosanitarnogo sostoyaniya pochvy i posevov ozimoy pshenitsy na chernoze-
makh vyshchelo-chennykhzony umerennogo uvlazhneniya Stavropol’skogo kraya [The
tillage method is a factor in regulating the phytosanitary state of the soil and winter wheat
crops on leached humus of the moderate humidification zone of the Stavropol Territory].
KubGAU, Krasnodar 04(68), 69–77 (2011). https://ej.kubagro.ru/2011/04/pdf/08. (in
Russian)
5. Tseplyaev V.A., Shaprov M.N., Tseplyaev A.N.: Optimizatsiya parametrov tekhnologich-
eskogo protsessa poverkhnostnoy obrabotki pochvy rotornym avtoprivodnym agregatom
[Optimization of the technological process parameters of the surface tillage by the rotary
automatic drive unit]. Izvestiya Nizhnevolzhskogo agrouniversitetskogo kompleksa: nauka i
vyssheye professional’noye obrazovaniye 1(25), 160–164 (2012). (in Russian)
1166 M. Abdulgalimov et al.
6. Schütte, G.: Herbicide resistance: promises and prospects of biodiversity for European
agriculture. Agric. Hum. Values 20(3), 217–230 (2003). https://doi.org/10.1023/A:
1026108900945
7. Bond, W., Grundy, A.C.: Non-chemical weed management in organic farming systems.
Weed Res. 41(5), 383–405 (2001). https://doi.org/10.1046/j.1365-3180.2001.00246.x
8. Izmaylov, A.YU., Khort, D.O., Smirnov, I.G., Filippov, R.A., Kutyrëv, A.I.: Analiz
parametrov raboty ustroystva dlya gidravlicheskogo udaleniya sornoy rastitel’nosti [Analysis
of Work Parameters of the Device for Hydraulic Removal of Weed Vegetation]
Inzhenernyye tekhnologii i sistemy. FGBOU VO «MGU im. N. P. Ogarëva». Saransk 29
(4), 614–634 (2019). https://doi.org/10.15507/2658-4123.029.201904.614-634. (in Russian)
9. Agricultural and Forestry Machinery. Catalogue of Exporters Czech Republic. Copyright: A.
ZeT, Brno (2005)
10. Bezhin, A.I.: Obosnovaniye parametrov i rezhimov raboty kul’tivatornogo agregata dlya
sploshnoy obrabotki pochvy [Rationale for parameters and operating modes of the cultivator
for complete tillage]. Ph.D. dissertation. Orenburg State Agrarian University, Orenburg
(2004),183 p. (in Russian)
11. Popay, I., Field, R.: Grazing animals as weed control agents. Weed Technol. 10(1), 217–231
(1996). https://doi.org/10.1017/S0890037X00045942
12. Astatkie, T., Rifai, M.N., Havard, P., Adsett, J., Lacko-Bartosova, M., Otepka, P.:
Effectiveness of hot water, infrared and open flame thermal units for controlling weeds. Biol.
Agric. Hortic. 25(1), 1–12 (2007). https://doi.org/10.1080/01448765.2007.10823205
13. Abdulgalimov, M.M.: Nestandartnyye resheniya i tekhnicheskiye sredstva dlya bor’by s
sornyakami v sadakh i vinogradnikakh [Non-standard solutions and technical means for
weed control in orchards and vineyards]. DagGAU im. M.M. Dzhambulatova. Makhachkala
4–6 (2015). (in Russian)
14. Blackshaw, R.E., Anderson, R.L., Lemerle, D.: Cultural weed management. In: Upadhyaya,
M.K., Blackshaw, R.E. (eds.) Non-Chemical Weed Management: Principles, Concepts and
Technology,1 edn.,CABI 2007, Wallingford, England, pp. 35–47. https://researchoutput.csu.
edu.au/en/publications/cultural-weed-management
15. Melikov, I., Kravchenko, V., Senkevich, S., Hasanova, E., Kravchenko, L.: Traction and
energy efficiency tests of oligomeric tires for category 3 tractors. In: IOP Conference Series:
Earth and Environmental Science, vol. 403, p. 012126 (2019). https://doi.org/10.1088/1755-
1315/403/1/012126
16. Senkevich, S., Ivanov, P.A., Lavrukhin, P.V., Yuldashev, Z.: Theoretical prerequisites for
subsurface broadcast seeding of grain crops in the conditions of pneumatic seed
transportation to the coulters. In: Handbook of Advanced Agro-Engineering Technologies
for Rural Business Development, pp. 28–64. IGI Global, Hershey (2019). https://doi.org/10.
4018/978-1-5225-7573-3.ch002
17. Lavrukhin, P., Senkevich, S., Ivanov, P.: Placement plants on the field area by seeding
machines: methodical aspects assessment rationality. In: Handbook of Research on Smart
Computing for Renewable Energy and Agro-Engineering, pp. 240–261. IGI Global, Hershey
(2020).https://doi.org/10.4018/978-1-7998-1216-6.ch010
18. Tokarev, N.A., Gar’yanova, E.D., Tokareva,, N.D., Gulyayeva, G.V.: Sposob bor’by s
sornyakami [Weed Control Method]. Zemledeliye. 2012. 8. p. 37–38. (in Russian)
19. Baerveldt, S., Ascard, J.: Effect of soil cover on weeds. Biol. Agric. Hortic. 17(2), 101–111
(1999). https://doi.org/10.1080/01448765.1999.9754830
20. Latsch, R., Anken, T., Herzog, C., Sauter, J.: Controlling Rumex obtusifolius by means of
hot water. Weed Res. 57(1), 16–24 (2017). https://doi.org/10.1111/wre.12233
21. Abdulgalimov, M.M.: Ognevoy kul’tivator [Flame Cultivator]. Byul. 14, 2016. RU Patent
2584481 C2. A01M 15/00
The Rationale for Using Improved Flame Cultivator 1167
22. Abdulgalimov, M.M., Magomedov, F.M., Senkevich, S.E., Umarov, R.D., Melikov, I.M.:
Sovershenstvovaniye tekhnologii i sredstv mekhanizatsii dlya bor’by s sornoy rastitel’nos-
t’yu [Improvement of technology and means of mechanization for weed control].
Sel’skokhozyaystvennyye mashiny i tekhnologii 5, 38–42 (2017). https://doi.org/10.
22314/2073-7599-2017-5-38-42. (in Russian)
23. Abdulgalimov, M.M.: Ognevoy kul’tivator [Flame Cultivator]. Byul. 7, 2019. RU Patent
187387 U1. A01M 15/00
The Lighting Plan: From a Sector-Specific
Urbanistic Instrument to an Opportunity
of Enhancement of the Urban Space
for Improving Quality of Life
Abstract. The urban space, its lighting, the environmental sustainability, the
energy efficiency, the attention to public spending, the innovation and the opti-
mization of public utility services, the light pollution, the energy and lighting
upgrading of public lighting systems and – most importantly – the growth
potential of urban living thanks to a new strategy for the cities of the future, the
creation of efficient and effective smart cities: these are the themes that innervate
the research whose results are presented in this article. These topics are very
different from each other but share the urban space with the features of com-
plexity and modernity. And it is on the city, on the urban space, that this study is
directed on since, as De Seta writes “nowadays over half the global population
lives in a city, and it is estimated that this percentage will rise to three quarters in
2050” [1].
The present research represents the evolution of what has been submitted by
the authors for the AIDI conference “XIX Congresso Nazionale AIDI, La luce
tra cultura e innovazione nell’era digitale” (The research work was submitted
and accepted at the XIX National Conference of AIDI, 21−22 May 2020,
Naples, Italy.).
1 Introduction
The complexity of a contemporary city led to the creation of new analysis, design and
planning tools in the discipline of urban technology. These above mentioned tools
relate to issues both widespread - including energy saving, water protection, waste
management and localization of broadcast media – and locally, such as traffic con-
gestion, pollution in its different forms, subsoil monitoring, etc.
These new tools fill a gap in the general planning method in its traditional meaning:
they listen to and interpret those specific territorial requirements and try to give them a
concrete answer, with the main aim of raising performance, functional and quality city
standards [2, 3].
These topics are fully covered by multiple types of plan in the sector-specific city
planning: the urban traffic and the acoustic renewal plan, the timetable/schedule and the
municipal energy plan, the color plan and the electro-smog renewal one and, last but
not least, the lighting plan.
To better understand the meaning of the latter tool, here’s a definition from an ENEA1
guideline:
«The lighting master plan is an urbanistic instrument to regulate all types of lighting
for the city, it is a veritable circuit of how the city is supposed to be planned in terms of
lighting technique. It has key advantages, as it allows to respect urban fabrics in general
correlating them with a type of proper lighting. The final result is the obtaining and the
optimizing of the municipal lighting network, depending on main requirements» [4].
The lighting of a city, therefore public, is a very
peculiar component in the structure of the urban
system, as it assumes two different identities
depending on the daytime and the night-time. By
day, its components are visible objects which have to
be mitigated as much as possible in the environ-
mental context, while by night, those same elements
play an essential role by lighting urban space,
ensuring that the whole city works as during the day
and that everything proceeds in terms of quality,
efficiency and reliability [5]. Not less important,
considering the technologically advanced and inno-
vative role that this engineerable and widespread
Fig. 1. Infographic of the report system can play, this may be considered the first real
by the McKinsey Global Institute step towards the formation of the smart city [6].
According to a report by the McKinsey Global
Institute, smart cities might reduce crime by 10% and healthcare costs by 15%, save
four days stuck in traffic and 80 L of water per day (Fig. 1) [7].
The real usefulness and importance of a dedicated sector-specific urbanistic
instrument – as is the lighting plan – becomes evident. Such an instrument, in fact, can
act as a guide in people’s lives within the night urban spaces [8]. In fact, almost all
Italian regional legislation – for about two decades now – imposes the obligation for
the Municipalities to adopt this sector-specific urbanistic instrument (unfortunately,
today it is still disregarded by the municipalities themselves). From an analysis con-
ducted by ENEA on municipalities involved in the Lumière Project Network, less than
20% of them had adopted a lighting plan at the time of the surveys [9].
1
(IT) National agency for new technologies, energy and sustainable economic development.
1170 C. B. Bellone and R. Ottavi
While, on one hand, there is no coherence of adoption and/or unique model of how
to cover issues revolving around urban space lighting, on the other hand, it is also true
that the sensitivity to urban lighting issues is increasing day by day; we are witnessing
the creation of a real process of change, which is felt not only in our country but all
over Europe, in response to an increasing citizens’ demand to see the quality of life in
urban public spaces improved [10, 11].
In order to scientifically analyze the different urban contexts and how the above
principles, within the lighting plans, may or may not be implemented in the public
lighting, a scale of values is proposed to represent a valid instrument for an objective
research.
The two levers of analysis will be quality and innovation. We’ll start investigating
what it means to generate quality in the urban space through light, and then understand
how it can represent a real innovation for the cities of the future.
The Lighting Plan: From a Sector-Specific Urbanistic Instrument 1171
Returning to the
above explained
methodological
approach, we report a
matrix chart “light
quality” (Fig. 2) high-
lighting the key points
and the implementation
practical connections
which can be planned,
in order to determine a
planning of the light
useful to men and their
space, so that light truly
becomes an instrument
Fig. 3. Matrix chart no.2 “Innovating with liht”
for the city planning, an
instrument to build the
city.
Here is the matrix chart “innovating with light” (Fig. 3), showing the objective key
points which will act as a guide in the planning research of the next paragraph.
The answers provided in Europe are quite varied2 and, among these ones, the French-
Belgian experience deserves special attention: in fact, the connection created there
between light, urban space and aesthetics led this reality to be an example in terms of
contemporary urban quality.
The plan lumière of the Région de Bruxelles-Capitale3 is particularly significant.
Also referred to as a « second generation plan » for its choice to put users, people
living the night urban space and the way they do it at the center of the study, rather than
making the city spectacular [12].
The merit of this plan is not confined to provide a global vision of lighting at
regional level, but also to accompany the actions carried out in the area defined as
priorities by the Brussels government in the 2014−2019 declaration4 with some
specific proposals for this areas of prime importance for the regional development.
In order to analyze the key points of this lighting plan, during the research project
some technical sheets have been drawn up to synthesize and frame the peculiarities of
this urbanistic instrument, in which the general data, the technical and regulatory
features and the purposes of the lighting plan itself have been reported.
2
In the European context, Brussels, Lion, Copenhagen, Eindhoven, Rotterdam and Barcelona stand
out.
3
Adopted by the municipality of Brussels in March 2016.
4
Accord de Majorité 2014−2019.
1172 C. B. Bellone and R. Ottavi
Even in Italy there are some virtuous examples of light planning5, among which the
P.C.I.P. (Public Lighting Municipal Plan) of Florence stands out, which gives priority
to the need and the concept of light quality6. What emerges from the analysis of the
Florence’s plan is indeed the pragmatization of the concept of quality, which can be
found in the issues covered such as plant engineering technological renewal of the
municipal lighting plan, modification of the road network and the urbanization
(pedestrian and cycling urban areas), energy saving, environmental protection and, not
least, green-procurement [13].
Similarly to the previous case study, a specific technical sheet of the urbanistic
instrument has been drawn up also for the city of Florence.
Both plans, both planning modes, express innovation of the subject by mutating the
very concept of lighting plan; despite implementing different approaches to this topic,
they both move in the direction of giving special attention to the most functional
themes of the cities of the future, with the aim of improving the urban space user’s
living standards.
In this regard, there are particularly significant elements that effectively show these
different approaches.
About the P.C.I.P. in Florence, project-based structure stands out immediately,
which means involving lighting mimic diagrams, reports, risk analysis and graphic
drawings: an important documentary writing that goes in determining what will be the
structure of the next project phases of technological upgrading and modernization.
While in the case of the plan lumière, even if a broader approach at an analysis context
level can be perceived, Brussels chose a more compact form and a very manual-like
representation of its urban space. The plan lumière of the Brussels-Capital Region is a
real guideline for the lighting of the urban space; everything is based on a very
innovative and efficient method, ready for use by light designers for a direct imple-
mentation on the territory.
Again, Florence chooses the theme of the user’s visual perception of the night
urban space [14] as the heart of its plan, all the planning follows this fil rouge with a
particular attention to protecting the artistic and architectural context of the space to be
illuminated: the city of Florence.
While the lighting plan of Florence shows innovation in planning the quality of
urban light in favor of the users, even imagining them as they are surrounded by a real
open-air night museum itinerary, Brussels adopts innovative layered planning tech-
niques, directly involving the citizens and all their sensations and desires for comfort
and wellbeing in the plan itself7.
5
Among the most relevant Italian experiences and which, in some cases, have even try to overcome
the limitations imposed by a concept of general development urbanistic instrument, there are
Florence, Udine, Turin, Bergamo, Rome and Milan.
6
Adopted by the municipality of Florence on 12 September 2016.
7
Organization’s program of the marche exploratoire nocturne (14/04/2016) – participatory study tool
involving a sort of night walk of the night urban space. (Source: Lighting Plan of the Brussels-
Capital Region, p. 125 - Partie Technique).
The Lighting Plan: From a Sector-Specific Urbanistic Instrument 1173
In the Plan Lumière of Brussels there is evidently a structure of the discipline of the
Light Urbanism and the methodology of the Light Master Plan8. It is also noteworthy
that the municipality of Brussels aims for a dynamic lighting through a chronobio-
logical approach (even creating moments of dark therapy) and a sensitivity focused on
people’s circadian rhythms, on how citizens can perceive the lighting, discovering and
understanding the functionality of the urban space around them. Finally, this special
attention towards the man also affects the surrounding environment, with the imple-
mentation of lighting control measures in harmony with flora and fauna [15].
What clearly emerges from the analysis and the comparison of these two plans is
the observation that both use consolidated techniques but at the same time innovative in
planning the urban lighting and, although they are two de facto capitals (one of the
Italian Renaissance and the other of a united Europe), it shows the lack of a magic
word, which is smart city, to converge on innovation, on the opportunities of lighting as
an “enabling technology” [16] capable of shaping the future of the smart city.
Both such planning experiences are examples of an organic, careful and innovative
planning modes at the same time.
On that basis, it is desirable that the culture of good lighting may be increasingly
extended and implemented by the local authorities, in order that lighting can actually be
considered as an element of characterization capable of deeply affecting the urban
space, no longer as a mere lighting plant [17] (already an important signal from the
world of standardization, with the birth of the new CIE9 234:2019 - A Guide to Urban
Lighting Masterplanning). In other words, it is proposed to imagine light as an
instrument capable of improving society in its everyday life, by redesigning the urban
space and making it more responsive to the needs of the community.
The results of the research have led to the definition of 15 guidelines standing as
good practices for the conception and implementation of innovative and high-quality
lighting plans focused on men and their relationship with the urban space in all its
forms, functions and conditions.
Here is an overview of the above guidelines, indicating the urgency of intervening
in this field with a regained approach of innovation and high quality:
(a) The choice of the model of plan and the implementation methodology (guidelines
1-2-3-15).
A model of plan shall be identified to combine functionality, aesthetics and
sociality; organic and proceduralized. The model shall begin from the pre-completion
status and present a layered structure, in order to make the elements overlapping and, at
the same time, analyzable separately.
8
The methodology of the Lighting Master Plan created by Roger Narboni, the lighting designer who
outlined the first plan of Brussels in 1997.
9
CIE: Commission Internationale de l'Éclairage.
1174 C. B. Bellone and R. Ottavi
References
1. De Seta, C. (ed.): La città: da Babilonia alla smart city. Rizzoli, Milano (2017)
2. Talia, M. (ed.): La pianificazione del territorio. Il sole 24 ore, Milano (2003)
3. Rallo, D. (ed.): Divulgare l’urbanistica. Alinea, Firenze (2002)
The Lighting Plan: From a Sector-Specific Urbanistic Instrument 1175
4. Cellucci, L., Monti, L., Gugliermetti, F., Bisegna, F.: Proposta di una procedura
schematizzata per semplificare la redazione dei Piani Regolatori di Illuminazione Comunale
(PRIC) (2012). http://www.enea.it
5. Enel-Federelettrica: Guida per l’esecuzione degli impianti di illuminazione pubblica, pp. 25–
26. Roma (1990)
6. Spaini, M.: Inquadramento normo-giuridico della pubblica illuminazione. Atto di convegno
AIDI-ENEA, Milano (2014)
7. McKinsey Global Institute: Smart cities: Digital solutions for a more livable future, rapporto
(2018). http://www.mckinsey.com
8. Süss, M.: I piani della luce – Obiettivi, finalità e opportunità per le Pubbliche
Amministrazioni. Atto di convegno AIDI-ENEA (2014)
9. Progetto Lumière ENEA - Report RdS/2010/250
10. Terzi, C. (ed.): I piani della luce. Domus « iGuzzini » , Milano (2001)
11. Ratti, C., con Claudel, M. (eds.): La città di domani. Come le reti stanno cambiando il futuro
urbano. Einaudi, Torino (2017)
12. Plan Lumière de la Régione de Bruxelles-Capitale (www.mobilite-mobiliteit.brussels/www.
radiance35.it/Bruxelles Mobilité)
13. Piano Comunale di Illuminazione Pubblica della Città di Firenze, (www.silfi.it/P.O. Man-
ufatti e Impianti Stradali comune di Firenze)
14. Gehl, J. (ed.): Vita in città. Maggioli, Rimini (2012)
15. Narboni, R. (ed.): italiana a cura di Palladino P.: Luce e Paesaggio. Tecniche Nuove, Milano
(2006)
16. Gianfrate, V., Longo, D. (eds.): Urban Micro-Design. FrancoAngeli, Milano (2017)
17. Frascarolo, M. (ed.): Manuale di progettazione illuminotecnica (parte a cura di E.Bordonaro)
Manco.su Edizioni Architectural Book and Review « TecnoTipo » , Roma (2011)
18. Dioguardi G. (ed.): Ripensare la città, p. 45. Donzelli, Roma (2001)
19. Bellone C., Ranucci P., Geropanta V.: The ‘Governance’ for smart city strategies and
territorial planning. In: (a cura di) Vasant, P., Zelinka, I, Weber, GM.: Intelligent Computing
& Optimization, Intelligent Computing & Optimization. AISC, pp. 76–86. Springer, Berlin
(2018)
20. Bellone C., Geropanta V.: The ‘Smart’ as a project for the city smart technologies for
territorial management planning strategies. In: (a cura di): Vasant, P., Zelinka, I., Weber,
GW. (eds.) Intelligent Computing & Optimization. AISC, vol. 866, pp. 66–75. Springer,
Berlin (2019)
PID Controller Design for BLDC Motor Speed
Control System by Lévy-Flight Intensified
Current Search
1 Introduction
Since 1970’s, the brushless direct current (BLDC) motor has been widely used in
different applications, for example, industrial automation, automotive, aerospace,
instrumentation and appliances [1]. This is because the BLDC motor retains the
characteristics of a brushed DC motor but eliminates commutator and brushes. It can be
driven by DC voltage source, while the current commutation is electronically done by
solid state switches [2]. From literature reviews, the BLDC motor have many advan-
tages over brushed DC motors. Among those, higher speed range, higher efficiency,
better speed versus torque characteristics, longer operating life, noiseless operation and
higher dynamic response can be found in BLDC usage [1–3].
The BLDC motor speed control can be efficiently operated under the feedback
control loop with the proportional-integral-derivative (PID) controller [4–6]. Moving
toward new era of control synthesis, an optimal controller design has been changed
from the conventional paradigm to the new framework based on modern optimization
by using some powerful metaheuristics as an optimizer [7, 8]. For example of the
BLDC motor speed control, optimal PID controller was successfully design by genetic
algorithm (GA) [6], tabu search [6], cuckoo search (CS) [9] and intensified current
search (ICuS) [10].
The Lévy-flight intensified current search (LFICuS), one of the most efficient
metaheuristics, is the lasted modified version of the ICuS. It utilizes the random number
drawn from the Lévy-flight distribution and the adjustable search radius mechanism to
enhance the search performance and speed up the search process. In this paper, the
LFICuS is applied to design an optimal PID controller for BLDC motor speed control
system. This paper consists of five sections. After an introduction is given in Sect. 1,
the BLDC motor control loop with the PID controller is described in Sect. 2. Problem
formulation of the LFICuS-based PID controller design is illustrated in Sect. 3. Results
and discussions are illustrated in Sect. 4. Finally, conclusions are given in Sect. 5.
The BLDC motor control loop with the PID controller can be represented by the block
diagram as shown in Fig. 1, where R(s) is reference input signal, C(s) is output signal,
E(s) is error signal between R(s) and C(s), U(s) is control signal and D(s) is disturbance
signal.
9
dia >
van ¼ Ra ia þ La þ ean >
>
>
dt >
>
=
dib
vbn ¼ Rb ib þ Lb þ ebn ð1Þ
dt >
>
>
>
>
þ ecn >
dic
vcn ¼ Rc ic þ Lc ;
dt
9
ean ¼ Ke xfa ðhe Þ >=
ebn ¼ Ke xfb ðhe 2p=3Þ ð2Þ
>
;
ecn ¼ Ke xfc ðhe 4p=3Þ
The electromagnetic torque Te depending on the current and back emf voltage is
expressed in (3), where Tea, Teb and Tec represent phase electric torque. The relation
between the speed and the torque is then stated in (4), where Tl is load torque, J is the
moment of inertia and B is the viscous friction.
In generally, a simple arrangement of a symmetrical (balanced) 3-phase wye
(Y) connection as shown in Fig. 2 can provide the per-phase concept. With symmet-
rical arrangement, the mechanical time constant sm and the electrical time constant se of
the BLDC can be formulated as expressed in (5) [5], where Kt is torque constant.
Commonly, the BLDC motor will be driven by DC input voltage via the particular
invertor (or power amplifier). Such the amplifier can be approximated by the first order
model. Therefore, the mathematical model of the BLDC motor can be formulated as the
transfer function stated in (6), where KA is amplifier constant and sA is amplifier time
constant. From (6), the BLDC motor model with power amplifier can be considered as
the third-order system.
XðsÞ KA 1=Ke b0
Gp ðsÞ ¼ ¼ ¼ ð6Þ
VðsÞ sA s þ 1 sm se s2 þ sm s þ 1 a 3 s 3 þ a 2 s 2 þ a1 s þ a0
In this work, BLDC motor model will be obtained by MATLAB system identifi-
cation toolbox [11]. The testing rig of the BLDC motor system including a blushless
DC motor of 350 W, 36 VDC, 450 rpm and a power amplifier (driver) is conducted as
shown in Fig. 3. The experimental speed data at 280, 320 and 360 rpm are tested and
recorded by a digital storage oscilloscope and PC. Speed data at 320 rpm as the
operating point will be used for model identification, while those at 280 and 360 rpm
will be used for model validation.
Based on the third-order model in (6), MATLAB system identification toolbox
provides the BLDC motor model as stated in (7). Results of model identification and
validation are depicted in Fig. 4. It was found that the model in (7) shows very good
agreement to actual dynamics (sensory data) of BLDC motor. The BLDC motor model
in (7) will be used as the plant model Gp(s) shown in Fig. 1.
Power Amp
(Driver)
+ 350W 36Vdc
10A
Power Supply
36VDC, 10A
BLDC Motor
350W 36VDC
-
450rpm.
Tachogenerator
Vout = 0-10V
Computor
Intel(R) Core(TM) i7-10750H -
CPU@2.60GHz
+
USB
DAQ
NI USB-6008
0:5725
Gp ðsÞ ¼ ð7Þ
s3 þ 3:795s2 þ 5:732s þ 0:5719
1180 P. Leeart et al.
Ki
Gc ðsÞ ¼ Kp þ þ Kd s ð9Þ
s
PID Controller Design for BLDC Motor Speed Control System 1181
3 Problem Formulation
The LFICuS-based PID controller design for the BLDC motor speed control system
can be formulated as represented in Fig. 5. Based on modern optimization, the
objective function f(Kp,Ki,Kd) is set as the sum-squared error (SSE) between the ref-
erence input r(t) and the speed output c(t) as stated in (10). The objective function f
() = SSE will be fed to the LFICuS block to be minimized by searching for the
appropriate values of the PID controller’s parameter, i.e., Kp, Ki and Kd, within the
corresponding search spaces and according to the constraint functions set from the
design specification as expressed in (11), where [Kp_min, Kp_max] is the lower and upper
bounds of Kp, [Ki_min, Ki_max] is the lower and upper bounds of Ki, [Kd_min, Kd_max] is
the lower and upper bounds of Kd, tr is the rise time, tr_max is the maximum allowance
of tr, Mp is the maximum percent overshoot, Mp_max is the maximum allowance of Mp,
ts is the settling time, ts_max is the maximum allowance of ts, ess is the steady-state error
and ess_max is the maximum allowance of ess.
Fig. 5. LFICuS-based PID controller design for BLDC motor speed control system.
X
N
Min f ðKp ; Ki ; Kd Þ ¼ ½ri ðtÞ ci ðtÞ2 ð10Þ
i¼1
9
Subject to tr tr max ; Mp Mp max ; >
>
>
>
ts ts max ; ess ess >
max ; >
=
Kp min Kp Kp max ; ð11Þ
>
>
Ki min Ki Ki max ; >
>
>
>
;
Kd min Kd Kd max
In this work, the LFICuS algorithm, the last modified version of the ICuS algorithm
[10], is applied to optimize the PID controller for the BLDC motor speed control
system. The LFICuS utilizes the random number drawn from the Lévy-flight distri-
bution to generate the neighborhood members as feasible solutions in each search
iteration. Such the Lévy-flight random distribution L can be approximated by (12) [12],
1182 P. Leeart et al.
where s is step length, k is an index and C(k) is the Gamma function as expressed in
(13). In the LFICuS algorithm, the adjustable search radius (AR) and adjustable
neighborhood member (AN) mechanisms are also conducted by setting the initial
search radius R = X (search space). The LFICuS algorithm for optimizing the PID
controller for the BLDC motor speed control system can be described step-by-step as
follows.
kCðkÞ sinðpk=2Þ 1
L 1þk ð12Þ
p s
Z1
CðkÞ ¼ tk1 et dt ð13Þ
0
Step-0 Initialize the objective function f() = SSE in (10) and constraint functions in
(11), search space X = [Kp_min, Kp_max], [Ki_min, Ki_max] and [Kd_min, Kd_-
max], memory lists (ML) W, Ck and N = ∅, maximum allowance of solution
cycling jmax, number of initial solutions N, number of neighborhood
members n, search radius R =X, k = j = 1.
Step-1 Uniformly random initial solution Xi= {Kp, Ki and Kd} within X. Evaluate f
(Xi) via (10) and (11), then rank and store Xi in W.
Step-2 Let x0 = Xk as selected initial solution. Set Xglobal = Xlocal= x0.
Step-3 Generate new solutions xi= {Kp, Ki and Kd} by Lévy-flight random in (12)
and (13) around x0 within R. Evaluate f(xi) via (10) and (11), and set the best
one as x*.
Step-4 If f(x*) < f(x0), keep x0 into Ck, update x0 = x* and set j = 1. Otherwise,
keep x* into Ck and update j = j + 1.
Step-5 Activate AR mechanism by R = qR, 0 < q < 1 and invoke AN mechanism
by n = an, 0 < a < 1.
Step-6 If j jmax, go back to Step-3.
Step-7 Update Xlocal = x0 and keep Xglobal into N.
Step-8 If f(Xlocal) < f(Xglobal), update Xglobal = Xlocal.
Step-9 Update k = k + 1 and set j = 1. Let x0 = Xk as selected initial solution.
Step-10 If k N, go back to Step-2. Otherwise, stop the search process and report
the best solution Xglobal= {Kp, Ki and Kd} found.
To optimize the PID controller of the BLDC motor speed control system, the LFICuS
algorithm was coded by MATLAB version 2017b (License No.#40637337) run on
Intel(R) Core(TM) i7-10510U CPU@1.80 GHz, 2.30 GHz, 16.0 GB-RAM. The
search parameters of the LFICuS are set from the preliminary study, i.e., initial search
radius R = X = [Kp_min, Kp_max], [Ki_min, Ki_max] and [Kd_min, Kd_max], step length
PID Controller Design for BLDC Motor Speed Control System 1183
s = 0.01, index k = 0.3, number of initial neighborhood members n = 100 and number
of search directions N = 50. Each search direction will be terminated by the maximum
iteration (Max_Iter) of 100. Number of states of the AR and AN mechanisms activation
h = 2, state-(i): at the 50th iteration, R = 50% of X and n = 50, state-(ii): at the 75th
iteration, R = 25% of X and n = 25. The constraint functions (11) are set as expressed
in (14). 50-trials are run for searching for the optimal parameters of the PID controller.
9
Subject to tr 2:50 sec.; Mp 5:00%; > >
>
>
ts 4:00 sec.; ess 0:01%; > >
=
0 Kp 10:00; ð14Þ
>
>
0 Ki 5:00; >
>
>
>
;
0 Kd 5:00
After the search process stopped, the LFICuS can successfully provide the PID
controller for the BLDC motor speed control system as stated in (15). The convergent
rates of the LFICuS for the PID design optimization over 50-trials run are depicted in
Fig. 6. The step-input and step-disturbance responses of the BLDC motor speed control
system without and with PID controller designed by the LFICuS are depicted in Fig. 7
and Fig. 8, respectively.
Fig. 6. Convergent rates of the LFICuS for the PID design optimization.
0:8164
Gc ðsÞ ¼ 8:5875 þ þ 1:8164s ð15Þ
s
Referring to Fig. 7, it was found that the BLDC motor without the PID controller
gives tr = 20.39 s., Mp = 0.00%, ts = 36.74 s., and ess = 0.00%. The BLDC motor
speed control system with the PID controller optimized by the LFICuS in (15) yields
tr = 2.44 s., Mp = 2.47%, ts = 3.28 s., and ess = 0.00%. From Fig. 8, it was found that
the BLDC motor without the PID controller cannot regulate the output response from
1184 P. Leeart et al.
the external disturbance. However, the BLDC motor speed control system with the PID
controller designed by the LFICuS in (15) can successfully regulate the output response
from the external disturbance with the maximum percent overshoot from regulation
Mp_reg = 10.43% and regulating time treg = 19.78 s.
5 Conclusions
The application of the LFICuS to design an optimal PID controller for the BLDC motor
speed control system has been presented in this paper. The LFICuS algorithm has been
developed from the ICuS as one of the most efficient metaheuristic optimization
techniques. Based on modern optimization, the sum-squared error between the refer-
ence input and the speed output of the BLDC motor speed control system has been set
as the objective function for minimization. In addition, the desired specification and
search bounds have been correspondingly set as the constraint functions. As results, the
optimal PID controller for the BLDC motor speed control system has been successfully
obtained by the LFICuS algorithm. The step-input and step-disturbance responses of
the BLDC motor speed controlled system are very satisfactory according to the design
specification. The advantage of the proposed design method is that users can utilize any
controller to be optimally designed for any plant of interest. However, the disadvantage
of this method is the boundary setting which is the problem-dependent. For future
research, the LFICuS algorithm will be conducted to design the PIDA, FOPID and
FOPIDA controllers for many real-world systems.
References
1. Yedamale, P.: Brushless DC (BLDC) motor fundementals. AN885 Michrochip Technol. Inc.
20, 3–15 (2003)
2. Vas, P.: Parameter Estimation, Condition Monitoring and Diagnosis of Electrical Machines.
Oxford University Press, Oxford (1993)
3. Tashakori, A., Ektesabi, M., Hosseinzadeh, N.: Modeling of BLDC motor with ideal back-
EMF for automotive applications. In: The World Congress on Engineering (2011)
4. Othman, A.S., Mashakbeh, A.: Proportional integral and derivative control of brushless DC
motor. Eur. J. Sci. Res. 35(4), 198–203 (2009)
5. Patel, V.K.R.S., Pandey, A.K.: Modeling and performance analysis of PID controlled BLDC
motor and different schemes of PWM controlled BLDC motor. Int. J. Sci. Res. Publ. 3, 1–14
(2013)
6. Boonpramuk, M., Tunyasirut, S., Puangdownreong, D.: Artificial intelligence-based optimal
PID controller design for BLDC motor with phase advance. Indonesian J. Electr. Eng.
Inform. 7(4), 720–733 (2019)
7. Zakian, V.: Control Systems Design: A New Framework, Springer-Verlag (2005)
8. Zakian, V., Al-Naib, U.: Design of dynamical and control systems by the method of
inequalities. IEEE Int. Conf. 120, 1421–1427 (1973)
9. Puangdownreong, D., Kiree, C., Kumpanya, D., Tunyasrirut, S.: Application of cuckoo
search to design optimal PID/PI controllers of BLDC motor speed control system. In: Global
Engineering & Applied Science Conference, pp. 99–106 (2015)
10. Puangdownreong, D., Kumpanya, D., Kiree, C., Tunyasrirut, S.: Optimal tuning of 2DOF–
PID controllers of BLDC motor speed control system by intensified current search. In:
Global Engineering & Applied Science Conference, pp. 107–115 (2015)
11. Ljung, L.: System Identification Toolbox for use with MATLAB. The MathWorks (2007)
12. Yang, X.S.: Flower pollination algorithm for global optimization. Unconventional Comput.
Nat. Comput. Lect. Notes Comput. Sci. 7445, 240–249 (2012)
Intellectualized Control System
of Technological Processes of an Experimental
Biogas Plant with Improved System
for Preliminary Preparation of Initial Waste
1 Introduction
individual stages are insufficient, which in some cases determines the low efficiency of
biogas plants, does not allow them to be controlled to the necessary extent, leads to
unjustified overstatement of construction volumes, increase in operating costs and,
accordingly, the cost of 1 m3 of biogas produced. This puts forward the tasks of
developing the most effective technological schemes for biogas plants, the composition
of their equipment, creating new designs and calculating their parameters, improving
the reliability of their work, reducing the cost and construction time, which is one of the
urgent problems in solving the issue of energy supply of agricultural production
facilities [1].
One of the most promising and energy-efficient methods of preparing a substrate for
fermentation is processing it in a vortex layer of ferromagnetic particles (vortex layer
apparatus (VLA)), which is created by the action of a rotating magnetic field [2].
Previously, the positive effect of processing various organic substrates in VLA on the
characteristics of methanogenic fermentation, in particular on the kinetics of
methanogenesis, the completeness of decomposition of organic matter, the content of
methane in biogas, and waste disinfection has been shown [3–6].
The purpose of this work is to develop an intellectualized control system of
technological processes of an experimental biogas plant with improved system for
preliminary preparation of initial waste, created to study the effectiveness of using of
VLA for the pretreatment of return biomass during anaerobic processing of organic
waste.
2 Background
3 Results of Investigations
Figure 1 shows a block diagram of the proposed method for increasing the efficiency of
anaerobic bioconversion of organic waste for the study of which an experimental setup
was created and an intellectualized process control system was developed.
Fig. 1. A block diagram of a method for increasing the efficiency of anaerobic bioconversion of
organic waste to produce gaseous energy and organic fertilizers based on biomass recycling using
a vortex layer apparatus
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
TIRCA1 LIRCA1
Fig. 2. Functional diagram of PLC integration into an intellectualized process control system
In Fig. 2, arrows show the directions of electrical signals between the simulated
controllers (temperature and level controllers, time relays) and the contacts of sensors
and actuators (the numbers of contacts in Fig. 2 correspond to the numbers of contacts
in Fig. 3). The designations of regulators and sensors comply with GOST 21.208-2013.
The locations of the sensors and actuators required for the implementation of the
processes of pretreatment of the mixture of substrates, anaerobic bioconversion of
organic matter, as well as for the recirculation of the thickened fraction of the anaer-
obically treated substrate, are shown in the schematic diagram of the intellectualized
process control system shown in Fig. 3.
The main parameters to be determined in the course of experimental research:
– hydraulic retention time in the anaerobic bioreactor;
– hydraulic retention time of the vortex layer apparatus;
– hydraulic retention time in the preparation reactor;
– the quantity and quality of the produced biogas;
– the rate of recirculation of the thickened fraction of the anaerobically treated
substrate;
– frequency of the electromagnetic field in the vortex layer apparatus;
– substrate temperature during anaerobic processing;
– initial substrate temperature.
1192 A. Kovalev et al.
3 M 1
QIE
8 M 10
11
LE1 5
2 7
1 4 6 6 LE2
2 9
14
TE1 TE2 9 10
4 5 TE3 TE4
3
12 13
8
16 15
12
7
11
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
EDT
UVDT EDT
On the panel
UVDT
EDT
UDIRCABiBo
4 Conclusion
The application of the proposed method of intensification allows to obtain the fol-
lowing effects:
1. The return of the condensed effluent fraction allows to increase the residence time of
the hardly decomposable components of organic waste in the reactor, which in turn
leads to an increase in the degree of decomposition of organic matter and,
accordingly, an increase in the biogas yield.
2. Biomass recycling leads to an increase in the concentration of methanogenic
microorganisms in the bioreactor, which will increase the productivity of the
bioreactor and ensure the stability of its operation when changing the characteristics
of the incoming raw materials.
Intellectualized Control System of Technological Processes 1193
3. Fine grinding, which allows to improve the rheological properties of the substrate,
to conduct partial hydrolysis of complex organic compounds, to improve the
availability of nutrients for microorganisms, to ensure heating of the substrate.
4. The introduction of the abrasive working body (steel needles) into the substrate of
ferromagnetic particles, which allows to shorten the launch time of the bioreactor,
increase the rate of formation and final yield of methane, provide more complete
decomposition of the substrate and reduce the required volume of the bioreactor,
increase the adaptive ability of the microbial community to adverse conditions (for
example excessive accumulation of volatile fatty acids (VFA) or H2, lower pH).
Thus, the use of the vortex layer apparatus for additional processing of the recir-
culated biomass of the condensed fraction of fermented organic waste mixed with
prepared source organic waste in a system for increasing the efficiency of anaerobic
bioconversion of organic waste to obtain a gaseous energy carrier based on biomass
recirculation will allow providing a synergistic effect of a set of methods for intensi-
fying the process of anaerobic biochemistry organic waste of the agricultural industry
in anaerobic bioreactors, namely:
– mechanical impact on the initial organic waste (grinding of the initial mass in VLA
before loading into the bioreactor);
– biochemical methods of influencing the fermentable substrate (introducing ferro-
magnetic particles into the substrate in VLA);
– electromagnetic processing of fermentable substrate;
– microbiological methods (biomass retention due to its recirculation).
The use of the developed intellectualized process control system makes it possible
to determine the main parameters of experimental research and maintain the specified
operating modes of the equipment of the experimental biogas plant, which in turn will
reduce the error in the subsequent mathematical processing of experimental data.
5 Acknowledgment
This work was supported by the Federal State Budgetary Institution “Russian Foun-
dation for Basic Research” as part of scientific project No. 18-29-25042.
References
1. Kovalev A.A.: Tekhnologii i tekhniko-energeticheskoye obosnovaniye proizvodstva biogaza
v sistemakh utilizatsii navoza zhivotnovodcheskikh ferm (Technologies and feasibility study
of biogas production in manure utilization systems of livestock farms). Dissertatsiya …
doktora tekhnicheskikh nauk (Thesis … doctor of technical sciences), p. 242, All-Russian
Research Institute of Electrification of Agriculture, Moscow (1998)
1194 A. Kovalev et al.
2. Kovalev, D., Kovalev, A., Litti, Y., Nozhevnikova, A., Katrayeva, I.: Vliyaniye nagruzki po
organicheskomu veshchestvu na protsess biokonversii predvaritel’no obrabotannykh
substratov anaerobnykh bioreaktorov (The Effect of the Load on Organic Matter on
Methanogenesis in the Continuous Process of Bioconversion of Anaerobic Bioreactor
Substrates Pretreated in the Vortex Layer Apparatus). Ekologiya i promyshlennost’ Rossii
(Ecology and Industry of Russia) 23(12), 9–13 (2019). https://doi.org/10.18412/1816-0395-
2019-12-9-13
3. Litti, Yu., Kovalev, D., Kovalev, A., Katraeva, I., Russkova, J., Nozhevnikova, A.:
Increasing the efficiency of organic waste conversion into biogas by mechanical pretreatment
in an electromagnetic mill. In: Journal Physics: Conference Series vol. 1111, no. 1,
p. 012013 (2018). https://doi.org/10.1088/1742-6596/1111/1/012013
4. Kovalev, D.A., Kovalev, A. A., Katraeva, I.V., Litti, Y.V., Nozhevnikova A.N.: Effekt
obezzarazhivaniya substratov anaerobnykh bioreaktorov v apparate vikhrevogo sloya (The
effect of disinfection of substrates of anaerobic bioreactors in the vortex layer apparatus).
Khimicheskaya bezopasnost’ (chemsafety) 3(1), 56–64 (2019) https://doi.org/10.25514/
CHS.2019.1.15004
5. Kovalev, A.A., Kovalev, D.A., Grigor’yev, V.S.: Energeticheskaya effektivnost’ pred-
varitel’noy obrabotki sinteticheskogo substrata metantenka v apparate vikhrevogo sloya
(Energy Efficiency of Pretreatment of Digester Synthetic Substrate in a Vortex Layer
Apparatus). 2020; 30(1):92–110. Inzhenernyye tekhnologii i sistemy (Engineering Tech-
nologies and Systems) 30(1), 92–110 (2020). https://doi.org/10.15507/2658-4123.030.
202001.092-110
6. Litti, Y.V., Kovalev, D.A., Kovalev, A.A., Katrayeva, I.V., Mikheyeva, E.R., Nozhevni-
kova, A.N.: Ispol’zovaniye apparata vikhrevogo sloya dlya povysheniya effektivnosti
metanovogo sbrazhivaniya osadkov stochnykh vod (Use of a vortex layer apparatus for
improving the efficiency of methane digestion of wastewater sludge) Vodosnabzheniye i
sanitarnaya tekhnika (Water supply and sanitary technique) 11, 32–40 (2019). https://doi.
org/10.35776/MNP.2019.11.05
7. Kovalev, A., Kovalev, D., Panchenko, V., Kharchenko, V., Vasant, P.: Optimization of the
process of anaerobic bioconversion of liquid organic wastes. In: Intelligent Computing &
Optimization. Advances in Intelligent Systems and Computing, vol. 1072, pp. 170–176
(2020). https://doi.org/10.1007/978-3-030-33585-4_17
8. Kovalev, A., Kovalev, D., Panchenko, V., Kharchenko, V., Vasant, P.: System of
optimization of the combustion process of biogas for the biogas plant heat supply. In:
Intelligent Computing & Optimization. Advances in Intelligent Systems and Computing,
vol. 1072, pp. 361–368 (2020). https://doi.org/10.1007/978-3-030-33585-4_36
9. Kalyuzhnyi, S., Sklyar, V., Fedorovich, V., Kovalev, A., Nozhevnikova, A.: The
development of biotechnological methods for utilisation and treatment of diluted manure
streams. In: Proceeding IV International Conference IBMER, Warshawa. (1998)
10. Mulata, D.G., Jacobi, H.F., Feilberg, A., Adamsen, A.P.S., Richnow, H.-H., Nikolausz, M.:
Changing feeding regimes to demonstrate flexible biogas production: effects on process
performance, microbial community structure and methanogenesis pathways. Appl. Environ.
Microbiol. 82(2), 438–449 (2016). https://doi.org/10.1128/AEM.02320-15
Way for Intensifying the Process of Anaerobic
Bioconversion by Preliminary Hydrolysis
and Increasing Solid Retention Time
1 Introduction
The negative impact of agricultural activities on the environment is associated not only
with the increasing consumption of natural resources, but also, to a greater extent, with
the formation of liquid and solid waste from agricultural and processing industries. In
particular, raising animals, processing meat and dairy products, producing beer, sugar,
starch, etc. are accompanied by the formation of a large amount of wastewater [1, 2].
Obtaining biogas is economically justified and is preferable in the processing of a
constant stream of waste, and is especially effective in agricultural complexes, where
there is the possibility of a complete ecological cycle [3].
The transition of animal husbandry to an industrial basis and the associated con-
centration of animals on large farms and complexes lead to a sharp increase in the
volume of organic waste that must be disposed of without polluting the environment.
One of the ways of processing organic waste is its anaerobic digestion in biogas
plants due to the vital activity of microorganisms (methanogenesis), when biogas is
obtained with the help of microbiological processing of biomass and used, in turn, as a
raw material for obtaining thermal and electrical energy [3].
Development, design and construction of new and modification of existing reactors
for biogas production from various agricultural waste are designed to solve a number of
significant energy and environmental problems. It allows to reduce anthropogenic load on
ecosystems by reduction harmful emissions of greenhouse gases and fully utilization and
recycling organic waste. In addition, the use of biogas plants can provide facilities with
uninterrupted electricity and heat supply for their own needs, as well as obtain, through
the use of various technological solutions, high-quality products from methanogenesis
waste, which can be used as fertilizer in greenhouses, feed additives or bedding for
livestock [3].
Despite the long-term use of biogas plants and an even longer period of studies of
the processes occurring in them, our ideas about its basic laws and mechanisms of
individual stages are insufficient, which in some cases determines the low efficiency of
biogas plants. It does not allow them to be controlled to the necessary extent, which
leads to unjustified overstatement of construction volumes, increase in operating costs
and, accordingly, the cost of 1 m3 of biogas produced. This puts forward the tasks of
developing the most effective technological schemes for biogas plants, the composition
of their equipment, creating new designs and calculating their parameters, improving
the reliability of their work, reducing the cost and construction time, which is one of the
urgent problems in solving the issue of energy supply of agricultural production
facilities [3].
In this regard, the main goal of the research was to modernize the technology using
one of the methods for intensifying processes in biogas plants - preliminary hydrolysis
of the initial organic waste together with recirculation of the thickened fraction of
fermented waste.
Thus, the purpose of the work is to develop the foundations of a way for inten-
sifying the process of anaerobic bioconversion of organic matter to obtain a gaseous
energy carrier. It is planned to intensify the process through the combined use of
recirculation of a thickened fraction of fermented waste and preliminary hydrolysis.
Recirculation of a thickened fraction of fermented waste makes it possible to increase
solid retention time and to retain biomass un bioreactor. At the same time, it is planned
to expose a mixture of an initial substrate and a thickened fraction of fermented waste
to preliminary hydrolysis.
2 Background
– stage of hydrolysis;
– fermentation stage;
– acetogenic stage;
– methanogenic stage [4].
The limiting stage of methane digestion of urban wastewater sludge (WWS) is the
hydrolysis of the solid phase, in particular, activated sludge, which consists mainly of
cells of microorganisms and a small amount of dissolved organic matter (OM). The
sediment of the primary settling tanks contains more dissolved OM and fewer active
microorganisms as compared to activated sludge. Soluble organic compounds, which
can be further converted into biogas, are formed in WWS in the course of hydrolysis.
Therefore, the biogas yield during WWS fermentation is in direct proportion to the
biodegradability of the sludge and, accordingly, the rate of hydrolysis. One of the
technological methods of increasing the bioavailability of sediments is their pretreat-
ment before fermentation in digesters. Pretreatment of sludge allows:
Thus, based on the degree of digestion of raw materials, such a retention time of
raw materials in the reactor is experimentally selected that ensures the most efficient
operation of the reactor (i.e., the maximum biogas yield at a relatively high degree of
decomposition of the substrate). A constant volume of raw materials is maintained in
the reactor by supplying new raw material to the reactor and removing the fermented
mass at regular intervals. Often the volume of added raw materials is more than the
volume of the removed fermented mass, because part of the total solids of the raw
material passes into biogas during the fermentation process. The volume of raw
materials in the reactor can be adjusted by adding a certain amount of liquid. The
fermented mass (substrate removed from the reactor after fermentation) consists of
water, including dissolved salts, inert materials and undecomposed OM. The fermented
mass also contains the biomass of microorganisms accumulated during the retention of
raw materials in the reactor.
Hydraulic retention time (HRT) is the average length of time that liquids and
soluble compounds remain in the reactor. An increase in HRT promotes a longer
contact between microorganisms and the substrate, but requires a slower supply of raw
materials (reactor loading) and / or a larger reactor volume. If the retention time is too
short, there is a great risk that the growth rate of the microorganisms will be lower than
the rate of removal of the raw materials from the reactor. Often, the doubling time of
methanogens in reactors is more than 12 days. Therefore, HRT should be longer than
this time, otherwise microorganisms will be removed out of the reactor during the
unloading of the fermented mass and the population will not be dense enough to
efficiently decompose the raw materials. For single-stage reactors, HRT ranges from 9
to 30 days; HRT for thermophilic reactors averages 66% of that for mesophilic reactors
(10–16 versus 15–25 days).
One of the ways to increase the efficiency of the process of anaerobic bioconversion
of organic waste to obtain a gaseous energy carrier and organic fertilizers is to increase
the time spent in the reactor on solid matter (solid retention time – srt).
Way for Intensifying the Process of Anaerobic Bioconversion 1199
3 Results of Investigations
Fig. 1. Technological scheme for using the proposed way for intensifying the process of
anaerobic bioconversion by preliminary hydrolysis and increasing solid retention time 1 -
hydrolyser; 2 - anaerobic bioreactor; 3 - effluent settler; 4 - pump for recirculation of the
thickened fraction; 5 - waterseal.
The anaerobic bioreactor (2) is equipped with both a stirring system and a heating
system with temperature control of the fermentation process. Also, the anaerobic
bioreactor (2) is equipped with systems for loading and unloading the substrate, as well
as a branch pipe for removing biogas. The design of the hydrolyser (1) repeats the
design of the anaerobic bioreactor (2) in a reduced form. As a consequence, the
hydrolyser (1) has the same systems as the anaerobic bioreactor (2). However, the heat
supply system of the hydrolyser (1) controls the temperature of the hydrolysis process,
1200 A. Kovalev et al.
and through the biogas removal pipe acid biogas is discharged, which is sent to the
anaerobic bioreactor. The effluent settler (3) is equipped with an overflow pipe, as well
as systems for loading the fermented substrate, removing the supernatant liquid and
draining the thickened fraction. The pump for recirculation of the thickened fraction
(4) delivers the thickened fraction of the fermented substrate to the hydrolyser (1). The
waterseal (5) is used to exclude air from entering the anaerobic bioreactor when
removing biogas, as well as to maintain the biogas pressure.
When developing a technological scheme for using the proposed way for intensi-
fying the process of anaerobic bioconversion by preliminary hydrolysis and increasing
solid retention time, both schemes of continuous supply of the initial substrate [8] and
schemes with discrete supply were taken into account [9]. In addition, it is proposed to
perform the heat supply scheme for the biogas plant in accordance with the scheme
described in the article [10].
Material Balance of Biogas Plant
The general view of the material balance of the hydrolyser is as follows:
where Ginit – specific feed of the initial substrate to the hydrolyser to pretreatment
for anaerobic digestion, kg/kginit, where kginit is the amount of organic matter in the
initial substrate;
k – proportion of recirculated thickened fraction of fermented waste;
Gthd – specific yield of the thickened fraction of fermented waste from the effluent
settler, kg/kginit;
Ginf – specific yield of the initial substrate prepared for anaerobic treatment,
kg/kginit;
Gabg – specific losses of organic matter with acid biogas during the preparation of
the initial substrate for anaerobic treatment, kg/kginit.
Specific losses of organic matter with acid biogas during the preparation of the
initial substrate for anaerobic treatment will be:
where uh – the degree of digestion of organic matter in the hydrolyser during the
preparation of the initial substrate for anaerobic treatment.
The general view of the material balance of the anaerobic bioreactor is as follows:
where Geff – specific yield of fermented waste from anaerobic bioreactor, kg/kginit;
Gbg – specific biogas yield from anaerobic bioreactor, kg/kginit.
Way for Intensifying the Process of Anaerobic Bioconversion 1201
where Gsup – specific yield of the supernatant liquid from the effluent settler,
kg/kginit.
The general view of the material balance of a biogas plant is as follows:
Figure 2 shows a block diagram of the material balance of a biogas plant with the
proposed way for intensifying the process of anaerobic bioconversion by preliminary
hydrolysis and increasing solid retention time.
Ginit Gbg
Anaerobic Effluent
Hydrolyser Ginf Geff Gsup
bioreactor settler
Gabg
k·Gthd
(1-k)·Gthd
Fig. 2. Block diagram of the material balance of a biogas plant with the proposed way for
intensifying the process of anaerobic bioconversion by preliminary hydrolysis and increasing
solid retention time
4 Conclusion
Application of the proposed method for intensifying the process of anaerobic bio-
conversion of organic matter will provide a synergistic effect of a combination of
methods for intensifying the process of anaerobic bioconversion.
The use of recirculation of the thickened fraction of fermented waste allows:
– to transfer a significant part of the organic matter of the substrate into a dissolved
state, thus improving the availability of nutrients to methanogenic microorganisms;
– to heat the initial substrate to the fermentation temperature in the anaerobic
bioreactor, which, in turn, allows you to maintain the optimal temperature mode in
the anaerobic bioreactor, avoiding temperature fluctuations that are harmful to
methanogenic microorganisms.
The supply of acid biogas formed during hydrolysis to the anaerobic bioreactor
allows:
– avoid the loss of organic matter of the original substrate with acid biogas formed
during hydrolysis;
– to intensify mixing in an anaerobic bioreactor, which, in turn, allows to improve
mass transfer and, as a consequence, the availability of nutrients of the fermented
mass to methanogenic microorganisms;
– to provide methanogenic microorganisms with additional nutrients in the gas phase.
Acknowledgment. This work was supported by the Federal State Budgetary Institution “Rus-
sian Foundation for Basic Research” as part of scientific project No. 18-29-25042.
References
1. Izmaylov, A.Y., Lobachevskiy, Y.P., Fedotov, A.V., Grigoryev, V.S., Tsench, Y.S.:
Adsorption-oxidation technology of wastewater recycling in agroindustrial complex
enterprises. Vestnik mordovskogo universiteta Mordovia Univ. Bull. 28(2), 207–221
(2018). https://doi.org/10.15507/0236-2910.028.201802.207-221
2. Artamonov, A.V., Izmailov, A.Y., Kozhevnikov, Y.A., Kostyakova, Y.Y., Lobachevsky, Y.
P., Pashkin, S.V., Marchenko, O.S.: Effective purification of concentrated organic
wastewater from agro-industrial enterprises, problems and methods of solution. AMA
Agric. Mech. Asia Afr. Latin Am. 49, 49–53 (2018)
3. Kovalev, A.A:. Tekhnologii i tekhniko-energeticheskoye obosnovaniye proizvodstva
biogaza v sistemakh utilizatsii navoza zhivotnovodcheskikh ferm (Technologies and
feasibility study of biogas production in manure utilization systems of livestock farms).
Dissertatsiya ... doktora tekhnicheskikh nauk (Thesis ... doctor of technical sciences), p. 242.
All-Russian Research Institute of Electrification of Agriculture, Moscow (1998)
Way for Intensifying the Process of Anaerobic Bioconversion 1203
1 Introduction
Improving the system of operation and maintenance of electric equipment will make it
possible to extend its lifespan, to save material expenditures for purchasing new electric
components and for repair of failed ones thus reducing associated technological
damage of agricultural enterprises [1–4]. Failure of electric motor or motor shutdown
due to a power outage lead to economic damage comprising its technological and
technical components arising due to, respectively, underproduction of livestock goods
and the need to replace failed electric motors [5]. Organizational concept, structure and
economic base of agricultural production have changed, during recent years. Modern
peasant households, farms and agricultural enterprises are provided with electrified
machinery and equipment, and new types of electric equipment are produced today
including electric motors and automatic tripping contactors [6, 7]. Evaluating damages
caused by failures of electric equipment, in livestock raising, is based on researches
[8–10] carried out earlier and dedicated to the methods of calculating damage to
agricultural production in case of power outages and electric equipment failures.
2 Technical Demages
Estimating damages shall be done on the basis of a complex approach to the solution of
the problem of improving operational reliability of rural electric installations with the
use of relevant data on damages caused by failure of electric drives, in technological
equipment. A convenient method for calculating damages is needed to be applied in
practice of electrical-technical services of agricultural enterprise that does not involve
insignificant components. It has to be supported with up-to-date reference materials on
electric equipment and agricultural products easy to use by service personnel of agri-
cultural enterprises.
In case of failure of electric motor, the plant or machine stops in which this motor is
installed. The extent of the damage is mainly defined by the period of time-out of
working mechanism [11, 12]. In the majority of cases, failed electric motors are not
subject to repair directly on the site. They are transported to a specialized service center
for complete repair in which case a spare electric motor has to be installed on the
working mechanism. Normally repair shops are located at a great distance from agri-
cultural production unit. That is why failed motors are stored on the site and sent to the
repair shops by batches of 10 to 15 pieces [12]. A certain majority of repair shops have
dedicated exchange fleets of electric motors in order to reduce the time of returning
repaired motors to the production sites. It means that repair shops receive failed electric
motors in exchange to repaired ones, in the same quantities, to maintain a bank of spare
components at the production sites [13].
Damage from replacement of failed electric motor comprises expenditures for
dismantling-assembling and the price of either a new electric motor or one that has
passed complete repair. Besides, it depends on capacity of electric motor, its size and
weight, design and methods of assembly for different types of machines characterized
by different operating conditions including convenience of dismantling-assembling.
That is why a uniform method of estimating the volumes of works and working
conditions of electrical installers cannot be reliably defined, in case of failure, for
particular type and capacity of electric motor. It is assumed that, in today’s operation
and maintenance conditions of electric motors, an approximate value of the average
share of technical component of economic damage associated with either ahead-of-time
repair of decommissioning of electric motor may amount to a half of its price.
1206 A. Nekrasov et al.
More specific values of technical damage due to electric motor failure can be
calculated in various ways depending on whether a failed electric motor has to be
replaced by either a new one of one that has passed the complete repair with the use of
fault-free components or it is intended to be repaired either at the site or in a repair shop
of agricultural enterprise. The unrecovered cost of electric motor and all other
expenditures and charges have to be taken into account which makes such calculations
relatively complicated.
As evidenced from the available information on the nomenclature of electric motors
supplied to agricultural consumers, electric equipment factories produce mainly motors
of general purpose industrial versions. It means that they are operated and applied in
conditions they are not intended for. With regard to low quality of complete repair the
lifespan of repaired motors does not, normally, exceed 1 to 1.5 years [8].
In Tables 1, 2, 3 and 4 the input data required for estimating the technical com-
ponent of economic damage is given including price, weight and complete repair
expenditures, for various types of asynchronous electric motors of unified series (in
Russia) most commonly applied in agricultural production [11, 12].
These data related to electric motors is intended for use while estimating economic
damages from failure of electric equipment, planning activities on technical mainte-
nance and improvement of its operational reliability to ensure high effectiveness of its
operation, in livestock-breeding farms.
Evaluation of Technical Damage Caused by Failures 1207
Technical damage from failures of electric motors depends on multiple factors [14].
That is why its correct values are difficult to define, in practice of agricultural enterprise
conditions. Technical damage DT associated with replacement of a failed electric motor
comprises the expenditures for dismantling-assembling and price of either new P0
electric motor or one that has passed complete repair PR, with the account of its
amortization.
In a general case, technical component of damage from ahead-of-time failure of
electric motor (that has not been operated during its rated lifespan) can be calculated as
follows [5]:
tF
DT ¼ PR þ KM 1 PD ð1Þ
TN
where PR are expenditures associated with replacement of the failed electric motor
with a new one (dismantling and assembling), KM is price of the failed electric motor,
PD is price of failed electric motor that has been decommissioned (metal scrap), TN and
tF are, respectively, rated and factual lifespan of electric motor.
In calculations, data on the factual lifespan shall be used that comes from agri-
cultural enterprises operating corresponding types of electric equipment in accordance
with the requirements of System of Scheduled-Preventive Repair and Technical
Maintenance of Electric Equipment in Agriculture. The rated lifespan of 7 years is
1208 A. Nekrasov et al.
Table 5. Calculation results for technical damage caused by failures of electric motors, in
livestock production
Electric motor model Technical damage (euros)
Lifespan tF (years)
1 2 3 4 5 6 7
AIR 0.55 kW 37 31 25 20 14 8 –
AIR 3.0 kW 80 66 53 40 27 14 –
AIR 5.5 kW 122 102 82 62 42 22 –
120 122
100 102
DT [euros]
80 80 82
66 62
60
53
40 40 42
37 27 22
20 31
25 14
20
0 14
8
0 1 2 3 4 5 6 7
tF [years]
Fig. 1. Dependence of technical damage caused by failure of electric motors on their lifespan,
for 0.55 kW, 3.0 kW and 5.5 kW.
Evaluation of Technical Damage Caused by Failures 1209
within a period not shorter than 5 to 6 years before the first complete repair and, at
least, 3 to 4 years after that. Electric motors shall be subject to decommissioning upon
termination of their rated lifespan determined by their depreciation rates. The values of
these rates are used while estimating the actual number of complete repairs and the
average values of technical damage.
In order to reduce material damages caused by failures of electrified equipment, at
agricultural enterprises, by decreasing the failure rate of electric motors it is necessary
to perform preventive activities on their technical maintenance and repair accompanied
by diagnostic checks of the technical status of electric motors. Besides, the required
sets of spare components and materials have to be kept for repair-maintenance pur-
poses, and the appropriate protection devices have to be applied to protect electric
motors, in alarm conditions. It is also important to improve operational efficiency of
repair teams and their provision with technical means for trouble-shooting and fault
handling.
A more reliable evaluation of the technical component of damage requires col-
lecting large volumes of data that can be only obtained in conditions of practical
operation and maintenance of electric equipment at agricultural enterprises. The latter
has to have a complete staff of energy service team capable to follow schedules of
preventive activities keeping accurate records related to the operation and maintenance
of electrified equipment. It is also important to know actual lifespan of spare electric
motors from the date of their manufacture by the factory till the moment of installation
on the working machine, as well as those of the failed electric motors from the moment
of manufacture by the factory till the date of failure on the corresponding working
machine. Their rated lifespan have to be also specified. In the process of operation of
electric motor, amortization charges related to electric motor’s initial price have to be
carried out, on the annual basis. These charges can be later involved for either repairing
the motor or its replacement by a new one. In case that a failed electric motor has not
operated during its whole rated lifespan before the complete repair or decommission-
ing, the enterprise will run up losses associated with repair or replacement that are a
part of technical damage DT.
In order to make practical evaluations of technical damage more convenient it is
advisable to apply summarizing coefficient of damage kD depending on actual lifespan
of electric motor. In the case of failure of electric motor followed by its decommis-
sioning or replacement by its analog, damage DT arises that is defined by the following
expression [12]:
DT ¼ PR kD ðtF Þ ð2Þ
1 1.0
0,8 0.8
0,6
kD [a.u.]
0.6
0,4 0.4
0,2 0.2
0.1
0
0 1 2 3 4 5 6 7
tF [years]
Fig. 2. Dependence of coefficient of damage on actual lifespan of electric motor, in livestock
production.
from Fig. 2, coefficient of damage equals to 0.5 to 0.6, for the average value of lifespan
of electric motor equal to 3 to 3.5 years, in livestock production.
Calculation results for the average value of technical damage DT caused by failure
of electric motors of type AIR depending on their capacity calculated with the use of
expression (2), for KD = 0.6, are presented in Table 6.
Table 6. Calculation results for technical damage caused by failure of electric motor of type
AIR, in livestock production, for KD = 0.6.
Capacity of electric motor (P) kW 1.1 2.2 3.0 4.0 5.5 7.5
Price of new electric motor (P0), euros 52.92 72.43 92.57 103.55 139.30 177.79
Price of installation (PI), euros 5.30 7.25 9.26 10.35 13.93 17.78
PR = P0 + PI, euros 58.21 79.61 101.83 113.90 153.23 195.56
Damage (DT), euros 34.92 47.76 61.10 68.33 91.86 117.34
4 Conclusions
damage arise from the reduction of actual lifespan and because of unexpended price of
electric motor. For practical application in conditions of livestock farm, the average
calculated value of technical damage due to emergent failure of single electric motor is
defined with the use of averaged coefficient of technical damage related to the price of a
new electric motor including installation expenditures. Calculations have been per-
formed, and dependences of technical damage on actual lifespan have been defined, for
failures of electric motors of various capacities applied in livestock sector of
agriculture.
The results of works will make it possible to estimate technical damage associated
with failures of electric motors, to define the extent of economical responsibility of
particular employed electric equipment, to make more specific schedules of preventive
maintenance and repair and to correct them, as well as to find room for improvement of
operational effectiveness of energy services at agricultural enterprises.
References
1. Gill, P.: Electrical Power Equipment Maintenance and Testing. CRC Press, Boca Raton
(2008)
2. Gracheva, E., Fedorov, V.: Forecasting reliability electrotechnical complexes of in-plant
electric power supply taking into account low-voltage electrical apparatuses. In: 2019
International Conference on Industrial Engineering, Applications and Manufacturing
(ICIEAM). IEEE, pp. 1–5 (2019)
3. Vasant, P., Zelinka, I., Weber, G.-W. (eds.): Intelligent computing and optimization. In:
Proceedings of the 2nd International Conference on Intelligent Computing and Optimization
2019 (ICO 2019). Springer (2019). ISBN 978–3–030–33585–4.
4. Vasant, P., Zelinka, I., Weber, G.-W. (eds.) Intelligent computing & optimization. In:
Conference proceedings ICO 2018. Springer, Cham (2018). ISBN 978-3-030-00978-6
5. Johannes, R., Schmidthaler, M., Schneider, F.: The value of supply security: the costs of
power outages to Austrian households, firms and the public sector. Energy Econ. 36, 256–
261 (2013)
6. Kharchenko, V., Vasant, P.: Friendly agricultural development. In: Handbook of Research
on Energy-Saving Technologies for Environmentally. IGI Global. https://doi.org/10.4018/
978-1-5225-9420-8
7. Kharchenko, V., Vasant, P.: Renewable Energy and Power Supply Challenges for Rural
Regions. IGI Global (2019). https://doi.org/10.4018/978-1-5225-9179-5
8. Syrykh, N.N., Ye Kabdin, N.: Theoretical Foundations of Operation of Electric Equipment.
Agrobiznestsentr Publ., Moscow (2007). 514 p. (in Russian)
9. Strebkov, D.S., Nekrasov, A.I., Nekrasov, A.A.: Maintenance of power equipment system
based on the methods of diagnosis and control of technical conditions. In: Handbook of
Research on Renewable Energy and Electric Resources for Sustainable Rural Development.
January 2018, Chapter 18, pages 535. pp. 421–448. https://doi.org/10.4018/978-1-5225-
3867-7
10. Nekrasov, A.I., Nekrasov, A.A.: Method for definition of required spare components set to
maintain electric equipment for rural applications. In: 2018 International Russian Automa-
tion Conference (RusAutoCon). IEEE (2018)
1212 A. Nekrasov et al.
11. Yu, S., Borisov, A., Nekrasov, A.N., Nekrasov, A.A.: Methodical recommendations on
forecasting and monitoring technical status of electric motors in agricultural production.
Moscow. GNU VIESH (2011). 108 p. (in Russian)
12. Yeroshenko, G.P., Bakirov, S.M.: Adaptation of operation of electric equipment to the
specifics of agricultural production. Saratov IC ‘Nauka’ (2011). p. 132. (in Russian)
13. Price of electric motors. https://electronpo.ru/price. Accessed 24 Jan 2020
14. LaVerne, S., Bodman, G., Schinstock, J.: Electrical wiring methods and materials for
agricultural buildings. Appl. Eng. Agric. 5(1), 2–9 (1989)
Development of a Prototype Dry Heat Sterilizer
for Pharmaceuticals Industry
1 Introduction
The importance of using sterilizer in healthcare sector is getting popularity due to its
capability of killing microorganisms. DHS is a necessary equipment for pharmaceu-
ticals as well as hospitals to facilitate sterilization process. Different type of cycles has
been used sterilization process to perform the function, such as drying, exhaust, pre-
sterilization (heating), stabilization, sterilization and cooling. Heat destroys bacterial
endotoxins (or pyroxenes) which are difficult to eliminate by other means, and this
property makes it applicable for sterilizing glass bottles and pharmaceuticals equip-
ment. In this paper, the design and implementation of a DHS using locally available
materials is presented. The performance analysis of the implemented DHS is conducted
and also presented.
Many researchers have suggested and prescribed many ways of sterilizer and
sterilization process. Pradeep and Rani [1] described the function of the sterilizer and
sterilization process. Their process is used for implementation of different type of
sterilizer. They also carried out detailed study on machine performance, to show the
evaluation process of the efficiency of micro-organism control. Kalkotwar et al. [2]
performed some research on the sterilization process for killing of microorganisms by
heat, which was a combination of time and temperature. Finally, they published papers
on methods and process of validation of sterilization process. Purohit and Gupta [3] has
analyzed the temperature mapping of DHS. Sultana [4] discussed about the sterilization
methods and principles, also have explain different type of sterilizer and sterilization
methods. She employs high temperatures in the range of 160–180 °C. Obayes [5] has
studied the construction and working principle of DHS. He applied different values of
time and temperature to perform the sterilization, such as 170 °C for 30 min, 160 °C
for 60 min, and 150 °C for 150 min. He also described the construction of the ster-
ilizer, he used conventional control system using switch and relay to perform the
functions of selecting electric heater for temperature rise and thermostat for measuring
temperature. Satarkar and Mankar [6] fabricated an autoclave sterilization machine for
reducing the microbial contamination of packaged products. They have described about
the moist heat with low range of temperature. Oyawale and Olaoye [7] designed and
fabricated a low cost conventional steam sterilizer. They performed the testing of their
design considering temperature of 120 °C for 12 min.
According to the above discussions, many researchers were designed and imple-
mented DHS and conducted the performance tests. In Bangladesh some switch- relay
controlled conventional machines are available that can be used only for oven purpose,
cannot use pharmaceuticals applications due to lack of accuracy, safety and operational
limitations. In Bangladesh, DHS are imported for industrial purpose which are large in
size and costly, and all sorts of control remained in the grip of them. Our engineers do
not have knowledge of controlling system; they only can operate. If there is any
problem the total system become useless.
In this project, a prototype DHS is designed and implemented using locally
available materials. For automatically control the operation of the DHS, PLC is used.
HMI is used for giving the instructions and for displaying all operation parameters.
The specific issues of this work can be summarized as follows:
a. To design and implement a PLC controlled DHS using locally available materials
for pharmaceuticals and experimental use.
b. To control the various parameters of the operation cycle and display the parameters
in the HMI display.
c. Analysis the performance of the implemented DHS and compare with the set val-
ues, and discuss the suitability of the implemented DHS for industrial use.
Figure 1 shows the operational block diagram of the proposed DHS comprising PLC,
HMI and other accessories. PLC is interfaced with the HMI for bidirectional control
and display.
Development of a Prototype Dry Heat Sterilizer for Pharmaceuticals Industry 1215
PLC will continuously monitor the inside temperature of the DHS. Door sensor and
emergency stop switch are used for safety purpose. Before starting operation, the door
of sterilizer should be closed, door sensor will ensure this. If door is not closed at the
beginning of operation cycle or any unwanted situation arises, buzzer will alarm. The
desired cycle parameters such as temperature, heating time, sterilizing time and cooling
time will be loaded in PLC through HMI. All parameters will be continuously moni-
tored by PLC and will be displayed on the HMI display. Heaters and blowers are used
for heating and cooling the chamber of DHS. Magnetic contractor is used to turn on/off
the blower and solid state relay is used to turn on/off the heater.
3 Implementation of DHS
The proposed DHS is implemented practically in the laboratory. Figure 2 shows the
control panel of the implemented DHS. Table 1 shows the list of materials used for
implementing the prototype DHS. Operational flowchart of the sterilization process in
shown in Fig. 3.
1216 Md. R. Ahmed et al.
5 Conclusion
High temperature for certain time is required to control or kill the microorganism.
A DHS is implemented in this project to perform the sterilization. Locally available
materials are used to implement the DHS, the cost of this machine is very low around
six hundred fifty US dollar.
Two setups are chosen to test the performance of the implemented DHS. It is seen
from the experimental results (Table 2 and 3) that the implemented DHS performed
welly as per setup operation cycle. Maximum deviation of temperature is found around
0.75 percent and time is around 0.31 percent. The results of completed cycles show that
the implemented DHS work perfectly. In this project, to construct the body of the DHS,
wooden material is used due to constructional limitations. Therefore, test cycles are run
considering comparatively low temperature. All parameters of the operation cycle of
DHS can be set through HMI, as a results the implemented DHS is user friendly. It is
expected that the DHS implemented in this project can be used for oven purpose in
pharmaceuticals and other industries and also can be used for experimental purpose.
This project will be helpful for producing low cost, customized design and user friendly
DHS locally, which can be used in industries.
Further research can be done by considering the high temperature resistance material
i.e. stainless steel (SS) for performing high temperature sterilization. Air velocity of
blowers did not calculate due to the use of small blowers and limitation of measurement
instrument. Rate of rise and fall of temperature did not measured in this project, further
research can be done in this direction.
References
1. Pradeep, D., Rani, L. : Sterilization protocols in dentistry – a review. J. Pharm. Sci. Res. 8(6),
558 (2016)
2. Kalkotwar, R.S., Ahire, T.K., Jadhav, P.B., Salve, M.B.: Path finder process validation of dry
heat sterilizer in parenteral manufacturing unit. Int. J. Pharm. Qual. Assur. 6(4), 100–108
(2015)
3. Purohit, I.K., Gupta, N.V.: temperature mapping of hot air oven (dry heat sterilizer). J. Pharm.
Res. 11(2), 120–123 (2017)
4. Sultana, D.Y.: Sterilization Methods and Principles, Faculty of Pharmacy, Jamia Hamdard,
Hamdard Nagar, New Delhi-110062, pp. 1–4, 11 July 2007
5. Obayes, S.AS.: Hot air oven for sterilization, definition and working principle. SSRN
Electron. J. (2018)
6. Satarkar, S.S., Mankar, A.R.: Fabrication and analysis of autoclave sterilization machine.
IOSR J. Eng. 05–08. ISSN (e): 2250–3021, ISSN (p): 2278–8719
7. Oyawale, F.A., Olaoye, A.E.: Design and construction of an autoclave. Pac. J. Sci. Technol. 8
(2), 224–230 (2007)
Optimization of Parameters of Pre-sowing Seed
Treatment in Magnetic Field
1 Introduction
2 Background
Many researchers have found a positive effect of the magnetic field on crop seeds,
which is manifested in improving seed sowing qualities [3], plant biometrics and yields
[4, 5], crop storage [6] and reducing plant morbidity [7], biochemical indicators [8] and
the quality of plant products [9].
However, all studies were performed at different values of magnetic induction and
treatment time (treatment dose), although it was found that crop yields and biometrics
depend on the dose of magnetic treatment. As a result, many different processing
modes are proposed, which sometimes differ significantly from each other. Since the
mechanisms of action of the magnetic field on the seeds did not have a clear expla-
nation, so not all the active factors in the magnetic treatment of seeds and their optimal
values were established.
Numerous studies suggest that seed treatment in a magnetic field may be an
alternative to chemical methods of pre-sowing treatment [10]. For successful intro-
duction of pre-sowing treatment of seeds in production it is necessary to establish mode
parameters of processing and their optimum values.
A common disadvantage of all existing methods of electromagnetic stimulation is
the lack of instrumental determination of the treatment dose. Its optimal value is usually
determined by yield, which largely depends on agro-climatic factors, soil fertility,
cultivation technology used and so on. Therefore, to determine the optimal modes of
pre-sowing seed treatment in a magnetic field, it is necessary to develop a method of
indicating its effect.
The purpose of the study is to establish the influence of the magnetic field on the
change of the biopotential of seeds of agricultural crops and to determine the optimal
parameters of magnetic seed treatment.
Various chemical and biochemical reactions take place in plant seeds, which are mainly
redox. Stimulation of seeds is associated with an increase in their speed, resulting in an
increase in the concentration of reaction products:
Under the action of a magnetic field changes the rate of chemical reactions [11]:
xm ¼ x exp mK 2 B2 þ 2KBvNa =2RT ; ð2Þ
where x – rate of chemical reaction without the action of a magnetic field, mol/(ls);
m – the mass of ions, kg; B – magnetic induction, T; v - velocity of ions, m/s; K –
coefficient that depends on the concentration and type of ions, as well as the number of
re-magnetizations, m/(sT); Na – Avogadro’s number, molecules/mol; R – universal gas
constant, J/(mol K); T – temperature, K.
To study biological objects, Albert Szent-Györgyi introduced the concept of
biopotential, which is associated with the redox potential (ORP) by a relation [8]:
RT
DORP ¼ 2; 3 ðlg C2 lg C1 Þ ð4Þ
zF
where z is the valence of the ion; F is the Faraday number, Cl/mol; C1 - concentration
of ions before magnetic treatment, mol/l; C2 - concentration of ions after magnetic
treatment, mol/l.
Taking into account (1)
RT
DORP ¼ ðlg x2 lg x1 Þ: ð5Þ
zF
Substituting in Eq. (4) the expression of the rate of chemical reaction (2), we
obtain:
mNa K KB2
DORP ¼ þ vB ð6Þ
zF 2
whence
mNa K KB2
DBP ¼ þ vB ð7Þ
zF 2
Experimental studies of the effect of the magnetic field on the seed biopotential
were performed with pea seed “Adagumsky”, beans seed “Hrybovsky”, rye seed
“Kharkivsky 98”, oats seed “Desnyansky”, barley seed “Solntsedar”, cucumber seed
“Skvyrsky”, sunflower seed “Luxe”.
The seeds were moved on a conveyor through a magnetic field created by a
multipolar magnetic system based on an induction linear machine (Fig. 1) [12].
а b
Fig. 1. Installation for pre-sowing treatment of seeds in a magnetic field: a - general view; b -
functional diagram: 1 – load device; 2 – conveyor; 3 – textolite inserts; 4 – permanent magnets; 5
– plate made of electrical steel; 6 – object of processing; 7 – container
Magnetic induction was adjusted by changing the distance between the magnets
and measured with a teslameter 43205/1. The velocity of movement of seeds through a
magnetic field was regulated by change of speed of rotation of the driving electric
motor of the conveyor by means of the frequency converter.
Seeds treated in a magnetic field were germinated and the ORP value was
measured.
A measuring electrode in the form of a platinum plate with a pointed end was
developed to measure ORP. The platinum electrode was inserted into the germinated
seed. A standard silver chloride electrode was used as an auxiliary electrode. Using the
ionomer И-160 M was determined by the ORP of germinated seeds.
The studies were performed using the experimental planning method. An orthog-
onal central-compositional plan was used for this purpose [13]. The values of the
upper, main and lower levels were taken for magnetic induction, respectively, 0, 0.065
and 0.13 T, for seed velocity - 0.4, 0.6 and 0.8 m/s, for response – the biopotential of
germinated seeds.
As a result of the conducted researches it is established that in the range of change
of magnetic induction from 0 to 0,065 T the seed biopotential increases, and at biger
values of magnetic induction the biopotential decreases (Fig. 2). At magnetic induc-
tion, which exceeds 0.13 T, the seed biopotential does not change, but exceeds its value
for the seed, untreated in a magnetic field.
1226 V. Kozyrsky et al.
Fig. 2. Dependence of change of biopotential of oat seeds on magnetic induction and speed of
movement of seeds
where a0, a1, a2, a12, a11 – coefficients, the values of which for different crops are
shown in Table 1.
The same optimal treatment regimes were established for magnetic treatment of
water and salt solutions [11]. This confirms the hypothesis of a significant role of water
in the mechanism of influence of the magnetic field on seeds.
The seed biopotential can determine not only the efficiency of pre-sowing treatment
in a magnetic field, but also the change in activation energy. The expression for the rate
of a chemical reaction in a magnetic field can be written as [14]:
Then
DE ¼ zFDBP ð13Þ
Thus, by formula (13) it is possible to determine the change in the activation energy
during seed treatment in a magnetic field according to experimentally determined
values of the change in biopotential.
The experimental dependences of the change in the activation energy on the
magnetic induction during seed treatment in a magnetic field are similar to the
dependences for the change in the seed biopotential during treatment in a magnetic
field. The activation energy changes the most at a magnetic induction of 0.065 T and a
velocity of 0.4 m/s. In this treatment mode, the activation energy changes by 2.4–
5.7 kJ/g-eq (Table 2).
Table 2. Change in activation energy during pre-sowing seed treatment in a magnetic field.
Agricultural culture Change in biopotential, mV Change in activation energy, kJ/g-eq
Rye 59 5,69
Barley 50 4,82
Sunflower 34 3.28
Oat 30 2,89
Pea 33 3,18
Bean 38 3,67
Cucumber 58 5,6
1228 V. Kozyrsky et al.
As follows from Table 2, the change in activation energy during seed treatment in a
magnetic field is less than the Van der Waals forces (10–20 kJ/mol), which characterize
the intermolecular interaction and the interaction between dipoles. Thus, the magnetic
field acts on the ions that are in the aqueous solution of the cell.
It is now established that the yield and biometric features of crops depend on the
dose of magnetic treatment, regardless of the method of creating a magnetic field [15].
Experimental studies of the change in biopotential during seed treatment in a
magnetic field made it possible to determine the energy dose of treatment.
The energy dose of treatment is determined by the formula [15]:
Z
W
D¼ dt ð14Þ
m
where W is the energy of the magnetic field, J; m – seed weight, kg; t – treatment time,
s;
or
Z
B2 dt
D¼ ð15Þ
2ll0 q
Fig. 3. Change of magnetic induction in the air gap along the conveyor belt
Optimization of Parameters of Pre-sowing Seed Treatment in Magnetic Field 1229
Using the dependence (Fig. 3), the integral (15) is determined by the method of
trapezoids:
ZL Zl=8 3L=8
Z 5L=8
Z
8Bm 2 8Bm 2 8Bm 2
B2 dl ¼ l dl þ 2Bm þ l dl þ 4Bm l dl
L L L
0 0 L=8 3L=8
7L=8
Z ZL
8Bm 2 8Bm 2 B2 L B2 L B2 L B2 L B2 L B2 L
þ 6Bm þ l dl þ 8Bm l dl ¼ m þ m þ m þ m þ m ¼ m ;
L L 24 12 12 12 24 3
5L=8 L=8
ð17Þ
where Bm is the maximum value of magnetic induction, which takes place in the plane
of installation of magnets, Tl, L – the path that passes the seed in a magnetic field, m
Then the energy dose of treatment
B2m L
D¼ ; ð18Þ
6ll0 qv
or
B2m ns
D¼ ; ð19Þ
6ll0 qv
Fig. 4. The dependence of the change in the biopotential of cucumber (1) and barley (2) seeds
on the energy dose of treatment in a magnetic field
4 Conclusion
The change in the seed biopotential during pre-sowing treatment in a magnetic field
depends on the square of the magnetic induction and the velocity of the seeds moving
in the magnetic field. By measuring the biopotential, it is possible to determine the
optimal treatment mode, which takes place at a magnetic induction of 0.065 T, fourfold
re-magnetization, pole division of 0.23 m and seed velocity in a magnetic field of
0.4 m/s.
The change in seed biopotential depends on the energy dose of treatment in a
magnetic field. The greatest seed biopotential during pre-sowing seed treatment in a
magnetic field was at an energy dose of treatment 1.7–3.8 J•s/kg.
The change in activation energy during seed treatment in a magnetic field is directly
proportional to the change in biopotential and at the optimal treatment mode is 2.4–
5.7 kJ/g-eq.
References
1. Vasilyev, A., Vasilyev, A., Dzhanibekov, A., Samarin, G., Normov, D.: Theoretical and
experimental research on pre-sowing seed treatment. In: IOP Conference Series: Materials
Science and Engineering, p. 791. 012078 (2020). https://doi.org/10.1088/1757-899x/791/1/
012078
2. Kutis, S.D., Kutis, T.L.: Elektromagnitnyye tekhnologii v rasteniyevodstve. 1. Elektromag-
nitnaya obrabotka semyan i posadochnogo materiala. [Electromagnetic technologies in crop
production. Part 1. Electromagnetic treatment of seeds and planting material], Moscow:
Ridero, p. 49 (2017)
3. Ülgen, C., Bİrinci Yildirim, A., Uçar Türker, A.: Effect of magnetic field treatments on seed
germination of melissa officinalis L. Int. J. Sec. Metabolite 4(3), 43–49 (2017)
Optimization of Parameters of Pre-sowing Seed Treatment in Magnetic Field 1231
4. Kataria, S., Baghel, L., Guruprasad, K.N.: Pre-treatment of seeds with static magnetic field
improves germination and early growth characteristics under salt stress in maize and
soybean. Biocatal. Agr. Biotechnol. 10, 83–90 (2017)
5. Maffei, M.E.: Magnetic field effects on plant growth, development, and evolution. Front.
Plant Sci. 5, 445 (2014)
6. Lysakov, A.A., Ivanov, R.V.: Vliyaniye magnitnogo polya na sokhrannost’ kartofelya
[Influence of the magnetic field on the preservation of potatoes]. Adv. Modern Natural Sci.
8, 103–106 (2014)
7. De Souza, A., Sueiro, L., Garcia, D., Porras, E.: Extremely low frequency non-uniform
magnetic fields improve tomato seed germination and early seedling growth. Seed Sci.
Technol. 38, 61–72 (2010)
8. Iqbal, M., ul Haq, Z., Jamil, Y., Nisar, J.: Pre-sowing seed magnetic field treatment influence
on germination, seedling growth and enzymatic activities of melon (Cucumis melo L.).
Biocatal. Agr. Biotechnol. 2016(6), 176–183 (2016)
9. Ramalingam, R.: Seed pretreatment with magnetic field alters the storage proteins and lipid
profiles in harvested soybean seeds. Physiol. Mol. Biol. Plants 24(2), 343–347 (2018)
10. Stange, B.C., Rowlans, R.E., Rapley, B.I., Podd, J.V.: ELF magnetic fields increase
aminoacid uptake into Vicia faba L. Roots and alter ion movement across the plasma
membrane. Bioelectromagnetics 23, 347–354 (2002)
11. Zablodskiy, M., Kozyrskyi, V., Zhyltsov, A., Savchenko, V., Sinyavsky, O., Spodoba, M.,
Klendiy, P., Klendiy, G. Electrochemical characteristics of the substrate based on animal
excrement during methanogenesis with the influence of a magnetic field. In: Proceedings of
the 40th International Conference on Electronics and Nanotechnology, ELNANO, pp. 530–
535 (2020)
12. Sinyavsky, O., Savchenko, V., Dudnyk, A.: Development and analysis methods of
transporter electric drive for electrotechnological complex of crop seed presowing by
electromagnetic field. In: 2019 IEEE 20th International Conference on Computational
Problems of Electrical Engineering (CPEE), pp. 1–6 (2019)
13. Adler, Yu.P., Markova, E.V., Granovskiy, Yu.V.: Planirovaniye eksperimenta pri poiske
optimal’nykh usloviy [Planning an experiment when searchingfor optimal conditions].
Moskow: Science, p. 278 (1976)
14. Kozyrskyi, V., Savchenko, V., Sinyavsky, O.: Presowing processing of seeds in magnetic
field. In: Handbook of Research on Renewable Energy and Electric Resources for
Sustainable Rural Development, IGI Global, USA, pp. 576–620 (2018)
15. Pietruszewski, S., Martínez, E.: Magnetic field as a method of improving the quality of
sowing material: a review. Int. Agrophys 29, 377–389 (2015)
Development of a Fast Response Combustion
Performance Monitoring, Prediction,
and Optimization Tool for Power Plants
1 Introduction
The coal-fired furnace is designed with the aim that there is sufficient amount of heat
transferred from the flames to the heat exchangers tubes [1]. This must be done to
ensure that the steam that entering turbine is at specific temperature value. At the same
time, the plant operators need to monitor the gas temperature to not exceeding the
design limit at the furnace exit. Higher gas temperature after the furnace zone will not
only damage the heat exchangers tubes but also lead to ash deposition problems and
environmental issues [2].
At the moment, it is a challenging task for plant operators to maintain plant
availability and efficiency due to wide variety of coal properties [3]. The properties of
coal that is received today will not be the same with the coal properties that will be
received in the next month. Hence, it will certainly results in different combustion
behavior. Furthermore, operators do not have sufficient information of combustion
inside the boiler where they can only evaluate the combustion behavior by looking at
certain parameters e.g. carbon monoxide (CO) emission and gas temperature [4].
Insufficient post combustion data and insights from critical furnace area pose a
challenging task for combustion performance monitoring [4]. The inability to predict
boiler performance will increase the operating cost due to higher coal consumption and
poor reliability that may jeopardize plant operation [5]. There has been an increasing
trend in power plant operational disruption across the globe and this has been con-
tributed by issues such as high furnace exit gas temperature (FEGT), increased emis-
sion level, and ash deposition [5].
In recent years, technologies for monitoring the quality of combustion in utility
boilers have been developed through online spatial measurements of oxygen and fuel
[4]. The measurements will be used as indications to balance the flow of air and fuel
and subsequently, the combustion performance. While experienced performance
engineers can relate these measurements to the combustion’s characteristics, and adjust
the boiler to improve combustion, the step in combustion tuning is still primarily a trial
and error procedure and even a highly experienced performance engineer could take a
number of steps to optimize the boiler performance.
Computational Fluid Dynamics (CFD) modelling has often been used for com-
bustion and flow field characterizations [6, 7]. One cannot deny the capability of CFD
modelling to facilitate combustion monitoring process by providing a visualization of
the combustion characteristics. Rousseau et al. (2020) [8] have developed CFD model
to predict the combustion behaviour of coals in coal-fired utility boilers. They also
demonstrated the viability of the computational approach as an effective tool for coal
burning optimization in full-scale utility boilers. General Electric (GE) Power Services
have also suggested an efficiency improvement for the coal-fired boiler guided by CFD
approaches via optimizing the flow characterization and combustion optimization [4].
Via CFD, the spatial visualization of combustibles in the flue gas, temperature, and also
the potential areas for slagging and fouling can be detected which could assist in the
combustion tuning process in order to achieve the required combustion quality.
Nonetheless, coal power plants, which are designed to operate within certain range
of coal properties, will require CFD modelling to be performed multiple times
depending on the properties of coal. To avoid the excessive calculation times, a CFD
feature namely Reduced Order Modelling (ROM) [9] can be used to enable continuous
monitoring through dynamic data simulation and this will allow multiple simulations to
be conducted at one time. Moreover, a system-level digital twin [4] can be generated
via integrating both CFD ROM and machine learning modules to assist not only in the
combustion monitoring, but also in predicting the upcoming combustion anomalies and
provide a certain guidance which will allow proactive troubleshooting to be carried out
to eliminate any potential issues. With a recent spike of popularity towards machine
learning, several publications have tested the capability of machine learning for opti-
mization purposes and the outcomes were found to be beneficial for the digital twin
perusal [10, 11].
The current paper discusses the general workflow of the proposed tool to provide a
guideline for optimizing combustion performance in thermal power plants, specifically
the coal-fired power plant, which is the location where the preliminary study was
carried out based on the plant’s historical data and boiler’s geometry.
1234 M. N. Rahman et al.
2 Methodology
2.1 Overview of the Proposed Application Architecture
The tool is based on the system-level digital twin which incorporate both CFD and
machine learning modules to provide a virtual model of combustion performance in a
boiler where it allows analysis of data and monitoring of combustion behavior to head
off any anomalies (out of range temperature and pollutants) before it occurs. Subse-
quently, operators could check and apply corrective action on the twin model platform
and monitor the predicted outputs prior to applying the corrective action on the actual
boiler. Refer to Fig. 1 below on the general application architecture of the tool.
Referring to Fig. 1, smart components in an actual boiler system which use sensors
to gather real-time status of the boiler operation will be integrated with the commercial
industrial connectivity platform in order to bridge the communication gap between the
Development of a Fast Response Combustion Performance Monitoring 1235
sensors in boiler system and the twin model platform. Once the real-time data has
passed through the platform, it will be sent to the twin model which act as the mon-
itoring platform to oversee the real-time parameters status, predicted outputs status, and
provide anomalies alert notifications. The twin model will also act as a bridge between
CFD and machine learning modules. The mentioned inputs are the list of parameters in
which the operators will have the flexibility to control in order to optimize the com-
bustion performance in boilers. The outputs are the outcome in which the operators
need to ensure that it is within the range of acceptable limit. There was a prior
engagement with the power plant where the study was held in order to get insights from
the plant personnel on the important combustion parameters which could massively
affect the plant’s reliability. Afterwards, the list of output parameters is determined
which includes the FEGT, flue gas concentration, and the rear pass temperature. These
three parameters are the common criteria to measure the combustion performance as
the negligence in monitoring and tuning these parameters will increase the potentiality
for unplanned outages and additional maintenance cost due to the formation of slagging
and fouling, along with the amount of emissions surpassing an acceptable limit.
The machine learning module will act as a “brain” where the machine learning
happened based on the incoming real-time data. Prior to the prediction stage, the raw
historical data (from January 2020 to August 2020) from the plant’s sensors was
cleaned and undergoes dimension reduction along with the parameter extraction. The
data was cleaned to remove any outliers since there are periods where the plant
experienced several outages. The raw historical data has a large number of parameters
which may not be affective since there are some parameters which might be redundant
and it will cause longer processing time. Moreover, there are several parameters that do
not contribute much on the specified output. From the raw historical data, there is a
total of 80 parameters. To determine the optimum number of parameters, the data was
trained using several values of parameter number including 70, 30, 25, 20, and 10.
Based on the values from the Root Mean Square Error (RMSE) and the Pearson
Correlation, the number of chosen parameters is 30 as the former and the later variables
have the lowest and the highest values as compared with the other tested parameter
numbers.
Once the validation of the analytics model is done, the output prediction was
executed where the twin model platform as mentioned before, will act as a bridge
between the predicted output from the analytics model and the CFD ROM database.
The CFD ROM database will received the real-time data from the plant’s sensors along
with the predicted outputs from the analytics model to visualize the real time com-
bustion behavior (5-min interval) and the predicted combustion behavior ahead of time.
From the prediction of combustion visualization in the CFD ROM interface, the
operators will have the flexibility to check and apply corrective action towards the
inputs of the boiler twin model prior to applying the corrective action on the actual
boiler.
of controlled inputs are reduced to 4 which include the flowrates of primary air
(PA) flowrate, secondary air (SA), over-fire air (OFA), and the fuel. The boiler
system under study is a 1000-MW boiler with an opposed wall-firing configuration, see
Fig. 2 (a).
The baseline case used the flowrates depicted in the pocket book for operator. The
subsequent FEGT from CFD simulation was compared with the design value of FEGT
from the pocket book, see Fig. 2 (b) on the visualization of FEGT.
[°C]
SA 1200
Fuel
1000
+ PA
800
600
(a) (b)
Fig. 2. A 1000 MW opposed-firing boiler; (a) Meshing model and (b) predicted FEGT.
Several assumptions were made on the operating conditions which were applied to
the CFD model to cater the efficiency and the accuracy of the model. The air and fuel
flow rates for the individual burners and OFA ports are assumed to be identical. The
swirl effect from the burners is also neglected. On the numerical sides, there are several
sub models implemented to characterize the coal combustion behavior, radiation,
gaseous reaction, and turbulence, all of which are utilized to dictate the source terms for
the mass, momentum, energy, and the species governing equations. The incoming coal
particles were tracked based on the Lagrangian scheme considering the turbulent
dispersion of particles [12]. The distribution of particle sizes was based on the Rosin-
Rammler distribution function where it was calculated based on the fineness test given
by the plant’s operators. For the devolatilisation of the coal, FG-DVC which is an
advanced coal network model, was used to get the volatiles composition along with the
respective rate constants [13]. The rate of devolatilisation was computed based on a
single Arrhenius equation model with an activation energy of 33.1 MJ/kmol and a pre-
exponential factor of 5799s−1 [12].
Development of a Fast Response Combustion Performance Monitoring 1237
Using too many features could degrade the prediction performance even when all
features are relevant and contain information about the response variable.
In the current scenario, a total of 80 parameters were collected and analyze by
observing the correlation matrix and the parameters were reduced to 30 relevant
parameters. Matlab Regression Learner software was used to train the machine learning
model. The software trains several regression models including the linear regression
models, regression trees, Gaussian process regression models, support vector machines,
and the ensembles of regression trees. The software can automatically train one or more
regression models, compare validation results, and choose the best model that works
for the regression problem.
[°C]
1200
1000
800
600
(a) (b)
Fig. 3. FEGT results from (a) the design and (b) the reduced over-fire air (OFA) conditions.
Development of a Fast Response Combustion Performance Monitoring 1239
The reduced OFA scenario as seen in Fig. 3 predicts the imbalance of temperature
distribution along with the excessive temperature at the furnace exit area. Higher gas
temperature after the furnace zone would not only affect the tubes of the heat
exchangers but also cause issues with ash deposition and environmental problems [2].
Figure 4 below shows the prediction and the actual data of FEGT from the his-
torical data test set. With the prediction ahead of the time-frame provided, plant
Temperature (°C)
Unplanned
outage
operators will have the capability to manage their work plan to avoid unwanted
conditions.
4 Conclusion
References
1. Speight, J.G.: Coal-Fired Power Generation Handbook, 1st edn. Scrivener Publishing LLC,
Massachusetts (2013)
2. Beckmann, A.M., Mancini, M., Weber, R., Seebold, S., Muller, M.: Measurements and CFD
modelling of a pulverized coal flame with emphasis on ash deposition. Fuel 167, 168–179
(2016)
3. Mat Zaid, M.Z.S., Wahid, M.A., Mailah, M., Mazlan, M.A.: Coal combustion analysis tool
in coal fired power plant for slagging and fouling guidelines. In: Editor, F., Editor, S. (eds.)
The 10th International Meeting of Advances in Thermofluids 2018, vol. 2062, AIP
Conference Proceedings (2019)
4. Zhou, W.: Coal fired boiler flow characterization, combustion optimization and Efficiency
improvement guided by computational fluid dynamics (CFD) modeling. Research Gate
(2017).
5. Achieving Better Coal Plant Efficiency and Emissions Control with Digital, GE (2017)
6. Laubscher, R., Rousseau, P.: Coupled simulation and validation of a utility-scale pulverized
coal-fired boiler radiant final-stage superheater. Thermal Sci. Eng. Progress 18, 100512
(2020)
7. Belosevic, S., Tomanovic, I., Crnomarkovic, N., Milicevic, A.: Full-scale CFD investigation
of gas-particle flow, interactions and combustion in tangentially fired pulverized coal
furnace. Energy 179, 1036–1053 (2019)
8. Rousseau, P., Laubscher, R.: Analysis of the impact of coal quality on the heat transfer
distribution in a high-ash pulverized coal boiler using co-simulation. Energy 198, 117343
(2020)
9. Rowan, S.L., Celik, I., Gutierrez, A.D., Vargas, J.E.: A reduced order model for the design
of oxy-coal combustion systems. J. Combustion 2015, 1–9 (2015)
10. Zhao, Y.: Optimization of thermal efficiency and unburned carbon in fly ash of coal-fired
utility boiler via grey wolf optimizer algorithm. IEEE Access 7, 114414–114425 (2019)
11. Sangram, B.S., Jagannath, L.M.: Modeling and optimizing boiler design using neural
network and firefly algorithm. J. Intell. Syst. 27, 393–412 (2018)
Development of a Fast Response Combustion Performance Monitoring 1241
12. Yang, J.-H., Kim, J.-E.A., Hong, J., Kim, M., Ryu, C., Kim, Y.J., Park, H.Y., Baek, S.H.:
Effects of detailed operating parameters on combustion in two 500-MWe coal-fired boilers of
an identical design. Fuel 144, 145–156 (2015)
13. Czajka, K.M., Modlinski, N., Kisiela-Czajka, A.M., Naidoo, R., Peta, S., Nyangwa, B.:
Volatile matter release from coal at different heating rates – experimental study and kinetic
modelling. J. Anal. Appl. Pyrol. 139, 282–290 (2019)
14. Yeoh, G.H., Yuen, K.K.: Computational Fluid Dynamics in Fire Engineering: Theory,
Modelling and Practice, 1st edn. Butterworth-Heinemann, USA (2009)
15. Joseph, D., Benedicte, C.: Discrete Ordinates and Monte Carlo Methods for Radiative
Transfer Simulation applied to CFD combustion modelling. Research Gate (2009)
16. Liu, H.: Encyclopedia of Machine Learning. Springer, Boston (2010)
Industry 4.0 Approaches for Supply Chains
Facing COVID-19: A Brief Literature Review
1 Introduction
The onset of COVID-19 in 2020 has detrimentally impacted countless lives and
industries across the globe. In terms of manufacturing and supply chain management,
global supply chains – networks of suppliers, manufacturers, retailers and distributors -
that had adapted lean manufacturing practices with streamlined inventories suddenly
found themselves crippled by shortages generated by lockdowns in China and in other
Southeast Asian suppliers. Shocks generated by the pandemic, such as overcompen-
sated ordering, lack of information sharing, and lack of collaboration between the
private and public sectors, have all forced various industries to seek alternative solu-
tions in order to survive.
Many researchers, however, have also identified these forced changes as oppor-
tunities for growth. While each industry experiences various difficulties across the
board due to inadequate visibility from upstream to downstream channels, many of the
available solution methods are also shared. Adaptation of prediction models used in
conjunction with automation and modelling techniques allow many manufacturing
factories to continue operations, and flexible supplier selection paired with collabora-
tive geographical tracking systems would have allowed manufacturers to predict dis-
ruptions and adjust their planning areas around impacted areas. In a few cases, some of
these digital solutions were preemptively prepared, which enabled supply chains in
these countries to continue operations. Such scenarios have led to a future outlook of
“resilient” supply chains, in which sourcing and manufacturing capacities are auto-
matically prepared for ahead of time across a variety of international channels, pro-
ceeding regardless of the situation at hand. It has also been proposed that most AI-
driven technologies and digital platforms used to solve issues created by the pandemic
would also strengthen the resiliency of their manufacturing networks against similar,
future disruptions.
Javaid et al. [1] provides some insight into certain technologies that present opportu-
nities for supply chains seeking to adopt Industry 4.0 practices: cloud computing
services such as Amazon AWS, Azure, and Google Cloud reduce operating costs and
increase efficiency. This is apparent through the close examination of supply chain
failures that gave rise to a new class of manufacturing technologies using self-learning
AI, which researchers hope will eventually enable smart supply chains to operate
autonomously. The authors maintain that Big Data methods can forecast the extent of
COVID-19, while AI can be used to predict and manage equipment manufacturing.
Ivanov [2] demonstrated that performing simulations to assess the supply chain effects
of COVID-19 could be used to demonstrate the effects of the pandemic across multiple
supply chain levels. Furthermore, while tracing diseases across geographic locales was
impossible before 2010, researchers such as Dong et al. [3] at John Hopkins university
have harnessed Global Imaging Systems (GIS) alongside data mining and machine
learning techniques to geographically trace the spread of the COVID-19 pandemic,
allowing for preemptive preparation. Furthermore, as detection techniques in other
systems were used to isolate the spread of the disease in localized areas, the spread of
the pandemic was tracked using live, real-time methods, with immediate and automated
alerts sent to key professionals. Wuest et al. [4] suggest further that the pandemic
provides strong justification for industries to adapt an “AI-Inspired Digital Transfor-
mation,” spearheaded by unmanned smart factories, automated supply chains and
logistics systems, AI-based forecasts for demand variation, shortages and bottlenecks,
and predicted maintenance routines to lessen the severity of internal disruptions.
Ivanov [2] further observed that supply chains are characterized by different challenges
corresponding to their respective objectives, noting that supply chains affected by
COVID-19 can be further categorized as humanitarian supply chains and commercial
supply chains. Altay et al. [5] found that post-disaster humanitarian logistics differed
greatly from pre-disaster logistics, or even commercial logistics.
1244 S. Reong et al.
Automotive Manufacturing
Pharmaceuticals Manufacturing
Aircraft Manufacturing
Defense Manufacturing
Fig. 1. Manufacturing and supply networks affected by COVID-19 at different stages, adapted
from Wuest et al. [4]
4 Solution Characteristics
Changes to manufacturing and supply chains in 2020 have been compared to a “global
reset,” according to the most recent meeting of the World Economic Forum’s Global
Lighthouse Network [6]. Out of the 54 leading advanced manufacturers in the world, a
performance survey found the relevant manufacturing changes to be (1) Agility and
customer centricity, (2) Supply chain resilience, (3) Speed and productivity, and
(4) Eco-efficiency. In terms of relevance to the COVID-19 pandemic, the former three
shifts are shown below (Table 1):
Table 1. Three necessary industry shifts: Adapted from the WEF 2020 Agenda.
Global changes Necessary industry shifts
Demand uncertainty and disruptions Agility; customer
centricity
National security interests, trade barriers and logistics disruption Supply chain resilience
Disruption of global manufacturing
Forced transition to remote management and digital collaboration Speed and productivity
Physical distancing regulations
Workforce displacement and unbalanced growth
Economic recession: costs must be reduced
1246 S. Reong et al.
Table 2. “Summary Strategies to Optimize the Supply of PPE during Shortages,” selected for
possible model objectives; adapted from the CDC guide on July 2020.
PPE type Conventional capacity Contingency capacity
All PPE • Use telemedicine whenever • Selectively cancel elective and non-
possible urgent procedures and
• Limit number of patients going to appointments where PPE is
hospital/outpatient settings typically used
• Limit face-to-face health care
professional encounters with
patients
N95 • Implement just-in time fit testing • Facilities communicate with local
Respirators • Extend the use of N95 respirators by healthcare coalitions and public
and wearing the same N95 for repeated health partners to identify
facemasks close-contact encounters with additional supplies
several patients (within reasonable • Track facemasks in a secure and
limits) monitored site and provide
• Restrict face mask usage to health facemasks to symptomatic patients
care professionals, rather than upon check-in at entry points
asymptomatic patients (who might
use cloth coverings) for source
control
Furthermore, Park et al. maintain that the main bottlenecks in the PPE supply chain
include raw materials shortages such as polypropylene, lack of production infrastruc-
ture, export bans, and transport constraints caused by quarantine measures or limited
workforce capacity. Thus, research on alternate sourcing solutions are a matter of
concern for all related industries. A segment of research in 2020, such as that of
Campos, et al. [23] suggests growing interest in the reuse of PPE masks and respirators.
In such a scenario, the local scale and collection methodologies of reverse logistics
activities can also be explored.
6 Conclusion
Due to its point of origin, scale, and difficulty for treatment, COVID-19 has created
some of the most significant and widespread global supply chain disruptions in modern
times. Supply chain managers and IT developers seeking to weather the effects of the
pandemic must first distinguish the different challenges faced by their humanitarian or
commercial enterprise, and then assess whether they are in a pre-disruption or dis-
ruption state.
If the supply chain is in a pre-disruption state, use of modeling systems and pre-
diction technologies will allow preparation for alternative material flow routes, which
increases supply chain resilience. Whether the supply chain will be able to adapt such a
disruption, however, will depend not only on the level of technological expertise
available, but also on information sharing attitudes in the industry and on the domestic
level of public-private cooperation.
1250 S. Reong et al.
Lastly, researchers seeking to model the COVID-19 vaccine supply chain are
suggested to investigate the cold chain, shelf life, and dosage requirements of the
leading candidates, which will determine relevance of traditional warehousing network
models or more innovative last-mile delivery solutions. PPE supply chain modelers
may find direction in alternate procurement, or in reverse logistics models that facilitate
the sterilization and reuse of protective equipment.
References
1. Javaid, M., Haleem, A., Vaishya, R., Bahl, S., Suman, R., Vaish, A.: Industry 4.0
technologies and their applications in fighting COVID-19 pandemic. Diabetes & Metabolic
Syndrome: Clinical Research & Reviews (2020)
2. Ivanov, D.: Predicting the impacts of epidemic outbreaks on global supply chains: a
simulation-based analysis on the coronavirus outbreak (COVID-19/SARS-CoV-2) case.
Transp. Res. Part E Logist. Transp. Rev. 136, 101922 (2020)
3. Dong, E., Du, H., Gardner, L.: An interactive web-based dashboard to track COVID-19 in
real time. Lancet. Infect. Dis. 20(5), 533–534 (2020)
4. Wuest, T., Kusiak, A., Dai, T., Tayur, S.R.: Impact of COVID-19 on Manufacturing and
Supply Networks—The Case for AI-Inspired Digital Transformation. Available at SSRN
3593540 (2020)
5. Altay, N., Gunasekaran, A., Dubey, R., Childe, S.J.: Agility and resilience as antecedents of
supply chain performance under moderating effects of organizational culture within the
humanitarian setting: a dynamic capability view. Prod. Plann. Control 29(14), 1158–1174
(2018)
6. Betti, F., De Boer, E.: Global Lighthouse Network: Four Durable Shifts for a Great Reset in
Manufacturing [Pdf]. World Economic Forum, Cologny (2020)
7. Swafford, P.M., Ghosh, S., Murthy, N.: The antecedents of supply chain agility of a firm:
scale development and model testing. J. Oper. Manage. 24(2), 170–188 (2006)
8. Sheffi, Y., Rice, J.B., Jr.: A supply chain view of the resilient enterprise. MIT Sloan Manage.
Rev. 47(1), 41 (2005)
9. Altay, N., et al.: Agility and resilience as antecedents of supply chain performance under
moderating effects of organizational culture within the humanitarian setting: a dynamic
capability view. Prod. Plann. Control 29(14), 1158–1174 (2018)
10. Ivanov, D., Dolgui, A.: A digital supply chain twin for managing the disruption risks and
resilience in the era of Industry 4.0. Production Planning & Control, pp. 1–14 (2020)
11. Lin, C., Braund, W.E., Auerbach, J., Chou, J.H., Teng, J.H., Tu, P., Mullen, J.: Early
Release-Policy Decisions and Use of Information Technology to Fight 2019 Novel
Coronavirus Disease, Taiwan (2020)
12. Dai, T., Zaman, M.H., Padula, W.V., Davidson, P.M.: Supply chain failures amid Covid-19
signal a new pillar for global health preparedness (2020)
13. Singh, G., Levi, J.: MiPasa project and IBM Blockchain team on open data platform to
support Covid-19 response, March 2020. https://www.ibm.com/blogs/blockchain/2020/03/
mipasa-project-and-ibm-blockchain-team-on-open-data-platform-to-support-covid-19-
response/Accessed Sept 2020
14. Intelligent Computing & Optimization, Conference proceedings ICO 2018, Springer, Cham,
ISBN 978–3–030–00978–6
Industry 4.0 Approaches for Supply Chains Facing COVID-19 1251
1 Introduction
Nowadays the synthesis of an efficient control system (CS) for a technological object
(TO) is still a creative process which is completely dependent on the personal pref-
erences of the CS designer. In the first place, it can be explained by the determinant part
of the designer’s subject domain initial knowledge and the empirical knowledge
obtained on its basis. Although the stages of developing an efficient control system for
a technological object have been formalized a long time ago [1, 2], they need to be re-
thought, given the current diversity of methods and approaches. In addition, techno-
logical objects operating in uncertain environment require the use of a generalized
methodology based on the life cycle (LC) of the control system for a technological
Objective set:
Object set Process set Design and monitoring
of the sub- of the sub- of the CS, compatible
ject domain ject domain objectives
Ontological system
Fig. 1. Components of the ontology of the subject domain of design and monitoring of the
efficient control systems operating in uncertain environment.
1254 N. Lutskaya et al.
Ontology forms the framework of the knowledge base in describing the basic
concepts of the subject domain and serves as the basis for the development of intel-
lectual decision support systems. Today, ontology can use different models of
knowledge presentation, for instance semantic or frame networks. Recently, a
descriptive logic sub-set and the OWL2 language dialects [10–12] become a popular
formal description for the development of the subject domain ontology. The means of
formal description and ontology development allow us to store, edit, verify, transmit
and integrate the developed ontology in different formats.
Technological regulations
(as the minimum and maximum values within
which there is the point of extremum)
Control system
Nonlinearity, uncer-
Technological Control
tainty and evolution of
object device
the object
C o n s e q u e n c e s
Decrease in
the quality of Readjustment
Loss of
the product of the control
rigidity
system
Increase in energy
consumption
Fig. 2. The problem domain of design and monitoring of the efficient control systems operating
in uncertain environment.
OnS ¼ \ OSD O; OP ; OT [ ð1Þ
The subject domain SD consists of two parts – the ontology of the objects O and the
ontology of the processes OP. The former defines the static terms, definitions and
relationships of the subject domain, while the second defines the dynamics (events and
duration) of the SD. The ontology of the processes OP can be constructed in accordance
with the operation of the objects of the subject domain, or it is possible to be done in
accordance with the objectives of the problem domain. The research work proposes to
develop the ontology of processes in accordance with the life cycle (LC) of the control
system (CS) for TO and the objectives of the problem domain OT.
Let us consider the CS for a TO in terms of its LC. Like any complex system, the
CS for a TO consists of at least three stages: origination, operation, and expiry. The
origination is connected with the CS design process and is included into the design and
operation LC of the industrial control system (ICS) of the technological process
(TP) and its parts, including automation equipment (AE). On the other hand, the LC of
the CS is associated with the operation of the TO, which also develops on the principle
of evolution. The incompleteness of the LC prevents from the optimality of the deci-
sion on designing the efficient CS for a TO, that is why when designing the efficient
CSs for TOs it is necessary to take it into account.
Let us describe the LC of the CS for a TO with the following tuple:
where P(LCCS) stands for the aim, requirement or assignment of the CS; {S} is the set
of stages of the life cycle of the CS; R is the result of the operation of the CS; T is the
life cycle time. Such dependence reflects the orientation of the CS for a TO both
towards the aim (assignment) of the system and towards the end result of its operation.
Thus, if the system loses its assignment or stops meeting the requirements, it goes to
the final stage of the life cycle, its expiry.
The LC of the efficient CS for a TO is divided into the following stages (Fig. 3).
The aim and criteria of control are selected and the limitations for the system are
determined at the stage of determining the requirements for the system. The input
documents are the technological regulations and the technical design specifications.
The idea of creating a system is substantiated and the input and output variables of
the CS are selected at the stage of formulating the concept of the CS. The end result of
this stage manifests in recommendations for the creation of the CS for the separate parts
of the TO with specified characteristics of each system, the sources and resource
limitation for its development and operation.
At the third and the subsequent stages of the LC, the issue of designing the CS for a
TO is solved using a number of the following traditional activities: the study of the TO;
identification of a mathematical model of the TO; design of the CS and the CS
modeling. However, the research work proposes to choose the CS from a variety of
alternative solutions, which includes different structures of the CS. The final choice is
the decision of a designer. In addition, the design process should be model-oriented,
1256 N. Lutskaya et al.
where the system model is used to design, simulate, verify and validate the programed
code (similar to the DO-178B standard for embedded systems).
Requirements Elimination of
from the upper Restriction the CS
levels for the АE
Unlike the previous stage, where the verification was only performed on the ready
built process models, the assumptions, calculations and conclusions, which have been
made at the previous stages are verified at the implementation stage. That is, the
reliability of the actual costs of the selected alternative solution is being assessed.
At the stage of operation and monitoring, the implemented CS for a TO is subjected
to final evaluation of theoretical and manufacturing research. The peculiarity of this
stage lies in monitoring the system and detecting its “aging”. The “aging” of the CS is
caused by lower efficiency of the CS for a TO and can eventually lead to a system
failure, the divergence of the time response being one of the manifestations of this
phenomenon for control systems. The evolution of the TO or its environment might be
the reason for the decrease in efficiency of such systems. A separate issue here is the
failure of individual elements of the AE [13].
Elimination of the CS is directly related to the termination of the TO operation
when the technological process is physically or morally outdated and its restoration is
futile for technical and economic reasons. The return to the previous stages of the LC is
predicated by the increase in flexibility and adaptability of the created system.
Thus, for the efficient design of the CS for a TO the ontological system is to be
considered in terms of ontological objectives which reflect the LC of the CS.
The proposed ontological system considers conducting ontological studies in order
to identify all the factors which affect the structure of the control system for the TO
operating in uncertain environment. The model also takes into account the LC of the
CS, as well as the peculiarities of developing the robust CSs for TOs, alternative
solutions for which will form the basis of the CS structures.
Ontological Aspects of Developing Robust Control Systems 1257
O ¼ hX; R; Fi ð3Þ
where X = Xi i = 1,n is a finite non-empty set of concepts (subject domain con-
cepts); R = ri i = 1,m is a finite set of semantically significant relationships between
concepts; F:X R is a finite set of interpretation functions which have been preas-
signed on concepts and relationships.
The basic concepts of the subject domain X have been determined as follows: a
technological object, standards, a life cycle, an individual, as well as a controlling part
of the ICS which consists of software and hardware facilities. These subject domain
concepts are substantiated by the direct impact on the subject domain and research
objectives.
According to semantic relations, significant relationships between concepts for the
set R, have been shown on the O (Fig. 4). The interpretation functions on the
ontologies have been shown by means of the corresponding arrows. In addition, the
model is divided into reference levels for better structuring of the categories and for
linking ontologies. The relationship between other schemes in a single ontological
system is carried out with the help of ovals (links) and rectangles (acceptance) with
numbers corresponding to the following pattern:
L0 Automation
L2 2 4 4 7 3
8
Metrology 1 Type of
National Designer Operator Software 3 8
IT standards process
standarts
standarts
of ICS 0
4 5
The nature
Systems Food CL P1.L5.07
of the
engineering quality design Hardware
7 3 functioning
standards standards 0 ICS
L3 2
5
7 5
2 8 4
0 LC TO 2
4
0
3
GOST 4
CRISP-
ISO 9001
34.601-90 DM Information Control
7 Field AE
support AE
8 5 Mathematical
GOST support
P1.L5.01
34.201-89 CL AE
ISO/IEC 5 Custom
7 5
15288:2005 software
9 6 Terms of
DSTU Technological reference 5
5
L4 2503-94... regulation 5 8
0 0 1
0 is a categorical relationship; 7
1 is an integer part; Logical
2 is a kind-of-relation; Requirements for part
Functional
3 is a class – subclass (set – sub-set); mathematical
requirements Regulatory
4 is a set – element; support part
5 is an attribute relationship;
6 is an equality relationship; T1(P1).L5.05
7 is an initialization relationship ; P1.L5.02 P1.L5.03 T1.L6.04
L5 2 5
8 is the relationship of the process behavior;
9 is the end of the process relationship.
The structure Parameters
L6 P1.L5.06
of the reg.part of the reg.part
processes at the lower L5 level correspond to the sequence of activities and operations
which contribute to the solution of the corresponding sub-tasks of the ontological
research.
Ontology of objectives
L0
0 0
0
1 0 5
Task 1.2.2.1 1 0
Preliminary
Synthesis 3
analysis of TO
methods of
9 Structural
5 Task 1.2.2.2 local regulators Synthesis
Identification of methods of
MM TO Task 1.2.2.3
Synthesis of robust control Mixed
P1.L5.11 5
СS TO Parametric
L5 3 3
9
The task of synthesizing the SC is divided into three subtasks (Fig. 5 - Task 1.2.2.1–
1.2.2.3): preliminary analysis of the TO, identification of the mathematical model of
TO and the synthesis of the control system. Each component of the task has corre-
sponding processes (Fig. 6) that must be performed upon its achievement. For exam-
ple, when using a basic TO model, it is necessary to perform six actions that will lead to
the receipt of MM TO. In contrast to obtaining MM TO, not considered in conditions
of uncertainty, action 6 was introduced - determining the region of uncertainty, which
1260 N. Lutskaya et al.
can be performed by the procedures described in [14]. The features of the proposed
approach to the synthesis of the regulatory part of the CS is testing of alternative
structures of the CS and the choice of an effective one based on the theory of decision
making.
Using the developed ontological model, it is possible to build a decision support
system based on the acquired knowledge of the problem area. This approach to the
development of the CS reduces the design time for the calculation of structured and
automated procedures that are embedded in the ontological model. When synthesizing
the CS, modern methods of robust TO control systems are taken into account, which
are formalized in the ontology of tasks and the main interrelationships with traditional
methods are indicated.
Stage 3. Development
О.L4.01 Action 5.
Action 2. Action 3. О.L3.07
Definition of regulated Evaluation Sort alternatives of
of CS
О.L4.01 (verification) MM
variables, control actions, О.L3.07 CS by criterions
basic disturbances T1.L5.13
Yes
Action 4.
Action 6. Determination of CS
T1.L5.12 Defining the area parameters and
T1.L5.05
of uncertainty MM modeling
Action 5.
Selection and
implementation of CS
No
Action 6.
CS verification О.L4.01
Yes
О.L5.05
L5
4 Conclusion
The analysis of the subject and problem domain of the synthesis of the CS for a TO in
uncertain environment has been carried out. On its basis, a system-ontological
approach to the efficient design and monitoring of the CS for a TO operating in
uncertain environment has been proposed. The ontological system consists of three
ontologies: objects, objectives, and processes. The conducted research revealed the
relationships between subject domain objects, objectives, methods for solving them and
the ontology of processes. The advantage of the given approach is the scientific novelty
of the knowledge presentation model, as well as practical importance for both the
Ontological Aspects of Developing Robust Control Systems 1261
References
1. McMillan, G.K., Considine, D.M.: Process/Industrial Instruments and Controls Handbook,
5th edn. McGraw-Hill Professional, New York (1999)
2. The Control Handbook: Control System Applications, 2nd edn. CRC Press, W.S. Levine
(2011)
3. Lutskaya, N., Zaiets, N., Vlasenko, L., Shtepa, V.: Effective robust optimal control system
for a lamellar pasteurization-cooling unit under the conditions of intense external
perturbations. Ukrainian Food J. 7(3), 511–521 (2018)
4. Korobiichuk, I., Lutskaya, N., Ladanyuk, A., et al.: Synthesis of optimal robust regulator for
food processing facilities, automation 2017: innovations in automation. Robot. Measurement
Techniques, Advances in Intelligent Systems and Computing, Springer International
Publishing 550, 58–66 (2017)
5. Takahara Y., Mesarovic M.: Organization Structure: Cybernetic Systems Foundation,
Springer Science & Business Media (2012).
6. Fernandez-Lopez, M., Gomez-Perez, A.: Overview and analysis of methodologies for
building ontologies. Knowl. Eng. Rev. 17(02), 129–156 (2003)
7. Baader, F., Calvanese, D., McGuinness, D.L., et al.: The Description Logic Handbook:
Theory. Implementation, Applications, Cambridge (2003)
8. Palagin, A., Petrenko, N.: System-ontological analysis of the subject area. Control Syst.
Mach. 4, 3–14 (2009)
9. Smith, B.: Blackwell guide to the philosophy of computing and information: Chapter
ontology. Blackwell 39, 61–64 (2003)
10. OWL 2 Web Ontology Language Document Overview, 2nd edn., W3C. 11 December 2012.
11. OWL Web Ontology Language Guide. W3C Recommendation, 10 February 2004. https://
www.w3.org/TR/owl-guide/
12. Protege Homepage. https://protege.stanford.edu/
13. Zaiets N., Vlasenko L., Lutska N., Usenko S.: System Modeling for Construction of the
Diagnostic Subsystem of the Integrated Automated Control System for the Technological
Complex of Food Industries/ICMRE 2019, Rome, Italy, pp. 93–98 (2019).
14. Lutska, N.M., Ladanyuk, A.P., Savchenko, T.V.: Identification of the mathematical models
of the technological objects for robust control systems. Radio Electron. Comput. Sci. Control
3, 163–172 (2019)
15. Voropai, N.I.: Multi-criteria decision making problems in hierarchical technology of electric
power system expansion planning. In: Intelligent Computing & Optimization. ICO 2018.
Advances in Intelligent Systems and Computing, vol. 866, pp. 362–368. Springer (2019)
16. Alhendawi, K.M., Al-Janabi, A.A., Badwan, J.: Predicting the quality of MIS characteristics
and end-users’ perceptions using artificial intelligence tools: expert systems and neural
network. In: Intelligent Computing and Optimization. ICO 2019. Advances in Intelligent
Systems and Computing, vol. 1072. pp. 18–30. Springer (2020)
A New Initial Basis for Solving the Blending
Problem Without Using Artificial Variables
1 Introduction
Over the past decade, the blending problem is one of the well-known optimization
problems related to the production process of a large number of raw materials to get
many types of products. It was first mentioned in 1952 by Chanes et al. [1], which
proposed a linear programming problem to find mix of fuels and chemicals in the
airline business. There is a lot of research involving the blending problem such as the
blending of tea, the blend of milk and coal (see more details in [2–4]).
Blending problem can be formulated as the linear programming model, and it can
be solved by the simplex method. To use this method, the transformation of the
canonical form to the standard form is performed. For converting it to the standard
form, slack and surplus variables are added. Then, the basic variable set must be chosen
and considered the possibility both of its inverse matrix and its feasibility. For solving a
large problem, it’s hard to choose the basic matrix that gives the feasible solution.
Thus, a main topic of research to improve the simplex method is to propose the method
to choose the initial basis or initial basic feasible solution.
The well-known methods that are used to find a basic feasible solution are Two-
Phase and Big-M method. These methods start by adding artificial variables to choose
the initial basis. However, adding the artificial variables to the problem not only makes
the problem size become larger but also increases the number of iterations. Conse-
quently, one of topics of the research to find the basic feasible solution without adding
the artificial variables called artificial-free technique is widely investigated.
In 1997, Arsham [5] proposed the method that has two phases which the first phase
starts with the empty set of basic variables. Then, each basic variable is chosen one by
one for entering the basic variable set until its set is full. The second phase is the
original simplex method to find the optimal solution which starts from the basic fea-
sible solution from the first phase. In 2015, Gao [6] gave the counterexample that
Arsham’s algorithm reports the feasible solution for the infeasible problem. Nowadays,
various artificial-free techniques are proposed [7–10].
In this research, we focus on solving blending problem by the simplex method.
However, since an origin point is not feasible solution, artificial variables are required.
This leads to the size of problem is expanded. Therefore, to avoid the step for finding a
basic feasible solution by adding artificial variables, we propose the technique for
choosing an initial basis without using artificial variables. First, the original form of the
blending problem is rewritten and divided into master problem and sub-problems.
Then, all sub-problems are considered to find a basic feasible solution. The mainly
topic of this research is not only suggesting the dividing of the original problem but
also propose the algorithm to choose the initial basis for each sub-problem which
avoids to add the artificial variables.
This paper is organized as follows: Sect. 2 is the briefly review of the blending
problem. Section 3 describes about detail of the proposed method. Section 4 gives a
numerical example showing the usage of the proposed method. And final section ends
with the conclusion.
2 Blending Problem
P P P P
max pj xij ci xij ð 1Þ
j2N i2M i2M j2N
P
s.t. xij si 8i 2 M ð 2Þ
j2N
P
xij aij xkj 8i 2 M; 8j 2 N ð 3Þ
k2M
P
xij aij xkj 8i 2 M; 8j 2 N ð 4Þ
k2M
P
xij ¼ dj 8j 2 N ð 5Þ
i2M
xij 0 8i 2 M; 8j 2 N ð 6Þ
In the above model, the objective function (1) aims to find the maximum profit.
Constraint (2) means that the amount of each raw material must not exceed the
available source.
h Constraints
i (3)–(4) aim to force the amount of materials must be on
the interval aij ; aij . Constraint (5) aims to force the amount of each raw material must
be enough to encounter the demand. Finally, constraint (6) indicates the domain of the
decision variables.
Before presenting the process of the proposed method, we rewrite the general form of
the blending problem (1)–(6) to the following form:
P P P P
max pj xij ci xij ð 7Þ
j2N i2M i2M j2N
P
s.t. xij si 8i 2 M ð 8Þ
j2N
P
aij xkj xij 0 8i 2 M; 8j 2 N ð 9Þ
k2M
P
xij aij xkj 0 8i 2 M; 8j 2 N ð10Þ
k2M
P
xij ¼ dj 8j 2 N ð11Þ
i2M
xij 0 8i 2 M; 8j 2 N: ð12Þ
In this paper, we reconstitute the above general model into the following model:
A New Initial Basis for Solving the Blending Problem 1265
P
n
max xj
cTj ^ ð13Þ
j¼1
s.t. AS x s ð14Þ
j
Aj Im ^x 0 8j ð15Þ
j
Im Aj ^x 0 8j ð16Þ
1m ^
x j
¼ dj 8j ð17Þ
xj
^ 0 8j; ð18Þ
max cT x
s.t. AS x s ð19Þ
x 0
max xj
cTj ^ ð20Þ
j
s.t. A j Im ^
x 0 ð21Þ
j
Im Aj ^x 0 ð22Þ
1m ^
x j
¼ dj ð23Þ
xj
^ 0: ð24Þ
To choose the initial basis for the blending problem, we will consider it from all
sub-problems. First, we transform Sub-problem j to a standard form and so 2m of slack
variables are added to this sub-problem. Let A j be the coefficient matrix of constraints
_ j
(21)–(22), A and bj are the coefficient matrix and the column vector of RHS of
constrains (21)–(23), respectively. So
1266 C. Boonmalert et al.
For choosing basic variables, all slack variables are chosen first. However, it is not
enough to construct the set of basic variables. Hence, to avoid adding an artificial
variable, one of decision variable is selected. We now present the new technique for
choosing this remaining basic variable. First, for each j 2 N, we let.
n o
lj ¼ arg max cTj ð25Þ
and . Since ,
Thus, the dual problem of Sub-problem j is feasible. The initial tableau can be
constructed as Table 1.
1 1
Since Blj bj l0 and cTBl Blj N l cTN l 0, the dual simplex can start without
j
using artificial variables for solving Sub-problem j. After the optimal solution to all
sub-problems is found, if it satisfies the master problem then it is the optimal solution to
the original blending problem.
A New Initial Basis for Solving the Blending Problem 1267
Table 1. Initial tableau of simplex method for sub-problem j of the blending problem.
z xBlj xN l RHS
T 1 1
z 1 0
cTBl Blj N l cTN l cTBl Blj bj
j j
1 1
xBlj 0 I2m þ 1
Blj Nl Blj bj
4 An Illustrative Example
In this section, we give an illustrative example showing the numerical results obtained
from our algorithm.
Example 1. Consider the following blending problem:
In Sub-problem 1, we let
A New Initial Basis for Solving the Blending Problem 1269
since l ¼ arg max cT1 ¼ 2. Thus, the basis can be constructed associated with
xB ¼ ½ x31 s1 s2 s3 s4 s5 s6 T as follows:
After the initial tableau is constructed, the dual simplex method is used for solving
each sub-problem. Thus, the optimal solutions to Sub-problem 1 and Sub-problem 2
are found at ^x1 ¼ ð42:7375; 183:9825; 98:28Þ and ^ x2 ¼ ð47:273; 237:103; 125:624Þ.
After checking this solution with the master problem, it satisfies the master problem.
Therefore, it is the optimal solution to the original problem. The number of iterations of
1270 C. Boonmalert et al.
each sub-problem is only two, while the simplex method uses 12 iterations. Moreover,
15 slack variables and 2 artificial variables are added before the simplex method starts.
The comparison of the simplex method and the proposed method of this example is
shown in Table 4.
Table 4. Comparison between the simplex method and the proposed method.
Method Simplex method Proposed method
Phase I Phase II Sub-problem 1 Sub-problem 2
Size of problem 17 23 17 21 7 10 7 10
Number of iterations 9 3 2 2
5 Conclusion
In this paper, the original blending problem is rewritten and divided into master
problem and sub-problems which the size of each sub-problem is smaller than the
original problem. After the problem is divided, sub-problems which depend on the
number of products are considered. To find the optimal solution of each sub-problem,
an initial basis is chosen to construct the initial simplex tableau for each sub-problem.
The algorithm for choosing the initial basic variables for each sub-problem is proposed
by selecting one variable that has the maximum coefficient of each objective function
and all slack variables. By this selection, the dual feasible solution can be obtained. So,
the dual simplex can be performed without using artificial variables. Based on the fact
that our algorithm doesn’t use artificial variables, the process for finding the initial
feasible solution is omitted. This brings to the advantage of our algorithm.
Acknowledgment. This work was supported by Thammasat University Research Unit in Fixed
Points and Optimization.
A New Initial Basis for Solving the Blending Problem 1271
References
1. Chanes, A., Cooper, W.W., Mellon, B.: Blending aviation gasolines—a study in
programming independent activities in an integrated oil company. Econometrica 20(2),
135–259 (1952)
2. Fomeni, F.D.: A multi-objective optimization approach for the blending problem in the tea
industrial. Int. J. Prod. Econ. 205, 179–192 (2018)
3. Marianov, V., Bronfman, V., Obreque, C., Luer-Villagra, A.: A milk collection problem
with blending. Transp. Res. Part E Logist. Transp. Rev. 94, 26–43 (2016)
4. Aurey, S., Wolf, D., Smeers, Y.: Using Column Generation to Solve a Coal Blending
Problem. RAIRO – University Press, New Jersey (1968)
5. Arsham, H.: An artificial-free simplex-type algorithm for general LP models. Math. Comput.
Model. 25(1), 107–123 (1997)
6. Gao, P.: Improvement and its computer implementation of an artificialree simplexype
algorithm by arsham. Appl. Math. Comput. 263, 410–415 (2015)
7. Huhn, P.: A counterexample to H. Arsham: Initialization of the Simplex Algorithm: An
Artificial Free Approach (1998)
8. Boonperm, A., Sinapiromsaran, K.: The artificial-free technique along the objective direction
for the simplex algorithm. J. Phys. Conf. Ser. 490, 012193 (2014)
9. Nabli, H., Chahdoura, S.: Algebraic simplex initialization combined with the nonfeasible
basis method. Eur. J. Oper. Res. 245(2), 384–391 (2015)
10. Phumrachat, M.T.: On the use of sum of unoccupied rows for the simplex algorithm
initialization. Doctoral dissertation, Thammasat University (2017)
Review of the Information that is Previously
Needed to Include Traceability in a Global
Supply Chain
Abstract. Caring for and preserving the quality of products throughout the
supply chain has become a challenge for global trade. So is the backup of
information so that it is available to the clients or authorities that require it.
Global chains are made up of different stages and processes that include
production, transformation, handling and transportation from origin to destina-
tion. These stages are executed by different actors and therefore an effective
communication is required on a continuous basis between them.
For this reason, the initial purpose of this research is to identify what prior
information is required to include traceability processes that serve as support for
an integrated supply chain and thus reduce potential risks to the integrity of the
marketed product. An analysis of available traceability technologies that have
been used in global supply chains is included.
The information was collected through a review of literature related to
traceability and quality with which it was possible to identify in a general way
the main stakeholders in supply chains, this provides us with a frame of refer-
ence on the possible limitations of the research and selection of a starting point
for the proposal of an efficient traceability process according to the client’s
requirements.
1 Introduction
In a market context with a tendency to expand in which logistics plays a key role in
constantly seeking to satisfy demand at the best possible cost without neglecting the
part of customer care and service that requires the delivery of their products in the best
possible quality and safety with the possibility of having access to the information
related to the product purchased, currently all this is possible thanks to the integration
of technological tools in supply chains in conjunction with the application of trace-
ability processes that have become a requirement in international trade for better
control and risk management, especially in agro-industrial chains. The food and
pharmaceutical industries must have traceability systems that allow the identification of
products and processes with a proper backup of such information.
It is necessary that our production chain complies with both local and international
regulations as well as the expected quality standards; also risks must be reduced as
much as possible by providing each participant the possibility of access to information
on the monitoring and management of their products, which is achieved when trace-
ability systems are integrated and they work properly. Tracking tools allow the product
to be tracked at each stage of the transformation process and during its distribution
throughout the supply chain by having information support, which includes the
implementation of new technologies.
Since our study is in the initial phase we can use the simulation tool to start doing
some tests. Evolutionary Prototyping is a popular system development technique and
can enable users to develop a concrete sense about software systems that have not yet
been implemented [1].
Traceability in the supply chain allows the identification of information related to
the monitoring and tracing of traceable units during the production and marketing
process [10]. Since a supply chain is made up of various actors according to each stage
that include producers, suppliers, packers, transporters, traders and consumers, trace-
ability system serves to support the integration of the chain to be able to satisfy demand
and if necessary take corrective actions.
All this in an effort to seek better marketing opportunities, ensuring compliance
with health and quality standards regardless of the type of transport used to reach your
destination.
All stages of the chain should be documented and managed so that it is possible to
trace them back to the origin of the product, step by step, and ensure that there has been
no contamination… This traceability minimizes the risk of fraud at all stages and is a
very important part of the inspection process of certifying organizations [5]. An
analysis of available methodologies and technologies is included that may be useful to
identify the risks involved in the chain and how the traceability processes allow to
ensure the quality of products by having determined which stages already have safe
handling and which are the vulnerable stages. Finally, the use of tools or technologies
that serve as support in the backup of information from the process documentation will
be proposed.
2 Background
Logistics is defined as the set of activities and processes necessary to ensure the
delivery of merchandise from the place of its production to the point where the product
is marketed or delivered to the final consumer. This involves inspection and quality
control of products and in terms of geographical scope it can be: urban, domestic or
foreign trade [2].
In this way, d that there are customers who are willing to pay more if the product
they are purchasing is of higher quality and they can stay informed.
1274 Z. M. Reyna Guevara et al.
– Ease of identifying defective products and the withdrawal of products in the event
of possible incidents.
– Support for product safety and quality control system.
– Helps reduce risks related to food safety.Helps reduce risks related to food safety.
3 Methodology
The factor rating method is an excellent tool for dealing with evaluation problems
such as country risk or for the selection of suppliers [11].
The methodology QFD (Quality function deployment) refers on first place to
determining what will satisfy the customer and second translating the customer’s
wishes into a target design. The idea is to gain a good understanding of the customer’s
wishes and to identify alternative process solutions. The QFD is used early in the
design process to help determine what will satisfy the customer and where to deploy
quality efforts [11].
TQM (Total Quality Management) refers to the emphasis that an entire organization
places on quality, from supplier to customer. TQM emphasizes management’s com-
mitment to continually direct the entire company toward excellence in all aspects of
products and services that are important to the customer [11].
The use of Hazard Analysis and Critical Control Point (HACCP) based quality
assurance has a well established place in controlling safety hazards in food supply
chains. It is an assurance system based on the prevention of food safety problems and is
1276 Z. M. Reyna Guevara et al.
The analysis of basic functions help us to identify the characteristics of the supply
chain starting from qualitative and quantitative evaluations that include interviews and
surveys applied to the members of the chain [23].
We can represent the current state of the supply chain through the geographic map,
thread diagram and process diagram used in the SCOR Model which describes the
existing process by identifying sources, manufacturing sites and distribution centers,
using the process categories [23].
Review of the Information that is Previously Needed to Include Traceability 1277
Object Management Group mentioned in 2010 that the BPMN tool is a notation
that has been specifically designed to coordinate the sequence of processes and the
messages that flow between the different participants of the process in a set of related
activities.
The analysis of a generic supply chain underscores that as a product passes from the
primary producer to the end customer, that product undergoes a series of transforma-
tions. Each transformation will involve different agents. Each agent must collect,
maintain and share information to enable food traceability. The registration and
management of information is facilitated through the use of an information system
generated from the BPMN model of the supply chain [25].
A well known technique of system modeling is simulation, which allows to imitate
the operation of various kinds of facilities and processes in the real world and since
there is one unique way to model by simulation, as this depends on the characteristics
of the system, the objectives of the study and the information available [26].
In the Sarli study published in 2018, distributed simulation appears as a more
appropriate configuration to run a simulation of a supply chain since it allows the reuse
of previously existing simulators in the members of the chain, thus maintaining
independence and avoiding need to build a single simulator that understands the
behavior of all its participants. In this way each member preserves its logic of business,
uses its simulation model and shares the minimum amount of information necessary,
which allows modifying its internal logic without affecting the rest [27].
Having mentioned the above, we can start the characterization of the supply chain
using the analysis of basic functions at first using qualitative or quantitative evaluations
and once that we have that information we can proceed to make a geographic map a
thread diagram and the process diagram from the SCOR Model.
Our next step can be a simulated supply chain, we considered it an appropriate
option for our study since it would allow us to make changes or new proposals in a
flexible way to the tests with the supply chain as appropriate.
4 Conclusions
We also need to find specific information about regulatory requirements. Also the
interest of those responsible for each phase throughout the supply chain is required to
achieve full traceability.
A good understanding of the business problem is required and also to have a clear
definition of the objectives to be achieved, since as logistics is not the same for all
organizations, traceability processes must be flexible and adapted to each case as
necessary, taking into account the requirements of the client, demand, capacity of
implementation and design of the supply chain.
In this order of ideas, we understand the level of implementation of traceability will
depend on the type of supply chain and the ability to adapt to the expectations and
Review of the Information that is Previously Needed to Include Traceability 1279
demands of the client with the best possible technology since global markets demand it
and there are also clients who they will be willing to pay more for a higher quality
product and be able to stay informed about it.
References
1. Zhang, X., Lv, S., Xu, M., Mu, W.: Applying evolutionary prototyping model for eliciting
system requirement of meat traceability at agribusiness level. Food Control 21, 1556–1562
(2010)
2. Montanez, L., Granada, I., Rodríguez, R., Veverka, J.: Guía logística. Aspectos conceptuales
y prácticos de la logística de cargas. Banco Interamericano de Desarrollo, Estados Unidos
(2015)
3. Barbero, J.: La logística de carga en América Latina y el Caribe, una agenda para mejorar su
desempeño. Banco Interamericano de Desarrollo, Estados Unidos (2010)
4. Besterfield, D.: Control de calidad, 8th edn. Pearson Educación, México (2009)
5. ITC: Guía del Exportador de Café, 3ra. Ed. Centro de Comercio Internacional, agencia
conjunta de la Organización Mundial del Comercio y las Naciones Unidas, Suiza (2011)
6. AECOC: La Asociación de Fabricantes y Distribuidores. Recuperado de https://www.aecoc.
es/servicios/implantacion/trazabilidad/. Accessed 10 Aug 2020
7. Dittman, J.: Gestión de riesgos en la cadena de suministro global, un informe de la Facultad
de Administración de la Cadena de Suministro en la Universidad de Tennessee. Patrocinado
por UPS Capital Corporation, Estados Unidos (2014)
8. Forero, N., González, J., Sánchez, J., Valencia, Y.: Sistema de trazado para la cadena de
suministro del café colombiano, Colombia (2019)
9. Sosa, C.: Propuesta de un sistema de trazabilidad de productos para la cadena de suministro
agroalimentaria, España (2017)
10. Rincón, D., Fonseca, J., Castro, J.: Hacia un marco conceptual común sobre trazabilidad en
la cadena de suministro de alimentos, Colombia (2017)
11. Heizer, J., Render, B.: Principios de administración de operaciones, 7th edn. Pearson
Educación, México (2009)
12. Stanley, R., Knight, C., Bodnar, F.: Experiences and challenges in the development of an
organic HACCP system, United Kingdom (2011)
13. Rodríguez, E.: Identificación de prácticas en la gestión de la cadena de suministro sostenible
para la industria alimenticia, Colombia (2018)
14. Badia-Melis, R., Mishra, P., Ruiz-García, L.: Food traceability: new trends and recent
advances. A review. Food Control 57, 393–401 (2015)
15. Catarinucci, L., Cuiñas, I., Expósito, I., Colella, R., Gay-Fernández, J., Tarricone, L.: RFID
and WSNs for traceability of agricultural goods from farm to fork: electromagnetic and
deployment aspects on wine test-cases. In: SoftCOM 2011, 19th International Conference on
Software, Telecommunications and Computer Networks, pp. 1–4. Institute of Electrical and
Electronics Engineers (2011)
16. Zou, Z., Chen, Q., Uysal, I., Zheng, L.: Radio frequency identification enabled wireless
sensing for intelligent food logistics. Philos. Trans. R. Soc. A 372, 20130302 (2014)
17. Mainetti, L., Patrono, L., Stefanizzi, M., Vergallo, R.: An innovative and low-cost gapless
traceability system of fresh vegetable products using RF technologies and EPCglobal
standard. Comput. Electron. Agric. 98, 146–157 (2013)
18. Mack, M., Dittmer, P., Veigt, M., Kus, M., Nehmiz, U., Kreyenschmidt, J.: Quality tracing
in meat supply chains. Philos. Trans. R. Soc. A 372, 20130302 (2014)
1280 Z. M. Reyna Guevara et al.
19. Zhao, Y., Zhang, B., Chen, G., Chen, A., Yang, S., Ye, Z.: Recent developments in
application of stable isotope analysis on agro-product authenticity and traceability. Food
Chem. 145, 300–305 (2014)
20. Consonni, R., Cagliani, L.: Nuclear magnetic resonance and chemometrics to assess
geographical origin and quality of traditional food products. Adv. Food Nutr. Res. 59, 87–
165 (2010)
21. De Mattia, F., Bruni, I., Galimberti, A., Cattaneo, F., Casiraghi, M., Labra, M.: A
comparative study of different DNA barcoding markers for the identification of some
members of Lamiacaea. Food Res. Int. 44, 693–702 (2011)
22. Casino, F., Kanakaris, V., Dasaklis, T., Moschuris, S., Rachaniotis, N.: Modeling food
supply chain traceability based on blockchain technology. IFAC PapersOnLine 52, 2728–
2733 (2019)
23. Martínez, K., Rivera, L., García, R.: Caracterización de la cadena de suministro de la
Asociación Ruta de la Carne en el departamento de Boyacá. Universidad Pedagógica y
Tecnológica de Colombia, Colombia (2018)
24. García-Cáceres, R., Perdomo, A., Ortiz, O., Beltrán, P., López, K.: Characterization of the
supply and value chains of Colombian cocoa. DYNA 81, 30–40 (2014). Universidad
Nacional de Colombia, Colombia
25. Pizzuti, T., Mirabelli, G., Sanz-Bobi, M., Goméz-Gonzaléz, F.: Food track & trace ontology
for helping the food traceability control. J. Food Eng. 120, 17–30 (2014)
26. Giraldo-García, J., Castrillón-Gómez, O., Ruiz-Herrera, S.: Simulación discreta y por
agentes de una cadena de suministro simple incluyendo un Sistema de Información
Geográfica (SIG). Información tecnológica 30, 123–136 (2019)
27. Sarli, J., Leone, H., Gutierrez, M.: SCFHLA: Un modelo de interoperabilidad semántica para
simulación distribuida de cadenas de suministro. RISTI, Revista lbérica de Sistemas y
Tecnologías de Información 30, 34–50 (2018)
Online Technology: Effective Contributor
to Academic Writing
1 Introduction
An academic writing without cut, copy, and paste has a greater implication in every
research and educational institutions where plagiarism-induced writing is intolerable [1,
2]. Online technology plays a significant role to produce quality writing. It facilitates
the provision of e-mentoring, e-library, and e-discussion forum [3, 4]. The most
common and popular forms of online technology include Dropbox, Google Drive,
YouTube, Facebook, Database from online, Reflective Journals, Blogs, Chat Room for
academic discussion, Wikis, Skype, WhatsApp, Zoom, and Social Media Group [5].
Students’ academic and research activities at the tertiary level are mostly influenced by
online technology because of its significant role in generating diversified learning
strategies [6]. It develops student presentation, writing, searching, concept
development, critical thinking, citation, and referencing skills [7]. Besides, mentees or
students can send assignments to their respective mentors, and academic supervisors or
teachers for assessment of their academic ask through Dropbox, Google Drive, and
WhatsApp. They can share their necessary files, documents, and relevant articles,
information, and data with other researchers, academicians, and policymakers by using
online technology. Proper utilization of online technology also helps to assist, select,
design, and develop the questionnaire, survey technique, study area, theory, concept,
research questions and objectives, research approach and philosophy, methodology,
and data management technique, and interview questions, and proofreading from
teachers, researchers, experts, and mentors. Thus, online technology has resulted in the
emergence of new forms of learning genres and literature that empowers students’ level
of cognition through collaboration among students, teachers, and researchers [8].
Considering the importance of online technology, a British proverb reveals that “if
students or learners do not effectively make a bridge between their ideas and online
technology, they might as well live in a cave or the ground shell of a space shuttle” [9,
p. 1]. Modern education and research always give emphasize on online technology for
significant academic writing [10].
In the present time, students are habituated to handling computer, internet, smart-
phone, and tab. Besides, they have proficiency in intelligent computing that leads to
motivating them using online and digital technologies. Optimum use of online tech-
nology in academic writing requires the proper utilization of online technology. Having
no prior application knowledge of online technology may lose students’ and
researchers’ level of interest and satisfaction towards it. This situation may narrow
down the scope of collaborative learning in a broader sense and restrict it to produce
quality academic writing. Besides, motivation towards more usage of online technol-
ogy in academic writing gets stuck when it faces a few challenges. Potential challenges
to using this technology include learners’ ability to purchase essential electronic gad-
gets and internet package, insufficient supports for designing online technology-
mediated curriculum and assessment techniques, motivation, and encouragement from
teachers for more usage of online technology in writing, students’ incapability to
handling online technology, and interrupted internet connection [11, 12]. Authenticity
verification, security, and fraud prevention are also another form of challenge to
continue online technology [13, 14].
Graduate college students of the National University (NU) of Bangladesh are not
habituated with online technology for their academic writing. They prefer verbal
instruction from their respective teacher, printed textbooks, newspapers, periodicals,
and predatory journals. Having no sound application knowledge of online technology
in academic writing, they prefer cut, copy, and paste practice to complete their term
papers and other academic assignments and fail to show their potentiality and capa-
bility in assignment and term paper [15]. Moreover, they fail to develop critical
thinking, constructive, innovative and creative ideas, and thus their academic assign-
ments lose scientific quality and originality. At the same time, they fail to present and
fail to present noble ideas and merge different research paradigm like epistemological
stance, and ontological beliefs. To promote online technology in academic writing, this
study tries to fulfill the following two objectives: detect potential contributors to more
Online Technology: Effective Contributor to Academic Writing 1283
usage of online technology at the graduate level under the NU of Bangladesh and
develop policy options for using online technology in academic writing.
Online technology has got popularity in academic writing in many universities and
research institutions due to its significant role in better learning, research, and educa-
tion. This study mainly focuses on the benefits of online learning in academic writing.
The study also provides an overview of online technology and academic writing. The
study is significant in several ways. It contributes to the current research regarding the
effect of online technology in academic writing in two ways. First. We provide
quantitative evidence on the effect of online technology in academic writing, which has
gained minor importance in the previous literature. Our findings highlight the effect of
online technology in writing. Second. We conduct a mini-experiment, which allows us
to consider two groups such as treated or experiment or study group and usual or
control group for proper empirical investigation. Hence, we can be able to measure the
causal effects of online technology in academic writing. In particular, the study gives a
fuller understanding of the use of online technology in academic writing and provides
an effective guideline for future researchers and policymakers in this field.
The remainder of this paper is organized as follows: Sect. 2 covers a brief literature
review. Section 3 presents the theoretical motivation, Sect. 4 highlights the method-
ology and research plan. The results and discussion are provided in Sect. 5. Section 6
concludes by offering recommendations and policy implications.
2 Literature Review
Online technology enhances students’ writing skills. It further helps to make strong
academic collaboration among students, teachers, researchers, mentors, and others.
Research work, communication, and social and cultural competencies get more force
when these are dealt with online technology [16, 17]. It works as a catalyst to reduce
plagiarism practice in academic writing and helps to seek research opportunities around
the world [18]. It is treated as a more effective, significant, and powerful tool that
facilitates concept building. Students get a good grade when they are involved in online
technology-mediated academic writing [19]. It deals the pedagogical tasks and con-
siders as a means and ways of inclusive and life-long learning. The principal objective
of this technology is to generate quality writing. On the other hand, students get easy
access to the learning platform through online technology that enhances higher-order
learning, thinking, and writing skills [20]. Based on the numerous benefits of online
technology, [21] gives more stress on online technology-induced academic writing
rather than the traditional modes of writing because it is a suitable platform for distance
mode learning. Students can get easy access to learning from anywhere and anytime
through online technology [22].
The use of online technology in academic writing through e-discussion forum,
social media group, wiki, skype, blog, and zoom are well recognized [23]. For instance,
Blog-mediated concepts and feedback play an important role in academic writing [24].
It is an important contributor to students’ learning and academic writing because it
provides critical comments and constructive feedback [25, 26]. It develops learners’
cognitive level, enhances capacity building in academic writing, and foster literacy
1284 Md. H. Iqbal et al.
skills [27]. It also develops the general concept of any particular issue, motivates
students to copy and paste free writing, and enhances students’ confidence along with
their perceived strengths and weakness of academic writing. Like the blogs, skype
generates a new dimension of the academic writing approach. Generally, it is seen that
traditional writing (such as library-oriented and class notes-based academic writing)
requires a longer time and a large number of academic resources (such as class note,
textbook, newspaper, and article) to finalize a manuscript [28]. Skype-mediated stu-
dents perform better in academic writing than traditional academic writing practices
[29]. Wikis have greater impacts on academic writing. It makes students pay closer
attention to the formal aspects of academic writing by collectively contributing or
mutually supportive approaches [30]. Wiki is an emergent technology responsible to
increase writing capacity. This web-based authoring instrument is used in academic
writing to execute in collaboration with others [31]. Wiki-assisted writing can con-
tribute to raise awareness of the audience and increase the use of interpersonal
metadiscourse [32]. On the other hand, the reflective journal has a vital role to eliminate
typos and errors free writing. In the provision of the reflective journal, students get
relevant and constructive feedback from researchers, scholars, key persons, and
teachers that meet their objectives [33]. Like the reflective journal, the chat room also
has positive impacts on concept building for academic writing. Online-based chat
rooms are functioning well in regular interactions with students and librarians for
essential, valid, and reliable bibliographic database, and desirable e-journals and books
[34]. Students who are habituated with the chat room for concept development perform
well in academic writing-mediated academic writing compared to those of students not
used chat reference service [35].
Assessment of the effectiveness of online technology in learning is highly required
for teachers, college administrators, the authority of NU of Bangladesh, and the gov-
ernment to implement, plan, and modify online technology-based learning for better
academic writing. Existing literature perfectly highlights the effectiveness of online
technology in learning in terms of different countries’ perspectives. A very few studies
focused on the effectiveness of online technology in academic writing in Bangladeshi
educational institutions like NU of Bangladesh. This study tries to cover this issue
significantly.
3 Theoretical Motivation
A society can utilize an alternative option when it fails to satisfy desirable condition
with its traditional or existing option. With this point of view, Kelvi Lancaster and
Richard G. Lipsey first formalized the theory of the second-best in their article entitled
“The General Theory of Second Best” in 1956 followed an earlier work by James E.
Meade [36]. In this viewpoint, online technology works as an effective contributor to
academic writing instead of traditional writing practice. The theory of the second-best
explains why students use online technology in their academic writing practice. This
theory is based on Pareto optimality (academic writing by online technology cannot
increase the writing efficiency without reducing traditional writing practice). A certain
social dividend can be gained by a movement from a Pareto non-optimal allocation to a
Online Technology: Effective Contributor to Academic Writing 1285
Pareto optimal allocation [37]. Therefore, the optimality of the Pareto condition is often
considered an efficient usage of online technology in academic writing.
For better understanding, the fundamental feature of this theory can be presented
for a single student case. The first-order condition for Pareto optimality is obtained by
maximizing the quality of the student’s academic writing subject to the associate
writing production function. From the Lagrangian function, it is possible to write the
following mathematical form:
V ¼ W ðx1 ; x2 ; . . .; xn Þ sF x1 ; x2 ; . . .; xn ; q0 ð1Þ
@V
¼ Wi sFi ; i ¼ 1; . . .; n ð2Þ
@x1
where Wi ¼ @W @F
@x1 and Fi ¼ @x1 . It follows that
Wi F i
¼ ; i; j ¼ 1; . . .; n ð3Þ
Wj F j
If Eq. (2) is satisfied, the rate of substitution between the traditional writing approach
and the online technology-mediated writing approach will equal the corresponding rate
of writing transformation.
Assume that an educational institution fails to promote an online technology-
mediated academic writing practice in its campus corresponding to one of the condi-
tions of Eq. (2). Under this practice, we can consider another assumption to eliminate
the implementation barriers of online technology-induced writing condition which can
be expressed as follows:
Wi kFi ¼ 0 ð4Þ
where k is a positive but arbitrary constant that will able to produce different optimal
values of s calculated from Eq. (2) and the writing production function.
The conditions for a second-best solution are obtained by maximizing online
technology-mediated academic writing practice subject to Eqs. (1) and (4). From the
Lagrangian function, we can write the following mathematical form:
V ¼ W ðx1 ; x2 ; . . .; xn Þ sF x1 ; x2 ; . . .; xn ; q0 cðWi kFi Þ ð5Þ
where s and s indicate undetermined and unknown multipliers. Take the partial
derivatives of V, setting them equal to zero
@V
¼ Wi sFi cðWi kFi Þ ¼ 0; i ¼ 1; . . .; n ð6Þ
@x1
1286 Md. H. Iqbal et al.
@V
¼ Fi x1 ; x2 ; . . .; xn ; q0 ¼ 0 ð7Þ
@s
@V
¼ ðWi kFi Þ ð8Þ
@c
Considering the condition c ¼ 0; it is not possible to get a solution from Eq. (6)
because of the violation of the assumption presented in Eq. (4). Moving to Eqs. (7) and
(8) to the right-hand side and dividing the ith equation by the jth:
It is not possible to know a priori about the signs of cross partial derivatives such as
Wi1 ; Wij ; F1i ; and F1j . Hence, anyone may not expect the usual Pareto conditions to be
required for the attainment of the second-best optimum [38]. People’s perception of
certain actions and their effectiveness are required in experiment, survey, observation,
interview, and judgment to attain the second-best optimum. Thus, the following section
will discuss students’ perceptions about online technology and its effectiveness in
academic writing through a few research plans such as experiment, focus group dis-
cussion (FGD), and questionnaire survey.
control group of this department. The first question was concerned to know about
online technology, and the other question was associated with the effectiveness of
online technology in academic writing. A total of 32 students answered “yes” to both of
the questions were classified into an experimental group and the rest 47 students
answered “no” to both of the questions were treated as a control group. The first group
was instructed to develop concept building and plagiarism-free academic writing by
using different tools of online technology (see Table 1 for more details) and the last
group was requested to follow traditional writing practice. Both groups were assigned
the same writing tasks such as group works in the classroom, home works (writing
summary, problem statement, review of literature, result and discussion, and recom-
mendations), and report writing on a particular topic. The grading scheme of the
writing assignment was carried a 10% mark in participation, a 20% mark in a home-
work assignment, and the rest 50% mark in short report writing. Our result showed that
the experimental group performs well in academic writing and they got a good grade in
their writing tasks (average score 82%) compared to those of the usual group (average
score 71%). This result motivated us to run the questionnaire survey.
Yi ¼ b0 þ b1 Xi þ ei ð9Þ
Equation (9) is known as the RPL model when we use only the attributes. It is also
known as the MNL model when we use both the attributes and SED characteristics
[43]. The use of online technology in academic writing (us_olt_ac_wri) is treated as an
outcome variable in both of the models.
Out of 151 students from seven NU affiliated colleges, 119 (78.807%) were male
students and 32 (21.192%) were female students in the survey. Among them, 92% of
respondents belonged to Muslim households and the rest 8% of respondents were from
Hindu households. About 100 (66%) of students argued that online technology could
play a significant role in academic writing, but 51 (34%) of respondents were not
interested in online technology due to low family income, scarcity of internet service,
inability to purchase smartphone (Android and iOS), the higher price of internet
package, lower speed of bandwidth, and the feeling of discomfort with online tech-
nology. All students in the survey process have strongly pointed out that they did not
get any support from their respective institutions for better utilization of online tech-
nology in academic writing. Table 1 outlines the summary statistics of the study.
Respondents’ average age and their parents’ monthly income were recorded at 22.33
and Bangladeshi currency Taka 43792 respectively. All respondents were studying at
undergraduate and graduate levels and their daily average expense was estimated Taka
71 per person. The majority of the respondents have more than 5 family members.
1290 Md. H. Iqbal et al.
us_olt_ac_wri 1
ins_inf_sup 0.831** 1
Mtv 0.518* 0.327 1
stu_resi_loc –0.623*** 0.442** –0.678 1
tut_se_on_te 0.891 ** 0.339* 0.589 * 0.348* 1
co_ef_in_pac 0.883 * 0.125** –0.423** 0.449 0.889 * 1
gad_sub_pri 0.567 ** 0.254*** –0.798*** 0.673** 0.805* 0.810*** 1
Note. * Significant at the 1% level.
**
Significant at the 5% level.
***
Significant at the 10% level.
This study explores the potential contributors to online technology in academic writing
in NU affiliated colleges of Bangladesh and develops a policy for using online tech-
nology. We used cross-sectional data collected from NU affiliated few colleges in the
Pabna district. FGDs help us to select a few attributes and design the questionnaire. We
also depend on mini-experiment and two types of regression models for proper
empirical assessment. We found that students were much aware of plagiarism-free
academic writing and they wanted to produce online technology-mediated writing. Our
empirical assessment also takes stand on the effectiveness of online technology based
on significant levels of institutional and infrastructural status, students’ residential
location, motivation towards more usage of online technology in academic writing,
tutorial session for using online technology, cost-effective internet package for students
1292 Md. H. Iqbal et al.
and provides electronic gadgets for students at the subsidized price. Sometimes few
SED characteristics work as an accelerator to move forward online technology in
writing [49].
The paradigm shift from the traditional writing approach to the online-mediated
writing approach is essential to develop the environment of online technology in any
educational institution. It requires appropriate guidelines, strategies, policies, and joint
efforts of other stakeholders like class teachers, college administrators, the authority of
NU of Bangladesh, and government intervention. For instance, the government can
formulate an online technology-mediated educational policy for college education. The
government can support training facilities for college teachers to habituate them to
online technology in education. The government can also provide free electronic
gadgets like tab, smartphone, and laptop to the students or provide them a subsidy to
purchase electronic gadgets at a lower price. Likewise, college authority can provide an
uninterrupted Wi-Fi facility, multimedia projector supported classroom, and a contin-
uous power supply facility for better access to online technology. The NU of Ban-
gladesh can redesign their syllabus and curriculum with special attention to online
technology-mediated academic writing. Lastly, the class teachers can encourage and
motivate their students to more use of online technology in academic writing. The class
teacher can arrange tutorial sessions for their students so that the students can effec-
tively using online technology in academic writing tasks.
The study is not free from certain limitations. As a new concept, it is essential to
undertake an in-depth study and a higher range of questionnaire surveys. For a better
assessment, it is essential to conduct further research in this field. Only a shorter range
of the survey in a few colleges may narrow down the scope, concept, and effectiveness
of online technology in academic writing. However, we successfully applied different
approaches that guarantee the validity, reliability, and consistency of our empirical
findings.
Acknowledgments. We greatly acknowledge Assistant Professor Dr. Renee Chew Shiun Yee,
School of Education, University of Nottingham, Malaysia Campus (UNMC) for her significant
comments and suggestions to develop the concept and methodology of this study. We are
thankful to the editor(s) and other anonymous referees of this journal paper for their valuable and
constructive suggestions for improving the draft.
References
1. Fahmida, B.: Bangladesh tertiary level students’ common errors in academic writing. BRAC
University, Dhaka (2020). https://dspace.bracu.ac.bd/xmlui/bitstream/handle/10361/252/
08163004.PDF?sequence=4
2. Iqbal, M.H.: E-mentoring: an effective platform for distance learning. E-mentor 2(84), 54–61
(2020)
3. Macznik, A.K., Ribeiro, D.C., Baxter, G.D.: Online technology use in physiotherapy
teaching and learning: a systematic review of effectiveness and users’ perceptions. BMC
Med. Educ. 15(160), 1–2 (2015)
4. Iqbal, M.H., Ahmed, F.: Paperless campus: the real contribution towards a sustainable low
carbon society. J. Environ. Sci. Toxicol. Food Technol. 9(8), 10–17 (2015)
Online Technology: Effective Contributor to Academic Writing 1293
5. Iqbal, M.H., Sarker, S., Mazid, A.M.: Introducing technology in Bangla written research
work: application of econometric analysis. SAMS J. 12(2), 46–54 (2018)
6. Hanson, C., West, J., Neiger, B., Thackeray, R., Barnes, M., Mclntyre, E.: Use and
acceptance of social media among health education. Am. J. Health Educ. 42(4), 197–204
(2011)
7. Kuzme, J.: Using online technology to enhance student presentation skills. Worces-
ter J. Learn. Teach. 5, 27–39 (2011)
8. Vlachopoulos, D., Makri, A.: The effect of games and simulations on higher education: a
systematic literature review. Int. J. High. Educ. 14(1), 22–34 (2017)
9. Adams, P.C.: Placing the anthropocene: a day in the life of an enviro-organism. Trans. Inst.
Br. Geogr. 41(1), 54–65 (2016)
10. Goldie, J.G.S.: Connectivism: a knowledge learning theory for the digital age? Med. Teach.
38(10), 1064–1069 (2016)
11. Shelburne, W.A.: E-book usage in an academic library: user attitudes and behaviors. Libr.
Collect. Acquisitions Tech. Serv. 33(2–3), 59–72 (2009)
12. Thornton, P., Houser, C.: Using mobile phones in English education in Japan. J. Comput.
Assisted Learn. 21(3), 217–228 (2005)
13. Bandyopadhyay, D., Sen, J.: Internet of things: applications and challenges in technology
and standardization. Wireless Pers. Commun. 58(1), 49–69 (2011)
14. Aebersold, R., Agar, J.N., Amster, I.J., Baker, M.S., Bertozzi, C.R., Boja, E., Ge, Y., et al.:
How many human proteoforms are there? Nat. Chem. Biol. 14(3), 206–217 (2018)
15. Click, A.B.: International graduate students in the United States: research process and
challenges. Libr. Inf. Sci. Res. 40(2), 153–162 (2018)
16. Aesaert, K., Vanderlinde, R., Tondeur, J., van Braak, J.: The content of educational
technology curricula: a cross-curricular state of the art. Educ. Tech. Res. Dev. 61(1), 131–
151 (2013)
17. Voogt, J., Roblin, N.P.: 21st century skills. Discussion Paper. University of Twente, Faculty
of Behavioral Science, University of Twente, Department of Curriculum Design and
Educational Innovation, Enschede (2010)
18. Amiel, T., Reeves, T.C.: Design-based research and educational technology: rethinking
technology and the research agenda. Educ. Technol. Soc. 11(4), 29–40 (2008)
19. Hendler, J., Shadbolt, N., Hall, W., Berners-Lee, T., Weitzner, D.: Web science: an
interdisciplinary approach to understanding the web. Commun. ACM 51(7), 60–69 (2008)
20. Levine, M., DiScenza, D.J.: Sweet, sweet science: addressing the gender gap in STEM
disciplines through a one-day high school program in sugar chemistry. J. Chem. Educ. 95(8),
1316–1322 (2018)
21. Schulman, A.H., Sims, R.L.: Learning in an online format versus an in-class format: an
experimental study. J. Technol. Horiz. Educ. 26(11), 54–64 (1999)
22. Murthy, D.: Digital ethnography: an examination of the use of new technologies for social
research. Sociology 42(5), 837–855 (2008)
23. Miyazoe, T., Anderson, T.: Learning outcomes and students’ perceptions of online writing:
simultaneous implementation of a forum, blog, and wiki in an EFL blended learning setting.
System. 38(2), 185–199 (2010)
24. Jimoyiannis, A., Schiza, E.I., Tsiotakis, P.: Students’ self-regulated learning through online
academic writing in a course Blog. In: Digital Technologies: Sustainable Innovations for
Improving Teaching and Learning, pp. 111–129. Springer, Cham (2018)
25. Hansen, H.E.: The impact of blog-style writing on student learning outcomes: a pilot study.
J. Polit. Sci. Educ. 12(1), 85–101 (2016)
26. Chong, E.K.: Using blogging to enhance the initiation of students into academic research.
Comput. Educ. 55(2), 798–807 (2010)
1294 Md. H. Iqbal et al.
27. Novakovich, J.: Fostering critical thinking and reflection through blog-mediated peer
feedback. J. Comput. Assisted Learn. 32(1), 16–30 (2016)
28. Hussein, N.O., Elttayef, A.I.: The impact of utilizing Skype as a social tool network
community on developing English major students’ discourse competence in the English
language syllables. J. Educ. Pract. 7(11), 29–33 (2016)
29. Lo Iacono, V., Symonds, P., Brown, D.H.: Skype as a tool for qualitative research
interviews. Sociol. Res. Online 21(2), 103–117 (2016)
30. Li, M., Zhu, W.: Patterns of computer-mediated interaction in small writing groups using
wikis. Comput. Assisted Lang. Learn. 26(1), 61–82 (2013)
31. Yusoff, Z.S., Alwi, N.A.N., Ibrahim, A.H.: Investigating students’ perception of using wikis
in academic writing. 3L Lang. Linguist. Lit. 18(3), 17–32 (2012)
32. Zhang, W., Cheung, Y.L.: Researching innovation in English language writing instruction: a
state-of-the art-review. J. Lang. Teach. Res. 9(1), 80–89 (2018)
33. Estaji, M., Salimi, H.: The application of wiki-mediated collaborative writing as a
pedagogical tool to promote ESP learners’ writing performance. Asian ESP J. 14(1), 112–
141 (2018)
34. Henter, R., Indreica, E.S.: Reflective journal writing as a metacognitve tool. Int. Sci.
Commit. 533, 133–141 (2014)
35. MacDonald, H.: Undergraduate students can provide satisfactory chat reference service in an
academic library. Evid. Based Libr. Inf. Pract. 13(2), 112–114 (2018)
36. Tudini, V.: Using native speakers in chat. Lang. Learn. Technol. 7(3), 141–159 (2003)
37. Lipsey, R.G., Lancaster, K.: The general theory of second best. Rev. Econ. Stud. 24(1), 11–
32 (1956)
38. Vaez-Glasemi, M., Moghadds, Z., Askari, H., Valizadeh, F.: A review of mathematic
algorithms and data envelopment analysis. Iran. J. Optim. 12(1), 115–137 (2020)
39. Dasgupta, P., Stiglitz, J.: On optimal taxation and public production. Rev. Econ. Stud. 39(1),
87–103 (1972)
40. Porter, W.W., Graham, C.R., Spring, K.A., Welch, K.R.: Blended learning in higher
education: institutional adoption and implementation. Comput. Educ. 75, 185–195 (2014)
41. Hsu, H., Wang, S.: The impact of using blogs on college students’ reading comprehension
and learning motivation. Literacy Res. Instr. 50(1), 68–88 (2010)
42. Peterson, P.W.: The debate about online learning: key issues for writing teachers. Comput.
Compos. 18(4), 359–370 (2001)
43. Iqbal, M.H.: Valuing ecosystem services of sundarbans approach of conjoint experiment.
J. Global Ecol. Conserv. 24, e01273 (2020)
44. Iqbal, M.H.: Telemedicine: an innovative twist to primary health care in rural Bangladesh.
J. Prim. Care Community Health 11, 2150132720950519 (2020)
45. Dillenburger, K., Jordan, J.A., McKerr, L., Keenan, M.: The Millennium child with autism:
early childhood trajectories for health, education and economic wellbeing. Dev. Neurore-
habilitation 18(1), 37–46 (2015)
46. Woodrich, M., Fan, Y.: Google Docs as a tool for collaborative writing in the middle school
classroom. J. Inf. Technol. Educ. Res. 16(16), 391–410 (2017)
47. Hemmert, G.A., Schons, L.M., Wieseke, J., Schimmelpfennig, H.: Log-likelihood-based
pseudo-R 2 in logistic regression: deriving sample-sensitive benchmarks. Sociol. Methods
Res. 47(3), 507–531 (2018)
48. Miaou, S.P., Lu, A., Lum, H.S.: Pitfalls of using R2 to evaluate goodness of fit of accident
prediction models. Transp. Res. Rec. 1542(1), 6–13 (1996)
49. Louviere, J.J., Hensher, D., Swait, J.: Stated Choice Methods: Analysis and Application.
University Press, Cambridge (2020)
A Secured Electronic Voting System
Using Blockchain
1 Introduction
Our modern representative democracy has been operated on the voting mecha-
nism since the 17th century [1]. The security of an election is the most fundamen-
tal issue and is a matter of national security in every democracy. Many voting
systems have been developed to conduct fair elections which are acceptable to all
form of people involved in it. With the prompt expansion of information technol-
ogy, electronic voting (e-voting) or online voting has become a new phenomenon.
For the last several years e-voting systems have been of growing interest in many
sectors as it provides much conveniences over traditional paper voting [2].
2 Related Works
in the domain of e-voting [5,6]. Ofori et al. presented OVIS based online vot-
ing system in [8]. Most of the earlier works were usually based on signature
mechanism and the votes were stored in a centralized server. These e-voting
systems are proven vulnerable to many threats like DDoS attack, sybil attack
[7]. Blockchain is the most suitable technology in storing the cast votes securely.
There exist very few research works on E-voting on blockchain. Some projects
of E-voting on blockchain are showed in [9].
In [10], Ayed proposed a method of storing the votes in the blocks in different
chains. In that work, the first transaction added to the block is a special transac-
tion that represents the candidate. Each chain is allocated for each candidate. If
there are n numbers of candidate, there will be n numbers of blockchains. There
will be much processing and storage overhead because of assigning a dedicated
blockchain for every candidate.
Smart contracts are irreversible and also traceable applications which exe-
cutes in decentralized environment like Blockchain. No one can change or manip-
ulate the code or its execution behavior once it is deployed. Hjalmarsson et al.
in [11] proposed election as a smart contract. The election process is represented
by a set of smart contracts. The smart contracts are instantiated by the election
administrators on the blockchain. Their implementation was based on Ethereum
platforms. Ethereum-based structures include Quorum and Geth For which con-
current execution of transactions is not supported, That limits scalability and
velocity. Exonium is a cryptocurrency based paid which makes it very expensive
for large scale implementation.
Lee et al. proposed a method where votes can be stored in the blockchain
preserving the anonymity of the voter in [12]. In this work they assumed that
there be a trusted authentication organization which authenticate the validity
of the voters. The organization will provide an unique key against each voter’s
identity. The voter has to vote using the given key. So the identity of the voter
and his or her information are not stored in the blockchain. The main limitation
of the work is that, the system completely assume that all the voters are eligible.
So a voter can cast his or her vote multiple time. The system validate the cast
vote at the end of the voting operation.
Yi proposed a blockchain-based e-voting scheme based on distributed ledger
technology (DLT) [13]. For providing authentication and non-repudiation, they
used a user credential model based on elliptic curve cryptography (ECC). This
system allows a voter to change the cast vote before deadline this can be used as
a way of vote alteration which violates the basic security of the e-voting system.
Bellini et al. proposed an e-voting service based on blockchain infrastructure [14].
In there system, service configuration defined by the end user is automatically
translated into a cloud-based deployable bundle.
Most of the existing works have directly adopted Bitcoin’s consensus mech-
anism, where nodes compete with each other to solve a cryptographic puzzle to
mine a block and get incentive. But in the case of public election, there is no
incentive necessary. So the competition among the nodes to mine a new block,
is quite redundant and will waste a lot of computing power of every node. Our
1298 Md. R. Rahman et al.
3 Methodology
The overall architecture of our proposed voting system is shown in Fig. 1. The
primary focus of this work is to store the voters’ votes securely in blockchain and
make sure that the voting information is immutable. Each center in the election
have a node in the network. Each time a vote is cast, first the validity of the vote
is checked. The system ensures that all the voters participating in the voting are
validate. Each center or node has its own copy of the blockchain, where all the
cast votes are stored.
Center n
Voter registration Authentic voter database
voter verification
& authentication Node n
Rest API
Vote key
Center 3
Rest API
• Election result
Block 1 Block 2 Block 3 Block N • Verify vote
• Query vote
Fig. 1. The system architecture overview of our proposed blockchain based voting
system
Our proposed voting system assumes that there must be an authentication orga-
nization to authenticate the eligible voters. The authentication organization han-
dles the registration of the voters. When a voter makes a vote request to the
system by entering the vote key given by the authentication organization, the
system first checks the vote key in the registered voter database provided by the
authentication organization. If the voter is valid and the voter has not yet voted
in the election, then the vote is considered as a valid vote and added to the open
transactions. The organization has all the information of the registered voters.
When voting, the voter has to give none of his private information but only the
vote key which is only known to that particular voter. Thus the anonymity of
the voter is well preserved. The voter is able to verify whether his or her vote
was cast by the vote key.
Each block in a blockchain has the only data field that is called its transactions.
The data which are to be stored in blockchain are in this transaction field. For our
case the votes are stored as data for the blocks in other words, votes are stored as
transactions in our blockchain. These data of the votes are very sensitive, these
data should not be trusted in the hand of third parties [15]. After casting a vote,
the vote is stored as the transaction in the block. The size of the transaction
field is variable. A block can store variable number of transactions (Fig. 3).
A valid Blockchain
casted vote Open transactions
Each center must have a public key and a private key. The keys must be loaded
before any vote is cast from any particular center. After creating or loading the
center. The voter will cast his or her vote with the voting key provided by the
authentication organization. The voter will input the key and select a candidate
to vote.
Load Center
Casting vote
Valid
Private key
Creates
Transaction
Created
Together
Signature
Each block has a field named previous hash. By this field one block is chained to
its prior block. Thus the integrity of the overall chain is preserved. The hashing
algorithm used for this hashing is Secure Hashing Algorithm (SHA-256). Mes-
sages up to 264 bits are transformed into digests or hash value of size 256 bits
or 32 bytes. It is quite impossible to forge the hash value.
64
SHA 256 : B 1 ∪ ............. ∪ B 2 → B 256
M →H
1302 Md. R. Rahman et al.
All the fields of the previous block are used as the input message strings
for the hashing function including all the transactions of the previous block.
The function maps all these information into a fixed message digests or hash of
size 256 bits or 32 bytes. Before mining any particular block the hash is gener-
ated every time then used as the previous hash field of the block being mined
(Fig. 7).
Blockchain has many use cases outside Bitcoin. In our work a methodology is
designed to store the votes securely insides block. The main distinction of the
work is, most of the other blockchain based voting systems, there is a reward
given to the mining node. Each and every node in the network tries to mine a
new block to the chain, the wining node has to solve a cryptographic puzzle to
find the nonce or proof which will produce the required hash for the block to
be mined. In our case there need not to be any mining competition among the
block. The procedure of our proposed consensus is showed in Algorithm 1.
Each node gets it turn Tn to mine a block and add it to the blockchain. The
mining node will generate a valid proof number which will be used to generate
a hash with specific condition and later used to verify the block by all the peer
nodes. Each transaction tx in the open transactions are checked for verification
with signature of that particular transaction. If all the transaction in the open
transactions are valid then proceed to the next step otherwise rejected. The hash
is generated with the Secure Hashing Algorithm (SHA-256). The information of
the previous block is used to generate the hash for the block. For each block
in the blockchain stored in this particular node (self chain), the validity is
checked. If all the blocks in the chain are valid then proceed to the next step
otherwise rejected. The new block is then broadcast to the connected peer nodes
(self.peer nodes). Finally if there exists no conflict situations the new block is
appended to the blockchain and all the open transactions are cleared.
A Secured Electronic Voting System Using Blockchain 1303
(i) Any of the nodes except the broadcasting node, has shorter chain.
Node
Node
Node
(a)
The Broadcasting node has shorter chain
Node
Conflict Occurs
Node
Node
(b)
Fig. 8. (a)Conflict when any node has shorter chain than the network, (b) Conflict
when the broadcasting node has shorter chain than the network
main goal of the resolving is “the longest chain wins”. The node which performs
the resolving task loop for all the chains in the connected network and repeatedly
compare the longest of chains. Ultimately the local chain of the node is replaced
by the longest chain in the network.
of the nodes is shown in Fig. 10. There exists so many challenges and security
requirements to develop a secure e-voting system. Some of the comprehensive
security requirements of contemporary e-voting systems are summarized in [16].
Our implemented blockchain based voting system satisfies several essential
criteria for transparent and trusted electronic voting. These are listed below:
Authentication: The system will only allow the authenticated voter who are
already registered to vote. Our system does not contain the registration process.
For simulating a election, we used about 1,000 voters in the voting process. We
tried to input same vote key from multiple nodes but the system automatically
detect this and does not permit multiple cast. The system is also able to verify
voter and makes sure that no two vote is cast from a single vote key.
Anonymity of voter: Voters in the system cast their vote by using they vote
key. No identity of the voter is revealed. The voter remains anonymous during
and after the election. We can see stored chain of a node in Fig. 10, where a vote
in block no 2 is shown. There is no information of the voter is revealed except the
key (QFYPLMBTLTDN). This ensures the anonymity of voter in our system.
Data integrity: The system makes sure that once a vote is cast, it can never
be changed. Each block is linked with hash to the prior block. To ensure that we
performed manual manipulation of the local chain of several nodes. An example
is shown in Fig. 11. We intentionally changed the proof number of a block from
17 to 23 shown in Fig. 11b. After the change, we tried to cast vote and mine the
block from the manipulated node, the system can detect the manipulation and
immediately mark the node as a violated node and reject mining Fig. 12.
Verifiability: There are 100 nodes in our implemented system and each nodes
stores a valid copy of the chain. Thus Each node knows how many vote is cast
in the system. When the final result of the election is counted, the system makes
sure all the valid nodes agree on the final count of the vote. So system is verifi-
able and makes sure that all votes are counted correctly.
A Secured Electronic Voting System Using Blockchain 1307
Fig. 12. The system rejects the mining request from the manipulated node
in the system. In this analysis we only considered the complexity of getting the
turn to mine a block. We assumed all the processing works in the blockchain are
constant.
5 Conclusion
In this work we proposed and implemented an electronic voting system based
on Blockchain. The votes are stored as transactions inside the blocks and the
blocks are chained together using hash. The system is decentralized so there is
no single point of failure. Each node mines a new block in the system based on
predetermined turn instead of all node competing to solve puzzle to get turn.
We have also mentioned two potential conflict situations and adopted a conflict
resolving methodology for our system. And have showed how conflicts can be
resolved in the system. There is scope of future research on this work. The system
can be employed to conduct real political or organizational elections to measure
its scalability and execution performance.
Our proposed voting system can be a solution to the security concern of the
electronic voting system and can be useful in conducting a transparent election
which is trusted by all. In this work, the consensus mechanism is a not exactly
the Bitcoin’s Proof of Work, rather it is a modified version of that. The method
is based on predefined turn on each node or vote center in the blockchain. Which
will save a lot of computational consumption.
References
1. Eijffinger, S., Mahieu, R., Raes, L.: Inferring hawks and doves from voting records.
Eur. J. Polit. Econ. 51, 107–120 (2018)
2. Taş, R., Tanrıöver, Ö.: A Systematic review of challenges and opportunities of
Blockchain for e-Voting. Symmetry. 12, 1328 (2020)
3. Neumann, P.G.: Security criteria for electronic voting. In: Proceedings of 16th
National Computer Security Conference, pp. 478-481. Maryland (1993)
4. Esteve, J.B., Goldsmith, B., Turner, J.: International Experience with E-
Voting. Available online: https://www.parliament.uk/documents/speaker/digital-
democracy/IFESIVreport.pdf (accessed on 20 August 2020)
5. Evans, D., Paul, N.: Election security: perception and reality. IEEE Secur. Privacy
Mag. 2(1), 24–31 (2004)
6. Chaum, D.: Secret-ballot receipts: true voter-verifiable elections. IEEE Secur. Pri-
vacy Mag. 2(1), 38–47 (2004). https://doi.org/10.1109/msecp.2004.1264852
7. Daramola, O., Thebus, D.: Architecture-centric evaluation of blockchain-based
smart contract e-voting for national elections. Informatics 7, 16 (2020)
A Secured Electronic Voting System Using Blockchain 1309
1 Introduction
Currently, the Russian Federation (RF) is one of the world's largest producers of milk
and dairy products, but has low share of commercial milk in total production, which is
57%. RF achieves about half milk productivity per cow compared to developed
countries.
The main two challenges the dairy industry in the Russian Federation faces are:
reducing the dairy industry's dependence on imported products; increased demand for
commercial milk.
The solution to the first problem stems from the country’s food security: the share
of domestic products should be 90% of the total volume of commodity resources, and
this can be achieved by replacing imported products with domestic products, by
increasing the production of commercial milk.
During solving the first problem, a natural solution to the second one appears, that
is, increasing the production of raw commercial milk, not losing, but increasing the
demand for it [1, 2].
In milk production, cooling is one of the most important factors. When cooled, the
bacteria in the milk do not die, but fall into suspended animation. When bacteria get
into comfortable temperature conditions, and this can occur at any time, including
during storage and transportation of milk, sometimes even 10 °C is enough, they begin
to procreate intensively, thereby affecting the quality of the product and, accordingly,
its cost [3–6].
Combining these two tasks, we can conclude that alternative methods and technical
means to kill bacteria are needed to solve them.
Analyzing the regulations of different economic zones and countries, we can see:
For the Russian Federation, Federal laws (FL) FL-88 and FL-163 “Technical
Regulations for Milk and Dairy Products” contain general and specific safety
requirements; State Standard (GOST) GOST R 52054–2003 “Cow’s milk raw. Spec-
ifications” contains general requirements for the production of all types of dairy
products [7].
For the European Economic Community (EEC) there is a Council Directive
92/46/EEC of 16 June 1992 laying down the health rules for the production and
placing on the market of raw milk, heat-treated milk and milk-based products; Council
Directive 2002/99/EC laying down the animal health rules governing the production,
processing, distribution and introduction of products of animal origin for human
consumption; Regulation (EC) No 852/2004 of the European Parliament and of the
Council of 29 April 2004 on the hygiene of foodstuffs; Regulation (EC) No 853/2004
of the European Parliament and of the Council of 29 April 2004 laying down specific
hygiene rules for food of animal origin; Regulation (EC) No 854/2004 of the European
Parliament and of the Council of 29 April 2004 laying down specific rules for the
organisation of official controls on products of animal origin intended for human
consumption; Regulation (EC) No 882/2004 of the European Parliament and of the
Council of 29 April 2004 on official controls performed to ensure the verification of
compliance with feed and food law, animal health and animal welfare rules [8, 9].
For the Eurasian Customs Union (EACU), there are technical regulations (TR):
“TR CU 033/2013. Technical Regulations of the Customs Union. On the Safety of
Milk and Dairy Products” [7].
The definition of the “quantity of mesophilic aerobic and facultative anaerobic
microorganisms” (QMFAnM) refers to the estimate of the size of the group of sanitary
indicative microorganisms. Different groups of microorganisms are part of the
QMFAnM: bacteria, fungi, yeast and others. Their total number indicates the sanitary
condition of the product and the degree of its insemination with microflora. The best
1312 G. N. Samarin et al.
Therefore, the goal of this work is to develop and justify the constructive and
technological parameters and modes of operation of the container-type milk decon-
tamination unit by the energy efficiency criteria.
Based on this goal, we have outlined the following objectives: to develop a tech-
nological scheme for milk decontamination with alternative variants allowing for
repeated exposure; to justify constructive and technological parameters and optimal
operation modes that ensure the required milk quality indicators at minimal energy
costs.
The description of the real processes of heat and mass exchange in the unit can be done
via mathematical modeling, therefore, the mathematical model will be defined by the
system of three Eqs. (1, 2, 3).
where W is energy consumption of the unit (pump and radiator), kWh; NP is power of
the drive of the pump supplying milk to the container (at small volumes of milk may
not be necessary), kW; s1 is operation time of the milk pump, s; RNu is the total power
consumed by the radiator of the unit, kW; s2 is total time of milk irradiation, s; f1, f2, f3
are functionalities; RGMi is the total fractional mass of milk, kg; DP is pressure drop
during pumping between the entrance and exit of the unit, Pa; FS is area of the cross-
section of the unit, m2; DnMB is the difference in the number of bacteria at the entrance
and exit of the unit, units; ηy0 is efficiency of the radiator; NG is power of the radiator
generator, kW; GA is mass of air that comes into contact with the unit, kg; tIM is initial
milk temperature, K; tIA is initial ambient air temperature, K; FV is surface area of the
container in the unit, m2; kP is the heat transfer coefficient, W/(m2K); ηt is thermal
efficiency of the radiator; lHi is fractional factor of freshly drawn milk; tKi is speed ratio
of souring of each fraction of milk when stored, 1/s; s3 is duration of milk storage, s;
GMi is mass of each fraction of milk, kg, hyI is depth of the radiator immersion; ηM is
bactericidal efficiency of the radiator.
As a target function of optimization of these parameters, we have set minimal
energy consumption while getting the longest duration of milk storage.
1314 G. N. Samarin et al.
3 Results
The purpose of this experiment was to determine the nature of milk souring in different
layers of liquid over time s. The study of layered freshly drawn milk souring in an open
container without mixing was conducted at ambient air temperature of about 20 °C.
During the experiment, we have measured three main parameters of milk: temperature
t, titratable acidity TA and number of CFU N in a unit volume of the liquid.
The results of studies of milk’s temperature and titratable acidity are presented in
Table 1.
The titratable acidity of milk is measured in degrees Therner, °Th.
Fig. 1. The dependence of the change in titratable acidity in the lower and upper milk layers in
the container at a certain temperature over time.
The results of temperature studies and CFU numbers in the unit volumes of liquid
are presented in Table 2.
Fig. 2. The dependence of changes in the number of bacteria in the lower and upper layers of
milk in the container at a certain temperature over time.
4 Discussion
The Eqs. (4) and (5) and the graphical interpretation (Fig. 1, 2) suggest that milk
souring in the container begins with the upper layer. Consequently, during the research
of milk treatment with ultrasound and microwave or ultraviolet radiation, the milk
should be treated top down.
The information obtained can be used in the following stages: the development of
new technological schemes for decontamination of milk using alternative methods and
new machines containing an optimized number of radiators, allowing milk to be
decontaminated repeatedly during processing.
Preconditions for Optimizing Primary Milk Processing 1317
5 Conclusions
In the course of the research conducted by the authors of the article, it was found that
the JSC “Krasnoe Znamya” farm located at the Novosokolniki district of the Pskov
region of the Russian Federation consumes 132 500 kWh during storage of 3 900 tons
of milk per year.
Milk in an open container begins to sour from its upper layer, therefore, the
required location of devices for preserving milk using alternative methods (ultrasound,
microwave radiation, ultraviolet and other options) was experimentally confirmed.
The milk souring process is described by regression Eqs. (4) and (5).
References
1. Samarin, G., Vasiliev, A.N., Vasiliev, A.A., Zhukov, A., Krishtopa, N., Kudryavtsev, A.:
Optimization of technological processes in animal husbandry. In: International Conference
on Efficient Production and Processing (ICEPP-2020), E3S Web Conferences, vol. 161,
p. 1094 (2020)
2. Samarin, G.N.: Energosberegayushchaya tekhnologiya formirovaniya sredy obitaniya
sel'skohozyajstvennyh zhivotnyh i pticy: monografiya [Energy-saving technology for
creating the habitat of farm animals and poultry: monograph]/G.N. Samarin. V. P. Goryachkin
Moscow state agrarian University, Moscow, 215 p. (2008) . (in Russian)
3. Boor, K.J.: Pathogenic Microorganisms of Concern to the Dairy Industry. . Dairy Food
Environ. Sanitation 17, 714–717 (1997)
4. Chambers, J.V.: The microbiology of raw milk. In: Robinson, R.K. (ed.) Dairy Microbiology
Handbook, 3rd edn, pp. 39–90 (2002). Wiley, New York
5. Ye, A., Cui, J., Dalgleish, D., et al.: Effect of homogenization and heat treatment on the
behavior of protein and fat globules during gastric digestion of milk. J. Dairy Sci. 100(1),
36–47 (2017)
6. Liang, L., Qi, C., Wang, X., et al.: Influence of homogenization and thermal processing on
the gastrointestinal fate of bovine milk fat: in vitro digestion study. J. Agric. Food Chem.
65(50), 11109–11117 (2017)
7. Coles, K.: Information Regarding US Requirements for the Importation of Milk and Dairy
Products/Washington Department of Agriculture. Food Safety Program. Milknews - News of
the dairy market [Electronic resource] – Electron. text data Moscow (2016). https://
milknews.ru/index/novosti-moloko_6294.html.
8. Bisig, W., Jordan, K., Smithers, G., Narvhus, J., Farhang, B., Heggum, C., Farrok, C.,
Sayler, A., Tong, P., Dornom, H., Bourdichon, F., Robertson, R.: The Technology of
pasteurisation and its effects on the microbiological and nutritional aspects of , p. 36. milk
International Dairy Federation, IDF, Brussels (2019)
9. Mack, G., Kohler, A.: Short and long-rung policy evaluation: support for grassland-based
milk production in Switzerland. J. Agric. Econ. 2018, 1–36 (2018)
10. Aiassa, E., Higgins, J.P.T., Frampton, G.K., et al.: Applicability and feasibility of systematic
review for performing evidence-based risk assessment in food and feed safety. Crit. Rev.
Food Sci. Nutr. 55(7), 1026–1034 (2015)
11. Finnegan, W., Yan, M., Holden, N.M., et al.: A review of environmental life cycle
assessment studies examining cheese production. Int. J. Life Cycle Assess. 23(9), 1773–
1787 (2018)
1318 G. N. Samarin et al.
12. Djekic, I., Miocinovic, J., Tomasevic, I., et al.: Environmental life-cycle assessment of
various dairy products. J. Cleaner Prod. 68, 64–72 (2014)
13. Depping, V., Grunow, M., van Middelaar, C., et al.: Integrating environmental impact
assessment into new product development and processing-technology selection: milk
concentrates as substitutes for milk powders. J. Cleaner Prod. 149, 1 (2017)
14. Bacenetti, J., Bava, L., Schievano, A., et al.: Whey protein concentrate (WPC) production:
environmental impact assessment. J. Food Eng. 224, 139–147 (2018)
15. Carlin, F.: Origin of bacterial spores contaminating foods. Food Microbiol. 28(2), 177–182
(2011). Special Issue SI
16. Coorevits, A., De Jonghe, V., Vandroemme, J., et al.: Comparative analysis of the diversity
of aerobic spore-forming bacteria in raw milk from organic and conventional dairy farms.
Syst. Appl. Microbiol. 31(2), 126–140 (2008)
17. Cusato, S., Gameiro, A.H., Corassin, C.H., et al.: Food safety systems in a small dairy
factory: implementation, major challenges, and assessment of systems’ performances.
Foodborne Pathog. Dis. 10(1), 6–12 (2013)
18. Doll, E.V., Scherer, S., Wenning, M.: Spoilage of Microfiltered and Pasteurized Extended
Shelf Life Milk Is Mainly Induced by Psychrotolerant Spore-Forming Bacteria that often
Originate from Recontamination. Front. Microbiol. 8, 135 (2017)
19. Boor, K.J., Murphy, S.C.: The microbiology of market milks. In: Robinson, R.K. (ed.) Dairy
Microbiology Handbook, 3rd edn, pp. 91–122. Wiley, New York, (2002)
20. Doyle, C.J., Gleeson, D., Jordan, K., et al.: Anaerobic sporeformers and their significance
with respect to milk and dairy products. Int. J. Food Microbiol. 197, 77–87 (2015)
21. O’Riordan, N., Kane, M., Joshi, L., et al.: Structural and functional characteristics of bovine
milk protein glycosylation. Glycobiology 24(3), 220–236 (2014)
22. Huck, J.R., Sonnen, M., Boor, K.J.: Tracking heat-resistant, cold-thriving fluid milk spoilage
bacteria from farm to packaged product. J. Dairy Sci. 91(3), 1218–1228 (2008)
23. Chavan, R.S., Chavan, S.R., Khedkar, C.D., et al.: UHT milk processing and effect of
plasmin activity on shelf life: a review. Compr. Rev. Food Sci. Food Saf. 10(5), 251–268
(2011)
24. Mafart, P., Leguerinel, I., Couvert, O., et al.: Quantification of spore resistance for
assessment and optimization of heating processes: a never-ending story. Food Microbiol.
27(5), 568–572 (2010)
25. Luecking, G., Stoeckel, M., Atamer, Z., et al.: Characterization of aerobic spore-forming
bacteria associated with industrial dairy processing environments and product spoilage. Int.
J. Food Microbiol. 166(2), 270–279 (2013)
26. Samarin, G.N., Vasilyev, A.N., Zhukov, A.A., Soloviev, S.V.: Optimization of microclimate
parameters inside livestock buildings. In: Vasant, P., Zelinka, I., Weber, G.W. (eds) Intel-
ligent Computing & Optimization. ICO 2018. Advances in Intelligent Systems and
Computing, vol 866. Springer, Cham (2018).
27. Özmen, A., Weber, G., Batmaz, İ.: The new robust CMARS (RCMARS) method. In:
International Conference 24th Mini EURO Conference “Continuous Optimization and
Information-Based Technologies in the Financial Sector” (MEC EurOPT 2010), Izmir,
Turkey, 23–26 June 2010 (2010).
Optimization of Compost Production
Technology
1 Introduction
One effective way to dispose of livestock waste is to produce organic composts based
on it. Peat, straw, sapropels, logging waste and organic waste from industrial plants, as
well as other ingredients, can be used as components of compost mixtures. The
additional costs of compost production are paid off by increasing crop yields and
improving soil fertility [1–5].
Modern composting is an aerobic process and consists of several stages: lag stage,
mesophilic, thermophilic and maturation. The timing of the compost maturation and the
quality of the fertilizers obtained depend on the nature of these processes. At the first
stage of composting, microorganisms “adapt” to the type of components of the mixture
and living conditions in the compost heap. At the same stage, organic components
begin to decompose, but the total population of microorganisms and temperature of the
mass are still small. In the second phase, the decomposition of substrates is increasing,
the number of microorganisms is constantly growing. At the beginning, simple sugars
and carbohydrates are decomposed, and from the moment of their depletion, bacteria
start processing cellulose proteins, while secreting a complex of organic acids neces-
sary to feed other microorganisms [6–11].
The third phase of composting is accompanied by a significant increase in tem-
perature caused by an increase in the number of thermophilic bacteria and their
metabolic products. The temperature of 55 °C is harmful for most pathogenic and
conditionally pathogenic microorganisms of humans and plants. However, it does not
affect aerobic bacteria, which continue the composting process: accelerated breakdown
of proteins, fats and complex carbohydrates. When food resources run out, the tem-
perature of compost mass begins to gradually decrease [7, 12–14].
In the fourth stage, mesophilic microorganisms begin to dominate the compost
heap, and the temperature indicates the onset of the maturation stage. In this phase, the
remaining organic matter forms complexes that are resistant to further decomposition.
They are called humic acids or humus. The resulting compost is easily loosened and
becomes free-flowing. Truck-mounted organic fertilizer spreaders can be used to dis-
perse it onto the fields [8, 14–17].
In order to produce quality organic composts and shorten their maturation, it is
necessary to thoroughly mix and grind the components of the mixture both during
laying down the heaps and at their ripening. In this case, the oxygen supply to different
areas of compost heaps is improved, the formation of “dead zones” is excluded,
microbiological processes are intensified. To perform these operations, special equip-
ment is needed: compost aerator-mixers [15, 18–20].
As promising ways of disposing of livestock waste we should note its composting
with the use of special technical means: aerator-mixers of compost heaps [21–23].
This sets us to formulate new goals of minimizing the cost of livestock products,
taking into account the reduction of negative impact on the environment, as well as the
development of promising ways and technical means for the utilization of livestock
waste.
In this article, the authors specify the prerequisites for optimizing compost production
technology in different ways [24–26]. The task of optimizing the economic parameters
of compost production technology in mathematical terms can be reduced to finding the
minimum value of the accepted target function: specific costs for the production of
compost SC′ [27–29]
3 Results
The results of the study of the parameters of the compost heaps agitator and the
experimental results are presented in Table 1.
Figure 2, 3, and 4 contain graphical representations of the experimental results.
1322 G. N. Samarin et al.
Table 1. The results of the study of the parameters of the aerator (agitator-mixer) of compost
heaps.
No Factor Natural factor Output parameter N, W h
variation values
levels
x n Vtr x n Vtr N1 N2 N3 Navg
x1 x2 x3 min−1 units km/h
1 1 1 0 230 6 0.2 62.98 64.29 69.54 65.60
2 −1 −1 0 170 2 0.2 48.60 49.60 52.10 50.10
3 1 −1 0 230 2 0.2 60.19 60.19 67.72 62.70
4 −1 1 0 170 6 0.2 52.28 51.74 57.67 53.90
5 1 0 1 230 4 0.25 77.72 74.58 83.21 78.50
6 −1 0 −1 170 4 0.15 45.70 46.65 50.46 47.60
7 1 0 −1 230 4 0.15 54.14 55.27 59.78 56.40
8 −1 0 1 170 4 0.25 61.75 63.05 70.20 65.00
9 0 1 1 200 6 0.25 71.38 68.50 76.43 72.10
10 0 −1 −1 200 2 0.15 45.01 44.08 50.11 46.40
11 0 1 −1 200 6 0.15 13.68 13.82 15.70 14.40
12 0 −1 1 200 2 0.25 3.84 3.72 4.19 3.92
13 0 0 0 200 4 0.2 5.92 6.05 6.54 6.17
14 0 0 0 200 4 0.2 6.30 6.23 6.55 6.36
15 0 0 0 200 4 0.2 5.93 6.05 6.35 6.11
Optimization of Compost Production Technology 1323
N, W h
80
75-80
75 70-75
70 65-70
65 60-65
55-60
60
50-55
55
b3 45-50
50 1
0,6 40-45
45
0,2
40 -0,2
-1 -0,6
-0.6
-0.2
0.2 -1
0.6
1 b1
Fig. 2. The dependence of the full power consumption of the compost agitator on the frequency
of drum rotation and the speed of mass supply.
N, W h
70 65-70
60-65
65
55-60
60 50-55
45-50
55
40-45
50
b2
1
45 0,6
0,2
40 -0,2
-1 -0,6
-0.6
-0.2
0.2 -1
0.6
1 b1
Fig. 3. The dependence of the full power consumption of the compost agitator on the frequency
of blades rotation and the number of tossing blades.
1324 G. N. Samarin et al.
N, W h
70-75
75
65-70
70 60-65
55-60
65
50-55
60 45-50
40-45
55
b3
50
1
45 0,6
0,2
40 -0,2
-1 -0,6
-0.6
-0.2
0.2 -1
0.6
1 b2
Fig. 4. The dependence of the full power consumption of the compost agitator on the number of
tossing blades and the speed of mass supply
4 Discussion
Taking into account the values of the coefficients of the obtained mathematical model,
analyzing the response surface (Fig. 2), we note that the increased influence on the
power consumed by the machine is exerted by the frequency of rotation of the agitator
drum and the rate of supply of compostable mass to the working body. With an
increase in these parameters, energy costs grow.
From the Fig. 3 we can see that the highest values of power consumed, required to
mix the compostable organic mass, are observed at the maximum rotation rate of the
agitator drum (230 min−1) and at the maximum rate of feeding the mass to the working
body (0.25 km/h).
Because of the dependence of energy costs (Fig. 4) on the frequency of the drum
rotation and the speed of feeding the compostable mass, the minimum energy intensity
is achieved at the minimum speed of movement of the unit and the minimum speed of
rotation of the agitator drum.
5 Conclusions
Based on the research conducted by the authors, the regression Eq. (2), which describes
the dependence of full power required for the process of mixing the components of
compost mixtures from the rotation rate x of the drum, the number of tossing blades n,
and the rate of feeding the organic mass onto the agitator drum.
The research will be used in the design of a three-dimensional model of the
upgraded aerator-mixer.
References
1. Abdalla, M.A., Endo, T., Maegawa, T., Mamedov, A., Yamanaka, N.: Effectiveness of
organic amendment and application thickness on properties of a sandy soil and sand
stabilization. J. Arid Environ. 183, 104273 (2020). https://doi.org/10.1016/j.jaridenv.2020.
104273
2. El-Haddad, M.E., Zayed, M.S., El-Sayed, G.A.M., Abd EL-Safer, A.M.: Efficiency of
compost and vermicompost in supporting the growth and chemical constituents of salvia
officinalis L. cultivated in sand soil. Int. J. Recycl. Org. Waste Agric. 9(1), 49–59 (2020).
https://doi.org/10.30486/IJROWA.2020.671209
3. Chen, T.M., Zhang, S.W., Yuan, Z.W.: Adoption of solid organic waste composting
products: a critical review. J. Cleaner Prod. 272, 122712 (2020). https://doi.org/10.1016/j.
jclepro.2020.122712
4. Alavi, N., Daneshpajou, M., Shirmardi, M., et al.: Investigating the efficiency of co-
composting and vermicomposting of vinasse with the mixture of cow manure wastes,
bagasse, and natural zeolite. Waste Manage. 69, 117–126 (2017)
5. Guo, H.N., Wu, S.B., Tian, Y.J., Zhang, J., Liu, H.T.: (2020). Application of machine
learning methods for the prediction of organic solid waste treatment and recycling processes:
a review. Bioresour. Technol. 319, p. 124114. https://doi.org/10.1016/j.biortech.2020.
124114.
1326 G. N. Samarin et al.
6. Kokunova, I.V., Kotov, E.G., Ruzhev, V.A.: Razrabotka klassifikacii tekhnicheskih sredstv
dlya proizvodstva organicheskih kom-postov (Development of classification of technical
means for the production of organic compost). In: The Role of Young Scientists in Solving
Urgent Problems of the Agro-Industrial Complex: Collection of Scientific Papers
International Scientific and Practical Conference. Saint Petersburg-Pushkin, pp. 179–182
(2018). (in Russian)
7. Akhtar, N., Gupta, K., Goyal, D. et al.: Recent advances in pretreatment technologies for
efficient hydrolysis of lignocellulosic biomass. Environ. Prog. Sustain. Energy 35(2),
pp. 489–511 (2016)
8. El-Sherbeny, S.E., Khalil, M.Y., Hussein, M.S. et al.: Effect of sowing date and application
of foliar fertilizers on the yield and chemical composition of rue (Ruta graveolens L.) herb.
Herba Polonica 54(1), 47–56 (2008)
9. Joshi, R., Singh, J., Vig, A.P.:Vermicompost as an effective organic fertilizer and biocontrol
agent: effect on growth, yield and quality of plants. Rev. Environ. Sci. Bio-Tech. 14(1), 137–
159 (2015)
10. Marinari, S., Masciandaro, G., Ceccanti, B. et al.: Influence of organic and mineral fertilizers
on soil biological and physical properties. Bioresour. Tech. 72(1), 9–17 (2000).
11. Hargreaves, J.C., Adl, M.S., Warman, P.R.: A review of the use of composted municipal
solid waste in agriculture. Agric. Ecosyst. Environ. 123(1–3), 14 (2008)
12. Ievinsh, G. (2011).Vermicompost treatment differentially affects seed germination, seedling
growth and physiological status of vegetable crop species. Plant Growth Regul. 65(1), 169–
181.
13. Bernstad, A.K., Canovas, A., Valle, R.: Consideration of food wastage along the supply
chain in lifecycle assessments: a mini-review based on the case of tomatoes. Waste Manag.
Res. 35(1), 29–39 (2017).
14. Buchmann, C., Schaumann, G.E.: The contribution of various organic matter fractions to
soil-water interactions and structural stability of an agriculturally cultivated soil. J. Plant
Nutr. Soil Sci. 181(4), 586–599.
15. Petrova, I.I., Kokunova, I.V.: Povyshenie effektivnosti vneseniya tverdyh organicheskih
udobrenij putem razrabotki novogo raspredelyayushchego ustrojstva dlya navozorazbrasy-
vatelya (Improving the efficiency of solid organic fertilizer application by developing a new
distribution device for the manure spreader of agricultural state), pp. 19–23 (2013). (in
Russian)
16. Hu, C., Xia, X., Chen, Y. et al.: Soil carbon and nitrogen sequestration and crop growth as
influenced by long-term application of effective microorganism compost. Chil. J. Agric. Res.
78(1), 13–22 (2018).
17. Case, S.D.C., Oelofse, M., Hou, Y., et al.: Farmer perceptions and use of organic waste
products as fertilizers - a survey study of potential benefits and barriers. Agric. Syst. 151,
84–95 (2017)
18. Kazimierczak, R., Hallmann, E., Rembialkowska, E.: Effects of organic and conventional
production systems on the content of bioactive substances in four species of medicinal
plants. Biol. Agric. Hortic. 31(2), 118–127 (2015)
19. Cerda, A., Artola, A., Font, X., et al.: Composting of food wastes: status and challenges.
Biores. Technol. 248, 57–67 (2018)
20. Dannehl, D., Becker, C., Suhl, J. et al.: Reuse of organomineral substrate waste from
hydroponic systems as fertilizer in open-field production increases yields, flavonoid
glycosides, and caffeic acid derivatives of red oak leaf lettuce (Lactuca sativa L.) much more
than synthetic fertilizer. J. Of Agric. Food Chem. 64(38), 7068–7075 (2016)
Optimization of Compost Production Technology 1327
21. Hwang, H.Y., Kim, S.H., Kim, M.S., Park, S.J., Lee, C.H...: Co-composting of chicken
manure with organic wastes: Characterization of gases emissions and compost quality. Appl.
Biol. Chem. 63(1), 3 (2020). https://doi.org/10.1186/s13765-019-0483-8
22. Awasthi, M.K., Pandey, A.K., Bundela, P.S. et al.: Co-composting of gelatin industry sludge
combined with organic fraction of municipal solid waste and poultry waste employing
zeolite mixed with enriched nitrifying bacterial consortium. Bioresour. Tech. 213, 181–189
(2016).
23. Cesaro, A., Belgiorno, V., Guida, M.: Compost from organic solid waste: quality assessment
and european regulations for its sustainable use. Resour. Conserv. Recycl. 94, 72–79 (2015)
24. Ben-Tal, A., Nemirovski, A.: Lectures on modern convex optimization: analysis, algorithms,
and engineering applications. MPR-SIAM Series on optimization. SIAM, Philadelphia
(2001)
25. Ben-Tal, A., Nemirovski, A.: Robust optimization - methodology and applications. Math.
Program. 92(3), 453–480 (2002)
26. Özmen, A., Weber, G.-W., Batmaz, İ: The new robust CMARS (RCMARS) method. In:
International Conference 24th Mini EURO Conference “Continuous Optimization and
Information-Based Technologies in the Financial Sector” (MEC EurOPT 2010), Izmir,
Turkey, 23–26 June 2010 (2010)
27. Samarin, G.N., Vasilyev, A.N., Zhukov, A.A., Soloviev, S.V.: Optimization of microclimate
parameters inside livestock buildings. In: Vasant, P., Zelinka, I., Weber, G.W. (eds) Intel-
ligent Computing & Optimization. ICO 2018. Advances in Intelligent Systems and
Computing, vol. 866. Springer, Cham (2018).
28. Samarin, G., Vasiliev, A.N., Vasiliev, A.A., Zhukov, A., Krishtopa, N., Kudryavtsev, A.:
Optimization of technological processes in animal husbandry. In: International Conference
on Efficient Production and Processing (ICEPP-2020), E3S Web Conferences, vol. 161,
p. 1094 (2020).
29. Carr, L., Grover, R., Smith, B., et al.: Commercial and on-farm production and marketing of
animal waste compost products Animal waste and the land-water interface, pp. 485–492.
Lewis Publishers , Boca Raton (1995)
30. Kranz, C.N., McLaughlin, R.A., Johnson, A., Miller, G., Heitman, J.L.: The effects of
compost incorporation on soil physical properties in urban soils - a concise review.
J. Environ. Manage. 261, 110209 (2020). https://doi.org/10.1016/j.jenvman.2020.110209
31. Beck-Broichsitter, S., Fleige, H., Horn, R.: Compost quality and its function as a soil
conditioner of recultivation layers - a critical review. Int. Agrophys. 32(1), 11–18 (2018).
Author Index
© The Editor(s) (if applicable) and The Author(s), under exclusive license
to Springer Nature Switzerland AG 2021
P. Vasant et al. (Eds.): ICO 2020, AISC 1324, pp. 1329–1332, 2021.
https://doi.org/10.1007/978-3-030-68154-8
1330 Author Index