Simulation of Relay Operation On Detection of Faulty Readings in Smart Meters
Simulation of Relay Operation On Detection of Faulty Readings in Smart Meters
A PROJECT REPORT
Submitted by
CATHERINE R 311116105015
SHOBANA P 311116105053
of
BACHELOR OF ENGINEERING
IN
CHENNAI
0
ANNA UNIVERSITY: CHENNAI 600 025
BONAFIDE CERTIFICATE
SIGNATURE:
DR.S.PRATHIBA
NUNGAMBAKKAM
CHENNAI-34
1
TABLE OF CONTENTS
ACKNOWLEDGEMENT 04
ABSTRACT 05
LIST OF TABLES 06
LIST OF FIGURES 07
LIST OF SYMBOLS 08
LIST OF ABBREVIATIONS 09
1 INTRODUCTION 1
1.3 BACKPROPAGATION 18
1.4 MATLAB 20
1
NON LINEAR AUTO REGRESSION WITH
1.9 EXOGENOUS INPUTS 29
ALGORITHM
THE NETWORK
NETWORK
PROCESSOR FUNCTIONS
NORMALISATION
AND SIMULATION
2
1.16.4 SETUP DIVISON OF DATA FOR TRAINING,
56
1.17 THE USE OF LM ALGORITHM OVER GD METHOD
58
1.18 EARLY PREDICTION NETWORK
59
1.19 GENERATING THE SIMULINK BLOCK FOR THE NET
61
1.20 THE NEURAL LIBRARY
64
1.21 BLOCK PARAMETERS
65
1.22 PREDICTED OUTPUT (DISPLAY BLOCK)
65
1.23 REMOVE ROWS BLOCK
67
1.24 OUTPUT PLOTS
68
1.25 RELAY STATE
2 LITERATURE REVIEW 69
3
ACKNOWLEDGEMENT
First and foremost, praises and thanks to the God, the Almighty,
for His showers of blessings throughout my project work to complete the
project successfully.
We express our sincere gratitude to our beloved Director
Rev.Dr.Maria Venish SJ, the Dean of Students Rev.Dr.Justin
Immanuel SJ, and the Dean of Engineering Education Mr Nicolas
Juhel for their continuous support and encouragement and also for
providing immense facilities in our Institution.
We proudly render our heartfelt thanks to our Principal Dr.Antony
Micheal Raj for the constant encouragement and support given by him
for the progress and completion of our project.
Our immense thanks to the Head of the Department
Ms.Dr.Prathiba for her effective leadership, encouragement and
guidance in this project.
I would like to express my deep and sincere gratitude to our project
supervisor Ms.Dr.Prathiba for her valuable ideas and constant
motivation throughout the course of our project.. It was a great privilege
and honor to work and do our project under her guidance.
We wish to extend our sincere thanks to all Faculty members and
the Lab Instructors of the Department of Electrical and Electronics
Engineering for their valuable suggestions and their kind cooperation for
the successful completion of our project.
Finally, our thanks go to all the people who have supported us to
complete our project work directly or indirectly.
4
ABSTRACT
Smart meters being the need for the electric utility in this
era as adoption of advanced metering infrastructure (AMI)
becomes more widespread, its appeal to cyber attackers will
undoubtedly increase, and addressing security vulnerabilities
across layers and by different stakeholders must be taken into
account from the outset. Smartgrid is increasingly benefiting
from intelligence-driven capability supported by machine
learning. The correct implementation of security best practices
prevents, as far as possible, breaches of complex systems, and
allows AMI component manufacturers to develop skills in the
area of cybersecurity which help protect customers from
potential and unwanted cyber-attacks. This project aims to use
neural networks, in detection of faulty or unusual readings of a
smart meter, and simulate using Matlab, the relay operation on
detection of faulty readings in the smart meter.
5
LIST OF TABLES :
6
LIST OF FIGURES:
7
LIST OF SYMBOLS:
1. ε - Error term
4. w - weight factor
5. v - Displacement factor
8
LIST OF ABBREVIATIONS:
4. AI - Artificial intelligence
9
CHAPTER-1
INTRODUCTION
1.1 SMART METER:
A smart meter (also known as advanced meters) is an electronic
device that records consumption of electric energy and communicates
the information to the electricity supplier for monitoring and billing.
Smart meters typically record energy hourly or more frequently, and
report at least daily. Grids with smart meters or “smart grids” attempt to
predict demand and react to rapid changes in demand and supply to
deliver efficient, reliable, and sustainable electric power.Smart meters
are part of a smart grid but the meters themselves do not comprise the
entire smart grid solution, but rather a part of the physical backbone of
the system.Such an advanced metering infrastructure (AMI) differs from
automatic meter reading (AMR) in that it enables two-way
communication between the meter and the supplier. Communications
from the meter to the network may be wireless, or via fixed wired
connections such as power line carrier (PLC). Wireless communication
options in common use include cellular communications, Wi-Fi,
wireless ad hoc networks over Wi-Fi, wireless mesh networks, low
power long range wireless , ZigBee , and Wi-SUN.
10
1.2 ARTIFICIAL INTELLIGENCE:
11
1.2.1 ARTIFICIAL NEURAL NETWORKS:
- Massive Parallelism
- Distributed representation
- Learning ability
- Generalization ability
- Fault tolerance
- Processing Units
- Topology
- Learning Algorithm
12
Artificial neural networks (ANN) or connectionist systems are
computing systems vaguely inspired by the biological neural networks
that constitute animal brains. Such systems "learn" to perform tasks by
considering examples, generally without being programmed with
task-specific rules. For example, in image recognition, they might learn
to identify images that contain cats by analyzing example images that
have been manually labeled as "cat" or "no cat" and using the results to
identify cats in other images. They do this without any prior knowledge
of cats, for example, that they have fur, tails, whiskers and cat-like faces.
Instead, they automatically generate identifying characteristics from the
examples that they process.
13
In ANN implementations, the "signal" at a connection is a real
number, and the output of each neuron is computed by some non-linear
function of the sum of its inputs. The connections are called edges.
Neurons and edges typically have a weight that adjusts as learning
proceeds. The weight increases or decreases the strength of the signal at
a connection. Neurons may have a threshold such that a signal is sent
only if the aggregate signal crosses that threshold. Typically, neurons are
aggregated into layers. Different layers may perform different
transformations on their inputs. Signals travel from the first layer (the
input layer), to the last layer (the output layer), possibly after traversing
the layers multiple times.A feedforward neural network is an artificial
neural network wherein connections between the nodes do not form a
cycle. As such, it is different from its descendant: recurrent neural
networks. Artificial neural networks (ANNs) are statistical models
directly inspired by, and partially modeled on biological neural
networks. They are capable of modeling and processing nonlinear
relationships between inputs and outputs in parallel.Artificial Neural
networks (ANN) or neural networks are computational algorithms. It
intended to simulate the behavior of biological systems composed of
“neurons”. A neural network is a machine learning algorithm based on
the model of a human neuron.
Applications:
- Pattern Classification
- Clustering/Categorization
- Function approximation
- Prediction/Forecasting
14
- Optimization
- Content-addressable Memory
- Control
1.2.2 FEEDFORWARD NEURAL NETWORK:
Data enters at the inputs and passes through the network, layer by
layer, until it arrives at the outputs. During normal operation, that is
when it acts as a classifier, there is no feedback between layers. This is
why they are called feedforward neural networks
15
threshold (typically 0) the neuron fires and takes the activated value
(typically 1); otherwise it takes the deactivated value (typically -1).
Neurons with this kind of activation function are also called artificial
neurons or linear threshold units.There are two main categories of
network architectures depending on the type of the connections between
the neurons, “feed-forward neural networks” and “recurrent neural
networks”. If there is no “feedback” from the outputs of the neurons
towards the inputs throughout the network, then the network is referred
to as “feed-forward neural network”. Otherwise, if there exists such
feedback,connection from the outputs towards the inputs then the
network is called a “recurrent neural network” .Usually, neural networks
are arranged in the form of “layers”. Feed-forward neural networks fall
into two categories depending on the number of the layers, either
1. Single layer feed-forward neural network
2. Multi-layer feed-forward neural network
16
A single-layer neural network represents the most simple form of
neural network, in which there is only one layer of input nodes that
send weighted inputs to a subsequent layer of receiving nodes, or
in some cases, one receiving node
17
1.2.3 RECURRENT NEURAL NETWORK:
A recurrent neural network (RNN) is a class of artificial neural
networks where connections between nodes form a directed graph along
a temporal sequence. This allows it to exhibit temporal dynamic
behavior. Derived from feedforward neural networks, RNNs can use
their internal state (memory) to process variable length sequences of
inputs. This makes them applicable to tasks such as unsegmented,
connected handwriting recognition or speech recognition.
The term "recurrent neural network" is used indiscriminately to
refer to two broad classes of networks with a similar general structure,
where one is finite impulse and the other is infinite impulse. Both
classes of networks exhibit temporal dynamic behavior. A finite impulse
recurrent network is a directed acyclic graph that can be unrolled and
replaced with a strictly feedforward neural network, while an infinite
impulse recurrent network is a directed cyclic graph that can not be
unrolled.
Both finite impulse and infinite impulse recurrent networks can
have additional stored states, and the storage can be under direct control
by the neural network. The storage can also be replaced by another
network or graph, if that incorporates time delays or has feedback loops.
Such controlled states are referred to as gated state or gated memory,
and are part of long short-term memory networks (LSTMs) and gated
recurrent units. This is also called Feedback Neural Network.
18
1.3 .3 BACKPROPAGATION:
In machine learning, specifically deep learning, backpropagation
(backprop, BP) is a widely used algorithm in training feedforward
neural networks for supervised learning. Generalizations of
backpropagation exist for other artificial neural networks (ANNs), and
for functions generally – a class of algorithms is referred to generically
as "backpropagation".Backpropagation is an algorithm used to calculate
derivatives quickly. Artificial neural networks use backpropagation as a
learning algorithm to compute a gradient descent with respect to
weights.The algorithm gets its name because the weights are updated
backwards, from output towards input.
Backpropagation is a short form for "backward propagation of
errors." It is a standard method of training artificial neural networks.
This method helps to calculate the gradient of a loss function with
respect to all the weights in the network. In deep learning,
backpropagation computes the gradient of the loss function with respect
to the weights of the network for a single input–output example, and
does so efficiently, unlike a naive direct computation of the gradient
with respect to each weight individually. This efficiency makes it
possible to use gradient methods for training multilayer networks,
updating weights to minimize loss; gradient descent, or variants such as
stochastic gradient descent, are commonly used. The backpropagation
algorithm works by computing the gradient of the loss function with
respect to each weight by the chain rule, computing the gradient one
layer at a time, iterating backward from the last layer to avoid redundant
calculations of intermediate terms in the chain rule; this is an example of
dynamic programming.
19
1.4 1.4 MATLAB:
MATLAB (matrix laboratory) is a multi-paradigm numerical
computing environment and proprietary programming language
developed by MathWorks. MATLAB allows matrix manipulations,
plotting of functions and data, implementation of algorithms, creation of
user interfaces, and interfacing with programs written in other
languages.
Although MATLAB is intended primarily for numerical
computing, an optional toolbox uses the MuPAD symbolic engine
allowing access to symbolic computing abilities. An additional package,
Simulink, adds graphical multi-domain simulation and model-based
design for dynamic and embedded systems.
MATLAB users come from various backgrounds of engineering,
science, and economics.
20
AUTOMOTIVE CONTROL SYSTEM
DESIGN AND ANALYSIS
21
COMPUTATIONAL CODE GENERATION
BIOLOGY
22
1.5 MACHINE LEARNING:
Machine learning is closely related to computational statistics,
which focuses on making predictions using computers. The study of
mathematical optimization delivers methods, theory and application
domains to the field of machine learning. In its application across
business problems, machine learning is also referred to as predictive
analytics.
Machine learning is the scientific study of algorithms and
statistical models that computer systems use to perform a specific task
without using explicit instructions, relying on patterns and inference
instead. It is seen as a subset of artificial intelligence. Machine
learning algorithms build a mathematical model based on sample data,
known as "training data", in order to make predictions or decisions
without being explicitly programmed to perform the task. Machine
learning algorithms are used in a wide variety of applications, such as
email filtering and computer vision, where it is difficult or infeasible to
develop a conventional algorithm for effectively performing the task.
Types of learning algorithms
1.Supervised learning
2.Unsupervised learning
3.Semi-supervised learning
4.Reinforcement learning
5.Self learning
1.Supervised learning:
23
Supervised learning algorithms build a mathematical model of a
set of data that contains both the inputs and the desired outputs. The data
is known as training data, and consists of a set of training examples.
Each training example has one or more inputs and the desired output,
also known as a supervisory signal.
2.Unsupervised learning:
3.Semi-supervised learning:
4.Reinforcement learning:
Reinforcement learning is an area of machine learning concerned
with how software agents ought to take actions in an environment so as
to maximize some notion of cumulative reward. Reinforcement learning
algorithms do not assume knowledge of an exact mathematical model of
the MDP, and are used when exact models are infeasible. Reinforcement
24
learning algorithms are used in autonomous vehicles or in learning to
play a game against a human opponent.
5.Self learning:
25
Figure 4 . Self learning block
26
Figure 5. Types of Machine Learning
27
1.8 SIMULATING THE DETECTION OF FAULT AND RELAY
OPERATION USING MATLAB
The following are the tools used for realisation of fault type
identification or in general anomaly detection in a smart power grid.
PROCESS FLOW
28
Figure 6 . Flow Diagram
1.9 NON LINEAR AUTO REGRESSION WITH EXOGENOUS
INPUTS:
I n time series modeling, a nonlinear autoregressive exogenous
model (NARX) is a nonlinear autoregressive model which has
exogenous inputs. This means that the model relates the current value of
a time series to both:
● past values of the same series; and
● current and past values of the driving (exogenous) series — that
is, of the externally determined series that influences the series
of interest.
In addition, the model contains:
● an "error" term
which relates to the fact that knowledge of other terms will not enable
the current value of the time series to be predicted exactly.
Such a model can be stated algebraically as
y_{t}=F(y_{t-1},y_{t-2},y_{t-3},\ldots,u_{t},u_{t-1},u_{t-2},u_{t-3},\l
dots )+\varepsilon _{t}}
Here y is the variable of interest, and u is the externally determined
variable. In this scheme, information about u helps predict y, as do
previous values of y itself. Here ε is the error term (sometimes called
noise). For example, y may be air temperature at noon, and u may be the
day of the year (day-number within year).
The function F is some nonlinear function, such as a polynomial. To test
for non-linearity in a time series, the BDS test
29
(Brock-Dechert-Scheinkman test) developed for econometrics can be
used.
NARX Network (narxnet, closeloop) in matlab
All the specific dynamic networks discussed so far have either been
focused networks, with the dynamics only at the input layer, or
feedforward networks. The nonlinear autoregressive network with
exogenous inputs (NARX) is a recurrent dynamic network, with
feedback connections enclosing several layers of the network. The
NARX model is based on the linear ARX model, which is commonly
used in time-series modeling. The next value of the dependent output
signal y(t) is regressed on previous values of the output signal and
previous values of an independent (exogenous) input signal.
30
for the approximation. This implementation also allows for a vector
ARX model, where the input and output can be multidimensional.
1.10 LEVENBERG-MARQUARDT ALGORITHM:
In mathematics and computing, the Levenberg–Marquardt
algorithm (LMA), also known as the Damped least-squares (DLS)
method, is used to solve non-linear least squares problems. These
minimization problems arise especially in least squares curve fitting.
The LMA is used in many software applications for solving
generic curve-fitting problems. However, as with many fitting
algorithms, the LMA finds only a local minimum, which is not
necessarily the global minimum. The LMA interpolates between the
Gauss–Newton algorithm (GNA) and the method of gradient descent.
The LMA is more robust than the GNA, which means that in many cases
it finds a solution even if it starts very far off the final minimum. For
well-behaved functions and reasonable starting parameters, the LMA
tends to be a bit slower than the GNA. LMA can also be viewed as
Gauss–Newton using a trust region approach.
The Levenberg-Marquardt algorithm is a very simple, but robust,
method for approximating a function. Basically, it consists in solving the
equation:
(Jt J + λI)δ = Jt E
31
our network weights to achieve a better solution. The Jt J matrix can also
be known as the approximated Hessian.
F(x,w) = y
Where x is the input vector presented to the network, w are the weights
of the network, and y is the corresponding output vector approximated
or predicted by the network. The weight vector w is commonly ordered
first by layer, then by neurons, and finally by the weights of each neuron
plus its bias.This view of network as an parameterized function will be
32
the basis for applying standard function optimization methods to solve
the problem of neural network training.
Where F(xi, w) is the network function evaluated for the i-th input vector
of the training set using the weight vector w and wj is the j- th element of
the weight vector w of the network.
33
Approximating the Hessian
H ≈ JtJ
approximated Hessian has not become singular. If this is the case, the
equation can still be solved by using a SVD decomposition.
34
After the equation is solved, the weights w are updated using δ and
network errors for each entry in the training set are recalculated. If the
new sum of squared errors has decreased, λ is decreased and the iteration
ends. If it has not, then the new weights are discarded and the method is
repeated with a higher value for λ.
35
1.11 BLOCK DIAGRAM FOR TRAINING USING LEVENBERG
MARQUARDT ALGORITHM:
36
1.12 NEURAL NETWORK TIME SERIES TOOL:
Syntax
ntstool
ntstool('close')
37
Description
ntstool opens the neural network time series tool and leads you through
solving a fitting problem using a two-layer feed-forward network.
ntstool('close')c loses the tool.
ntstool can be used to solve three different kinds of time series problems.
● In the first type of time series problem, you would like to predict
future values of a time series y( t) from past values of that time
series and past values of a second time series x( t) . This form of
prediction is called nonlinear autoregressive with exogenous
(external) input, or NARX, and can be written as follows:
y( t) = f(y( t – 1), ..., y( t – d) , x(t – 1), ..., (t – d) )
This model could be used to predict future values of a stock or
bond, based on such economic variables as unemployment rates,
GDP, etc. It could also be used for system identification, in which
models are developed to represent dynamic systems, such as
chemical processes, manufacturing systems, robotics, aerospace
vehicles, etc.
● In the second type of time series problem, there is only one series
involved. The future values of a time series y(t) are predicted only
from past values of that series. This form of prediction is called
nonlinear autoregressive, or NAR, and can be written as follows:
y( t) = f(y( t – 1), ..., y( t – d) )
This model could also be used to predict financial instruments, but
without the use of a companion series.
● The third time series problem is similar to the first type, in that two
series are involved, an input series x( t) and an output/target series
y( t). Here you want to predict values of y( t) from previous values
38
of x( t), but without knowledge of previous values of y(t). This
input/output model can be written as follows:
y( t) = f(x( t – 1), ..., x( t – d) )
The NARX model will provide better predictions than this
input-output model, because it uses the additional information
contained in the previous values of y( t). However, there may be
some applications in which the previous values of y( t) would not
be available. Those are the only cases where you would want to
use the input-output model instead of the NARX model.
39
4)V3 – Voltage of phase 3 (V)
5)W1 – Electricity consumption of phase 1 (Wh)
6)W2 – Electricity consumption of phase 2 (Wh)
7)W3 – Electricity consumption of phase 3 (Wh)
8)Virtual Apartment- Virtual Apartment ID
9)Date – Date in YYYY-MM-DD
10)Time – Time in HH:MM:SS
11)Power- Sum of W1 + W2 + W3 (Wh)
40
The following BibTex can be used to cite the dataset
@inproceedings{mammen2018want,
title={Want to Reduce Energy Consumption, Whom should we call?},
author={Mammen, Priyanka Mary and Kumar, Hareesh and
Ramamritham, Krithi and Rashid, Haroon},
booktitle={Proceedings of the Ninth International Conference on Future
Energy Systems},
pages={12–20},
year={2018},
organization={ACM}
}
41
42
The data is imported to the workspace and saved.
Targets are selected as ‘data’ with the time steps along a matrix
column and samples for each timestep along matrix row
43
1.14 NEURAL NETWORK TRAINING TOOL:
Syntax
nntraintool
nntraintool('close')
Description
nntraintool opens the training window GUI.
This is launched automatically when TRAIN is called.
44
1.15 VIEWING THE NEURAL NETWORK
● The NAR neural network considered here for training contains 2
layers with 10 hidden layers and 8 sublayers in the output layer.
● The samples are row vectors each of size 8 X 1.
● The input layer has a delay of 1.
SYNTAX
view(net);
45
Solve an Autoregression Time-Series Problem with a NAR
Neural Network
46
1.16.1 CHOOSE FEEDBACK PRE/POST-PROCESSING
FUNCTIONS:
Settings for feedback input are automatically applied
to feedback output
net.inputs{1}.processFcns
={'removeconstantrows','mapminmax'};
47
x2 = [rand(1,20)*5-1; rand(1,20)*20-10;
rand(1,20)-1];
y2 = mapminmax.apply(x2,settings)
48
Choose a Training Function
net.trainFcn = 'trainlm'; % Levenberg-Marquardt
net.trainFcn = 'traingdx'; %gradient descent method
Choose a Performance Function
net.performFcn = 'mse'; % Mean squared error
Choose Plot Functions
net.plotFcns={'plotperform','plottrainstate','plotresp
onse','ploterrcorr', 'plotinerrcorr'};
Train the Network
[net,tr]=train(net,inputs,targets,inputStates,layerSta
tes);
49
Test the Network
outputs = net(inputs,inputStates,layerStates);
errors = gsubtract(targets,outputs);
50
Figure 11. Performance plot
51
Plots figure, plotperform(tr)
52
figure, plottrainstate(tr)
figure, plotresponse(targets,outputs)
figure, ploterrcorr(errors)
53
plotinerrcorr(inputs,errors)
54
1.17 THE USE OF LM ALGORITHM OVER GD
METHOD:
At each iteration, the Levenberg–Marquardt Algorithm chooses
either the gradient descent or GN and updates the solution.
55
Disadvantages:
56
1.18 Early Prediction Network :
For some applications it helps to get the prediction a timestep early.The
original network returns predicted y(t+1) at the same time it is given
y(t+1).
For some applications such as decision making, it would help to have
predicted y(t+1) once y(t) is available, but before the actual y(t+1)
occurs.
The network can be made to return its output a timestep early by
removing one delay so that its minimal tap delay is now 0 instead of 1.
The new network returns the same outputs as the original network, but
outputs are shifted left one timestep.
nets = removedelay(net);
view(nets)
[xs,xis,ais,ts] = preparets(nets,{},{},targetSeries);
ys = nets(xs,xis,ais);
closedLoopPerformance = perform(net,tc,yc)
57
1.19 GENERATING THE SIMULINK BLOCK FOR THE
NETWORK
gensim
Generate a Simulink block diagram of a neural network.
gensim(net,'name','nntsnet');
58
'Name' - A string to be the new system's name.
59
1.20 THE NEURAL LIBRARY:
60
THE SIMULATION OF RELAY OPERATION ON DETECTING
FAULTY READING:
BLOCKS USED
Table:1 The table shows the list of blocks used for realizing the simulation.
61
Figure 15. Simulation block
62
1.21 BLOCK PARAMETERS:
RELAY STATE (VectorScope) :
The scope shows the plot of the relay state at a time span of 1000
frames.
63
1.22 PREDICTED OUTPUT(DISPLAY BLOCK):
64
FUNCTION BLOCK PARAMETERS:
RELAY:
The relay is switched on for threshold voltage of 200 volts and switched
off for any value below 10 volts.
65
1.24 OUTPUT PLOT:
66
1.25 RELAY STATE:
Figure 17 .Plot of the state of relay from time of start to 0.16 min.
The relay state transition occurs at time 0.065 min after detecting fault.
67
CHAPTER-2
2.1 LITERATURE REVIEW
● HD Beale, HB Demuth, MT Hagan - Pws, Boston, 1996 -
pdfs.semanticscholar.org-Designing Neural Networks - Evolving
Artificial Neural Network Proceedings of the IEEE, 1999 -
ieeexplore.ieee.org-Recurrent Neural Network and time series
prediction JT Connor, RD Martin, LE Atlas - … on neural
networks, 1994 - ieeexplore.ieee.org-Training Recurrent Neural
Networks With the Levenberg–Marquardt Algorithm.
● X Fu, S Li, M Fairbank, DC Wunsch… - IEEE transactions on …,
2014 - ieeexplore.ieee.org-Levenberg–Marquardt Backpropagation
Training of Multilayer Neural Networks for State Estimation of a
Safety-Critical Cyber-Physical System
● C Lv, Y Xing, J Zhang, X Na, Y Li, T Liu… - IEEE Transactions
2017 - ieeexplore.ieee.org-Neural networks in process fault
diagnosis
● T Sorsa, HN Koivo, H Koivisto - IEEE Transactions on systems
…, 1991 - ieeexplore.ieee.org-Smart attacks in smart grid
communication networks
● PY Chen, SM Cheng, KC Chen - IEEE Communications …, 2012
- ieeexplore.ieee.org
● A Very Brief Introduction to Machine Learning With Applications
to Communication Systems
● O Simeone - IEEE Transactions on Cognitive Communications …,
2018 - ieeexplore.ieee.org
● Artificial Neural Networks-Based Machine Learning for Wireless
Networks M Chen, U Challita, W Saad, C Yin… - … Surveys &
Tutorials, 2019 - ieeexplore.ieee.org
68
● Learning capability and storage capacity of two-hidden-layer
feedforward networks GB Huang - IEEE transactions on neural
networks, 2003 - ieeexplore.ieee.org
● AN EFFICIENT ELECTRICITY THEFT AND FAULT
DETECTION SCHEME IN DISTRIBUTION SYSTEM :
Muhammad Salik, Muhammad Fahad Zia, Fahad Ali School of
Science and Technology, University of Management and
Technology
● Rezaei, N., & Haghifam, M. R. (2008). Protection scheme for a
distribution system with distributed generation using neural
networks. International Journal of Electrical Power & Energy
Systems, 30(4), 235-241.
● NARMAX time series model prediction: feedforward and
recurrent fuzzy neural network approaches Y Gao, MJ Er - Fuzzy
sets and systems, 2005 - Elsevier
● A delay damage model selection algorithm for NARX neural
networks
● TN Lin, CL Giles, BG Horne… - IEEE Transactions on …, 1997 -
ieeexplore.ieee.org
69