Computational Neural Networks Driving Complex Analytical Problem Solving
Computational Neural Networks Driving Complex Analytical Problem Solving
Computational Neural Networks Driving Complex Analytical Problem Solving
Grady Hanrahan
KANJANA PATCHARAPRASERTSOOK
lytical problem solving abilities to meet the demands of
modern chemical research. (To listen to a podcast about
this article, please go to the Analytical Chemistry multi-
mediapageatpubs.acs.org/page/ancham/audio/index.html.)
a ) f(Wx + b)
NETWORK ARCHITECTURES
group demonstrated that neural networks with appropriate
Knowledge of network connectivity and layer arrangement comple-
architecture could be used to develop linear and nonlinear
ments our understanding of the basic operation of a computational
calibration models that performed as well as those developed
neuron and its ability to model complex analytical processes. The
by principal component regression or partial least squares.
modified Rosenblatt’s model of a neuron (Figure 2) is simple in
The form of nonlinear mapping that MLPs offer lends itself
structure but has great deal of computing potential. However, its
applications are limited because it only generates a binary output well to pattern recognition tasks common to both classification
with fixed weight and threshold values. Connecting multiple and regression. Traditionally, classification is accomplished using
neurons feeding forward to one output layer reveals the true one or several well-established techniques, including statistical
computing power of neural networks. Single-layer perceptrons are discriminant analysis, principal component analysis (PCA), K-NN,
easy to set up, train, and have explicit links to statistical models, SIMCA, and hierarchical clustering.19 A comprehensive review
e.g., sigmoid output functions that allow a link to posterior by Brown et al.20 noted, “The most novel research in pattern
probabilities. Nonetheless, they do have their limitations, including recognition involved work with artificial neural networks.” Work
the inability to solve exclusive disjunction (XOR) binary functions. by Baczek et al.21 used a neural network based on a MLP for
In 1969, Minsky and Papert created the multilayer perceptron evaluation of peptide MS/MS spectra in proteomics data. The
(MLP), the first solution to the XOR problem, by combining appropriately trained and validated neural network demonstrated
perceptron unit responses into a much more practical structure efficient classification of the manually interpreted MS/MS spectra
using a second layer of units.6 More recent work by Pina et al.15 as “good” or “bad” and enabled the automatic processing of large
developed simple (photo)chemical systems that, when performing amounts of MS/MS spectra, hence proving the utility of neural
as threshold devices and as XOR logic gates, were capable of networks in processing routine proteomic data. In terms of
integrating the effects of two inputs into a single output. calibration models, the previously mentioned study by Garcı́a-
MLPs consist of an input layer, one or more hidden layers, Reiriz et al.7 successfully used unfolded PCA, residual bilinear-
and an output layer and can be employed to approximate any ization, and neural networks based on radial basis functions to
continuous function. The generalized structure of a MLP with examine four-way kinetic-excitation-emission fluorescence data to
multiple inputs/outputs is presented in Figure 3. Close examina- determine levels of malonaldehyde in olive oil samples.
tion reveals one neuron in the input layer for each predictor In contrast to the strictly feedforward approach described
variable. If categorical variables are considered, N-1 neurons are above, recurrent neural networks (RNNs) have at least one
used to signify the N categories of the variable. The input layer feedback (closed loop) connection. Such networks have activation
standardizes the individual input values (x1.. .xR) to a range of feedback that symbolizes short-term memory, which allows them
-1 to 1. This layer distributes the values to each of the neurons to perform sequence recognition, sequence reproduction, and
in the hidden layer. In addition to the predictor variables, a temporal association. Feedback is modified by a set of weights to
constant input of 1.0 (bias) is fed to the hidden layer. In the enable automatic adaptation through a learning process (e.g.,
hidden layer, the value from each input neuron is multiplied backpropagation). A state layer is updated with the external input
by a given weight, and the resulting weighted values are of the network, as well as with activation from the previous forward
summed. The weighted sum is then fed into a selected transfer propagation. Internal states retain previous information and use
function with final distribution to the output layer. As an this historical record, or “memory”, when functioning under new
example study, Gemperline and colleagues compared weighted inputs. Interest in such memories has been spurred by the seminal
and unweighted linear calibration methods with second-order work of Hopfield in the early 1980s, which showed how a simple
calibration methods using quadratic principal component scores discrete nonlinear dynamical system could exhibit associative
and nonlinear calibration methods employing MLPs.18 The recall of stored binary patterns through collective computing.22,23
Analytical Chemistry, Vol. 82, No. 11, June 1, 2010 4309
reflectance NIR spectroscopy and SOMs. Latino and Aires-de-
Sousa31 performed classification of photochemical and metabolic
reactions by Kohonen SOMs and random forests by inputting the
difference between the 1H NMR spectra of the products and
the reactants.