Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Research On CDR

Download as pdf or txt
Download as pdf or txt
You are on page 1of 24

Conditional Random Fields

Rahul Gupta∗
(under the guidance of Prof. Sunita Sarawagi, KReSIT, IIT Bombay)

Abstract
In this report, we investigate Conditional Random Fields (CRFs), a family of conditionally trained
undirected graphical models. We give an overview of linear CRFs that correspond to chain-shaped
models and show how the marginals, partition function and MAP-labelings can be computed. Then,
we discuss various approaches for training such models - ranging from the traditional method of
maximizing the conditional likelihood or its variants like the pseudo likelihood to margin maximiza-
tion. For the margin-based formulation, we look at two approaches - the SMO algorithm and the
exponentiated gradient algorithm. We also discuss two other training approaches - one that attempts
at removing the regularization term and other that uses a kind of boosting to train the model.
Apart from training, we look at topics like the extension to segment level CRFs, inducing features
for CRFs, scaling them to large label sets, and performing MAP inferencing in the presence of
constraints.
From linear CRFs, we move on to arbitrary CRFs and discuss exact algorithms for performing
inferencing and the hardness of the problem. We go over a special class of models - Associative Markov
Networks, which are applicable in some real-life scenarios and which permit efficient inferencing. We
then look at collective classification as an application of general undirected models.
Finally, we very briefly summarize the work that could not be covered in this report and look at
possible future directions.

1 Undirected Graphical Models


Let X = X1 , . . . , Xn be a set of n random variables. Assume that p(X) is a joint probability distribution
over these random variables. Let XA and XB be two subsets of X which are known to be conditionally
independent, given XC . Then, p(.) respects this conditional independence statement if

p(XA |XB , XC ) = p(XA |XC ) (1)

or alternatively,

p(XA , XB , XC )
p(XA , XB |XC ) =
p(XC )
p(XA |XB , XC )p(XB , XC )
=
p(XC )
= p(XA |XC )p(XB |XC ) (2)

The shorthand notation for such a statement is : XA ⊥ XB |XC .


Given X and a list of such conditional independence statements, we would like to characterize the
family of joint probability distributions over X that satisfy all these statements. To achieve this, consider
an undirected graph G = (X, E) whose vertices correspond to our set of random variables. We would
construct the edge set E in such a manner that the following property holds: If the deletion of all vertices in
XC from the graph results in the removal of all paths from XA to XB , then XA ⊥ XB |XC . Conversely,
given an undirected graph G = (X, E), we can exhaustively enumerate all conditional independence
∗ grahul@it.iitb.ac.in

1
statements represented by it. However, note that the number of such statements can be exponential in
the number of vertices.
Let us restrict our attention to ’Markovian’ probability distributions. A probability distribution p(.)
is said to be Markovian w.r.t G and a set of vertices S if

p(S|S̄) = p(S|N (S)) (3)

where N (S) is the set of those neighbours of vertices in S which lie outside S. N (S) is often called the
Markovian blanket of S.
If p(.) is Markovian for all singleton sets S = {Xi }, then p(.) is said to be locally Markovian. If
p(.) is Markovian for all sets S ∈ 2X , then p(.) is globally Markovian. Trivially, a globally Markovian
distribution is also locally Markovian.
Hammersley and Clifford proved the following two theorems regarding Markovian distributions. The
proofs are available in [Cli90]. Here C is the set of all cliques in the graph.

Theorem 1. A locally Markovian distribution is also globally Markovian.

Theorem 2. P is Markovian iff it can be written in the form


X
P (X) ∝ exp( Q(C, X))
C∈C

In Theorem 2, Q(.) is an arbitrary real valued function that judges how likely is an assignment of
values to the random variables that form the clique vertices.
By summing over all possible assignments, we can remove the proportionality sign and write P (X)
as P
exp( C∈C Q(C, X))
P (X) = P P (4)
X exp( C∈C Q(C, X))

The denominator in Equation 4 is denoted as Z and is called the partition function.


The exponential form in Equation 4 allows us to write P (X) as a product :
Q
ψC (X)
P (X) = C (5)
Z
where ψC (X) = exp(Q(C, X)) is called the potential function for clique C.

Note: There is a slight abuse of notation here. Both Q and ψC do not take the entire assignment
X as input, but only the assignment restricted to the vertices in C.
The potential functions can be intuitively seen as preference functions over assignments to clique
vertices. A more probable assignment X = (x1 , . . . , xn ) is likely to have better contributions from most
of the constituent potential functions than a less probable assignment. However, the potential function
of a clique should not be confused with its marginal distribution. Infact, as we will see in Section 5.1,
potential function is just one of the terms that the marginal is proportional to.
This is one of the areas where undirected models score over directed models like MEMMs and HMMs.
Directed models have a ’probability mass conservation constraint’ that forces the local distributions to
be normalized to 1. Hence, they suffer from the the label bias problem ([LMP01]). In undirected models,
the local potential functions are unnormalized, and instead, global normalization is done using Z.

1.1 Conditional Random Fields


Consider a scenario where a hidden process is generating observables. Assume that the structure of the
hidden process is known. For example, in NER and POS tagging tasks, we make the assumption that a
particular POS tag (or named entity tag) depends only on the current word and the immediately previous
and the immediately next tags. This corresponds to an undirected graphical model in the shape of a
linear chain. Another example is the classification of a set of hyperlinked documents. The label of a

2
document can be assumed to be dependent upon the document itself and the labels of the documents
that link into it or out of it.
Two tasks arise in these scenarios:

1. Learning: Given a sample set of the observables {x1 , . . . , xN } along with the values of the hidden
labels {y1 , . . . , yN }, learn the best possible potential functions such that some criteria is maximized.

2. Inference: Given a new observable x, find the most likely set of hidden labels y∗ for x, i.e. compute
(exactly or approximately):
y∗ = arg max P (y|x) (6)
y

Here, the graphical model would have some nodes (say Yi ’s) and edges corresponding to the labels
and the dependencies between them and atleast one more node (say X) corresponding to the observable
x, along with some edges of the kind (X, Yi ). The joint probability distribution can thus be written as
1 Y
P (x, y1 , . . . , yM ) = ψ{X} (x) ψC (x, y) (7)
Z
C∈C,C6={X}

Learning this joint distribution is both intractable (because the ψ{X} (.) function is hard to approximate
without making naive assumptions) as well as useless (because x is already provided to us). Thus, it
makes sense to learn the following conditional distribution:
1 Y
P (y1 , . . . , yM |x) = ψC (x, y) (8)
Zx
C∈C,C6={X}

Note that the normalizer is now observable-specific.


The undirected graph with the set of nodes {X} ∪ Y and the relevant Markovian properties is called a
conditional random field (CRF). From now on, we will assume that C excludes the singleton clique {X}.

1.2 CRFs for sequence labeling


Before we move further, let us look at a special kind of CRFs, one where all the nodes in the graph
form a linear chain. Such models are extensively used in POS tagging, NER tasks and shallow parsing
([LMP01], [SP03]). For these models, the set of cliques, C, is just the set of all cliques of size 1 (viz. the
nodes) and the set of all cliques of size 2 (the edges). Thus, the conditional probability distribution can
be written as:
1 Y
P (Y1 , . . . , YM |X) = (ψi (Yi , X)ψi0 (Yi , Yi−1 , X)) (9)
Zx i

where ψ (.) acts over single labels and ψ 0 (.) acts over edges. Most sequence labeling applications param-
eterize ψi (.) and ψi0 (.) in a log-linear fashion.
X
ψi (.) = exp( θk sk (yi , x, i)) (10)
k
X
ψi0 (.) = exp( λj tj (yi−1 , yi , x, i)) (11)
j

where sk is a state feature function that uses only the label at a particular position, and tj is a transition
feature function that depends on the current and the previous label. Examples of some such functions
are: ”is the label NOUN and the current word capitalized?” and ”was the previous label SALUTATION,
current label PERSON and the current word in the dictionary of proper nouns?”. The parameters (Θ, Λ)
denote the importance of each of the features and are learnt during the learning phase by maximing some
criteria like conditional log likelihood.
For ease of notation, we will merge the node features with the edge features and use fj to denote the
j th feature function. Assume that there are a total of k feature functions. All the learnt parameters will

3
be merged into a single Λ vector (k × 1). Now consider the k × n matrix F where Fji = fj (yi , yi−1 , x, i).
Thus, the conditional probability of a given label sequence can be succintly written as
exp(ΛT F1n×1 )
P (y1 , . . . , yn |x) = (12)
Zx
The vector F1n×1 is called the global feature vector and is denoted as F(y, x). f (yi , yi−1 , x, i) will denote
the local feature vector at the ith position. The quantities exp(ΛT f (y, y 0 , x, i)) are often represented using
matrices Mi ’s whose rows and columns are indexed by labels.
Note that the normalizer of the conditional probability is independent of y, so during inferencing, we
have to compute y∗ such that :
y∗ = arg max ΛT .F(y, x) (13)
y

1.2.1 Forward and backward vectors

Since the space of possible label sequences is exponentially large in the size of the input, techniques like
dynamic programming are used, both in training as well as inferencing. Suppose that we are interested
in tagging a sequence only partially, say till the position i. Also, lets assume that the last label in this
partial labeling is some arbitrary but fixed y. Denote the unnormalized probability of a partial labeling
ending at position i with label y by α(y, i). Similarly, denote the unnormalized probability of a partial
segmentation starting at position i + 1 assuming a label y at position i by β(y, i).
α and β can be computed via the following recurrences:
X
α(y, i) = α(y 0 , i − 1). exp(ΛT f (y, y 0 , x, i)) (14)
y0
X
β(y, i) = β(y 0 , i + 1). exp(ΛT f (y 0 , y, x, i + 1)) (15)
y0

where f (., ., ., i) is the feature vector evaluated at the ith sequence position. The base cases are:

α(y, 0) = Jy = ‘start0 K (16)


β(y, n + 1) = Jy = ‘stop0 K (17)

α and β are called the forward and backward vectors respectively. We can now write the marginals and
partition function in terms of these vectors.

P (Yi = y|x) = α(y, i)β(y, i)/Zx (18)


P (Yi = y, Yi+1 = y 0 |x) = α(y, i) exp(ΛT f (y 0 , y, x, i + 1))β(y 0 , i + 1)/Zx (19)
X X
Zx = α(y, |x|) = β(y, 1) (20)
y y

1.2.2 Inference in linear CRFs using the Viterbi algorithm

In CRFs, training and inference are often interleaved. At each iteration during training, the system
computes its best estimate for labeling the training data and updates the model based on the error in
that estimate. Given the parameter vector Λ, the best labeling for a sequence can be found exactly using
the Viterbi algorithm.
For each tuple of the form (i, y), the Viterbi algorithm maintains the unnormalized probability of
the best labeling ending at position i with the label y. The labeling itself is also stored along with the
probability. Denoting the best unnormalized probability for (i, y) by V (i, y), the recurrence is:
(
maxy0 (V (i − 1, y 0 ). exp(ΛT f (y, y 0 , x, i))) (i > 0)
V (i, y) = (21)
Jy = startK (i = 0)
y max V (n,y)
The normalized probability of the best labeling is given by Zx and the labeling itself is given by
arg maxy V (n, y). Thus, if y can range over a set of m labels, then the runtime of the Viterbi algorithm
is O(nm2 ).

4
2 Training
The various methods used to train CRFs differ mainly in the objective function they try to optimize. We
look at the following methods to train a CRF.

1. The penalized log-likelihood criteria.

2. Pseudo log-likelihood.

3. Voted perceptron.

4. Margin maximization.

5. Gradient tree boosting.

6. Logarithmic pooling.

2.1 Penalized log-likelihood


The conditional log-likelihood of a set of training instances (xk , yk ) using parameters Λ is given by:
X
LΛ = ΛT .F(yk , xk ) − log ZΛ (xk ) (22)
k

The gradient of the log-likelihood is given by

F(y, xk ) exp(ΛT F(y, xk ))


P
y
X
k k
∇LΛ = (F(y , x ) − )
ZΛ (xk )
k
X X
= (F(yk , xk ) − F(y, xk )P (y|xk ))
k y
X
= (F(y , x ) − EP (y|xk ) [F(y, xk )])
k k
(23)
k

where E [.] is the expected value of the global feature vector under the conditional probability distribution.
Note that putting the gradient equal to zero corresponds to the maximum entropy constraint. This
is expected because CRFs can be seen as a generalization of logistic regression. Recall that for logistic
regression, the conditional distribution that maximizes the log-likelihood also has the maximum entropy,
assuming that the statistics in the training data are preserved. In both cases, this is made possible
because of the exponential form of the distribution, which is the only family of distributions to posess
such characteristics ([Ber]).
Like logistic regression, CRFs too suffer from the bane of overfitting. Thus, we impose a penalty on
large parameter values. The most popular technique imposes a zero prior on all the parameter values.
The penalized log-likelihood is given by (upto a constant):
X kΛk2
LΛ = (ΛT .F(yk , xk ) − log ZΛ (xk )) − (24)
2σ 2
k

and the gradient is given by


X Λ
∇LΛ = (F(yk , xk ) − EP (y|xk ) [F(y, xk )]) − (25)
σ2
k

The tricky term in the gradient is the expectation EP (y|xk ) [F(y, xk )] those computation requires the
enumeration of all the y sequences. Let us look at the j th entry in this vector, viz. Fj (.). Fj (y, xk ) is

5
fj (yi , yi−1 , xk , i). Therefore, we can rewrite EP (y|xk ) [Fj (y, xk )] as
P
equal to i
X
EP (y|xk ) [Fj (y, xk )] = EP (y|xk ) [ fj (yi , yi−1 , xk , i)]
i
X
= EP (y|xk ) [fj (yi , yi−1 , xk , i)]
i
T
f (y,y 0 ,xk ,i)
XX
= α(i − 1, y 0 ).fj (y, y 0 , xk , i).eΛ .β(i, y)
i y 0 ,y
X
T
= αi−1 Qi βi (26)
i

where αi , βi are the forward and backward vectors at position i, indexed by labels and Qi is a matrix
T 0 k
s.t. Qi (y 0 , y) = fj (y, y 0 , xk , i).eΛ f (y,y ,x ,i) .
Thus, after all the α, β vectors and Q matrices have been computed (only O(mn + km2 ) values), the
gradient can be easily obtained. Various iterative methods have been used to maximize the log-likelihood.
Some of them are :

1. Iterative Scaling and its variants like Improved Iterative Scaling, Generalized Iterative Scaling etc.

2. Conjugate Gradient Descent and its variants like Preconditioned Conjugate Gradient Descent and
Mixed Conjugate Gradient Descent.

3. Limited Memory Quasi Newton method (L-BFGS).

L-BFGS is a scalable second order method and has thus become the tool of choice in the past few years.
We briefly go over the basic algorithm. An outline of the other methods, as applied to CRFs, can be seen
in [LMP01], [Wal02] and [SP03].

2.1.1 L-BGFS

The standard Newton method uses second order derivatives to update the current guess of the optimum.
Using Taylor’s expansion, a function f can be approximated in a local neighbourhood of x as :
1
f (x + ∆) ≈ f (x) + ∆T ∇|x + ∆T H|x ∆ (27)
2
where ∇|x and H|x are the gradient and Hessian at x. Optimizing w.r.t. ∆, we get the Newton update
rule:
xk+1 = xk − ηH−1k ∇k (28)
The step-size η is computed via line-search methods or taken to be 1 for quadratic optimization problems.
However, when the dimensionality is large, computing the inverse of the Hessian is not feasible. So we
need methods to approximate the inverse and update this approximation at each iteration. Denoting
H−1
k by Bk , the BFGS update step gives such an approximation :

sk sTk ykT Bk yk 1
Bk+1 = Bk + ( + 1) − T (sk ykT Bk + Bk yk sTk ) (29)
ykT sk ykT sk yk sk

where yk = ∇k − ∇k−1 and sk = xk − xk−1 . B0 is usually taken to be a positive-definite diagonal matrix.


The BFGS update does away with the inverse computation, but we still have to store all the sk and yk
vectors of the previous iterations. The L-BFGS algorithm solves this problem by storing only θ(m) such
vectors, corresponding to the last m iterations. At the (m + i)th iteration, the vectors corresponding to
ith iteration are thrown away. To see this, note that the BFGS update step can be re-written as :
1
Bk+1 = (I − ρk sk ykT )Bk (I − ρk yk sTk ) + ρk sk sTk (where ρk = )
ykT s
= vkT Bk vk + ρk sk sTk (30)

6
Algorithm 1 ComputeDirection(k, {(si , yi )| k − 1 ≤ i ≤ k − m})
dk ← ∇k
for k − 1 ≤ i ≤ k − m do
βi ← ρi dTk si
dk ← dk − βi yi
end for
dk ← B0 dk
for k − m ≤ i ≤ k − 1 do
dk ← dk + (βi − ρi dTk yi )si
end for
return dk

Discarding the old vectors at the (m + i)th iteration is equivalent to making vi = I and ρi si sTi = 0n×n .
But we are not interested in explictly approximating Bk+1 . Rather, we just need to compute the direction
in which to update xk , viz. Bk ∇k (= dk say). Algorithm 1 shows how dk can be computed using the
stored values of si and yi .
L-BFGS has been experimentally shown to be a very practical second-order optimization algorithm on
real life problems. It has been shown to be considerably faster than conjugate gradient methods, which
are first order. In addition to the basic L-BFGS algorithm, a host of improvements have been suggested
to make it converge even faster. Some of them are :

1. After the direction dk is computed, the step-length η is computed using Wolfe conditions :

f (xk + ηdk ) ≤ f (xk ) + µη∇Tk dk (Objective decreases a lot)


|∇xk +ηdk | ≥ ν|∇k dk | (Curvature Condition)

Here µ and ν are pre-specified constants such that 0 ≤ µ ≤ 1 and µ ≤ ν ≤ 1. Usually a value of
η = 1 is checked for compliancy with Wolfe conditions before proceeding with line-search.
T
yk sk
2. In Algorithm 1, instead of B0 , a scaled version Bk0 = kyk k2 B0 is used.

2.2 Voted Perceptron Method


Perceptron uses an approximation of the gradient of the unregularized log-likelihood function. Recall
that the gradient is given by :
X
∇LΛ = (F(yk , xk ) − EP (y|xk ) [F(y, xk )]) (31)
k

Perceptron-based training considers one misclassified instance at a time, along with its contribution to
the gradient viz. (F(yk , xk ) − EP (y|xk ) [F(y, xk )]). The feature expectation is further approximated by a
point estimate of the feature vector at the best possible labeling. The approximation for the k th instance
can be written as :

∇LΛ ≈ (F(yk , xk ) − F(yk∗ , xk )) (yk∗ = arg max ΛT F(y, xk )) (32)


y

Note that this approximation is analogous to approximating a Bayes-optimal classifier with a MAP-
hypothesis based classifier. Using this approximate gradient, the following first order update rule can be
used for maximization :

Λt+1 = Λt + F(yk , xk ) − F(yk∗ , xk ) (33)


This update step is applied once for each misclassified instance xk in the training set and multiple
passes are made over the training corpus. However, it has been reported that the final set of parameters

7
obtained suffer from overfitting ([Col02]). To solve this, [Col02] suggests a voting scheme, where, in a
particular pass of the training data, all the updates are collected and their unweighted average is applied
as an update to the current set of parameters. The voted perceptron scheme has been shown to achieve
much lower errors in a much less number of iterations than the non-voted perceptron.

2.3 Pseudo Likelihood


So far we have been interested in maximizing the conditional probability of joint labelings. For a training
instance xi , yi , if the trained model predicts a labeling y other than yi then an error is said to have
occured. However, in many scenarios, we are willing to assign different error values to different labelings
y. For example, in case of POS tagging, a labeling which matches the training data labeling in all
positions except one is better than a labeling which matches in only a few positions.
Thus, for these scenarios, it makes sense to maximize the marginal distributions P (yti |x) instead of
P (yi |x). This objective is called the pseudo-likelihood and for the case of linear CRFs, it is given by :
i
X|
X t=|x
0
L (Λ) = log P (yti |xi , Λ) (34)
i t=1

The marginal distribution P (yti |xi , Λ) is given by :


X exp(ΛT F(y, xi ))
P (yti |xi , Λ) = (35)
i
ZΛ (xi )
y:yt =yt

and the gradient of L0 is


T
F(y,xi )
F(y, xi )eΛ X eΛT F(y,xi )
P
y:yt =yti
XX
∇ = ( − )
eΛT F(y,xi ) ZΛ (xi )
P
i t y:yt =yti y
XX
= (EP (y|x,Λ,yti ) [F(y, xi )] − EP (y|x,Λ) [F(y, xi )]) (36)
i t

The second expectation, which arises from the gradient of log ZΛ (xi ), can be computed as in the case of
log-likelihood, using forward and backward vectors. The k th component of the first expectation can be
rewritten as :
X X
EP (y|x,Λ,yti ) [ fk (yj , yj−1 , xi )] = EP (y|x,Λ,yti ) [fk (yj , yj−1 , xi )]
j j
X
= EP (yj ,yj−1 |x,Λ,yti ) [fk (yj , yj−1 , xi )]
j

The second identity holds because of the fact that Ep(A,B) [g(A)] = Ep(A) [g(A)]. Now, P (yj , yj−1 |x, Λ, yti )
can be computed directly using three recursively computed vectors viz. the α, β vectors and a new vector,
say γ. γ is defined as the partial unnormalized probability of starting at state i with label y and ending
at state j with label y 0 . Thus, γ can be computed as :
( P P 0 00
00 k λk fk (y ,y ,j−1,x)
0 y 00 γ(i, j − 1, y, y )e (i < j)
γ(i, j, y, y ) = (37)
Jy = y 0 K (i = j)

Note that γ can also be computed in a backward fashion. Using α, β and γ, we can obtain P (yj , yj−1 |x, Λ, yti )
as  P i
 α(t,yti )γ(t,j−1,yti ,yj−1 )e k λk fk (yj ,yj−1 ,j,x ) β(j,yj ) (t ≤ j − 1)
i i Z (xi)
P (yj , yj−1 |x , Λ, yt ) = P Λ
i (38)
 α(j−1,yj−1 )e k λk fk (yj ,yj−1 ,j,x ) γ(j,t,yj ,yti )β(t,yti ) (t ≥ j)
ZΛ (x)

However, computing these probabilities for all instances in a training corpus can require anywhere
from O(mn) to O(m2 n2 ) γ values. For a large or varied corpus, this is prohibitive and an alternate
Pt=|xi |
mechanism, as outlined in [KTR02], is used to directly compute t=1 P (yj , yj−1 |xi , Λ, yti ).

8
2.4 Max Margin Method
In this section, we look at an approach to train CRFs in a max-margin sense. Recall that the margin
is a measure of a classifier’s ability to contain any loss that it incurs while labeling data with a wrong
label. A classifier that achieves a larger margin while training is less likely to make errors than one with
a smaller margin.
In CRFs, we are dealing with structured classification, so it doesn’t make much sense to use a 0 − 1
loss function that penalizes all wrong labelings alike. Instead, a Hamming loss function that counts
the number of mislabelings is more intuitive. This loss function has the added advantage of being
decomposable. Now, let us define the margin criteria as follows:

ΛT (F(xi , yi ) − F(xi , y)) ≥ γL(i, y) ∀i, y 6= yi (39)


Here, γ is the margin that we want to be as high as possible and L(i, y) is the loss incurred when we
mislabel xi with y. As a shorthand, we will denote the differnce in global feature vector by ∆Fi,y . Thus,
we can write our optimization program as:

max γ s.t. ΛT ∆Fi,y ≥ γL(i, y) ∀i, y 6= yi (40)

or equivalently,
ΛT Λ
min s.t. ΛT ∆F(i, y) ≥ L(i, y) ∀i, y (41)
2
This is similar to the problem formulation in the case of SVMs for separable data. Carrying this analogy
forward to inseparable data, the quadratic program (QP) can be written as:

ΛT Λ X
min +C ξi
2
i
s.t. ΛT ∆F(i, y) ≥ L(i, y) − ξi ∀i, y (42)

ξi is the slack associated with the ith data instance. The correspond dual is given by:
X 1 X
max αi,y L(i, y) − | αi,y ∆Fi,y |2
i,y
2 i,y
X
s.t. αi,y = C ∀i
y
αi,y ≥ 0 ∀i, y (43)

The primal and dual optima are related to each other by:
X
Λ∗ = αi,y ∆Fi,y
i,y
X X
= F(xi , yi ) − αi,y F(xi , y) (44)
i i,y

Now because y has a structure, the number of primal constraints (and the dual variables) can be ex-
ponentially large. So, we cannot directly apply any optimization techniques to the primal or the dual
program. It is here that the decomposability of the loss function and ∆F comes to our rescue. Recall
that the global feature vector is just a sum over local feature vectors. Note that the first term in the dual
objective can be written as:
X XX X X
αi,y L(i, y) = αi,y L(i, yj ) = L(i, yj ) αi,y (45)
y y j j,yj y∼[yj ]

Here, y ∼ [yj ] means all those labelings y which assign a label yj to the j th position. Now note that
because of the dual program constraints, the α values behave like probabilities (that sum to C instead
P
of 1). So, the quantity y∼[yj ] αi,y can be seen as the marginal probability of having the label yj at the

9
j th position. We will denote this marginal by µi (yj ). Similarly, the second term in the dual objective
P
can be rewritten because of the decomposability of the global feature vector (∆Fi,y = j,k ∆Fi,yj ,yk ).
P
In this case, we have the pairwise marginals: µi (yj , yk ) = y∼[yj ,yk ] αi,y . The original dual can thus be
rewritten as:
X 1X X X 0
max µi (yj )L(i, yj ) − µi (yj , yk )µi0 (yj 0 , yk0 )f (xi , yj , yk )T f (xi , yj 0 , yk0 )
i,j,y
2 0
j 0 0i,i (j,k) (j ,k )
yj ,yk yj 0 ,yk0
X X
s.t. µi (yj , yk ) = µi (yk ), µi (yj ) = C, µi (yj , yk ) ≥ 0 (46)
yj yj

f (.) is the local feature vector that arises because of the decomposition of F(.). Hence, if there were N
training instances of length M each and |Y| possible labels for a particular word, then the original dual
with N |Y|M variables has been reduced to an equivalent form with just N M |Y|2 variables. Further, the
optimal solution for the primal can be computed from the optimal dual solution via:
X
Λ∗ = µi (yj , yk )∆f (xi , yj , yk ) (47)
i,(j,k),yj ,yk

Looking at Equation 46, it is clear that we can use the standard kernel trick as in SVMs, to compute the
dot product of the feature vectors as projected in a very high (possibly infinite) dimensional space. We
now briefly discuss two approaches to solve the max-margin formulation.

2.4.1 SMO Algorithm

The SMO algorithm for SVMs considers two α variables at a time, keeping their sum constant, so as
to obey the dual constraints. At each iteration, the algorithm optimally redistributes the mass between
the two chosen dual variables, keeping the other dual variables fixed. The next pair of dual variables are
chosen through a heuristic.
In our case, we cannot afford to materialize an exponential number of dual variables. So, we run a
variant of SMO as follows: we choose two µ variables based on some criteria. Then, using these two, we
generate two α variables. Due to the many-one dependence between α and µ, there are multiple choices for
the α vector. We choose a vector α which is consistent with the µ variables and has the maximum entropy.
The SMO algorithm modifies the generated pair of α’s and updates the corresponding µ variables.
If we choose to generate αi,y1 and αi,y2 and shift a mass  to the first variable, then the effect on an
explicit dual variable µi (yj , yk ) is:

µnew
i (yj , yk ) = µold 1 2 2 2
i (yj , yk ) + Jyj = yj , yk = yk K − Jyj = yj , yk = yk K (48)

The optimal value of  can be found in closed form and used to update the µ dual variables. The next
pair of variables can be chosen using any heuristic.

2.4.2 Exponentiated Gradient Algorithm

The generic exponentiated gradient algorithm is used to solve QPs with a positive-semidefinite coefficient
matrix. It applies positive multiplicative updates to the variables, thus ensuring their non-negativity all
the way. Consider the following QP (α = {α1,y1 , . . . , α2,y1 , . . . , αn,y1 , . . .}):

1
min J(α) = αT Aα + bT α
2
X
s.t. αi,y = 1 ∀i, αi,y ≥ 0 ∀i, y (49)
y

Algorithm 2 outlines the exponentiated gradient approach to solve this QP.


Note that this is a slightly different formulation from the one we saw earlier. Here, the αi variables
sum upto 1 rather than C. It is easy to outline the one-one correspondence between the formulation and

10
Algorithm 2 ExponentiatedGradient(A, b)
Choose any learning rate η > 0.
α1 ← Any feasible solution
for 1 ≤ t ≤ T do
∇t ← Aαt + b (gradient)
t+1 αti,y exp(−η∇ti,y )
∀i, y αi,y = P
αti,y0 exp(−η∇ti,y0 )
y0
end for
return αT +1

the quantities required by Algorithm 2.


X C X
J(α) = −C( αi,y Li,y − | αi,y ∆Fi,y |2 )
i,y
2 i,y
X C X X
= −C( αi,y Li,y − | F(xi , yi ) − αi,y F(xi , y)|2 )
i,y
2 i i,y
X
∇i,y = −CLi,y − C 2 F(xi , y)T (F(xi , yi ) − αi,y F(xi , y))
i,y

= −C(Li,y + ΛT F(xi , y)) (See Equation 44)

In the last identity we used the fact that Λ is our current estimate of the optima and we absorbed C
because the α values have been scaled down.
Now we still have the old problem of facing an exponential number of α variables, and once again,
the decomposability of the global feature vector and the loss function saves us. Also, note that because
of the exponential updates, it helps if we parameterize αi,. themselves in an exponential form.
P
exp( r∈R(xi ,y) θi,r )
αi,y = P P (50)
y0 exp( r∈R(xi ,y) θi,r )

Here R(xi , y) is the set of parts the loss function and the global feature vector decompose over. In the
case of linear CRFs, R(xi , y) is the set of nodes and edges of the chain whose labelings and local features
are consistent with y and F(xi , y) respectively. The number of θ variables is much less than those of the
α variables (the dominant term is governed by the size of the biggest part).
Instead of multiplicatively updating the α variables, we can additively update (a potentially much
less number of) θ variables at each iteration of Algorithm 2. The only hitch is computing the gradient,
or rather, computing Λt = C( i F(xi , yi ) − i,y αi,y t
F(xi , y)). The second term can be rewritten
P P
i i
P P P P
as i,y αi,y r∈R(xi ,y) f (x , r) = i,r∈R(xi , ) µi,r f (x , r). If we can calculate µi,r = i,y:r∈R(xi ,y) αi,y
easily, then the gradient can be efficiently computed. For the case of linear CRFs, µi,r is the marginal
probability of observing a particular label (or label pair) at a node (or an edge), using the current weight
vector.
Experimental evidence [BCTM04] shows that the exponentiated gradient algorithm ends up with a
better objective and doesn’t plateau out as much as the SMO algorithm.

2.5 Gradient Tree Boosting


The potential functions of CRFs belong to the exponential family of functions:

ψ(y, y 0 , x, i) = exp(φ(y, y 0 , x, i))

Gradient tree boosting learns φ(.)’s using functional gradient ascent. Functional gradient ascent allows
us to see how the objective function behaves as a function of φ. We begin with an initial guess of the
φ() functions (and thus, the feature weights). At each step of functional gradient ascent, we add a ’delta’
function to the current approximation of φ().

11
This ’delta’ function has no closed form, instead it is represented using regression trees. At the end
of M iterations, the functional approximation of a particular φ() is given by:
φM (y, y 0 , x, i) = φ0 (y, y 0 , x, i) + ∆1 + . . . + ∆M (51)
A big advantage of this approach is that it allows efficient induction of conjunctive features. In Section
4.2, we will look at a greedy feature induction mechanism. However, functional gradient ascent learns
one regression tree per iteration, and thus induces numerous simultaneous features per iteration.
The core issue in gradient tree boosting is estimating the delta function in each iteration. For a fixed
training sample, (x, y), the delta function’s value at (x, y) is the functional gradient of the conditional
likelihood of the sample.

∆m (x, y) = log P (y|x) (52)
∂φm−1
Given the value of ∆m at many such points, we can arrive at a representation of ∆m by learning a
regression tree hm that minimizes the square error i (hm (xi , yi ) − ∆m (xi , yi ))2 . One way to learn such
P

a regression tree is to use a variant of the CART algorithm. Overfitting can be avoided by stopping the
procedure at L leaves, where L is a preset parameter ([Fri01]).
In our scenario, the functional gradient of the conditional likelihood can easily be simplified ([DAB04]):
∂ X
0
(φ(yt , yt−1 , x, t) − log Z(x)) = Jyi−1 = y 0 & yi = yK − P (yi−1 = y 0 , yi = y|x)(53)
∂φ(y, y , x, i) t

where the probability term is equal to α(i − 1, y 0 ) exp(φ(y, y 0 , x, i))β(i, y)/Z(x). Note that the gradient’s
value is simply the error in our current estimation of the pairwise marginal.
Computationally, if we have N training samples of size n each, then we generate N |Y|2 n samples
to learn the delta functions. To scale the algorithm, [Fri01] suggests using sampling and discarding
small-valued delta-samples to cut down on the computational costs.
After learning the regression trees for all the φ’s, at testing time, given a sample x, we can compute
its best labeling by running a modified version of the Viterbi algorithm.

2.6 Logarithmic Pooling


Strictly speaking, logarithmic pooling is not an alternate way of training, rather an alternate way to
regularize CRFs. The standard way to avoid overfitting in CRFs is to impose a prior on the feature weights
(usually 01×k ) , and to penalize any deviation from this prior according to the Euclidean distance. The
intuition is that CRFs, like logistic regression, have a tendency to assign arbitrary large feature weights
when unregularized. Hence, like SVMs and logistic regression, a penalty term to counter this is included
in the objective function.
However, there are two issues while dealing with this kind of regularization. The penalty term is
2
usually of the form (Λ−Λσ2
0)
, thus forcing the user to select k + 1 parameters before starting the training.
Usually k, the number of features, is very large, and so searching through the hyperparameter space is
very difficult, even with cross-validation.
Logarithmic pooling ([SCO05]) tackles this problem by training multiple unregularized CRFs (in the
conventional manner) on the training data. At inference time, the predictions of these individual ’experts’
are combined using previously learnt weights. The combination is done by taking a weighted geometric
mean of the individual distributions:

(pi (y|x))wi
Q
p(y|x) = i (54)
ZLOP (x)
where pi (.) is the conditional probability provided by the ith expert and wi is the expert’s weight. As
before, ZLOP (x) is the partition function obtained by summing the probabilities to 1.
When p(.) is pooled using the above form, it can be shown that its KL distance from the true
distribution p∗ (.) is given by :
X X
KL(p∗ , p) = wi KL(p∗ , pi ) − wi KL(p, pi ) (55)
i i

12
Thus, to obtain a small distance between p and p∗ , we need individual component distributions that are
close to the true distribution but are diverse amongst themselves (and thus to p(.)). This is analogous to
bagging, where we need diverse classifiers in order for the combination distribution to be good.
Also, note that because the individual experts are combined using a product form, the resultant
distribution can be seen as one coming from a single CRF.
wi
Q
i (pi (y|x))
p(y|x) =
ZLOP (x)
Λi T Fi (y, x))
Q
i exp(wiQ
=
ZLOP (x) i ZΛi (x)wi
exp( i wi Λi T Fi (y, x))
P
= (56)
Z(x)

where Z(x) = ZLOP (x) i Zi (x)wi = y exp( i wi Λi T Fi (y, x)) and Fi is the global feature vector of
Q P P

the ith expert. Thus, the combined distribution has a feature vector of length M , where M is the number
of experts. The ith feature value is Λi T Fi (y, x). All the useful statistics, like feature value expectations,
can be computed using the usual dynamic programming based techniques that are used for normal CRFs.
Similarly, a minor variant of Viterbi can be used at inference time. The details for these are omitted,
and instead we focus on learning the weight vectors.

2.6.1 Learning component weights

The weights wi are learnt by maximing the log-likelihood of the training data. Note that before learning
these weights, we have already learnt the feature weights for the individual experts. The log-likelihood
of the training data is given by :
XX X
LL(w) = wi Λi T Fi (yj , xj )) − log Z(xj ) (57)
j i j

The ith component of the gradient is :


X
∇i = ΛTi Fi (yj , xj ) − Ep [ΛTi Fi (yj , xj ] (58)
j

Again, the gradient can be computed using the dynamic programming techniques mentioned before. The
standard gradient based methods such as conjugate gradient or LBFGS can now be used.
One thing to note here is that there is no regularization term in the likelihood function. Experiments
with the regularization of the weights using a Dirichlet prior suggest that very little is gained by imposing
any prior onto the weights. One reason why overfitting may not occur here inspite of the absence of
regularization is that the number of experts is typically low as compared to the number of features in
the original CRFs. Therefore the extent of overfitting seen during the training of feature weights may
not happen here.
Logarithmically pooled CRFs have been experimentally shown to be comparable or slightly better
than their regularized counterparts in NER and POS tagging tasks.

2.6.2 Choice of experts

As in bagging, the experts can be chosen in a variety of ways :

1. By exposing varying random subsets of the feature set available to the trainer.

2. By partitioning the features such that each set deals with either events only behind the current
position or only in the present or only in the future.

3. Partitioning the features by label instead of by sequence position.

13
3 Semi-Markov CRFs
Till now we have been dealing with CRFs that use features that depend only the previous label yi−1
(Markovian assumption), the current label yi , input string x and the position i. However for typical NER
and POS tasks, it often turns out that we wish to simultaneously assign the same label to a contiguous
chunk of words. Note that if we use the traditional word-based CRFs, then we cannot encode long-range
label dependencies without violating the Markovian assumption. A pragmatic solution is to mark an
entire segment of words together with the same label.
For that to happen, the trained model has to decide how long the chunk has to be and what label
will it get. As we are still in the realm of first-order Markovian dependence, the label of a segment can
depend only on the segment features and the label of the previous segment. Note that the family of
segment-level features is much more powerful than that of word-level features. Binary functions such as
J3 ≤ SegmentLength ≤ 5K cannot be encoded using word-level features.
Some notation before we proceed further. A segment si is denoted by a tuple (ti , ui , yi ) where ti and
ui are the start and end offsets and yi is the segment’s label. A segmentation comprises of consecutively
labeled segments and is denoted by s. Thus a feature function evaluation at the j th segment is a function
of (yj , yj−1 , x, tj , uj ). All partial segmentations that end at position i in the string and label the last
segment with y will be denoted by si:y . Similarly, all partial segmentations that start at position i + 1,
given that the label of the segment ending at i is y will be denoted by si:y . Since we are segmenting and
labeling simultaneously, we will use s to denote model output rather than y.
From here on we assume that the model rejects any segmentation whose maximum segment size is
more than L. The training time and Viterbi run-time will clearly be a function of L.

3.1 Forward and backward vectors


The forward and backward vectors can be naturally extended to work with segments.
X T 0
α(y, i) = eΛ F(s ,x)
s0 ∈si:y
T
F(s0 ,x)
X
β(y, i) = eΛ
s0 ∈si:y

Note that since s0 is a partial segmentatation, the global feature vector F(.) is computed only till the ith
position for α’s and only beyond i for β’s. The vectors can be calculated using a small variant of the
recursion discussed in Section 1.2.1.
min(L,i)
T
f (y,y 0 ,x,i−d+1,i)
X X
α(y, i) = α(y 0 , i − d)eΛ
d=1 y0
min(L,|x|−i)
T
f (y 0 ,y,x,i+1,i+d)
X X
β(y, i) = eΛ β(y 0 , i + d)
d=1 y0

The base cases are :

α(y, 0) = 1
β(y, |x| + 1) = 1
P
As before, the value of the partition function equals y α(y, |x|) and the marginal P ((i, j, y)|x) is equal
to y0 α(i − 1, y 0 ) exp(ΛT f (y, y 0 , i, j, x))β(j, y).
P

3.2 Inferencing
The Viterbi algorithm can be modified to look at all candidate segments and labels at each step. Denoting
the unnormalized probability of the best partial segmentation ending at position i with label y by V (i, y),

14
the recursion can be written as :

0 ΛT f (y,y 0 ,x,i−d+1,i)
 maxy0 ,d=1···L V (i − d, y ).e
 (i ≥ 1)
V (i, s) = 0 (i = 0) (59)

 −∞ (i < 0)

Ofcourse, we can work with log V (.) instead of V (.), in which case, we can replace the product with sum
and do away with the exponentation. The best overall segmentation is the one with score maxy V (|x|, y).
From the recursion, it is obvious that at each step we are doing L times more work as compared to
word-level CRFs. Since L ≤ |x|, we are still performing polynomial time inferencing.

3.3 Training
Using the penalized log-likelihood criteria, the gradient can be written as :
T
F(xj ,sj )
F(xj , sj )eΛ
P
X
s0 Λ
∇Λ = (F(xj , sj ) − )− (60)
j
ZΛ (xj ) σ2

The k th component of the second term is the expectation of the k th function’s global value. Consider
the partial unnormalized expectation ηk (i, y) = s0 fk (s0 , x) exp(ΛT F(s0 , x)), where the segmentations
P

s0 are restricted to the first i words. The required expectation is y ηk (|x|, y)/ZΛ (x). We can recursively
P

compute ηk as :
X T 0
ηk (i, y) = (ηk (i − d, y 0 )eΛ f (y,y ,x,i−d+1,i)
d,y 0
T
F(s0 ,x)+ΛT f (y,y 0 ,x,i−d+1,i)
X
+ eΛ fk (y, y 0 , x, i − d + 1, i))
s0 ∈s(i−d):y0
T
f (y,y 0 ,x,i−d+1,i)
X
= (ηk (i − d, y 0 ) + α(i − d, y 0 )fk (y, y 0 , x, i − d + 1, i))eΛ
d,y 0

4 Miscellaneous topics on linear CRFs


4.1 Scaling CRFs for medium-sized label sets
In the previous sections we had seen that computing any important statistic like the forward-backward
vectors, probability of the best labeling, or any marginal probability requires time proportional to O(|Y|2 ).
When the label set is large, this can significantly slow everything down, including training. [CSO05]
presents an approach to remove the |Y|2 term by training multiple CRFs over binary label sets. The
label set is represented by a group of possibly overlapping subsets. Let T1 , . . . , Ts be these subsets. A
label y is assigned the code b1 · · · bs where bi is one if y ∈ Ti . Now s different CRFs are trained to output
binary labels. For this, the training data is transformed into s different versions. In the ith version, a
label y is replaced by 1 if y ∈ Ti and 0 otherwise. Note that the runtimes of all these CRFs is independent
of |Y|2 .
The value of s is still set through engineering. If s is low then the number of binary CRFs is too
low to distinguish between |Y| labels. If s is too large, then many pairs of binary CRFs will be highly
correlated.
Consider the case where all strings are of length one. Then the problem reduces to multi-class
classification. In that case, the outputs of all the binary CRFs can be used to construct an s-bit string.
The final output is the label whose code has the least Hamming distance to this bit string. If the code of
the correct label is atleast a Hamming distance l from all other codes, then this allows upto l/2 individual
binary CRFs to be wrong.
We can extend this to the sequence labeling scenario in multiple ways:

15
1. The Viterbi labeling is obtained from each of the CRFs. Each such labeling is a binary string. At
each position j of the input string x, we construct an s-bit string from the corresponding j th bits
of the Viterbi strings. The j th position is assigned the label whose code is nearest to this bit string.
This approach is very efficient but doesn’t incorporate the confidence scores of the individual CRFs.

2. At each position j, all the CRFs output the marginal P (yj = 1|x), ∀j. This vector of marginal
is compared with all the codes (e.g. using the L1 -metric) and the label with the closest code is
outputted for the position j.

3. For each labeling y, each CRF Ci outputs its confidence P (bi (y)|x), bi (y) being the bit string
Q
conversion of y, specific to Ci . The overall best labeling is the one that maximizes i P (bi (y)|x).
The maxima can be found using a minor variant of Viterbi. Note that this is a special form of
logarithmic pooling (ref. Section 2.6), where all the weights are set to one.

Experimental evidence ([CSO05]) suggests that the last approach gives the lowest test error. However,
one open question is to efficiently select a good set of subsets. A greedy approach to select subsets in a
forward manner is infeasible because of the exponential size of the problem. [CSO05] discusses a code
selection heuristic that picks a code length which minimizes an upper bound on the error of the overall
classifier.

4.2 Efficient feature induction


So far we had assumed that the feature set has been fixed apriori for us. However, for many applications,
selecting a good subset of features is a highly non-trivial problem. Consider the case of entity extraction
or POS tagging, where linear CRFs can be used. Any feature depends on a potential label of a token, the
input string and optionally, the labels of the two neighbouring tokens. Typically, the features are boolean
which fire only when the current word and the corresponding labels fulfill some criteria, e.g. capitalization,
dictionary presence etc. .
More complex features can be constructed using the conjuncts of simpler features. However, the space
of even the simple features is very large (e.g. dictionary features) and so, the number of conjuncts is larger
than we can handle. One possible way is to use forward feature selection ([McC03]) that greedily learns
a good subset of features (atomic as well as conjuncts) for a given training set.
At each step, the feature selection scheme ranks the features by how much they will increase the
likelihood of the training data when added to the feature set. This ranking is updated at each step and
the algorithm stops when even the top ranked features don’t add much to the likelihood.
Such schemes have been used in text classification, where the ranking metric is not the increase in
likelihood but the increase in some classification score (e.g. F1). When we attempt at using this scheme
in the domain of linear CRFs, the following issues arise:

1. The number of features can be in millions or even more. So we can’t afford to select just one feature
per iteration. Also, the rank scores have to be computed very efficiently.

2. CRFs learn a weight vector for the feature set. On adding a new feature, a new vector has to be
learnt from scratch.

For ranking a feature, we need to compute the increase in likelihood on adding it. Strictly speaking,
the two likelihoods (before and after) use entirely different weight vectors, as produced by the training
algorithm. But for the sake of efficiency, the weights of the old features are assumed to be fixed and we
only optimize over the new feature’s weight. The ’gain’ score of a new feature g (with associated weight
µ) is thus given by:
GΛ (g) = max LLΛ+(g,µ) − LLΛ − µ2 /2σ 2 (61)
µ

16
In order to make the gain computations tractable, the likelihood is approximated by the pseudo
likelihood ([McC03]):
Y
PΛ+(g,µ) (y|x) ≈ PΛ+(g,µ) (yj |x)
j
P (y |x) exp(µg(yj , x, j))
P Λ j 0
Y
= 0
j y0 PΛ (yj |x) exp(µg(yj , x, j))
j

The optimal value of µ can now be computed using any iterative method. The algorithm in [McC03] uses
other optimizations to cut down on the computation costs. It calculates the gain only over those tokens
which were misclassified by the current weight vector. It also selects multiple features per iteration.
Further, after adding new feature(s), while retraining the CRF, it discards the last few iterations of the
LBFGS (or whatever method is used) to avoid overfitting.

4.3 Constrained Inferencing


So far we have performed unconstrained inferencing on CRFs. However, in many scenarios, we are inter-
ested in computing the best segmentation or labeling that satisfies a given set of constraints. Examples
of such constraints are:
• The number of occurences of a particular label y should be atleast (or atmost) k.

• If a particular label y is present then some other label y 0 should be present (or absent).

• A label ya should be followed be the label yb .

4.3.1 Constrained Viterbi

Only some of the constraints can be incorporated into the Viterbi algorithm. Specifically, constraints
that deal with adjacent labels can be dealt with by modifying the matrices Mi (ref. Section 1.2). For
example, the third constraint can be enforced by using the modified matrix Mi0 :
(
0 0 Mi (y 0 , y) J y 0 = ya & y = yb K
Mi (y , y) =
0 otherwise
In general, this approach can only work with ’local’ constraints that affect the labels of adjacent positions.
We next look at integer linear programs (ILP) that deal with both local as well as global constraints in
a natural way.

4.3.2 Integer linear programming based inferencing

We introduce ILP variables of the kind zi,yy0 (1 ≤ i ≤ |x|, y, y 0 ∈ Y) which is set to 1 if the word at
position i is labeled y 0 , with the previous label being y and 0 otherwise. These variables can be seen as
the nodes of a graph (called the trellis graph), where there is an edge between every pair of nodes of the
form (zi,yy0 , zi+1,y0 y00 ), with weight ΛT f (y 00 , y 0 , x, i + 1).
Computing the best constraint labeling for the string x is equivalent to finding the maxmimum weight
path from the start node to the end node in this trellis.
Any constraints, global or local, can be represented as linear constraints on the zi,yy0 ’s. The uncon-
strained labeling problem can easily be shown to be equivalent to the following ILP ([RY05]):
X
max ΛT f (y 0 , y, x, i)zi,yy0
z
i,y,y 0

X 00
X
s.t. zi−1,yy0 = zi,y0 y00 ∀i, y 0
y y
X X
z0,0y = 1, z|x|+1,y0 = 1
y y
zi,yy0 ∈ {0, 1}

17
Note that the LP relaxation of this ILP will still have an integral optima because of the total uni-
modularity of the constraint matrix.
User-defined constraints can be appended to this ILP to yield the best constrained solution. However,
the total unimodularity of the matrix can no longer be guaranteed. In that case, either the ILP must be
solved optimally or the solution to the LP relaxation be rounded intelligently.
As an example, the linear inequalities corresponding to the constraints discussed before are given by:
P P
1. y 0 zi,y y ≤ k.
0
i
P P
2. y 00 zi,y 00 y ≤ j,y 00 zj,y 00 y 0 ∀i.
P
3. y zi,yya ≤ zi+1,ya yb ∀i.

Examples of more complex constraints are given in [RY05]. Non ILP-based techniques like the A*
algorithm can also be used to perform constrained inferencing. However, they too suffer from having no
bounds on the running time.

5 Exact inference in arbitrary undirected models


Given an undirected model G = (V, E), we are often interested in computing the following quantities:
1. Marginal probability of a subset of nodes, P (YS ) where S ⊆ V .

2. Maximum a-posteriori configuration of the model (MAP configuration), i.e. arg maxy P (YV = y).

3. Conditional probability P (YS |YT ).


The task of computing conditional probabilities is reduced to that of computing marginals so let us focus
only on the first two problems.
Inference algorithms for general graphs fall into three major categories - (a) exact methods, (b)
sampling based methods, and (c) variational methods. Here, we look at two of the exact methods
viz. Sum Product algorithm and the Junction Tree algorithm.

5.1 Sum Product Algorithm


Let us first look at the case where the given graph is a tree (or a forest). In the later sections, we will
see how to adapt the inference techniques to general graphs.
Now, since the graph is a tree, we will work only with potentials over nodes and edges. The Sum-
Product algorithm (also called the message passing algorithm) begins by defining an ordering of the
vertices by rooting the tree at a node r. Then ’belief’ messages are passed from child nodes to their
parent nodes. A node v sends messages to its (unique) parent only when v has in turn received messages
from all its children. The marginal probability of the root node is proportional to the messages it receives
from its children. The message from node u to its parent v is denoted by muv (yv ). One can think that u
sends a message to v for each value of yv or sends a single message which is a function of yv . The message
is given by:
X Y
muv (yv ) = ψu (yu )ψuv (yu , yv ) mwu (yu ) (62)
yu w∈Ch(u)

where Ch(u) denotes the children nodes of u. On ’unrolling’ the messages, it can be seen that the message
passed by a node u to its parent v is simply the unnormalized contribution of the subtree rooted at u to
the marginal P (Yv = yv ). Therefore, once the root receives all the messages, its marginal is proportional
to their product.
Y
p(yr ) ∝ ψr (yr ) mur (yr )
u∈Ch(r)
Q
ψr (yr ) u∈Ch(r) mur (yr ) X Y
= , Z= ψr (yr ) mur (yr ) (63)
Z yr u∈Ch(r)

18
Let |Y| denote the number of possible values of a single label yu . Then it is obvious that θ(|E||Y|2 )
work is performed while sending all the messages. Also, the number of iterations taken by the algorithm
depends on the diameter of the tree.
The key thing to note here is that the intermediate messages once computed can be stored at their
destination nodes and be reused for computing any other marginal. Infact after running the message
passing algorithm over the entire tree, we can compute the marginal of any arbitrary node without doing
any extra work!

5.1.1 Viterbi Algorithm

The Sum Product algorithm can also be used to compute the MAP configuration (where it is called the
Viterbi algorithm). To achieve this, simply replace the sum by the max operator. Whenever node u
communicates with node v, instead of passing a message that quantifies its ’consolidated belief’ in yv , the
node will pass a message that says whats the best that it can do to be consistent with yv . The modified
message is: Y
muv (yv ) = max ψu (yu )ψuv (yu , yv ) mwu (yu ) (64)
yu
w∈Ch(u)

When these messages reach the root r, it knows what it is best possible configuration of other nodes given
yr . Thus, by iterating over yr , we can find the probability of the most probable global configuration (the
choice of root node doesn’t matter). Along with the max, we can also store the argmax, which can be
used to find the best assignment.
Note that in the case where the graph is a linear chain, the Viterbi algorithm reduces to the algorithm
given in Section 1.2.2. Further, note how the forward and backward vectors correspond to the belief
messages.

5.2 Junction Tree Algorithm


Let us see why the Sum Product algorithm fails on general graphs. Consider the graph G which is a
cycle in four nodes Y1 , . . . , Y4 . ’Rooting’ at node 1, the message that the root receives is:
X X X X
p(y1 ) ∝ ψ1 (y1 )[ ψ12 (y1 , y2 )ψ2 (y2 ) ψ3 (y3 )ψ23 (y2 , y3 )][ ψ14 (y1 , y4 )ψ4 (y4 ) ψ3 (y3 )ψ43 (y4 , y3 )]
y2 y3 y4 y3
(65)
while if we use the definition of marginal probability to compute p(y1 ), we get:
X Y
p(y1 ) ∝ ψ12 (y1 , y2 )ψ23 (y2 , y3 )ψ34 (y3 , y4 )ψ41 (y1 , y4 ) ψi (yi )
y2,3,4 i
X
= ψ1 (y1 )ψ2 (y2 )ψ4 (y4 )ψ12 (y1 , y2 )ψ14 (y1 , y4 )m3 (y2 , y4 )
y2,4
= ···

Clearly the two are not equal. The key problem is caused by the presence of multiple paths. The potential
of some nodes may reach their ancestors more than once (e.g. ψ3 (y3 ) in the example). Further, due to
the lack of acyclicity, there is no stopping criterion for the message passing algorithm. Although, one can
go ahead and execute the algorithm and empirically demonstrate acceptable convergence, examples can
be given where the algorithm doesn’t converge at all.
To tackle this problem, the natural approach is to first make the graph acyclic and then run the Sum-
Product algorithm on the transformed graph. The most intuitive step in this direction is to construct a
hypertree, whose nodes correspond to vertices of maximal cliques in the original graph. Two nodes in the
tree are adjacent if the corresponding cliques share vertices. Clique-wise division is performed because
the potentials are defined over cliques (or portions thereof), which helps us in defining the potential of a
node of a hypertree. The potential of a clique is the product of potentials defined over its subgraphs. It
should be ensured that a given potential is not assigned to two different clique nodes.

19
Note that two distant nodes V1 and V2 in the hypertree can have some common vertices, say T , in
the original graph. After we run the message passing algorithm on this tree, it is imperative that the
P P
marginals of T be equal, viz. P (T ) = V1 \T P (V1 ) = V2 \T P (V2 ). However, the algorithm ensures
only local consistency, viz. between adjacent nodes. Therefore, to obtain global consistency from local
consistency, we constrain the hypertree to have the running intersection property. This property states
that if two non-adjacent hypertree nodes V1 and V2 share a set of vertices T , then T should be present
in all the nodes in the (unique) path from V1 to V2 . Any hypertree that satisfies this property is called a
Junction-Tree.
It can be shown that the family of undirected graphs that possess a Junction-Tree is the family
of triangulated graphs (graphs where each cycle of length more than four has a chord). For example,
consider a 4-node cycle, with edge potentials specified. It is easily seen that any hypertree of this graph
will not satisfy the running intersection property. However, after the introduction of a chord, we have
two cliques of size three each. This can be trivially represented as a hypertree. Note that triangulation
can increase the size of the maximal clique of the graph. From now on, we can assume that the graph is
triangulated apriori.
The message passing protocol on a Junction-Tree works as follows:
X Y
mcc0 (yc0 ) = ψc (yc ) mc1 c (yc ) (66)
c c1 ∈Ch(c)

where ψc includes all the product of all the potentials associated with the clique c. The Viterbi algorithm
for this case is analogous:
Y
mcc0 (yc0 ) = max0 ψc (yc ) mc1 c (yc ) (67)
yc ∼yc
c1 ∈Ch(c)

where yc ∼ yc0 denotes that yc and yc0 assign the same values to the common vertices.
Junction-Tree algorithm’s running time depends on the size of the biggest clique in the triangulated
graph (or the treewidth, which is defined as one less than the size of the biggest clique). The runtime
is exponential in the treewidth, and consequently the onus of doing efficient inferencing lies on the
triangulation performed. Unfortunately, it is NP-hard to find a triangulation that achieves minimum
treewidth. Hence heuristics are used to come up with good enough triangulations.

5.3 Linear Programming based inferencing


Computation of the MAP labeling can also be done using Integer Linear Programming. Let C denote
the set of maximal cliques and c ∈ C be any maximal clique. Then, the binary variable µc (yc ) denotes
whether the clique vertices of c have the labeling yc or not. For a valid global assignment y, we need the
µ variables to be consistent. Moreover, for a fixed clique, only one of its µ variables can be 1. Thus, the
ILP for MAP labeling can be written as:
X
max µc (yc ) log ψc (yc )
c,y0 c
X
∀c, yc : µc (yc ) ∈ {0, 1}, µc (yc ) = 1
yc
X
∀c, yc , c0 ⊃ c : µc0 (y0 c0 ) = µc (yc ) (68)
y0 c0 :y0 c0 ∼yc

The LP relaxation of this program is obtained by letting the µ’s to vary between 0 and 1. Looking at
the type of constraints, it is clear that the µ variables behave like marginal probabilities over the clique
vertices. Assume that the graph was triangulated and the original ILP had a unique solution. In that
case, relaxing the ILP doesn’t change the optimal solution, viz. the marginals have all their ’probabilistic’
mass located at either 1 or 0. One possible way to prove this would be to demonstrate that the constraint
matrix is totally unimodular (determinant of every square submatrix is 0 or ±1). In case of multiple MAP
labelings, any convex combination of the optimal labelings would also be optimal for the LP relaxation.

20
Since c ranges over elements of C, one can immediately see the parallels between this formulation
and the Junction-Tree algorithm. For example, the consistency constraint is equivalent to the running
intersection property. However, while the ILP is valid for any arbitrary graph, the LP relaxation admits
invalid marginals as feasible solutions in the case of untriangulated graphs. This is analogous to the
absence of any Junction-Tree for untriangulated graphs. Hence, to deal with such cases, we have to
triangulate the graph, which translates to adding more µ variables (over potentially bigger cliques) and
more consistency constraints. Again, the number of extra variables (and constraints) created by this
process depends on the tree-width and solving the LP becomes more and more intractable with increasing
tree-width.

6 Associative Markov Networks


Recently, [TCK04] looked a restricted class of graphical models, called Associative Markov Networks
(AMN). In such models, the potential function over a clique favours common labeling of all the vertices in
that clique. Approximate inference in such models can be done very efficiently ([TCK04]), and polynomial
time exact inferencing is possible for the case of two labels. Note that such potentials can arise in real-life
tasks like classifying hyperlinked documents. In this case, linked documents would tend to have same
topic labels and the corresponding potential would be high. AMNs can be trained using the standard
max-margin formulation, so we focus on inferencing here. Consider potential functions of the form:
(
ηkc (yc = [k k . . . k k])
ψc (yc ) = (69)
1 otherwise

where ηkc ≥ 1. For AMNs, we can replace the generic inference variable µc (yc ) by µc (k), which is true
only if all the vertices in the clique c are labeled k. Thus, the global potential of a labeling y is given by:
Y Y
ψ(y) = ψv (i)µv (i) ψc (k)µc (k) (70)
v,i c,k

Also, note that if u, v ∈ c, then µc (k) ≤ µu (k) ∧ µv (k). Thus, for an AMN, the MAP labeling can be
computed using the following ILP:
X X
max µv (i) log ψv (i) + µc (k) log ψv (k)
v∈V,i c∈C\V,k
X
∀v, c, k : µv (k), µc (k) ∈ {0, 1}, ∀v : µv (i) = 1
i
∀c, v ∈ c, k : µc (k) ≤ µv (k) (71)

It can be shown that if the number of labels (= m) is 2, then the ILP can be relaxed and the optima
would still lie at an integral point. Infact, the LP relaxation for m = 2 can be reduced to a graph min-cut
which can be solved very efficiently using combinatorial methods. If m > 2, then the relaxation has an
1
optimum whose value is atleast |cmax | times the ILP optimum, |cmax | being the size of the biggest clique
in C.
For m > 2, we can perform an iterative min-cut. The algorithm begins with any arbitrary labeling of
the vertices. At step i, an ’i-expansion’ step is performed, where we do a min-cut on the vertex set. The
first partition of the min-cut preserves its current label while the vertices in the second partition switch to
label i. The min-cut is performed so that we achieve a maximum improvement in the objective. Repeated
min-cuts are performed by varying i, until we can no longer improve the objective. Experimental evidence
in [TCK04] shows that only a few iterations are sufficient for convergence in practice.

7 Collective Classification
Collective classification is the task of jointly assigning labels to a set of correlated entities. One example
is assigning topic labels to a set of hyperlinked webpages ([TAK02]) . The labels of linked webpages

21
are obviously correlated. Another example is information extraction while exploting the fact that the
individual extractions are correlated ([BM04]).
To achieve this, [TAK02] defines the notion of a clique template. A clique template chooses a set of
entities according to a user defined criteria. A clique between these entites is introduced in the Markov
field. For example, in the webpage classification task, we may introduce a clique between every pair
of documents that have a hyperlink between them, based on the belief that labels of linked pages are
similar. Similarly, we can introduce a clique between every pair of documents that are pointed to by
a common page. For the information extraction scenario, we may be interested in introducing a clique
between every pair of words (or phrases) which have a high textual similarity ([BM04]), mirroring our
belief that similar words should get similar labels..
Training these models through likelihood maximization leads to the same gradient as in Equation 25,
where the global feature vector F(.) is given by:
X
F(x, y) = f (xC , yC , C) (72)
C

Again, the tricky part is to compute the expectation of F() based on the current weight vector. In the
case of sequential models, we computed it using the forward-backward vectors. For acyclic models, we
can use the Sum-Product algorithm to compute marginal probabilities exactly. However, in the presence
of cycles (which is usually the case in such applications), the exact probabilities can be computed using
the Junction-Tree algorithm (ref. Section 5.2). The Junction-Tree algorithm provides us with marginals
of each clique which can be used to obtain the expectaton.
However, if the cliques are large, then the Junction-Tree algorithm is very expensive to use. In this
scenario, we can directly run a slightly modified version of the Sum-Product algorithm on the original
model. This variant is called loopy belief propagation ([Pea88]). Loopy belief propagation runs the
Sum-Product algorithm by initializing all messages to 1 and repeatedly passing the messages through the
cycles in the graph. Although there is no theoretical guarantee on the convergence of this algorithm, it
has been found to provide good approximations to the true marginals in experiments.
Both [TAK02] and [BM04] report a significant accuracy improvement by using undirected models
as against labeling the data using techniques like SVMs and logistic regression that don’t exploit label
correlation.

8 Other work
In this section, we briefly discuss the work that could not be covered in this report. From a training point
of view, there are atleast two other approaches. [QSM05] presents a method for Bayesian estimation of
the model parameters. The posteriors are approximated using a variant of the Expectation-Propagation
([Min01]) algorithm. Another method ([SM05]) divides the training task into multiple pieces of the graph.
Each piece is trained separately, using its own partition function. At testing time, the learned weights
are combined into a global weight vector.
Inference is a key problem in the usage of graphical models. Some of the training methods, like the
cutting plane algorithm for the max-margin formulation, rely on efficient inference algorithms for their
feasibility. However, for arbitrarily large tree-width graphs, inference is NP-hard. So, fast approximate
inferencing algorithms are of paramount importance. Broadly, the set of inexact algorithms can be divided
into three major categories.
The first family comprises of sampling based algorithms, like Monte Carlo Markov Chains, importance
sampling, or Gibbs Sampling. Gibbs sampling picks a random vertex at each iteration, and sets its
conditional probabilities, given its neighbours.
The second approach, called the variational approach, transforms the inference problem into an opti-
mization problem. The optimization is done using an objective like KL-distance, and it is done over a set
of simplified probability distributions. This gives a bound on the desired probability. Another transform

22
must be used to obtain the matching bound. The quality of bounds depends on the transformation and
the simplification enforced in the optimization.
The last set of approaches comes from the theory community. One approach ([NB04]) deals with
computing a subgraph of the given model, that is optimal in terms of KL-divergence and belongs to a
fixed family of graphs, e.g. subgraphs with d fewer edges than the original graph. In [KT99], a MAP
inference algorithm is presented for the case when the potentials are pairwise and metrics. The algorithm
has an approximation ratio of O(log k log log k), where k is the number of labels. The ratio further drops
to 2 if all the potentials are uniform.

References
[BCTM04] P. L. Bartlett, M. Collins, B. Taskar, and D. McAllester. Exponentiated gradient algorithms
for large-margin structured classification. In NIPs, 2004. http://www.stat.berkeley.edu/
~bartlett/papers/bcmt-lmmsc-04.ps.gz.

[Ber] A. Berger. A brief maxent tutorial. http://www.cs.cmu.edu/afs/cs/user/aberger/www/


html/tutorial/tutorial.

[BM04] R. Bunescu and R. Mooney. Collective information extraction with relational markov net-
works. In ACL, 2004. http://www.cs.stanford.edu/~btaskar/papers/bcmt-lmmsc-04.
ps.gz.

[Cli90] P. Clifford. Markov random fields in statistics. In G.R. Grimmett and D.J.A. Welsh (Eds)
Disorder in Physical Systems, J.M. Hammersley Festschrift, pages 19–32. Oxford University
Press, 1990. http://www.statslab.cam.ac.uk/~grg/books/hammfest/3-pdc.ps.

[Col02] M. Collins. Discriminative training methods for hidden markov models: Theory and ex-
periments with perceptron algorithms. In EMNLP, 2002. http://citeseer.ist.psu.edu/
collins02discriminative.html.

[CSO05] T. Cohn, A. Smith, and M. Osborne. Scaling conditional random fields using error correcting
codes. In ACL, 2005. http://www.cs.mu.oz.au/~tacohn/acl05_scaling.pdf.

[DAB04] T. Dietterich, A. Ashenfelter, and Y. Bulatov. Training conditional random fields via gradient
tree boosting. In ICML, 2004. http://citeseer.ist.psu.edu/dietterich04training.
html.

[Fri01] J. H. Friedman. Greedy function approximation: A gradient boosting machine. In Annals of


Statistics, volume 29, 2001. http://www-stat.stanford.edu/~jhf/ftp/trebst.ps.

[KT99] J. Kleinberg and E. Tardos. Approximation algorithms for classification problems with
pairwise relationships: Metric labeling and markov random fields. In FOCS, 1999. http:
//www.cs.cornell.edu/home/kleinber/focs99-mrf.ps.

[KTR02] S. Kakade, Y. Teh, and S. Roweis. An alternate objective function for markovian fields. In
ICML, pages 275–282, 2002. http://www.cs.berkeley.edu/~ywteh/research/newcost/
icml2002.pdf.

[LMP01] J. Lafferty, A. McCallum, and F. Pereira. Conditional random fields: Probabilistic models for
segmenting and labeling sequence data. In ICML, pages 282–289, 2001. http://citeseer.
ist.psu.edu/lafferty01conditional.html.

[McC03] A. McCallum. Efficiently inducing features of conditional random fields. In UAI, 2003.
http://citeseer.ist.psu.edu/mccallum03efficiently.html.

[Min01] T. Minka. Expectation propagation for approximate bayesian inference. In UAI, 2001. http:
//research.microsoft.com/~minka/papers/ep/minka-ep-uai.pdf.

23
[NB04] M. Narasimhan and J. Bilmes. Optimal sub-graphical models. In NIPS, 2004. http://ssli.
ee.washington.edu/~mukundn/pubs/nips2004.pdf.

[Pea88] J. Pearl. Probabilistic reasoning in intelligent systems. Morgan Kaufmann, San Francisco,
1988.

[QSM05] Y. Qi, M. Szummer, and T. P. Minka. Bayesian conditional random fields. In AISTATS,
2005. http://people.csail.mit.edu/alanqi/papers/Qi-Bayesian-CRF-AIstat05.pdf.

[RY05] D. Roth and W. Yih. Integer linear programming inference for conditional random fields. In
ICML, pages 737–744, 2005. http://l2r.cs.uiuc.edu/~danr/Papers/RothYi05.pdf.

[SCO05] A. Smith, T. Cohn, and M. Osborne. Logarithmic opinion pools for conditional random fields.
In ACL, 2005. http://www.iccs.informatics.ed.ac.uk/~osborne/papers/acl05a.pdf.

[SM05] C. Sutton and A. McCallum. Piecewise training for undirected models. In UAI, 2005. http:
//www.cs.umass.edu/~mccallum/papers/piecewise-uai05.pdf.

[SP03] F. Sha and F. Pereira. Shallow parsing with conditional random fields. In NAACL, 2003.
http://www.cis.upenn.edu/~feisha/pubs/shallow03.pdf.

[TAK02] B. Taskar, P. Abbeel, and D. Koller. Discriminative probabilistic models for relational data.
In UAI, 2002. http://www.cs.berkeley.edu/~taskar/pubs/rmn.ps.

[Tas04] B. Taskar. Learning Structured Prediction Models: A Large Margin Approach. PhD thesis,
Stanford University, 2004. http://www.cs.berkeley.edu/~taskar/pubs/thesis.pdf.

[TCK04] B. Taskar, V. Chatalbashev, and D. Koller. Learning associative markov networks. In ICML,
2004. http://www.cs.berkeley.edu/~taskar/pubs/mmamn.ps.

[Wal02] H. Wallach. Efficient training of conditional random fields. Master’s thesis, University of
Edinburgh, 2002. http://citeseer.ist.psu.edu/wallach02efficient.html.

24

You might also like