Markov random field
In the domain of physics and probability, a Markov random field, Markov network or undirected graphical model is a set of random variables having a Markov property described by an undirected graph. In other words, a random field is said to be a Markov random field if it satisfies Markov properties. The concept originates from the Sherrington–Kirkpatrick model.
A Markov network or MRF is similar to a Bayesian network in its representation of dependencies; the differences being that Bayesian networks are directed and acyclic, whereas Markov networks are undirected and may be cyclic. Thus, a Markov network can represent certain dependencies that a Bayesian network cannot ; on the other hand, it can't represent certain dependencies that a Bayesian network can. The underlying graph of a Markov random field may be finite or infinite.
When the joint probability density of the random variables is strictly positive, it is also referred to as a Gibbs random field, because, according to the Hammersley–Clifford theorem, it can then be represented by a Gibbs measure for an appropriate energy function. The prototypical Markov random field is the Ising model; indeed, the Markov random field was introduced as the general setting for the Ising model. In the domain of artificial intelligence, a Markov random field is used to model various low- to mid-level tasks in image processing and computer vision.
Definition
Given an undirected graph, a set of random variables indexed by form a Markov random field with respect to if they satisfy the local Markov properties:The Global Markov property is stronger than the Local Markov property, which in turn is stronger than the Pairwise one. However, the above three Markov properties are equivalent for positive distributions.
The relation between the three Markov properties is particularly clear in the following formulation:
- Pairwise: For any not equal or adjacent,.
- Local: For any and not containing or adjacent to,.
- Global: For any not intersecting or adjacent,.
Clique factorization
Given a set of random variables, let be the probability of a particular field configuration in —that is, is the probability of finding that the random variables take on the particular value. Because is a set, the probability of should be understood to be taken with respect to a joint distribution of the.
If this joint density can be factorized over the cliques of as
then forms a Markov random field with respect to. Here, is the set of cliques of. The definition is equivalent if only maximal cliques are used. The functions are sometimes referred to as factor potentials or clique potentials. Note, however, conflicting terminology is in use: the word potential is often applied to the logarithm of. This is because, in statistical mechanics, has a direct interpretation as the potential energy of a configuration.
Some MRF's do not factorize: a simple example can be constructed on a cycle of 4 nodes with some infinite energies, i.e. configurations of zero probabilities, even if one, more appropriately, allows the infinite energies to act on the complete graph on.
MRF's factorize if at least one of the following conditions is fulfilled:
- the density is strictly positive
- the graph is chordal
Exponential family
Any positive Markov random field can be written as exponential family in canonical form with feature functions such that the full-joint distribution can be written aswhere the notation
is simply a dot product over field configurations, and Z is the partition function:
Here, denotes the set of all possible assignments of values to all the network's random variables. Usually, the feature functions are defined such that they are indicators of the clique's configuration, i.e. if corresponds to the i-th possible configuration of the k-th clique and 0 otherwise. This model is equivalent to the clique factorization model given above, if is the cardinality of the clique, and the weight of a feature corresponds to the logarithm of the corresponding clique factor, i.e., where is the i-th possible configuration of the k-th clique, i.e. the i-th value in the domain of the clique.
The probability P is often called the Gibbs measure. This expression of a Markov field as a logistic model is only possible if all clique factors are non-zero, i.e. if none of the elements of are assigned a probability of 0. This allows techniques from matrix algebra to be applied, e.g. that the trace of a matrix is log of the determinant, with the matrix representation of a graph arising from the graph's incidence matrix.
The importance of the partition function Z is that many concepts from statistical mechanics, such as entropy, directly generalize to the case of Markov networks, and an intuitive understanding can thereby be gained. In addition, the partition function allows variational methods to be applied to the solution of the problem: one can attach a driving force to one or more of the random variables, and explore the reaction of the network in response to this perturbation. Thus, for example, one may add a driving term Jv, for each vertex v of the graph, to the partition function to get:
Formally differentiating with respect to Jv gives the expectation value of the random variable Xv associated with the vertex v:
Correlation functions are computed likewise; the two-point correlation is:
Unfortunately, though the likelihood of a logistic Markov network is convex, evaluating the likelihood or gradient of the likelihood of a model requires inference in the model, which is generally computationally infeasible.
Examples
Gaussian
A multivariate normal distribution forms a Markov random field with respect to a graph if the missing edges correspond to zeros on the precision matrix :such that