Jump to content

Markov logic network

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Felix QW (talk | contribs) at 13:50, 2 February 2024 (Stub for a semantics section). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

A Markov logic network (MLN) is a probabilistic logic which applies the ideas of a Markov network to first-order logic, defining probability distributions on possible domains of a given signature. Markov logic networks generalize first-order logic, in the sense that, in a certain limit, all unsatisfiable statements have a probability of zero, and all tautologies have probability one.

History

In 2002, Ben Taskar, Pieter Abbeel and Daphne Koller introduced relational Markov networks as templates to specify Markov networks abstractly and without reference to a specific domain.[1][2] Work on Markov logic networks began in 2003 by Pedro Domingos and Matt Richardson.[3][4]

Description

Briefly, it is a collection of formulas from first-order logic, to each of which is assigned a real number, the weight. Taken as a Markov network, the vertices of the network graph are atomic formulas, and the edges are the logical connectives used to construct the formula. Each formula is considered to be a clique, and the Markov blanket is the set of formulas in which a given atom appears. A potential function is associated to each formula, and takes the value of one when the formula is true, and zero when it is false. The potential function is combined with the weight to form the Gibbs measure and partition function for the Markov network.

The above definition glosses over a subtle point: atomic formulas do not have a truth value unless they are grounded and given an interpretation; that is, until they are ground atoms with a Herbrand interpretation. Thus, a Markov logic network becomes a Markov network only with respect to a specific grounding and interpretation; the resulting Markov network is called the ground Markov network. The vertices of the graph of the ground Markov network are the ground atoms. The size of the resulting Markov network thus depends strongly (exponentially) on the number of constants in the domain of discourse.

Semantics

Together with a given domain, a Markov logic network defines a probability distribution on the set of all interpretations of its predicates on the given domain. It does so by inducing a Markov network whose nodes are ground atomic formulas.

Inference

The goal of inference in a Markov logic network is to find the stationary distribution of the system, or one that is close to it; that this may be difficult or not always possible is illustrated by the richness of behaviour seen in the Ising model. As in a Markov network, the stationary distribution finds the most likely assignment of probabilities to the vertices of the graph; in this case, the vertices are the ground atoms of an interpretation. That is, the distribution indicates the probability of the truth or falsehood of each ground atom. Given the stationary distribution, one can then perform inference in the traditional statistical sense of conditional probability: obtain the probability that formula A holds, given that formula B is true.

Inference in MLNs can be performed using standard Markov network inference techniques over the minimal subset of the relevant Markov network required for answering the query. These techniques include Gibbs sampling, which is effective but may be excessively slow for large networks, belief propagation, or approximation via pseudolikelihood.

See also

Resources

  1. ^ Cozman, Fabio Gagliardi (2020), "Languages for Probabilistic Modeling Over Structured and Relational Domains", A Guided Tour of Artificial Intelligence Research, Cham: Springer International Publishing, pp. 247–283, doi:10.1007/978-3-030-06167-8_9, ISBN 978-3-030-06166-1
  2. ^ Taskar, Ben; Abbeel, Pieter; Koller, Daphne (2002-08-01). "Discriminative probabilistic models for relational data". Proceedings of the Eighteenth conference on Uncertainty in artificial intelligence. UAI'02. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc.: 485–492. doi:10.5555/2073876.2073934. ISBN 978-1-55860-897-9. {{cite journal}}: Check |doi= value (help)
  3. ^ Domingos, Pedro (2015). The Master Algorithm: How machine learning is reshaping how we live. p. 246-7.
  4. ^ Richardson, Matthew; Domingos, Pedro (2006). "Markov Logic Networks" (PDF). Machine Learning. 62 (1–2): 107–136. doi:10.1007/s10994-006-5833-1.