Markov logic network
From Wikipedia, the free encyclopedia
A Markov logic network (or MLN) is a first-order knowledge base with a real number, or weight, attached to each formula. The weights associated with the formulas in an MLN jointly determine the probabilities of those formulas (and vice versa) via a log-linear model. An MLN defines a probability distribution over Herbrand interpretations (sometimes referred to as "possible worlds"), and can be thought of as a template for constructing Markov networks. Inference in MLNs can be performed using standard Markov network inference techniques over the minimal subset of the relevant Markov network required for answering the query. These techniques include Gibbs sampling, which is effective but may be excessively slow for large networks, loopy belief propagation, or approximation via pseudolikelihood.
Markov logic networks have been studied extensively by members of the Statistical Relational Learning group at the University of Washington.