Constrained conditional model

A constrained conditional model (CCM) is a machine learning and inference framework that augments the learning of conditional (probabilistic or discriminative) models with declarative constraints. The constraint can be used as a way to incorporate expressive[clarification needed] prior knowledge into the model and bias the assignments made by the learned model to satisfy these constraints. The framework can be used to support decisions in an expressive output space while maintaining modularity and tractability of training and inference.

Models of this kind have recently[when?] attracted much attention[citation needed] within the natural language processing (NLP) community. Formulating problems as constrained optimization problems over the output of learned models has several advantages. It allows one to focus on the modeling of problems by providing the opportunity to incorporate domain-specific knowledge as global constraints using a first order language. Using this declarative framework frees the developer from low level feature engineering while capturing the problem's domain-specific properties and guarantying exact inference. From a machine learning perspective it allows decoupling the stage of model generation (learning) from that of the constrained inference stage, thus helping to simplify the learning stage while improving the quality of the solutions. For example, in the case of generating compressed sentences, rather than simply relying on a language model to retain the most commonly used n-grams in the sentence, constraints can be used to ensure that if a modifier is kept in the compressed sentence, its subject will also be kept.

Motivation

edit

Making decisions in many domains (such as natural language processing and computer vision problems) often involves assigning values to sets of interdependent variables where the expressive dependency structure can influence, or even dictate, what assignments are possible. These settings are applicable not only to Structured Learning problems such as semantic role labeling, but also for cases that require making use of multiple pre-learned components, such as summarization, textual entailment and question answering. In all these cases, it is natural to formulate the decision problem as a constrained optimization problem, with an objective function that is composed of learned models, subject to domain- or problem-specific constraints.

Constrained conditional models form a learning and inference framework that augments the learning of conditional (probabilistic or discriminative) models with declarative constraints (written, for example, using a first-order representation) as a way to support decisions in an expressive output space while maintaining modularity and tractability of training and inference. These constraints can express either hard restrictions, completely prohibiting some assignments, or soft restrictions, penalizing unlikely assignments. In most applications of this framework in NLP, following,[1] Integer Linear Programming (ILP) was used as the inference framework, although other algorithms can be used for that purpose.

Formal Definition

edit

Given a set of feature functions   and a set of constraints  , defined over an input structure   and an output structure  , a constraint conditional model is characterized by two weight vectors, w and  , and is defined as the solution to the following optimization problem:

 .

Each constraint   is a boolean mapping indicating if the joint assignment   violates a constraint, and   is the penalty incurred for violating the constraints. Constraints assigned an infinite penalty are known as hard constraints, and represent unfeasible assignments to the optimization problem.

Training paradigms

edit

Learning local vs. global models

edit

The objective function used by CCMs can be decomposed and learned in several ways, ranging from a complete joint training of the model along with the constraints to completely decoupling the learning and the inference stage. In the latter case, several local models are learned independently and the dependency between these models is considered only at decision time via a global decision process. The advantages of each approach are discussed in [2] which studies the two training paradigms: (1) local models: L+I (learning + inference) and (2) global model: IBT (Inference based training), and shows both theoretically and experimentally that while IBT (joint training) is best in the limit, under some conditions (basically, ”good” components) L+I can generalize better.

The ability of CCM to combine local models is especially beneficial in cases where joint learning is computationally intractable or when training data are not available for joint learning. This flexibility distinguishes CCM from the other learning frameworks that also combine statistical information with declarative constraints, such as Markov logic network, that emphasize joint training.

Minimally supervised CCM

edit

CCM can help reduce supervision by using domain knowledge (expressed as constraints) to drive learning. These settings were studied in [3] and.[4] These works introduce semi-supervised Constraints Driven Learning (CODL) and show that by incorporating domain knowledge the performance of the learned model improves significantly.

Learning over latent representations

edit

CCMs have also been applied to latent learning frameworks, where the learning problem is defined over a latent representation layer. Since the notion of a correct representation is inherently ill-defined, no gold-standard labeled data regarding the representation decision is available to the learner. Identifying the correct (or optimal) learning representation is viewed as a structured prediction process and therefore modeled as a CCM. This problem was covered in several papers, in both supervised[5] and unsupervised [6] settings. In all cases research showed that explicitly modeling the interdependencies between representation decisions via constraints results in an improved performance.

Integer linear programming for natural language processing applications

edit

The advantages of the CCM declarative formulation and the availability of off-the-shelf solvers have led to a large variety of natural language processing tasks being formulated within the framework, including semantic role labeling,[7] syntactic parsing,[8] coreference resolution,[9] summarization,[10][11][12] transliteration,[13] natural language generation[14] and joint information extraction.[15][16]

Most of these works use an integer linear programming (ILP) solver to solve the decision problem. Although theoretically solving an Integer Linear Program is exponential in the size of the decision problem, in practice using state-of-the-art solvers and approximate inference techniques [17] large scale problems can be solved efficiently.

The key advantage of using an ILP solver for solving the optimization problem defined by a constrained conditional model is the declarative formulation used as input for the ILP solver, consisting of a linear objective function and a set of linear constraints.

Resources

edit
edit

References

edit
  1. ^ Dan Roth and Wen-tau Yih, "A Linear Programming Formulation for Global Inference in Natural Language Tasks." Archived 2017-10-25 at the Wayback Machine CoNLL, (2004).
  2. ^ Vasin Punyakanok and Dan Roth and Wen-Tau Yih and Dav Zimak, "Learning and Inference over Constrained Output." Archived 2017-10-25 at the Wayback Machine IJCAI, (2005).
  3. ^ Ming-Wei Chang and Lev Ratinov and Dan Roth, "Guiding Semi-Supervision with Constraint-Driven Learning." Archived 2016-03-03 at the Wayback Machine ACL, (2007).
  4. ^ Ming-Wei Chang and Lev Ratinov and Dan Roth, "Constraints as Prior Knowledge." Archived 2016-03-03 at the Wayback Machine ICML Workshop on Prior Knowledge for Text and Language Processing, (2008).
  5. ^ Ming-Wei Chang and Dan Goldwasser and Dan Roth and Vivek Srikumar, "Discriminative Learning over Constrained Latent Representations." Archived 2017-10-25 at the Wayback Machine NAACL, (2010).
  6. ^ Ming-Wei Chang Dan Goldwasser Dan Roth and Yuancheng Tu, "Unsupervised Constraint Driven Learning For Transliteration Discovery."[permanent dead link] NAACL, (2009).
  7. ^ Vasin Punyakanok, Dan Roth, Wen-tau Yih and Dav Zimak, "Semantic Role Labeling via Integer Linear Programming Inference." Archived 2017-08-09 at the Wayback Machine COLING, (2004).
  8. ^ Kenji Sagae and Yusuke Miyao and Jun’ichi Tsujii, "HPSG Parsing with Shallow Dependency Constraints." ACL, (2007).
  9. ^ Pascal Denis and Jason Baldridge, "Joint Determination of Anaphoricity and Coreference Resolution using Integer Programming." Archived 2010-06-21 at the Wayback Machine NAACL-HLT, (2007).
  10. ^ James Clarke and Mirella Lapata, "Global Inference for Sentence Compression: An Integer Linear Programming Approach." Archived 2013-05-10 at the Wayback Machine Journal of Artificial Intelligence Research (JAIR), (2008).
  11. ^ Katja Filippova and Michael Strube, "Dependency Tree Based Sentence Compression."[permanent dead link] INLG, (2008).
  12. ^ Katja Filippova and Michael Strube, "Sentence Fusion via Dependency Graph Compression." EMNLP, (2008).
  13. ^ Dan Goldwasser and Dan Roth, "Transliteration as Constrained Optimization." Archived 2017-08-11 at the Wayback Machine EMNLP, (2008).
  14. ^ Regina Barzilay and Mirrela Lapata, "Aggregation via Set Partitioning for Natural Language Generation." NAACL, (2006).
  15. ^ Dan Roth and Wen-tau Yih, "A Linear Programming Formulation for Global Inference in Natural Language Tasks." Archived 2017-10-25 at the Wayback Machine CoNLL, (2004).
  16. ^ Yejin Choi and Eric Breck and Claire Cardie, "Joint Extraction of Entities and Relations for Opinion Recognition." EMNLP, (2006).
  17. ^ André F. T. Martins, Noah A. Smith, and Eric P. Xing, "Concise Integer Linear Programming Formulations for Dependency Parsing ." ACL, (2009).