Skip to content

mihail911/CoolPapers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 

Repository files navigation

CoolPapers

List of cool academic papers I've read with summaries

[A Few Useful Things to Know About Machine Learning] (http://homes.cs.washington.edu/~pedrod/papers/cacm12.pdf)

[GloVe: Global Vectors for Word Representation] (http://www-nlp.stanford.edu/projects/glove/glove.pdf)

[Parsing Natural Scenes and Natural Language with Recursive Neural Networks] (http://www-nlp.stanford.edu/pubs/SocherLinNgManning_ICML2011.pdf)

[From Frequence to Meaning: Vector Space Models of Semantics] (http://www.jair.org/media/2934/live-2934-4846-jair.pdf)

[Semi-Supervised Recursive Autoencoders for Predicting Sentiment Distributions] (http://www.socher.org/uploads/Main/SocherPenningtonHuangNgManning_EMNLP2011.pdf)

[A Joint Model of Language and Perception for Grounded Language Learning] ( )

  • Task of extracting representations of language tied to physical world
  • New grounded concepts from a set of scenes containing only sentences, images, and indications of what objects being referred to
  • System includes:
    • Semantic parsing model
      • Defines distribution over logical meaning representations for each given sentence
    • Set of visual attribute classifiers for each possible object in scene
    • Joint model learning mapping from logical constants in logical form to set of visual attribute classifiers
    • Extracted depth and RGB values from images as features (shape and color attributes)

[Learning Distributions over Logical Forms for Referring Expression Generation] ()

  • Efficiently find fraction of referring expressions for scenes that are used; estimate associated likelihoods
  • Learn probability distribution over set of logical expressions that select a target set of objects in a world state
  • Model as globally normalized log-linear model using features of logical form z
  • Distinction for plural entities in generated logical forms
  • globally optimized log-linear model, conditioned on state S and set of target objects G
  • Three kinds of features: logical expression structure features, situated features, and a complexity feature
  • learning two models--one for a global logical form for world-state model; and one learning a series of classifiers for each
  • learn codebooks and associated sparse codes

#[A Game-Theoretic Approach to Generating Spatial Descriptions] ()

  • Generate spatial references to objects where listener much accurately identify object described by the speaker
  • Given target, speaker generates utterance according to a distribution, listener generates guess according to another distribution and both agents get a utility of 1 iff the guess matches the target
  • Train log-linear model for speaker and listener respectively

#[ Recursive Neural Networks Can Learn Logical Semantics] (http://arxiv.org/pdf/1406.1827v3.pdf)

  • General Problem/ Task Definition

    • The paper wished to see whether distributed word representations could be used in a machine learning setting to achieve good performance in the task of natural language inference (also called recognizing textual entailment).
  • Summary

    • Paper investigates the use of two neural architectures for identifying the entailment and contradiction logical relationships. In particular, the paper uses tree-based recursive neural networks and tree-based recursive neural tensor networks relying on the notion of compositionality to codify natural language word order and semantic meaning. The models are first tested on a reduced artificial data set organized around a small boolean structure world model where the models are tasked with learning the propositional relationships. Afterwards these models are tested on another more complex artificial data set where simple propositions are converted into more complex formulas. Both models achieved solid performance on these datasets although in the latter data set the RNTN seemed to struggle when tested on larger-length expressions. The models were also tested on an artificial dataset where they were tasked with learning how to correctly interpret various quantifiers and negation in the context of natural logics. Finally, the models were tested on a freely available textual entailment dataset called SICK (that was supplemented with data from the Denotation Graph project). The models achieved reasonably good performance on the SICK challenge, showing that they have the potential to accurately learn distributed representations on noisy real-world data.
  • Future Work

    • The neural models proposed seem to show particular promise for achieving good performance on very natural language logical semantics tasks. It is firmly believed that given enough data, the neural architectures proposed have the potential to perform even better on the proposed task. This makes acquisition of a more comprehensive and diverse dataset a natural next step in pursuing this modeling approach. Further even the more powerful RNTN seems to show rapidly declining performance on larger expressions which leaves the question of whether stronger models or learning techniques can be used to improve performance on considerably-sized expressions. In addition, there is still the question as to how these architectures actually encode the natural logics they are being asked to learn.

#[ An Extended Model of Natural Logic] (http://www.aclweb.org/anthology/W09-3714)

  • Problem Statement

    • Extend work on natural logic in semantic containment by providing support for semantic exclusion and implicativity. Derives series of atomic edits linking premise to hypothesis, converts each atomic edit into lexical semantic relation, and propagates these relations through semantic composition tree.
  • Summary

    • Approach proposed in paper seeks to deviate from traditional first-order-logic approaches to determining inferential validity. Natural logic sidesteps need to achieve full semantic interpretation of a phrase by characterizing syntactic forms. Paper developed a model of natural language inference focusing on seven classes designated basic semantic relations. Defines rules for the semantic join of certain relations. Paper also defined a series of rules for deriving lexical semantic relations following atomic edits. Describes semantic composition in terms of lexical semantic relation transformations propagated up constituents tree. Proposed a general method for establishing semantic relation between premise and hypothesis: 1) Find a sequence of atomic edits which transforms premise into hypothesis 2) for each atomic edit, determine the lexical semantic relation associated with that edit and project that relation upward through the semantic composition tree 3) join the semantic relations across the sequence of edits. This method suffers from having to find an appropriate edit sequence connecting the premise and hypothesis, the tendency of the join operation toward less informatic semantic relations, and the inability to combine information from multiple premises. Hence the method has less deductive power than first-order logic. The model was implemented in the NatLog system. NatLog suffers from finding appropriate edit sequence connecting premise and hypothesis which it does by relying on edit sequences from other sources. It uses machine learning techniques to learn the lexical semantic relation for each edit.
    • NatLog was evaluted on the FraCaS test suite of 346 NLI problems, achieving roughly 70% accuracy. It was also tested on the RTE3 test suite, achieving 59% accuracy.
  • Future Work More work needs to be done in establishing proper projectivity signatures for a broader range of quantifiers, verbal constructs, implicatives and factives, logical connections, and other semantic functions.

#[The PASCAL Recognising Textual Entailment Challenge] (http://download-v2.springer.com/static/pdf/580/bok%253A978-3-540-33428-6.pdf?token2=exp=1430025100~acl=%2Fstatic%2Fpdf%2F580%2Fbok%25253A978-3-540-33428-6.pdf*~hmac=c050ffd46337c67706a3e17accb3fb348d1a14980319bbddf584b8d93f82f0f7)

  • General Problem
    • Paper proposes an abstract generic task to frame the textual entailment problem.
  • Summary
    • Generated a dataset of text snippets from general news domain, annotated by humans with entailment properties. Annotators generated hypotheses for certain text corpora by converting questions and text phrases across various domains including QA, information extraction, reading comprehension, machine translation, and paraphrase acquisition. Sixteen submissions made to the challenge encompassing a wide variety of entailment inference systems. Basic kinds of features for the system include stemming, lemmatization, POS tagging, and some sort of statistical weighting. Other features included making use of higher-level lexical relationships via Wordnet or evaluating distance between syntactic structures of hypothesis and premise.
  • Future Work
    • Wished to improve the challenge by dealing with multi-valued annotation, relaxing assumptions on assumed background knowledge, providing entailment subtasks, and offering a wider variety of inference scope.

#[NaturalLI: Natural Logic Inference for Common Sense Reasoning] (http://nlp.stanford.edu/pubs/angeli2014-emnlp-naturalli.pdf)

  • General Task

    • Seeks to tackle database completion using Maccartney's natural logic. Approach does not require explicit alignment between premise and query and allows imprecise inferences at an associated cost learned from data. Casts transformation from query to supporting premise as a unified search problem where each step may have associated with it a cost reflecting confidence in the step. System allows for unstructured text as input to database, without a need to specify a schema or domain of text.
  • Summary

    • Represent Maccartney's inference model as a finite state machine that can be collapsed into three states. Represent acquisition of new premises in knowledge base as search over FSA, with nodes representing candidate facts and edges as mutations of these facts with associated costs. The confidence is a path is then computed using the cost vector and associated feature vector (representing the distance between the endpoints of two transitions). Model was tested on FraCaS entailment corpus as well as a corpus of OpenIE extractions.
  • Future Work

    • Search process does not have full access to the parse tree so struggles with issues of alignment.

#[Learning Dependency-Based Compositional Semantics] (http://cs.stanford.edu/~pliang/papers/dcs-acl2011.pdf)

  • Supervised semantic parsers
  • First must map questiosn into logical forms and this requires data with manually labeled semantic forms
  • all we really care about is resulting denotation for a given input, so are free to choose how we represent logical forms
  • introduce new semantic representation: dependency-based compositional semantics
  • represent logical forms as DCS trees where nodes represent predicates (State, Country, Genus, ...) and edges represent relations
  • such a form allows for a transparency between syntactics and semantics and hence a streamlined framework for program induction
  • denotation at root node
  • trees mirror syntactic dependency structure, facilitating parsing but also enable efficient computation of denotations defined on a given tree
  • to handle divergence between syntactic and semantic scope in some more complicated expressions, mark nodes low in tree with mark relation (E, Q, or C) and then invoke it higher up with execute relation to create desired semantic scope
  • discriminative semantic parsing model placing a log-linear distribution over the set of permissible DCS trees given an utterance

#[Bringing Machine Learning and Compositional Semantics Together] (http://www.annualreviews.org/doi/pdf/10.1146/annurev-linguist-030514-125312)

  • logical approaches rely on techniques from proof theory and model-theoretic semantics; primarily concerned with inference, ambiguity, vagueness, and compositional interpretation
  • statistical approaches derive their tools from algorithms and optimization and tend to focus on word meanings, vector space models, and other broad notions of semantic content
  • principle of compositionality: meaning of complex syntactic phrase is function of meanings of its constituent phrases
  • heart of grammar is its lexicon
  • generate a grammar exponential in the length of the sentence; dynamic programming can mitigate problems for parsing
  • learning via denotations in general results in increased computational complexity
  • learning from denotations offers advantage of being able to define features on denotations
  • semantic representations can also be distributed representations (rather than logical forms)

#[Semantic Parsing on Freebase from Question-Answer Pairs] (http://cs.stanford.edu/~pliang/papers/freebase-emnlp2013.pdf)

  • traditional semantic parsers get around need for annotated logical forms by increasing number of logical predicates
  • semantic parser must combine predicates into coherent logical form
  • parser define few simple composition rules which over-generate and use model features to simulate soft rules and categories
  • use POS tag features and features on denotations of predicted logical forms
  • database is queried using logical language: lambda-dependency-based compositional semantics
  • given utterance, semantic parser constructs a distribution over possible derivations, each derivation specifying application of a set of rules that culminates in the logical form at root of tree
  • derivations constructed recursively by mapping natural language phrases to knowledge base predicates and small set of composition rules
  • to produce manageable set of predicates per utterance, construct a lexicon that maps natural language phrases to logical predicates by aligning large text corpus to Freebase; also generate logical predicates compatible with neighboring predicates using bridging operation

About

List of cool academic papers I've read with summaries

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published