Language and Natural Reasoning
Knowing what a text means involves drawing inferences based in the information in the text. Our group works on inferential properties of linguistic expressions to enable automated reasoning for NL understanding.
Our group works on inferential properties of linguistic expressions to enable automated reasoning for NL understanding.
We want to contribute to the theoretical understanding of how language and reasoning interact and to the computational modeling of such interactions.
Currently we concentrate on:
- the linguistic encoding of temporal and spatial information,
- the linguistic encoding of modality,
- local textual inferences,
- natural logic,
- deriving logical forms that allow interaction with structured information and computational reasoners.
Meaning in Context (MIC 3)
Bi-annual workshop bringing together linguists, computer scientists, psychologists and philosophers to discuss problems in the intersection of semantics, pragmatics, and cognition.
Participants:
- Roger Levy
- Felix Hill
- Maja Rudolph
- Noah Smith
- Ron Kaplan
- Jussi Karlgren
- David Israel
- Michael Frank
- Jay McClelland
- Juan Ignacio Cases Martin
- Thomas F Icard III
- Ido Dagan
- Louise McNally
- Pentti Kanerva
- Jason Baldridge
- Dan Lassiter
- Rami Al-Rfou
- Kelvin Guu
- Valeria de Paiva
- Hinrich Schuetze
- Annie Zaenen
- Angeliki Lazaridou
- Lauri Karttunen
- Chara Tsoukala
- Marc Pickett
- Dan Cer
- Volker Tresp
- Arianne Yuan
- Katrin Erk
- Andrew Kyle Lampinen
- Tal Linzen
- Noah Goodman
- Mengting Fang
Workgroup 1: Tuesday/Wednesday
Embeddings 2.0: The Lexicon as Memory, animated by Hinrich Schütze and Jay McClelland
for more information about this work group see: http://cis.lmu.de/schuetze/wgmemory.pdf
In NLP, we need knowledge about words. This knowledge today often comes from embeddings learned by word2vec, FastText etc. But these embeddings are limited. We will explore models of memory and learning from cognitive science and machine learning to come up with a foundation for better word representations.
Workgroup 2: Thursday/Friday
Exploring grounded and distributional learning of language, animated by Noah Goodman and Mike Frank
for more information see: http://web.stanford.edu/~azaenen/MIC/MikeandNoah.pdf
Learning language from text corpora ('distributional') allows easy access to very large data, on the other hand learning from situated, or 'grounded', utterances seems closer to what language is actually used for and closer to the human language acquisition problem. a lot of progress has been made recently on both approaches. here we will try to understand the relationship between them, and how they can be combined. for instance we will ask: do (Bayesian) theories of language use prescribe the relationship between grounded and corpus data? can representations (deep-) learned from corpora help to learn grounded meanings?
Workgroup 3: Tuesday/Wednesday
Coping with polysemy, animated by Louise McNally
for more information about this workgroup see: http://web.stanford.edu/~azaenen/MIC/wgpolysemy.pdf
Although distributional/distributed systems have proven fairly good at handling the resolution of polysemy in context, one might wonder whether they could be improved by incorporating a key premise of the Rational Speech Act (RSA) model of utterance meaning: namely, that speakers and hearers carry out probabilistic reasoning about the meaning conveyed by a given expression in context under specific assumptions about a generally limited set of alternative expressions that could have been chosen in that context. The goal of this working group is to consider how distributed representations might be dynamically modulated with information about the contextually-salient alternatives that conversational agents are considering, or, alternatively, how distributed representations could be exploited specifically in interaction with the utterance options component of an RSA model.
Workgroup 4: Thursday/Friday
Neural networks and Textual Inference, animated by Lauri Karttunen and Ignacio Cases
for more information about this workgroup see: http://web.stanford.edu/~azaenen/MIC/inferences.pdf
Recent work by Bowman et al. (2015), Rocktäschel et. al. (2016) has shown that given a sufficiently large data set such as the Stanford Natural Language Inference Corpus (SNLI) neural networks can match the performance of classical RTE systems (Dagan et al. 2006) that rely on NLP pipelines with many manually created components and features. The goal of this group is to explore whether neural networks can learn general properties of inference relations that are not represented in existing data sets such as SNLI and SICK, for example, that entailment is a reflexive and transitive relation and contradiction is symmetric.
Email or contact info for questions and comments: Annie Zaenen azaenen@stanford.edu
People
- Senior Researcher
- Director Emeritus, CSLI
- Senior Researcher