Semantically Aligned Sentence-Level Embeddings for Agent Autonomy and Natural Language Understanding
Abstract
Many applications of neural linguistic models rely on their use as pre-trained features for downstream tasks such as dialog modeling, machine translation, and question answering. This work presents an alternate paradigm: Rather than treating linguistic embeddings as input features, we treat them as common sense knowledge repositories that can be queried using simple mathematical operations within the embedding space, without the need for additional training. Because current state-of-the-art embedding models were not optimized for this purpose, this work presents a novel embedding model designed and trained specifically for the purpose of "reasoning in the linguistic domain".Our model jointly represents single words, multi-word phrases, and complex sentences in a unified embedding space. To facilitate common-sense reasoning beyond straightforward semantic associations, the embeddings produced by our model exhibit carefully curated properties including analogical coherence and polarity displacement. In other words, rather than training the model on a smorgaspord of tasks and hoping that the resulting embeddings will serve our purposes, we have instead crafted training tasks and placed constraints on the system that are explicitly designed to induce the properties we seek. The resulting embeddings perform competitively on the SemEval 2013 benchmark and outperform state-of- the-art models on two key semantic discernment tasks introduced in Chapter 8.The ultimate goal of this research is to empower agents to reason about low level behaviors in order to fulfill abstract natural language instructions in an autonomous fashion. An agent equipped with an embedding space of sucient caliber could potentially reason about new situations based on their similarity to past experience, facilitating knowledge transfer and one-shot learning. As our embedding model continues to improve, we hope to see these and other abilities become a reality.
Degree
PhD
College and Department
Physical and Mathematical Sciences; Computer Science
Rights
http://lib.byu.edu/about/copyright/
BYU ScholarsArchive Citation
Fulda, Nancy Ellen, "Semantically Aligned Sentence-Level Embeddings for Agent Autonomy and Natural Language Understanding" (2019). Theses and Dissertations. 7550.
https://scholarsarchive.byu.edu/etd/7550
Date Submitted
2019-08-01
Document Type
Dissertation
Handle
http://hdl.lib.byu.edu/1877/etd10949
Keywords
Nancy Fulda, dissertation, knowledge representation, linguistic embeddings, word2vec, FastText, universal sentence encoder, BERT embeddings, sentence embedding, common-sense reasoning
Language
english