Abstract
This paper demonstrates how transformer language models can be improved by giving them access to relevant structured data extracted from a knowledge base. The knowledge base preparation process and modifications to transformer models are explained. We evaluate these methods on language modeling and question answering tasks. These results show that even simple additional knowledge augmentation leads to a reduction in validation loss by 73%. These methods also significantly outperform common ways of improving language models such as increasing the model size or adding more data.
Degree
MS
College and Department
Physical and Mathematical Sciences; Computer Science
Rights
https://lib.byu.edu/about/copyright/
BYU ScholarsArchive Citation
Morain, Robert Kenneth, "Symbolic Semantic Memory in Transformer Language Models" (2022). Theses and Dissertations. 9380.
https://scholarsarchive.byu.edu/etd/9380
Date Submitted
2022-03-16
Document Type
Thesis
Handle
http://hdl.lib.byu.edu/1877/etd12017
Keywords
natural language processing, knowledge base, data set augmentation, semantic memory
Language
english