event

PhD Defense by Yuval Pinter

Primary tabs

Title: integrating Distributional, Compositional, and Relational Approaches to Neural Word Representations

 

 

 

Yuval Pinter

 

Ph.D. Student

 

School of Interactive Computing

 

Georgia Institute of Technology

 

www.yuvalpinter.com

 

 

Date: Thursday, April 23rd, 2020

 

Time: 3:00 pm to 5:00 pm (EDT)

 

Location: *No Physical Location*

 

BlueJeans: https://bluejeans.com/976150455

 

 

 

Committee:

 

Dr. Jacob Eisenstein (advisor), School of Interactive Computing, Georgia Institute of Technology

 

Dr. Mark Riedl, School of Interactive Computing, Georgia Institute of Technology

 

Dr. Diyi Yang, School of Interactive Computing, Georgia Institute of Technology

 

Dr. Dan Roth, Department of Computer and Information Science, University of Pennsylvania

 

 

 

Abstract:

 

When the field of natural language processing (NLP) entered the era of deep neural networks, the task of representing basic units of language using low-dimensional real vectors, or embeddings, became crucial. The dominant technique to perform this task has for years been to segment input text sequences into space-delimited words, for which embeddings are trained over a large corpus by means of leveraging distributional information: a word is reducible to the set of contexts it appears in.

This approach is powerful but imperfect; words not seen during the embedding learning phase, known as out-of-vocabulary words (OOVs), emerge in any plausible application where embeddings are used. One approach applied in order to combat this and other shortcomings is the incorporation of compositional information obtained from the surface form of words, enabling the representation of morphological regularities and increasing robustness to typographical errors. Another approach leverages word-sense information and relations curated in large semantic graph resources, offering a supervised signal for embedding space structure and improving representations for domain-specific rare words.

 

In this thesis, I offer several analyses and remedies for the OOV problem based on the utilization of character-level compositional information in multiple languages and the structure of semantic knowledge in English. In addition, I provide two novel datasets for the continued exploration of vocabulary expansion in English: one with a taxonomic emphasis on novel word formation, and the other generated by a real-world data-driven use case in the entity graph domain. Finally, recognizing the shift in NLP towards context-dependent representations of sub-word tokens, I describe the form in which the OOV problem still appears in these methods, and propose a new integrative compositional model to address it.

Status

  • Workflow Status:Published
  • Created By:Tatianna Richardson
  • Created:04/10/2020
  • Modified By:Tatianna Richardson
  • Modified:04/10/2020

Categories

Keywords