Incidental Supervision for Natural Language Understanding

Loading...
Thumbnail Image

Degree type

Doctor of Philosophy (PhD)

Graduate group

Computer and Information Science

Discipline

Computer Sciences

Subject

Funder

Grant number

License

Copyright date

2023

Distributor

Related resources

Contributor

Abstract

Acquiring human annotations for natural language understanding (NLU) tasks is often labor-intensive and demands significant domain-specific expertise, making it crucial to obtain supervision from indirect signals to improve target task performance. This dissertation presents a novel approach for enhancing NLU by harnessing on the power of incidental supervision signals, which are present in the data and the environment, regardless of the specific tasks being considered. The primary aim of this research is to deepen our understanding of incidental supervision signals and to develop efficient algorithms for their acquisition, selection, and usage in NLU tasks. This problem presents numerous challenges, including the intricate nature of natural language and the inherent disparities between incidental supervision signals and target tasks. To tackle these challenges, this dissertation employs a multifaceted approach. First, we demonstrate the feasibility of utilizing cost-effective signals to enhance various target tasks. Specifically, we retrieve signals from sentence-level question-answer pairs to help NLU tasks via two types of sentence encoding approaches, depending on whether the target task involves single- or multi-sentence input. Second, we introduce a unified informativeness measure to quantitatively assess the effectiveness of diverse incidental supervision signals for a given target task. This approach offers a promising way to determine, ahead of learning, which supervision signals would be beneficial. Finally, we present a suite of efficient algorithms for exploiting distinct types of incidental supervision signals, including a weighting strategy to enhance sample efficiency in cross-task learning and a post-processing technique for faithful large language model inference with external knowledge.

Date of degree

2023

Date Range for Data Collection (Start Date)

Date Range for Data Collection (End Date)

Digital Object Identifier

Series name and number

Volume number

Issue number

Publisher

Publisher DOI

Journal Issues

Comments

Recommended citation