Label-Aware Pseudo-Training Sample Generation for Text Classification
Owen Rambow (owen.rambow@stonybrook.edu)
Main Article Content
Abstract
Deep learning models excel in various Natural Language Processing (NLP) tasks, but their performance (excluding approaches like zero-shot learning or few-shot learning) relies on ample data, posing challenges in fields with limited datasets. To address the poverty in the size of training data, a number of approaches could be taken, such as multi-task learning and data augmentation. Aiming to leverage Large Language Models (LLMs), we propose a data augmentation...
