Leveraging Moment Injection for Enhanced Semi-supervised Natural Language Inference with Large Language Models
- Authors
- 박서연
- Issue Date
- Apr-2025
- Publisher
- Association for Computational Linguistics
- Citation
- Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, v.2, pp 641 - 648
- Pages
- 8
- Indexed
- FOREIGN
- Journal Title
- Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
- Volume
- 2
- Start Page
- 641
- End Page
- 648
- URI
- https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/126244
- DOI
- 10.18653/v1/2025.naacl-short.54
- Abstract
- Natural Language Inference (NLI) is crucial for evaluating models’ Natural Language Understanding (NLU) and reasoning abilities. The development of NLI, in part, has been driven by the creation of large datasets, which require significant human effort. This has spurred interest in semi-supervised learning (SSL) that leverages both labeled and unlabeled data. However, the absence of hypotheses and class labels in NLI tasks complicates SSL. Prior work has used class-specific fine-tuned large language models (LLMs) to generate hypotheses and assign pseudo-labels but discarded many LLM-constructed samples during training to ensure the quality. In contrast, we propose to leverage all LLM-constructed samples by handling potentially noisy samples by injecting the moments of labeled samples during training to properly adjust the level of noise. Our method outperforms strong baselines on multiple NLI datasets in low-resource settings.
- Files in This Item
-
Go to Link
- Appears in
Collections - COLLEGE OF COMPUTING > ERICA 컴퓨터학부 > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.