Recent Publications
Publications The latest 10 papers published or under reviewDSLR: Document Refinement with Sentence-Level Re-ranking and Reconstruction to Enhance Retrieval-Augmented Generation
Taeho Hwang, Soyeong Jeong, Sukmin Cho, SeungYoon Han, and Jong C. Park
The Third workshop on knowledge-augmented methods for NLP at ACL 2024 (KnowledgeNLP@ACL 2024), Aug 16, 2024.
The Third workshop on knowledge-augmented methods for NLP at ACL 2024 (KnowledgeNLP@ACL 2024), Aug 16, 2024.
Ask LLMs Directly, "What shapes your bias?": Measuring Social Bias in Large Language Models
Jisu Shin, Hoyun Song, Huije Lee, Soyeong Jeong, and Jong C. Park
Findings of the Association for Computational Linguistics: ACL 2024 (Findings of ACL), Aug 11-16, 2024.
Findings of the Association for Computational Linguistics: ACL 2024 (Findings of ACL), Aug 11-16, 2024.
Retrieval-Augmented Generation through Zero-shot Sentence-Level Passage Refinement using LLMs
Taeho Hwang, Soyeong Jeong, Sukmin Cho, and Jong C. Park
Proceedings of the Korea Computer Congress (KCC 2024), June 26-28, 2024.
(selected as the outstanding paper)
Proceedings of the Korea Computer Congress (KCC 2024), June 26-28, 2024.
(selected as the outstanding paper)
Enhancing Sign Language Recognition with Pose-Based Data Augmentation: Focusing on Hand Keypoints
SeungYoon Han, Aujin Kim, KyungGeun Roh, and Jong C. Park
Proceedings of the Korea Computer Congress (KCC 2024), June 26-28, 2024.
Proceedings of the Korea Computer Congress (KCC 2024), June 26-28, 2024.
Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity
Effective Pre-processing on Hand Keypoints for Sign Language Recognition
KyungGeun Roh
MS Thesis, KAIST, 2024
MS Thesis, KAIST, 2024
A Gloss-free Sign Language Production with Discrete Representation
Preprocessing Mediapipe Keypoints with Keypoint Reconstruction and Anchors for Isolated Sign Language Recognition
Augmentation of Sign Language Poses by Including the Understanding of the Sign Language Domain by Body Part
Aujin Kim
MS Thesis, KAIST, 2023.
MS Thesis, KAIST, 2023.
Capturing Ambiguity in Natural Language Understanding Tasks with Information from Internal Layers
Hancheol Park
PhD Dissertation, KAIST, 2023.
Show abstract
PhD Dissertation, KAIST, 2023.
Show abstract
In natural language understanding (NLU) tasks, there are a large number of ambiguous samples where veracity of their labels is debatable among annotators. Recently, researchers have found that even when additional annotators evaluate such ambiguous samples, they tend not to converge to single gold labels. It has been also revealed that, even when they are assessed by different groups of annotators, the degree of ambiguity is similarly reproduced. Therefore, it is desirable for a model used in an NLU task not only to predict a label that is likely to be considered correct by multiple annotators for a given sample but also to provide information about the ambiguity, indicating whether other labels could also be correct. This becomes particularly crucial in situations where the outcomes of decision-making can lead to serious problems, as information about ambiguity can guide users to make more cautious decisions and avoid risks. In this dissertation, we discuss methods for capturing ambiguous samples in NLU tasks. Due to the inherent ambiguity in NLU tasks, numerous samples with different labels can exist among those that share similar features. Therefore, it is highly likely that the model has learned information within its internal layers about which features are associated with various labels, and consequently, whether or not they exhibit ambiguity. Based on this assumption, our investigation of the representations for samples at each internal layer has revealed that information about the ambiguity of samples is more accurately represented in lower layers. Specifically, in lower layers, ambiguous samples are represented closely to samples with relevant labels in their embedding space. However, this tendency is no longer observed in the higher layers. Based on these observations, we propose methods for capturing ambiguous samples using the distribution or representation information from lower layers of encoder-based pre-trained language models (PLMs) or decoder-based large language models (LLMs). Recently, these two types of models have been predominantly used for NLU tasks. More specifically, we introduce various approaches, including using layer pruning that removes upper layers close to the output layer to utilize information from lower layers, knowledge distillation that distills distribution knowledge from lower layers, and methods utilizing internal representations from lower layers. Through experiments with NLU datasets from various domains and tasks, we demonstrate that information from internal layers, particularly from lower layers, is valuable for capturing the ambiguity of samples. We also show that our proposed methods, which use the information from lower layers, significantly outperform existing methods.