Sukmin Cho

ORCID: 0000-0003-1290-1570
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Topic Modeling
  • Natural Language Processing Techniques
  • Multimodal Machine Learning Applications
  • Domain Adaptation and Few-Shot Learning
  • Intelligent Tutoring Systems and Adaptive Learning
  • Digital Media Forensic Detection
  • Text and Document Classification Technologies
  • Particle accelerators and beam dynamics
  • Advanced Malware Detection Techniques
  • Web Data Mining and Analysis
  • Technology Assessment and Management
  • Advanced Image and Video Retrieval Techniques
  • Algorithms and Data Compression
  • Online Learning and Analytics
  • Expert finding and Q&A systems
  • Biblical Studies and Interpretation
  • Data Quality and Management
  • Advanced Database Systems and Queries
  • Image Retrieval and Classification Techniques
  • Pentecostalism and Christianity Studies
  • Data Management and Algorithms
  • Handwritten Text Recognition Techniques
  • Innovation Policy and R&D
  • Advanced Data Compression Techniques
  • Library Science and Information Systems

Korea Advanced Institute of Science and Technology
2022-2023

University of British Columbia
2022

National Institute of Advanced Industrial Science and Technology
2022

Bocconi University
2022

Tianjin University
2022

University of Cambridge
2022

New York University
2022

Technical University of Darmstadt
2022

Institute of Informatics of the Slovak Academy of Sciences
2022

Heidelberg University
2022

Retrieval-Augmented Large Language Models (LLMs), which incorporate the non-parametric knowledge from external bases into LLMs, have emerged as a promising approach to enhancing response accuracy in several tasks, such Question-Answering (QA). However, even though there are various approaches dealing with queries of different complexities, they either handle simple unnecessary computational overhead or fail adequately address complex multi-step queries; yet, not all user requests fall only...

10.48550/arxiv.2403.14403 preprint EN arXiv (Cornell University) 2024-03-21

Accelerating inference in Large Language Models (LLMs) is critical for real-time interactions, as they have been widely incorporated into real-world services. Speculative decoding, a fully algorithmic solution, has gained attention improving speed by drafting and verifying tokens, thereby generating multiple tokens single forward pass. However, current strategies usually require significant fine-tuning or inconsistent performance across tasks. To address these challenges, we propose...

10.48550/arxiv.2502.05609 preprint EN arXiv (Cornell University) 2025-02-08

Dense retrieval models, which aim at retrieving the most relevant document for an input query on a dense representation space, have gained considerable attention their remarkable success. Yet, models require vast amount of labeled training data notable performance, whereas it is often challenging to acquire query-document pairs annotated by humans. To tackle this problem, we propose simple but effective Document Augmentation Retrieval (DAR) framework, augments representations documents with...

10.18653/v1/2022.acl-short.48 article EN cc-by 2022-01-01

Re-rankers, which order retrieved documents with respect to the relevance score on given query, have gained attention for information retrieval (IR) task. Rather than fine-tuning pre-trained language model (PLM), large-scale (LLM) is utilized as a zero-shot re-ranker excellent results. While LLM highly dependent prompts, impact and optimization of prompts are not explored yet. Along highlighting re-ranker, we propose novel discrete prompt method, Constrained Prompt generation (Co-Prompt),...

10.18653/v1/2023.findings-acl.61 article EN cc-by Findings of the Association for Computational Linguistics: ACL 2022 2023-01-01

Re-rankers, which order retrieved documents with respect to the relevance score on given query, have gained attention for information retrieval (IR) task. Rather than fine-tuning pre-trained language model (PLM), large-scale (LLM) is utilized as a zero-shot re-ranker excellent results. While LLM highly dependent prompts, impact and optimization of prompts are not explored yet. Along highlighting re-ranker, we propose novel discrete prompt method, Constrained Prompt generation (Co-Prompt),...

10.48550/arxiv.2305.13729 preprint EN cc-by arXiv (Cornell University) 2023-01-01

Large language models (LLMs) enable zero-shot approaches in open-domain question answering (ODQA), yet with limited advancements as the reader is compared to retriever. This study aims at feasibility of a that addresses challenges computational cost and need for labeled data. We find LLMs are distracted due irrelevant documents retrieved set overconfidence generated answers when they exploited readers. To tackle these problems, we mitigate impact such via Distraction-aware Answer Selection...

10.18653/v1/2023.findings-emnlp.207 article EN cc-by 2023-01-01

Information retrieval models that aim to search for the documents relevant given query have shown many successes, which been applied diverse tasks. However, provided by user is oftentimes very short, challenges retrievers correctly fetch documents. To tackle this, existing studies proposed expanding with a couple of additional (user-related) features related query. Yet, they may be suboptimal effectively augment query, though there plenty information available it in relational database....

10.48550/arxiv.2406.16013 preprint EN arXiv (Cornell University) 2024-06-23

Sukmin Cho, Soyeong Jeong, Wonsuk Yang, Jong Park. Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Architectures. 2022.

10.18653/v1/2022.deelio-1.3 article EN cc-by 2022-01-01

The robustness of recent Large Language Models (LLMs) has become increasingly crucial as their applicability expands across various domains and real-world applications. Retrieval-Augmented Generation (RAG) is a promising solution for addressing the limitations LLMs, yet existing studies on RAG often overlook interconnected relationships between components or potential threats prevalent in databases, such minor textual errors. In this work, we investigate two underexplored aspects when...

10.48550/arxiv.2404.13948 preprint EN arXiv (Cornell University) 2024-04-22

Recent language models have shown remarkable performance on natural understanding (NLU) tasks. However, they are often sub-optimal when faced with ambiguous samples that can be interpreted in multiple ways, over-confidently predicting a single label without consideration for its correctness. To address this issue, we propose novel self-knowledge distillation method enables to learn distributions more accurately by leveraging knowledge distilled from their lower layers. This approach also...

10.48550/arxiv.2406.09719 preprint EN arXiv (Cornell University) 2024-06-14

Recent advancements in Large Language Models (LLMs) have significantly improved their performance across various Natural Processing (NLP) tasks. However, LLMs still struggle with generating non-factual responses due to limitations parametric memory. Retrieval-Augmented Generation (RAG) systems address this issue by incorporating external knowledge a retrieval module. Despite successes, however, current RAG face challenges failures and the limited ability of filter out irrelevant information....

10.48550/arxiv.2407.03627 preprint EN arXiv (Cornell University) 2024-07-04

We introduce EXIT, an extractive context compression framework that enhances both the effectiveness and efficiency of retrieval-augmented generation (RAG) in question answering (QA). Current RAG systems often struggle when retrieval models fail to rank most relevant documents, leading inclusion more at expense latency accuracy. While abstractive methods can drastically reduce token counts, their token-by-token process significantly increases end-to-end latency. Conversely, existing but rely...

10.48550/arxiv.2412.12559 preprint EN arXiv (Cornell University) 2024-12-17

Dense retrieval models, which aim at retrieving the most relevant document for an input query on a dense representation space, have gained considerable attention their remarkable success. Yet, models require vast amount of labeled training data notable performance, whereas it is often challenging to acquire query-document pairs annotated by humans. To tackle this problem, we propose simple but effective Document Augmentation Retrieval (DAR) framework, augments representations documents with...

10.48550/arxiv.2203.07735 preprint EN other-oa arXiv (Cornell University) 2022-01-01

Recent instruction-finetuned large language models (LMs) have achieved notable performances in various tasks, such as question-answering (QA). However, despite their ability to memorize a vast amount of general knowledge across diverse they might be suboptimal on specific tasks due limited capacity transfer and adapt target tasks. Moreover, further finetuning LMs with labeled datasets is often infeasible absence, but it also questionable if we can smaller having only unlabeled test data. In...

10.48550/arxiv.2310.13307 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Large language models (LLMs) enable zero-shot approaches in open-domain question answering (ODQA), yet with limited advancements as the reader is compared to retriever. This study aims at feasibility of a that addresses challenges computational cost and need for labeled data. We find LLMs are distracted due irrelevant documents retrieved set overconfidence generated answers when they exploited readers. To tackle these problems, we mitigate impact such via Distraction-aware Answer Selection...

10.48550/arxiv.2310.17490 preprint EN cc-by arXiv (Cornell University) 2023-01-01

Recent instruction-finetuned large language models (LMs) have achieved notable performances in various tasks, such as question-answering (QA). However, despite their ability to memorize a vast amount of general knowledge across diverse they might be suboptimal on specific tasks due limited capacity transfer and adapt target tasks. Moreover, further finetuning LMs with labeled datasets is often infeasible absence, but it also questionable if we can smaller having only unlabeled test data. In...

10.18653/v1/2023.findings-emnlp.1033 article EN cc-by 2023-01-01
Coming Soon ...