본문 바로가기

NLP

(111)
[논문이해] Why Can GPT Learn In-Context? Language Models Implicitly Perform Gradient Descent as Meta-Optimizers 논문명: Why Can GPT Learn In-Context? Language Models Implicitly Perform Gradient Descent as Meta-Optimizers 논문링크: https://arxiv.org/abs/2212.10559 Why Can GPT Learn In-Context? Language Models Implicitly Perform Gradient Descent as Meta-Optimizers Large pretrained language models have shown surprising in-context learning (ICL) ability. With a few demonstration input-label pairs, they can predict t..
[논문이해] Ground-Truth Labels Matter: A Deeper Look into Input-Label Demonstrations 논문명: Ground-Truth Labels Matter: A Deeper Look into Input-Label Demonstrations 논문링크: https://arxiv.org/abs/2205.12685 Ground-Truth Labels Matter: A Deeper Look into Input-Label Demonstrations Despite recent explosion of interests in in-context learning, the underlying mechanism and the precise impact of the quality of demonstrations remain elusive. Intuitively, ground-truth labels should have as..
[개발하며 깨닫는 것] 라이브러리 버전의 중요성 Github 에선 requirements.txt 를 통해서 라이브러리 버전을 알려준다. https://github.com/ArrowLuo/CLIP4Clip GitHub - ArrowLuo/CLIP4Clip: An official implementation for "CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Ret An official implementation for "CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval" - GitHub - ArrowLuo/CLIP4Clip: An official implementation for "CLIP4Clip: A..
[논문이해] Z-ICL: Zero-Shot In-Context Learning with Pseudo-Demonstrations 논문명:Z-ICL: Zero-Shot In-Context Learning with Pseudo-Demonstrations 논문링크: https://arxiv.org/abs/2212.09865 Z-ICL: Zero-Shot In-Context Learning with Pseudo-Demonstrations Although large language models can be prompted for both zero- and few-shot learning, performance drops significantly when no demonstrations are available. In this paper, we introduce Z-ICL, a new zero-shot method that closes th..
[논문이해] Rethinking the Role of Demonstrations: What Makes In-Context Learning Work? 논문명: Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?논문 링크: https://arxiv.org/abs/2202.12837 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?Large language models (LMs) are able to in-context learn -- perform a new task via inference alone by conditioning on a few input-label pairs (demonstrations) and making predictions for new inputs. Howev..
[논문이해] Noisy Channel Language Model Prompting for Few-Shot Text Classification 논문명: Noisy Channel Language Model Prompting for Few-Shot Text Classification https://arxiv.org/abs/2108.04106 Noisy Channel Language Model Prompting for Few-Shot Text Classification We introduce a noisy channel approach for language model prompting in few-shot text classification. Instead of computing the likelihood of the label given the input (referred as direct models), channel models compute..
[논문이해] LARGER LANGUAGE MODELS DO IN-CONTEXT LEARNING DIFFERENTLY 논문명: LARGER LANGUAGE MODELS DO IN-CONTEXT LEARNING DIFFERENTLY https://arxiv.org/abs/2303.03846 Larger language models do in-context learning differently We study how in-context learning (ICL) in language models is affected by semantic priors versus input-label mappings. We investigate two setups-ICL with flipped labels and ICL with semantically-unrelated labels-across various model families (GP..
[논문이해] A Survey on In-context Learning 논문명: A Survey on In-context Learning 논문 링크: https://arxiv.org/abs/2301.00234 A Survey on In-context Learning With the increasing ability of large language models (LLMs), in-context learning (ICL) has become a new paradigm for natural language processing (NLP), where LLMs make predictions only based on contexts augmented with a few examples. It has been a new tren arxiv.org 논문 선정 이유 비교적 최근인 23년 6..