본문 바로가기

llm

(12)
[논문이해] DPO 손실함수는 어떻게 탄생했는가 이 글을 이해하려면, RLHF에 대한 이해가 필요하니 아래 블로그부터 읽으면 좋다. https://heygeronimo.tistory.com/122 [논문이해] training language models to follow instructions with human feedback논문을 이해하고 싶다면 아래 글을 읽으세요. 너무 잘 써서 이것보다 더 잘 쓸 자신이 없어요. https://taeyuplab.tistory.com/10 [논문 리뷰] InstructGPT: Training language models to follow instructions with human feedbackheygeronimo.tistory.com 이 글의 관심사는 오로지 수학적인 유도 과정이다. 나처럼 인공지능을 머신러닝이..
[논문이해] LORA-FA: MEMORY-EFFICIENT LOW-RANK ADAPTATION FOR LARGE LANGUAGE MODELS FINE-TUNING 논문명: LORA-FA: MEMORY-EFFICIENT LOW-RANK ADAPTATION FOR LARGE LANGUAGE MODELS FINE-TUNING논문 링크: https://arxiv.org/abs/2308.03303 LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuningThe low-rank adaptation (LoRA) method can largely reduce the amount of trainable parameters for fine-tuning large language models (LLMs), however, it still requires expensive activation m..
[Insight] Successful language model evals 글 제목: Successful language model evals글 링크: https://www.jasonwei.net/blog/evals Successful language model evals — Jason WeiEverybody uses evaluation benchmarks (“evals”), but I think they deserve more attention than they are currently getting. Evals are incentives for the research community, and breakthroughs are often closely linked to a huge performance jump on some evalwww.jasonwei.net 거대언어모델의..
[논문이해] Adapt in Contexts: Retrieval-Augmented Domain Adaptation via In-Context Learning 논문명: Adapt in Contexts: Retrieval-Augmented Domain Adaptation via In-Context Learning논문링크: https://arxiv.org/abs/2311.11551 Adapt in Contexts: Retrieval-Augmented Domain Adaptation via In-Context LearningLarge language models (LLMs) have showcased their capability with few-shot inference known as in-context learning. However, in-domain demonstrations are not always readily available in real scen..
[논문이해] Compressing Context to Enhance Inference Efficiency of Large Language Models 논문명: Compressing Context to Enhance Inference Efficiency of Large Language Models논문 링크: https://arxiv.org/abs/2310.06201 Compressing Context to Enhance Inference Efficiency of Large Language ModelsLarge language models (LLMs) achieved remarkable performance across various tasks. However, they face challenges in managing long documents and extended conversations, due to significantly increased co..
[논문이해] Lost in the Middle: How Language Models Use Long Contexts 논문명: Lost in the Middle: How Language Models Use Long Contexts 논문 링크: https://arxiv.org/abs/2307.03172 Lost in the Middle: How Language Models Use Long Contexts While recent language models have the ability to take long contexts as input, relatively little is known about how well they use longer context. We analyze the performance of language models on two tasks that require identifying relevant..
[Insight] Some intuitions about large language models 블로그명: Some intuitions about large language models 블로그 링크: https://www.jasonwei.net/blog/some-intuitions-about-large-language-models Some intuitions about large language models — Jason Wei An open question these days is why large language models work so well. In this blog post I will discuss six basic intuitions about large language models. Many of them are inspired by manually examining data, wh..
[논문이해] unlearn what you want to forget efficient unlearning for llms 논문명: unlearn what you want to forget efficient unlearning for llms 논문 링크: https://arxiv.org/abs/2310.20150 Unlearn What You Want to Forget: Efficient Unlearning for LLMs Large language models (LLMs) have achieved significant progress from pre-training on and memorizing a wide range of textual data, however, this process might suffer from privacy issues and violations of data protection regulatio..