본문 바로가기

전체 글

(99)
[논문이해] Adapt in Contexts: Retrieval-Augmented Domain Adaptation via In-Context Learning 논문명: Adapt in Contexts: Retrieval-Augmented Domain Adaptation via In-Context Learning논문링크: https://arxiv.org/abs/2311.11551 Adapt in Contexts: Retrieval-Augmented Domain Adaptation via In-Context LearningLarge language models (LLMs) have showcased their capability with few-shot inference known as in-context learning. However, in-domain demonstrations are not always readily available in real scen..
[논문이해] Compressing Context to Enhance Inference Efficiency of Large Language Models 논문명: Compressing Context to Enhance Inference Efficiency of Large Language Models논문 링크: https://arxiv.org/abs/2310.06201 Compressing Context to Enhance Inference Efficiency of Large Language ModelsLarge language models (LLMs) achieved remarkable performance across various tasks. However, they face challenges in managing long documents and extended conversations, due to significantly increased co..
인공지능 잘 하는 법: 영어와 일본어를 하세요 저는 감자입니다.저는 학생이고요,제 의견은 근거도 없어요.하지만 써보고 싶어서요.OpenAI가 아시아에 세운 지점은 일본이었다https://openai.com/blog/introducing-openai-japan Introducing OpenAI JapanWe are excited to announce our first office in Asia and we’re releasing a GPT-4 custom model optimized for the Japanese language.openai.com아시아 국가 중에서 AI에 관심이 많고 시장성이 좋은 나라는 중국이다하지만 미국은 중국의 기술력과 성장을 제어하는 정책을 펼치기 때문에 중국 지점을 세우긴 어려울 것이다그 다음이 솔직히 한국일 줄 알았는데, ..
[논문이해] RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augmentation 논문명: RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augmentation논문 링크: https://arxiv.org/abs/2310.04408 RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective AugmentationRetrieving documents and prepending them in-context at inference time improves performance of language model (LMs) on a wide range of tasks. However, these documents, often spanning h..
SAM dataset 사용 방법 https://arxiv.org/abs/2304.02643 Segment Anything We introduce the Segment Anything (SA) project: a new task, model, and dataset for image segmentation. Using our efficient model in a data collection loop, we built the largest segmentation dataset to date (by far), with over 1 billion masks on 11M license arxiv.org 작년을 강타한 논문이 있다. 위 논문인데, 사실 필요없고 데이터셋 사용법을 알아야 한다. 문제점 https://ai.meta.com/dataset..
[논문이해] Lost in the Middle: How Language Models Use Long Contexts 논문명: Lost in the Middle: How Language Models Use Long Contexts 논문 링크: https://arxiv.org/abs/2307.03172 Lost in the Middle: How Language Models Use Long Contexts While recent language models have the ability to take long contexts as input, relatively little is known about how well they use longer context. We analyze the performance of language models on two tasks that require identifying relevant..
[논문이해] ShareGPT4V: Improving Large Multi-Modal Models with Better Captions 논문명: ShareGPT4V: Improving Large Multi-Modal Models with Better Captions 논문 링크: https://arxiv.org/abs/2311.12793 ShareGPT4V: Improving Large Multi-Modal Models with Better Captions In the realm of large multi-modal models (LMMs), efficient modality alignment is crucial yet often constrained by the scarcity of high-quality image-text data. To address this bottleneck, we introduce the ShareGPT4V d..
[논문이해] ReZero is All You Need: Fast Convergence at Large Depth 논문명: ReZero is All You Need: Fast Convergence at Large Depth 논문링크: https://arxiv.org/abs/2003.04887 ReZero is All You Need: Fast Convergence at Large Depth Deep networks often suffer from vanishing or exploding gradients due to inefficient signal propagation, leading to long training times or convergence difficulties. Various architecture designs, sophisticated residual-style networks, and initi..