๋ชฉ๋ก์ „์ฒด ๊ธ€ (89)

SJ_Koding

[LLM] ์ ์€ ๋ฐ์ดํ„ฐ๋กœ fine-tuningํ•˜์ž! LIMA: Less Is More for Alignment ๋ฆฌ๋ทฐ (Meta, 2023) - ไธ‹ํŽธ

ํ•ด๋‹น ํฌ์ŠคํŒ…์€ ์ด์ „ ๊ธ€๊ณผ ์ด์–ด์ง„ ํฌ์ŠคํŒ…์ž…๋‹ˆ๋‹ค. ๊ฐœ์ธ์ ์œผ๋กœ ไธ‹ํŽธ์— ์žฌ๋ฐŒ๋Š” ๋‚ด์šฉ์ด ๋งŽ์€ ๊ฒƒ ๊ฐ™์Šต๋‹ˆ๋‹ค. ๋…ผ๋ฌธ ์ „์ฒด๋‚ด์šฉ์„ ์ œ๊ฐ€ ์ดํ•ดํ•œ๋Œ€๋กœ ๋น ์ง์—†์ด ๊ธฐ์ž…ํ•˜๊ธฐ ๋•Œ๋ฌธ์—, ๋‚ฎ์€ ํ™•๋ฅ ๋กœ ์ž˜๋ชป๋œ ๋‚ด์šฉ์ด ํฌํ•จ๋˜์–ด์žˆ์„ ์ˆ˜ ์žˆ์œผ๋ฉฐ ๊ธ€์ด ๊ธธ๊ฒŒ ๋‚˜์—ด๋˜์–ด์žˆ์Šต๋‹ˆ๋‹ค. ๊ฐ€๋…์„ฑ์„ ์œ„ํ•ด section์ด๋‚˜ ์ค‘์š”๋ถ€๋ถ„์€ ์ปฌ๋Ÿฌ๋กœ ํ‘œ์‹œํ•ฉ๋‹ˆ๋‹ค. 2024.05.03 - [LLM] - [LLM] ์ ์€ ๋ฐ์ดํ„ฐ๋กœ fine-tuningํ•˜์ž! LIMA: Less Is More for Alignment ๋ฆฌ๋ทฐ (Meta, 2023) - ไธŠํŽธ [LLM] ์ ์€ ๋ฐ์ดํ„ฐ๋กœ fine-tuningํ•˜์ž! LIMA: Less Is More for Alignment ๋ฆฌ๋ทฐ (Meta, 2023) - ไธŠํŽธLLM์„ ํŒŒ์ธํŠœ๋‹ ํ•  ์ผ์ด ์ƒ๊ฒผ๋Š”๋ฐ, ๋ฌด์—‡๋ณด๋‹ค ์ƒˆ๋กœ์šด ๋„๋ฉ”์ธ์—์„œ ์ž˜ ์ถ”๋ก ํ•  ์ˆ˜ ์žˆ..

LLM 2024. 5. 11. 11:56
[LLM] ์ ์€ ๋ฐ์ดํ„ฐ๋กœ fine-tuningํ•˜์ž! LIMA: Less Is More for Alignment ๋ฆฌ๋ทฐ (Meta, 2023) - ไธŠํŽธ

LLM์„ ํŒŒ์ธํŠœ๋‹ ํ•  ์ผ์ด ์ƒ๊ฒผ๋Š”๋ฐ, ๋ฌด์—‡๋ณด๋‹ค ์ƒˆ๋กœ์šด ๋„๋ฉ”์ธ์—์„œ ์ž˜ ์ถ”๋ก ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜๊ธฐ์œ„ํ•ด ๋ฐ์ดํ„ฐ์…‹์ด ๋‹น์—ฐํžˆ ๋งŽ์•„์•ผ ํ•œ๋‹ค๊ณ  ์ƒ๊ฐํ–ˆ์—ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด Function calling๊ธฐ๋Šฅ์„ ์ˆ˜ํ–‰ํ•  ๋•Œ, Function์„ ์–ด๋А prompt์—์„œ ํ˜ธ์ถœํ•  ์ง€ ์ž˜ ์•Œ๊ธฐ ์œ„ํ•ด์„œ๋Š” ๋‹น์—ฐํžˆ ์ˆ˜๋งŽ์€ ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ ์ด๋ฅผ ๊ตฌ๋ณ„์‹œ์ผœ์•ผ ํ•œ๋‹ค๊ณ  ์ƒ๊ฐํ–ˆ๋‹ค.๊ทธ๋Ÿฐ๋ฐ, ์ด ์ƒ๊ฐ์ด ํŽธํ–ฅ๋œ ์ƒ๊ฐ์ž„์„ ๊นจ๋‹ซ๊ฒŒ ๋œ ๋…ผ๋ฌธ์ด Meta์—์„œ ๋ฐœํ‘œํ•œ LIMA: Less Is More for Alignment(2023) ๋…ผ๋ฌธ์ด๋‹ค. Abstract์ €์ž๋Š” LLM์ด ํ›ˆ๋ จ๋˜๋Š” ๋‘ ๋‹จ๊ณ„์— ๋Œ€ํ•ด ์„ค๋ช…ํ•˜๊ณ  ์žˆ๋‹ค. ์ฒซ ๋ฒˆ์งธ๋Š” unsupervised pretraining์œผ๋กœ raw text๋กœ๋ถ€ํ„ฐ general-purpose representations์„ ํ•™์Šตํ•œ๋‹ค๋Š” ๊ฒƒ์ด๊ณ , ๋‘..

LLM 2024. 5. 3. 11:06
LLaMA-1๋ฅผ ์•Œ์•„๋ณด์ž - 4ํŽธ, Instruction Finetuning๊ณผ Bias๋ฐ Toxicity, Misinformation

LLaMA์˜ Instruction finetuning๊ฒฐ๊ณผ์™€ bias, toxicity, misinformation๋“ฑ LLM์˜ ๋Œ€ํ‘œ์ ์ธ ๋ฌธ์ œ์  ์ •๋„๋ฅผ ๋””ํ…Œ์ผํ•˜๊ฒŒ ํ‰๊ฐ€ํ•œ๋‹ค. LLaMA-1์€ ๋ฌธ์ œ์ ์„ ํ•ด๊ฒฐํ•œ ํ…Œ์Šคํฌ๋„ ์ผ๋ถ€ ์žˆ์ง€๋งŒ, ์—ฌ์ „ํžˆ ํ•ด๊ฒฐํ•˜์ง€ ๋ชปํ•œ ํ…Œ์Šคํฌ๋„ ์กด์žฌํ–ˆ๋‹ค. ํ•ด๋‹น ํฌ์ŠคํŒ…์€ ์ด์ „ ๊ธ€๋“ค๊ณผ ์ด์–ด์ง€๋Š” ๋‚ด์šฉ์ด๋‹ค. LLaMA: Open and Efficient Foundation Language Models๋ฅผ ์•Œ์•„๋ณด์ž - 3ํŽธ, Main ResultLLaMA์˜ ์„ฑ๋Šฅ ๋น„๊ต์‹คํ—˜ ๊ฒฐ๊ณผ๋ฅผ ์ž์„ธํžˆ ๊ธฐ์ˆ ํ•œ๋‹ค. ํ…Œ์Šคํฌ๋ณ„๋กœ ํ•˜์œ„ ์„น์…˜์„ ๋‚˜๋ˆ„์—ˆ์œผ๋ฉฐ ์–ด๋–ค์‹์œผ๋กœ ์‹คํ—˜์„ ๊ตฌ์„ฑํ–ˆ๋Š”์ง€ ์ž˜ ์„ค๋ช…๋˜์–ด์žˆ๋‹ค. LLaMA๊ฐ€ ๋‹น์‹œ ์™œ ๊ฐ๊ด‘๋ฐ›์•˜๋Š”์ง€ ์•Œ ์ˆ˜ ์žˆ๋Š” ์„น์…˜์ธ ๊ฒƒ ๊ฐ™๋‹ค.Intsjkoding.tistory.com 4. Instruction..

LLM 2024. 4. 27. 14:00
LLaMA: Open and Efficient Foundation Language Models๋ฅผ ์•Œ์•„๋ณด์ž - 2ํŽธ, Approch

์ด์ „ ๊ธ€์— ์ด์–ด Approch์— ๋Œ€ํ•œ ๋‚ด์šฉ์ด๋‹ค. ์ด์ „๊ธ€๊ณผ ๋‹ค๋ฅด๊ฒŒ ์ง€๊ธˆ๋ถ€ํ„ฐ๋Š” ํ•ต์‹ฌ๋งŒ ์š”์•ฝํ•œ๋‹ค. LLaMA-1์˜ Pre-training, Architecture, Optimizer, Efficient implementation์„ ์ •๋ฆฌํ•œ๋‹ค. LLM ๋ชจ๋ธ์—์„œ ์–ด๋–ค์‹์œผ๋กœ ๋ฐ์ดํ„ฐ์…‹์„ ๊ตฌ์ถ•ํ•˜๊ณ , ์–ผ๋งŒํผ์˜ ์ž์›์„ ์‚ฌ์šฉํ•˜๋Š”์ง€, ์–ด๋–ค์‹์œผ๋กœ ํ•™์Šตํ•˜๋Š”์ง€๋ฅผ ํŒŒ์•…ํ•  ์ˆ˜ ์žˆ๋Š” ์„น์…˜์ด๋‹ค. ํ•ด๋‹น ์„น์…˜์„ ๋ฆฌ๋ทฐํ•˜๋ฉด์„œ ๋Œ€๊ฐ• LLM์˜ ์ „๋ฐ˜์ ์ธ ์ ‘๊ทผ๋ฐฉ์‹์„ ํŒŒ์•…ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. LLaMA: Open and Efficient Foundation Language Models๋ฅผ ์•Œ์•„๋ณด์ž - 1ํŽธ, Introduction ํ•ด๋‹น ๋…ผ๋ฌธ์„ ๋ณด๋ฉด์„œ LLM ์—ฐ๊ตฌ์˜ ํฐ ํ๋ฆ„์„ ๋Œ€๊ฐ•์ด๋ผ๋„ ํŒŒ์•…ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์ตœ๊ทผ์— LLaMA2์— ๋น„ํ•ด ๋น„์•ฝ์ ์œผ๋กœ ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œํ‚จ (L..

LLM 2024. 4. 23. 17:21
LLaMA: Open and Efficient Foundation Language Models๋ฅผ ์•Œ์•„๋ณด์ž - 1ํŽธ, Introduction

ํ•ด๋‹น ๋…ผ๋ฌธ์„ ๋ณด๋ฉด์„œ LLM ์—ฐ๊ตฌ์˜ ํฐ ํ๋ฆ„์„ ๋Œ€๊ฐ•์ด๋ผ๋„ ํŒŒ์•…ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์ตœ๊ทผ์— LLaMA2์— ๋น„ํ•ด ๋น„์•ฝ์ ์œผ๋กœ ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œํ‚จ (LLaMA3-8B๊ฐ€ LLaMA2-70B๋ฅผ ์ด๊น€;;) LLaMA3์˜คํ”ˆ์†Œ์Šค๊ฐ€ hugging face์— ๊ณต๊ฐœ๋˜๋ฉด์„œ ๋”์šฑ ๊ถ๊ธˆ์ฆ์ด ์ƒ๊ฒผ๋‹ค. LLM์„ ํ•  ์ผ์ด ์ƒ๊ฒผ๋Š”๋ฐ, Vision์€ ์ž ์‹œ ์ ‘์–ด๋‘๊ณ  LLM ๊ณต๋ถ€์— ํˆฌ์žํ•ด์•ผ๊ฒ ๋‹ค. LLaMA๋ชจ๋ธ์€ Meta์—์„œ ๋ฐœํ‘œํ•œ ๋ชจ๋ธ๋กœ ์ ์€ ํŒŒ๋ผ๋ฉ”ํ„ฐ ์ˆ˜(7B)์™€ ๋Œ€๊ทœ๋ชจ ์–ด๋””์„œ๋“  ์ ‘๊ทผ ๊ฐ€๋Šฅํ•œ ๋ฐ์ดํ„ฐ์…‹(์ˆ˜์กฐ ๊ฐœ token)๋งŒ์„ ์‚ฌ์šฉํ•˜์—ฌ SOTA๋ฅผ ๋‹ฌ์„ฑํ•œ ๋ชจ๋ธ์ด๋‹ค. ์‚ฌ์ „ ์ง€์‹์ด ๋ถ€์กฑํ•˜๊ธฐ ๋•Œ๋ฌธ์—, Introduction ๋งŒํผ์€ ํ•œ ์ค„ ํ•œ ์ค„ ์ž์„ธํžˆ ์‚ดํŽด๋ณด๊ณ , ์ธ์šฉ๋œ ์ค‘์š”ํ•œ ๋…ผ๋ฌธ์„ ๋Œ€๊ฐ• ํ›‘์–ด ์ •๋ฆฌํ•ด๋ณธ๋‹ค. Introduction Large Languages Mo..

LLM 2024. 4. 22. 16:46
tqdm ๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ ์‚ฌ์šฉ๋ฒ• (+ Train loss ์‹ค์‹œ๊ฐ„ ์ถœ๋ ฅ ๋ฐฉ๋ฒ•)

๋ฐ์ดํ„ฐ๋ฅผ ๋งŒ๋“ค์–ด๋‚ด๊ฑฐ๋‚˜, ํ•™์Šต์„ ๋Œ๋ฆฌ๋Š” ๊ฒƒ ์ฒ˜๋Ÿผ ๋ฐ˜๋ณต์ ์ธ ์ฝ”๋“œ ํ๋ฆ„์ด ์ง„ํ–‰ ๋  ๋•Œ, ๋ฌด์ž‘์ • ๊ธฐ๋‹ค๋ฆฌ๋Š” ๊ฒƒ ๋ณด๋‹จ ์ง„ํ–‰๋„๋ฅผ ์•Œ ์ˆ˜ ์žˆ์œผ๋ฉด ์ข‹๋‹ค. ํŠนํžˆ AI๋ฅผ ๊ณต๋ถ€ํ•˜๊ณ  ์žˆ์„๋•Œ ๊ฑฐ์˜ ํ•„์ˆ˜์ ์œผ๋กœ ์‚ฌ์šฉ๋˜๋Š”๋ฐ, ๋‹จ์ˆœ ์ง„ํ–‰๋ฅ ๋งŒ ๋ณด๋Š”๊ฒŒ ์•„๋‹ˆ๋ผ ์‹ค์‹œ๊ฐ„ loss, smooth loss๋ฅผ ๋ณด๋ฉฐ ํ•™์Šต์ด ์ž˜ ๋˜๊ณ ์žˆ๋Š”์ง€ ์‹ค์‹œ๊ฐ„์œผ๋กœ ํ™•์ธํ•˜๋Š” ์ฝ”๋“œ๋„ ํฌํ•จํ•œ๋‹ค. tqdm ์ด๋ž€?tqdm์€ 'taqaddum'์˜ ์•ฝ์ž์ด๋ฉฐ ์•„๋ž์–ด๋ผ๊ณ  ํ•œ๋‹ค. '์ง„ํ–‰'์ด๋ผ๋Š” ์˜๋ฏธ๋ฅผ ๊ฐ€์ง€๋ฉฐ ํ”„๋กœ๊ทธ๋ž˜๋จธ์—๊ฒŒ ์–ด๋– ํ•œ ํ”„๋กœ์„ธ์Šค์˜ ์ง„ํ–‰ ์ƒํ™ฉ์„ ์‹œ๊ฐ์ ์œผ๋กœ ๋ณด์—ฌ์ฃผ๋Š” ๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ์ด๋‹ค. tqdm ๊ธฐ๋ณธ์ ์ธ ์‚ฌ์šฉ๋ฐฉ๋ฒ•1. tqdm์„ค์น˜pip install tqdm # ๋…ธํŠธ๋ถ ์ƒ์—์„œ ์„ค์น˜ํ•˜๋ ค๋ฉด ์•ž์— !๋ฅผ ๋ถ™์—ฌ์•ผํ•จ 2. tqdm ๋ถˆ๋Ÿฌ์˜ค๊ธฐ & ์‚ฌ์šฉ๋ฐฉ๋ฒ•from tqdm import tqdmfor i..

Google Colab ๋Ÿฐํƒ€์ž„ ๋Š๊น€, ์•ˆ๋Š๊ธฐ๊ฒŒ ๋ฐฉ์ง€ํ•˜๋Š” ๋ฒ•(2024 ์ตœ์‹ )

๊ตฌ๊ธ€ ์ฝ”๋žฉ์€ ๋ฌด๋ฃŒ๋ฒ„์ „ ์ตœ๋Œ€ 12์‹œ๊ฐ„, Pro๋ฒ„์ „ ์ตœ๋Œ€ 24์‹œ๊ฐ„ (๋ณ€๋™๋  ์ˆ˜ ์žˆ์Œ) ์—ฐ๊ฒฐ์ด ์ง€์†๋˜๋ฉฐ ์ด๋•Œ 90๋ถ„๊ฐ„ ์–ด๋– ํ•œ ์ด๋ฒคํŠธ๊ฐ€ ์—†์„ ๊ฒฝ์šฐ ๋Ÿฐํƒ€์ž„์„ ์ข…๋ฃŒ์‹œ์ผœ๋ฒ„๋ฆฐ๋‹ค.ํ•™์Šต์ด ์˜ค๋ž˜๊ฑธ๋ฆฌ๋Š” AI ๋ชจ๋ธ์„ Google Colab์—์„œ ํ•™์Šต์‹œํ‚ฌ๋•Œ, ์ด๋ฒคํŠธ๋ฅผ ์ง€์†์ ์œผ๋กœ ๋‚ ๋ ค์ค˜์•ผํ•˜๋Š”๋ฐ ๋งค๋ฒˆ ๊ทธ๋Ÿด ์ˆ˜๋„ ์—†๋‹ค.์ด๋ฏธ ๋Ÿฐํƒ€์ž„ ๋Š๊น€ ์˜ˆ๋ฐฉ ๋ฐฉ๋ฒ•์€ ๋งŽ์ง€๋งŒ, ๊ตฌ๊ธ€์ฝ”๋žฉ์˜ ์ง€์†์ ์ธ ์—…๋ฐ์ดํŠธ๋กœ ๋„๋ฆฌ ํผ์ ธ์žˆ๋Š” ๊ฒƒ๋“ค์ด ๋ฌด์šฉ์ง€๋ฌผ์ด ๋˜๋ฒ„๋ ธ๋‹ค.์Šคํƒ์˜ค๋ฒ„ํ”Œ๋กœ์šฐ์— ์–ด๋А ํ•œ ์œ ์ €๊ฐ€ ์ตœ์‹ ์ฝ”๋“œ๋ฅผ ๊ณต์œ ํ–ˆ๋‹ค. How can I prevent Google Colab from disconnecting?Is there a way to programmatically prevent Google Colab from disconnecting on a timeout? The fo..