๋ชฉ๋ก2024/05/03 (1)

SJ_Koding

[LLM] ์ ์€ ๋ฐ์ดํ„ฐ๋กœ fine-tuningํ•˜์ž! LIMA: Less Is More for Alignment ๋ฆฌ๋ทฐ (Meta, 2023) - ไธŠํŽธ

LLM์„ ํŒŒ์ธํŠœ๋‹ ํ•  ์ผ์ด ์ƒ๊ฒผ๋Š”๋ฐ, ๋ฌด์—‡๋ณด๋‹ค ์ƒˆ๋กœ์šด ๋„๋ฉ”์ธ์—์„œ ์ž˜ ์ถ”๋ก ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜๊ธฐ์œ„ํ•ด ๋ฐ์ดํ„ฐ์…‹์ด ๋‹น์—ฐํžˆ ๋งŽ์•„์•ผ ํ•œ๋‹ค๊ณ  ์ƒ๊ฐํ–ˆ์—ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด Function calling๊ธฐ๋Šฅ์„ ์ˆ˜ํ–‰ํ•  ๋•Œ, Function์„ ์–ด๋А prompt์—์„œ ํ˜ธ์ถœํ•  ์ง€ ์ž˜ ์•Œ๊ธฐ ์œ„ํ•ด์„œ๋Š” ๋‹น์—ฐํžˆ ์ˆ˜๋งŽ์€ ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ ์ด๋ฅผ ๊ตฌ๋ณ„์‹œ์ผœ์•ผ ํ•œ๋‹ค๊ณ  ์ƒ๊ฐํ–ˆ๋‹ค.๊ทธ๋Ÿฐ๋ฐ, ์ด ์ƒ๊ฐ์ด ํŽธํ–ฅ๋œ ์ƒ๊ฐ์ž„์„ ๊นจ๋‹ซ๊ฒŒ ๋œ ๋…ผ๋ฌธ์ด Meta์—์„œ ๋ฐœํ‘œํ•œ LIMA: Less Is More for Alignment(2023) ๋…ผ๋ฌธ์ด๋‹ค. Abstract์ €์ž๋Š” LLM์ด ํ›ˆ๋ จ๋˜๋Š” ๋‘ ๋‹จ๊ณ„์— ๋Œ€ํ•ด ์„ค๋ช…ํ•˜๊ณ  ์žˆ๋‹ค. ์ฒซ ๋ฒˆ์งธ๋Š” unsupervised pretraining์œผ๋กœ raw text๋กœ๋ถ€ํ„ฐ general-purpose representations์„ ํ•™์Šตํ•œ๋‹ค๋Š” ๊ฒƒ์ด๊ณ , ๋‘..

LLM 2024. 5. 3. 11:06