kakaocorp

20 models โ€ข 2 total models in database
Sort by:

kanana-1.5-2.1b-instruct-2505

๐Ÿค— 1.5 HF Models &nbsp | &nbsp ๐Ÿ“• 1.5 Blog &nbsp | &nbsp ๐Ÿ“œ Technical Report - โœจ`2025/05/23`: Published a blog post about `Kanana 1.5` models and released ๐Ÿค—HF model weights. - ๐Ÿ“œ`2025/02/27`: Released Technical Report and ๐Ÿค—HF model weights. - ๐Ÿ“•`2025/01/10`: Published a blog post about the development of `Kanana Nano` model. - ๐Ÿ“•`2024/11/14`: Published blog posts (pre-training, post-training) about the development of `Kanana` models. - โ–ถ๏ธ`2024/11/06`: Published a presentation video about the development of the `Kanana` models. - Kanana 1.5 - Performance - Base Model Evaluation - Instruct Model Evaluation - Contributors - Citation - Contact `Kanana 1.5`, a newly introduced version of the Kanana model family, presents substantial enhancements in coding, mathematics, and function calling capabilities over the previous version, enabling broader application to more complex real-world problems. This new version now can handle up to 32K tokens length natively and up to 128K tokens using YaRN, allowing the model to maintain coherence when handling extensive documents or engaging in extended conversations. Furthermore, Kanana 1.5 delivers more natural and accurate conversations through a refined post-training process. > [!Note] > Neither the pre-training nor the post-training data includes Kakao user data. Kanana-1.5-2.1B 56.30 45.10 77.46 52.44 47.00 55.95 Kanana-Nano-2.1B 54.83 44.80 77.09 31.10 46.20 46.32 Models MT-Bench KoMT-Bench IFEval HumanEval+ MBPP+ GSM8K (0-shot) MATH MMLU (0-shot, CoT) KMMLU (0-shot, CoT) FunctionChatBench Kanana-1.5-2.1B 7.01 6.54 68.61 68.90 65.08 81.43 60.62 53.87 32.93 53.70 Kanana-Nano-2.1B 6.40 5.90 71.97 63.41 62.43 72.32 29.26 52.48 38.51 26.10 > [!Note] > \ Models released under Apache 2.0 are trained on the latest versions compared to other models. Contributors - Language Model Training: Yunju Bak, Doohae Jung, Boseop Kim, Nayeon Kim, Hojin Lee, Jaesun Park, Minho Ryu - Language Model Alignment: Jiyeon Ham, Seungjae Jung, Hyunho Kim, Hyunwoong Ko, Changmin Lee, Daniel Wontae Nam - AI Engineering: Youmin Kim, Hyeongju Kim Contact - Kanana LLM Team Technical Support: [email protected] - Business & Partnership Contact: [email protected]

NaNK
llama
9,702
33

kanana-nano-2.1b-instruct

NaNK
llama
6,387
70

kanana-1.5-8b-instruct-2505

๐Ÿค— 1.5 HF Models &nbsp | &nbsp ๐Ÿ“• 1.5 Blog &nbsp | &nbsp ๐Ÿ“œ Technical Report - โœจ`2025/05/23`: Published a blog post about `Kanana 1.5` models and released ๐Ÿค—HF model weights. - ๐Ÿ“œ`2025/02/27`: Released Technical Report and ๐Ÿค—HF model weights. - ๐Ÿ“•`2025/01/10`: Published a blog post about the development of `Kanana Nano` model. - ๐Ÿ“•`2024/11/14`: Published blog posts (pre-training, post-training) about the development of `Kanana` models. - โ–ถ๏ธ`2024/11/06`: Published a presentation video about the development of the `Kanana` models. - Kanana 1.5 - Performance - Base Model Evaluation - Instruct Model Evaluation - Processing 32K+ Length - Contributors - Citation - Contact `Kanana 1.5`, a newly introduced version of the Kanana model family, presents substantial enhancements in coding, mathematics, and function calling capabilities over the previous version, enabling broader application to more complex real-world problems. This new version now can handle up to 32K tokens length natively and up to 128K tokens using YaRN, allowing the model to maintain coherence when handling extensive documents or engaging in extended conversations. Furthermore, Kanana 1.5 delivers more natural and accurate conversations through a refined post-training process. > [!Note] > Neither the pre-training nor the post-training data includes Kakao user data. Models MT-Bench KoMT-Bench IFEval HumanEval+ MBPP+ GSM8K (0-shot) MATH MMLU (0-shot, CoT) KMMLU (0-shot, CoT) FunctionChatBench Kanana-1.5-8B 7.76 7.63 80.11 76.83 67.99 87.64 67.54 68.82 48.28 58.00 Kanana-8B 7.13 6.92 76.91 62.20 43.92 79.23 37.68 66.50 47.43 17.37 > [!Note] > \ Models released under Apache 2.0 are trained on the latest versions compared to other models. Processing 32K+ Length Currently, the `config.json` uploaded to HuggingFace is configured for token lengths of 32,768 or less. To process tokens beyond this length, YaRN must be applied. By updating the `config.json` with the following parameters, you can apply YaRN to handle token sequences up to 128K in length: Contributors - Language Model Training: Yunju Bak, Doohae Jung, Boseop Kim, Nayeon Kim, Hojin Lee, Jaesun Park, Minho Ryu - Language Model Alignment: Jiyeon Ham, Seungjae Jung, Hyunho Kim, Hyunwoong Ko, Changmin Lee, Daniel Wontae Nam - AI Engineering: Youmin Kim, Hyeongju Kim Contact - Kanana LLM Team Technical Support: [email protected] - Business & Partnership Contact: [email protected]

NaNK
llama
5,338
52

kanana-1.5-v-3b-instruct

NaNK
โ€”
4,403
48

kanana-2-30b-a3b-instruct

NaNK
โ€”
2,469
32

kanana-nano-2.1b-base

NaNK
llama
1,761
39

kanana-1.5-15.7b-a3b-instruct

NaNK
โ€”
1,672
21

kanana-2-30b-a3b-thinking

NaNK
โ€”
1,283
26

Kanana Safeguard 8b

๋ชจ๋ธ ์ƒ์„ธ์„ค๋ช… Kanana Safeguard๋Š” ์นด์นด์˜ค์˜ ์ž์ฒด ์–ธ์–ด๋ชจ๋ธ์ธ Kanana 8B๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•œ ์œ ํ•ด ์ฝ˜ํ…์ธ  ํƒ์ง€ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค. ์ด ๋ชจ๋ธ์€ ๋Œ€ํ™”ํ˜• AI ์‹œ์Šคํ…œ ๋‚ด ์‚ฌ์šฉ์ž ๋ฐœํ™” ๋˜๋Š” AI ์–ด์‹œ์Šคํ„ดํŠธ์˜ ๋‹ต๋ณ€์œผ๋กœ๋ถ€ํ„ฐ ๋ฆฌ์Šคํฌ ์—ฌ๋ถ€๋ฅผ ๋ถ„๋ฅ˜ํ•˜๋„๋ก ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ๋ถ„๋ฅ˜ ๊ฒฐ๊ณผ๋Š” <SAFE> ๋˜๋Š” <UNSAFE-S4> ํ˜•์‹์˜ ๋‹จ์ผ ํ† ํฐ์œผ๋กœ ์ถœ๋ ฅ๋ฉ๋‹ˆ๋‹ค. ์—ฌ๊ธฐ์—์„œ S4๋Š” ์‚ฌ์šฉ์ž ๋ฐœํ™” ๋˜๋Š” AI ์–ด์‹œ์Šคํ„ดํŠธ ๋‹ต๋ณ€์ด ์œ„๋ฐ˜ํ•œ ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ์˜ ์ฝ”๋“œ๋ฅผ ์˜๋ฏธํ•ฉ๋‹ˆ๋‹ค. ๋ฆฌ์Šคํฌ ๋ถ„๋ฅ˜ ์ฒด๊ณ„ ๋ณธ ๋ชจ๋ธ์˜ ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ๋Š” MLCommons ๋ถ„๋ฅ˜์ฒด๊ณ„์— ๊ธฐ๋ฐ˜ํ•˜๊ณ  ์žˆ์œผ๋ฉฐ, ์—ฌ๊ธฐ์— ํ•œ๊ตญ ๋กœ์ปฌ ํŠน์„ฑ์— ๋งž๋Š” ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ๋ฅผ ์ถ”๊ฐ€ํ•จ์œผ๋กœ์จ ์•„๋ž˜์™€ ๊ฐ™์ด ์ด 7๊ฐ€์ง€ ์นดํ…Œ๊ณ ๋ฆฌ๋กœ ๊ตฌ์„ฑ๋œ ๋ฆฌ์Šคํฌ ๋ถ„๋ฅ˜์ฒด๊ณ„๋ฅผ ์ˆ˜๋ฆฝํ•˜์˜€์Šต๋‹ˆ๋‹ค. ๋ณธ ๋ชจ๋ธ์—์„œ โ‘ ์‚ฌ์šฉ์ž์˜ ๋ฐœํ™”์™€ โ‘กAI ์–ด์‹œ์Šคํ„ดํŠธ์˜ ๋‹ต๋ณ€์€ ๋™์ผํ•œ ๋ฆฌ์Šคํฌ ๋ถ„๋ฅ˜์ฒด๊ณ„์— ์˜ํ•ด ํŒ๋ณ„๋ฉ๋‹ˆ๋‹ค. S1 ์ฆ์˜ค ์ถœ์‹ , ์ธ์ข…, ์™ธ์–‘, ์žฅ์•  ๋ฐ ์งˆ๋ณ‘ ์œ ๋ฌด, ์‚ฌํšŒ ๊ฒฝ์ œ์  ์ƒํ™ฉ ๋ฐ ์ง€์œ„, ์ข…๊ต, ์—ฐ๋ น, ์„ฑ๋ณ„ยท์„ฑ ์ •์ฒด์„ฑยท์„ฑ์  ์ง€ํ–ฅ ๋˜๋Š” ๊ธฐํƒ€ ์ •์ฒด์„ฑ ์š”์ธ ๋“ฑ์„ ์ด์œ ๋กœ ํŠน์ • ๋Œ€์ƒ์„ ์ฐจ๋ณ„ํ•˜๊ฑฐ๋‚˜, ์ด๋Ÿฌํ•œ ์ฐจ๋ณ„์— ๊ธฐ๋ฐ˜ํ•ด ๊ฐœ์ธ ๋˜๋Š” ์ง‘๋‹จ์„ ๊ณต๊ฒฉํ•˜๋Š” ๋ฐœํ™” S2 ๊ดด๋กญํž˜ ํƒ€์ธ์—๊ฒŒ ๋ถˆ์พŒ๊ฐ์ด๋‚˜ ๊ตด์š•๊ฐ์„ ์ฃผ๊ฑฐ๋‚˜, ์œ„ํ˜‘์ ์ด๊ฑฐ๋‚˜, ํŠน์ • ๋Œ€์ƒ์— ๋Œ€ํ•œ ๊ดด๋กญํž˜์„ ๋ถ€์ถ”๊ธฐ๋Š” ๋ฐœํ™” S3 ์„ฑ์  ์ฝ˜ํ…์ธ  ์„ฑ์  ํ–‰์œ„๋‚˜ ์‹ ์ฒด๋ฅผ ๋ฌ˜์‚ฌ/์•”์‹œํ•˜๊ฑฐ๋‚˜, ์„ฑ์  ์ˆ˜์น˜์‹ฌ/ํ˜์˜ค๊ฐ์„ ์ผ์œผํ‚ฌ ์ˆ˜ ์žˆ๋Š” ๋ฐœํ™” (์„ฑ๊ต์œก ๋ฐ ์›ฐ๋น™ ์ œ์™ธ) S4 ๋ฒ”์ฃ„ ๋ถˆ๋ฒ•์ ์ธ ํ–‰์œ„(์˜ˆ: ํญ๋ ฅโˆ™๋น„ํญ๋ ฅ ๋ฒ”์ฃ„, ์„ฑ๋ฒ”์ฃ„, ๋ฌด๊ธฐ ์ œ์ž‘ยท์กฐ๋‹ฌ)๋ฅผ ๊ธฐํšํ•˜๊ณ  ์ค€๋น„ํ•˜๋Š” ๊ณผ์ •์„ ๋‹ด์€ ๋ฐœํ™” S5 ์•„๋™ ์„ฑ์ฐฉ์ทจ ์•„๋™ ๋Œ€์ƒ์˜ ์„ฑ์  ํ•™๋Œ€์™€ ๊ด€๋ จ๋œ ์„ค๋ช…, ๊ฒฉ๋ ค, ์ง€์ง€ ๋“ฑ์˜ ๋ฐœํ™” (์˜ˆ: ๊ทธ๋ฃจ๋ฐ, CSAM ๊ด€๋ จ ํ…์ŠคํŠธ ๋“ฑ) S6 ์ž์‚ด ๋ฐ ์žํ•ด ์˜๋„์ ์œผ๋กœ ์ž์‹ ์˜ ์ƒ๋ช…์„ ๋Š๊ฑฐ๋‚˜ ์ž์‹ ์˜ ์‹ ์ฒด๋ฅผ ์˜๋„์ ์œผ๋กœ ํ•ด์น˜๋Š” ํ–‰์œ„๋ฅผ ๋ฌ˜์‚ฌํ•˜๊ฑฐ๋‚˜ ์œ ๋„ํ•˜๋Š” ๋ฐœํ™” ๋น ๋ฅธ ์‹œ์ž‘ ๐Ÿค— HuggingFace Transformers - ๋ชจ๋ธ์„ ์‹คํ–‰ํ•˜๋ ค๋ฉด `transformers>=4.51.3` ๋˜๋Š” ์ตœ์‹  ๋ฒ„์ „์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค. Kanana Safeguard์˜ ํ•™์Šต ๋ฐ์ดํ„ฐ๋Š” ์ˆ˜๊ธฐ ๋ฐ์ดํ„ฐ์™€ ํ•ฉ์„ฑ ๋ฐ์ดํ„ฐ๋กœ ๊ตฌ์„ฑ๋˜๋ฉฐ ํ•œ๊ตญ์–ด ๋ฐ์ดํ„ฐ๋กœ๋งŒ ๊ตฌ์„ฑ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค. ์ˆ˜๊ธฐ ๋ฐ์ดํ„ฐ๋Š” ๋‚ด๋ถ€์ •์ฑ…์— ๋ถ€ํ•ฉํ•˜๋„๋ก ์ „๋ฌธ ๋ผ๋ฒจ๋Ÿฌ๊ฐ€ ์ง์ ‘ ์ƒ์„ฑํ•˜๊ณ  ๋ผ๋ฒจ๋งํ•œ ๋ฐ์ดํ„ฐ์ž…๋‹ˆ๋‹ค. ํ•ฉ์„ฑ ๋ฐ์ดํ„ฐ๋Š” LLM ๊ธฐ๋ฐ˜ ํ‘œํ˜„ ๋ณ€ํ™˜๊ณผ ๋…ธ์ด์ฆˆ ์‚ฝ์ž… ๋“ฑ ๋‹ค์–‘ํ•œ ๋ฐ์ดํ„ฐ ์ฆ๊ฐ• ๊ธฐ๋ฒ•์„ ํ†ตํ•ด ์ƒ์„ฑ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค. ํ•™์Šต ๋ฐ์ดํ„ฐ์—๋Š” ์•ˆ์ „ํ•˜์ง€ ์•Š์€ ๋ฐœํ™” ๋ฐ์ดํ„ฐ ์™ธ์—๋„, ๋ชจ๋ธ์˜ ๊ฑฐ์ง“ ์–‘์„ฑ(false positive) ๋น„์œจ์„ ์ค„์ด๊ธฐ ์œ„ํ•ด ์œ ํ•ดํ•œ ์งˆ๋ฌธ์— ๋Œ€ํ•ด ์•ˆ์ „ํ•˜๊ฒŒ ์‘๋‹ตํ•œ AI ์–ด์‹œ์Šคํ„ดํŠธ์˜ ๋Œ€ํ™” ๋ฐ์ดํ„ฐ๊ฐ€ ํฌํ•จ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค. ํ‰๊ฐ€ Kanana Safeguard๋Š” SAFE/UNSAFE ์ด์ง„ ๋ถ„๋ฅ˜ ๊ธฐ์ค€์œผ๋กœ ์„ฑ๋Šฅ์„ ํ‰๊ฐ€ํ–ˆ์Šต๋‹ˆ๋‹ค. ๋ชจ๋“  ํ‰๊ฐ€๋Š” UNSAFE๋ฅผ ์–‘์„ฑ(positive) ํด๋ž˜์Šค๋กœ ๊ฐ„์ฃผํ•˜๊ณ , ๋ชจ๋ธ์ด ์ถœ๋ ฅํ•œ ์ฒซ ๋ฒˆ์งธ ํ† ํฐ์„ ๊ธฐ์ค€์œผ๋กœ ๋ถ„๋ฅ˜ํ–ˆ์Šต๋‹ˆ๋‹ค. ์™ธ๋ถ€ ๋ฒค์น˜๋งˆํฌ ๋ชจ๋ธ์€ ๊ฐ ๋ชจ๋ธ์˜ ์ถœ๋ ฅ๊ฐ’์— ๋Œ€ํ•ด ๋‹ค์Œ๊ณผ ๊ฐ™์€ ๋ฐฉ์‹์œผ๋กœ ํ‰๊ฐ€ํ•˜์˜€์Šต๋‹ˆ๋‹ค. LlamaGuard๋Š” SAFE/UNSAFE ํ† ํฐ์„ ๊ทธ๋Œ€๋กœ ํ™œ์šฉํ•ด ๊ฒฐ๊ณผ๋ฅผ ํŒ์ •ํ–ˆ์Šต๋‹ˆ๋‹ค. ShieldGemma๋Š” ์ž„๊ณ„์น˜๋ฅผ 0.5๋กœ ์„ค์ •ํ•˜์—ฌ ์ด์ง„ ๋ถ„๋ฅ˜๋ฅผ ์ˆ˜ํ–‰ํ–ˆ์Šต๋‹ˆ๋‹ค. GPT-4o๋Š” ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ ๊ธฐ๋ฐ˜ ๋ถ„๋ฅ˜ ํ”„๋กฌํ”„ํŠธ๋ฅผ zero-shot ๋ฐฉ์‹์œผ๋กœ ์ž…๋ ฅํ•˜๊ณ , ์ถœ๋ ฅ ๋‚ด์šฉ์ด ํŠน์ • ์ฝ”๋“œ๋กœ ๋ถ„๋ฅ˜๋œ ๊ฒฝ์šฐ UNSAFE๋กœ ๊ฐ„์ฃผํ•˜์—ฌ ์ด์ง„ ๋ถ„๋ฅ˜๋ฅผ ์ˆ˜ํ–‰ํ–ˆ์Šต๋‹ˆ๋‹ค. ๊ทธ ๊ฒฐ๊ณผ ์ž์ฒด์ ์œผ๋กœ ๊ตฌ์ถ•ํ•œ ํ•œ๊ตญ์–ด ํ‰๊ฐ€ ๋ฐ์ดํ„ฐ์…‹์—์„œ Kanana Safeguard์˜ ๋ถ„๋ฅ˜ ์„ฑ๋Šฅ์ด ํƒ€ ๋ฒค์น˜๋งˆํฌ ๋ชจ๋ธ ๋Œ€๋น„ ๊ฐ€์žฅ ์šฐ์ˆ˜ํ•œ ์„ฑ๋Šฅ์„ ๋‚˜ํƒ€๋ƒˆ์Šต๋‹ˆ๋‹ค. ๋ชจ๋“  ๋ชจ๋ธ์€ ๋™์ผํ•œ ํ‰๊ฐ€ ๋ฐ์ดํ„ฐ์…‹๊ณผ ๋ถ„๋ฅ˜ ๊ธฐ์ค€์œผ๋กœ ํ‰๊ฐ€๋˜์—ˆ์œผ๋ฉฐ, ์ •์ฑ… ๋ฐ ๋ชจ๋ธ ๊ตฌ์กฐ ์ฐจ์ด์— ๋”ฐ๋ฅธ ์˜ํ–ฅ์„ ์ตœ์†Œํ™”ํ•˜๊ณ , ๊ณต์ •ํ•˜๊ณ  ์‹ ๋ขฐ๋„ ๋†’์€ ๋น„๊ต๊ฐ€ ๊ฐ€๋Šฅํ•˜๋„๋ก ์„ค๊ณ„๋˜์—ˆ์Šต๋‹ˆ๋‹ค. Kanana Safeguard๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์€ ํ•œ๊ณ„์ ์ด ์žˆ์œผ๋ฉฐ, ์ด๋Š” ํ–ฅํ›„ ์ง€์†์ ์œผ๋กœ ๊ฐœ์„ ํ•ด๋‚˜๊ฐˆ ์˜ˆ์ •์ž…๋‹ˆ๋‹ค. ๋ณธ ๋ชจ๋ธ์€ 100% ์™„๋ฒฝํ•œ ๋ถ„๋ฅ˜๋ฅผ ๋ณด์žฅํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. ํŠนํžˆ, ๋ชจ๋ธ์˜ ์ •์ฑ…์€ ์ผ๋ฐ˜์ ์ธ ์‚ฌ์šฉ์‚ฌ๋ก€์— ๊ธฐ๋ฐ˜ํ•˜์—ฌ ์ˆ˜๋ฆฝ๋˜์—ˆ๊ธฐ ๋•Œ๋ฌธ์— ํŠน์ •ํ•œ ๋„๋ฉ”์ธ์—์„œ๋Š” ์ž˜๋ชป ๋ถ„๋ฅ˜๋  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ๋ณธ ๋ชจ๋ธ์€ ์ด์ „ ๋Œ€ํ™” ์ด๋ ฅ์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๋ฌธ๋งฅ์„ ์œ ์ง€ํ•˜๊ฑฐ๋‚˜ ๋Œ€ํ™”๋ฅผ ์ด์–ด๊ฐ€๋Š” ๊ธฐ๋Šฅ์€ ์ œ๊ณตํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. ๋ณธ ๋ชจ๋ธ์€ ์ •ํ•ด์ง„ ๋ฆฌ์Šคํฌ๋งŒ์„ ํƒ์ง€ํ•˜๋ฏ€๋กœ ์‹ค์‚ฌ๋ก€์˜ ๋ชจ๋“  ๋ฆฌ์Šคํฌ๋ฅผ ํƒ์ง€ํ•  ์ˆ˜๋Š” ์—†์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์˜๋„์— ๋”ฐ๋ผ Kanana Safeguard-Siren(๋ฒ•์  ๋ฆฌ์Šคํฌ ํƒ์ง€ ๋ชจ๋ธ), Kanana Safeguard-Prompt(ํ”„๋กฌํ”„ํŠธ ๊ณต๊ฒฉ ํƒ์ง€ ๋ชจ๋ธ)์™€ ํ•จ๊ป˜ ์‚ฌ์šฉํ•˜๋ฉด ์ „์ฒด์ ์ธ ์•ˆ์ „์„ฑ์„ ๋”์šฑ ๋†’์ผ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. Contributors JeongHwan Lee, Deok Jeong, HyeYeon Cho, JiEun Choi

NaNK
llama
953
25

kanana-2-30b-a3b-base

NaNK
โ€”
853
10

kanana-1.5-2.1b-base

๐Ÿค— 1.5 HF Models &nbsp | &nbsp ๐Ÿ“• 1.5 Blog &nbsp | &nbsp ๐Ÿ“œ Technical Report - โœจ`2025/05/23`: Published a blog post about `Kanana 1.5` models and released ๐Ÿค—HF model weights. - ๐Ÿ“œ`2025/02/27`: Released Technical Report and ๐Ÿค—HF model weights. - ๐Ÿ“•`2025/01/10`: Published a blog post about the development of `Kanana Nano` model. - ๐Ÿ“•`2024/11/14`: Published blog posts (pre-training, post-training) about the development of `Kanana` models. - โ–ถ๏ธ`2024/11/06`: Published a presentation video about the development of the `Kanana` models. - Kanana 1.5 - Performance - Base Model Evaluation - Instruct Model Evaluation - Contributors - Citation - Contact `Kanana 1.5`, a newly introduced version of the Kanana model family, presents substantial enhancements in coding, mathematics, and function calling capabilities over the previous version, enabling broader application to more complex real-world problems. This new version now can handle up to 32K tokens length natively and up to 128K tokens using YaRN, allowing the model to maintain coherence when handling extensive documents or engaging in extended conversations. Furthermore, Kanana 1.5 delivers more natural and accurate conversations through a refined post-training process. > [!Note] > Neither the pre-training nor the post-training data includes Kakao user data. Kanana-1.5-2.1B 56.30 45.10 77.46 52.44 47.00 55.95 Kanana-Nano-2.1B 54.83 44.80 77.09 31.10 46.20 46.32 Models MT-Bench KoMT-Bench IFEval HumanEval+ MBPP+ GSM8K (0-shot) MATH MMLU (0-shot, CoT) KMMLU (0-shot, CoT) FunctionChatBench Kanana-1.5-2.1B 7.01 6.54 68.61 68.90 65.08 81.43 60.62 53.87 32.93 53.70 Kanana-Nano-2.1B 6.40 5.90 71.97 63.41 62.43 72.32 29.26 52.48 38.51 26.10 > [!Note] > \ Models released under Apache 2.0 are trained on the latest versions compared to other models. Contributors - Language Model Training: Yunju Bak, Doohae Jung, Boseop Kim, Nayeon Kim, Hojin Lee, Jaesun Park, Minho Ryu - Language Model Alignment: Jiyeon Ham, Seungjae Jung, Hyunho Kim, Hyunwoong Ko, Changmin Lee, Daniel Wontae Nam - AI Engineering: Youmin Kim, Hyeongju Kim Contact - Kanana LLM Team Technical Support: [email protected] - Business & Partnership Contact: [email protected]

NaNK
llama
796
8

kanana-nano-2.1b-embedding

NaNK
license:cc-by-nc-4.0
687
27

kanana-1.5-8b-base

๐Ÿค— 1.5 HF Models &nbsp | &nbsp ๐Ÿ“• 1.5 Blog &nbsp | &nbsp ๐Ÿ“œ Technical Report - โœจ`2025/05/23`: Published a blog post about `Kanana 1.5` models and released ๐Ÿค—HF model weights. - ๐Ÿ“œ`2025/02/27`: Released Technical Report and ๐Ÿค—HF model weights. - ๐Ÿ“•`2025/01/10`: Published a blog post about the development of `Kanana Nano` model. - ๐Ÿ“•`2024/11/14`: Published blog posts (pre-training, post-training) about the development of `Kanana` models. - โ–ถ๏ธ`2024/11/06`: Published a presentation video about the development of the `Kanana` models. - Kanana 1.5 - Performance - Base Model Evaluation - Instruct Model Evaluation - Processing 32K+ Length - Contributors - Citation - Contact `Kanana 1.5`, a newly introduced version of the Kanana model family, presents substantial enhancements in coding, mathematics, and function calling capabilities over the previous version, enabling broader application to more complex real-world problems. This new version now can handle up to 32K tokens length natively and up to 128K tokens using YaRN, allowing the model to maintain coherence when handling extensive documents or engaging in extended conversations. Furthermore, Kanana 1.5 delivers more natural and accurate conversations through a refined post-training process. > [!Note] > Neither the pre-training nor the post-training data includes Kakao user data. Models MT-Bench KoMT-Bench IFEval HumanEval+ MBPP+ GSM8K (0-shot) MATH MMLU (0-shot, CoT) KMMLU (0-shot, CoT) FunctionChatBench Kanana-1.5-8B 7.76 7.63 80.11 76.83 67.99 87.64 67.54 68.82 48.28 58.00 Kanana-8B 7.13 6.92 76.91 62.20 43.92 79.23 37.68 66.50 47.43 17.37 > [!Note] > \ Models released under Apache 2.0 are trained on the latest versions compared to other models. Processing 32K+ Length Currently, the `config.json` uploaded to HuggingFace is configured for token lengths of 32,768 or less. To process tokens beyond this length, YaRN must be applied. By updating the `config.json` with the following parameters, you can apply YaRN to handle token sequences up to 128K in length: Contributors - Language Model Training: Yunju Bak, Doohae Jung, Boseop Kim, Nayeon Kim, Hojin Lee, Jaesun Park, Minho Ryu - Language Model Alignment: Jiyeon Ham, Seungjae Jung, Hyunho Kim, Hyunwoong Ko, Changmin Lee, Daniel Wontae Nam - AI Engineering: Youmin Kim, Hyeongju Kim Contact - Kanana LLM Team Technical Support: [email protected] - Business & Partnership Contact: [email protected]

NaNK
llama
515
11

kanana-1.5-15.7b-a3b-base

๐Ÿค— 1.5 HF Models &nbsp | &nbsp ๐Ÿ“• Kanana-1.5-15.7B-A3B Blog &nbsp - โœจ`2025/07/24`: Published a blog post about `Kanana-1.5-15.7B-A3B` models and released ๐Ÿค—HF model weights. - ๐Ÿ“•`2025/05/23`: Published a blog post about `Kanana 1.5` models and released ๐Ÿค—HF model weights. - ๐Ÿ“œ`2025/02/27`: Released Technical Report and ๐Ÿค—HF model weights. - ๐Ÿ“•`2025/01/10`: Published a blog post about the development of `Kanana Nano` model. - ๐Ÿ“•`2024/11/14`: Published blog posts (pre-training, post-training) about the development of `Kanana` models. - โ–ถ๏ธ`2024/11/06`: Published a presentation video about the development of the `Kanana` models. - Kanana-1.5-15.7B-A3B - Performance - Base Model Evaluation - Instruct Model Evaluation - Contributors - Citation - Contact Introducing `Kanana-1.5-15.7B-A3B`, the first Mixture-of-Experts (MoE) model in our Kanana family, engineered for exceptional efficiency and powerful performance. `Kanana-1.5-15.7B-A3B`, which has sparse architecture, delivers capabilities comparable to the `Kanana-1.5-8B` dense model while utilizing only 37% of the FLOPS per token, making it a highly inference-efficient and cost-effective solution for real-world applications. Furthermore, `Kanana-1.5-15.7B-A3B` is powered by our newly enhanced post-training strategy, which includes on-policy distillation followed by reinforcement learning. > [!Note] > Neither the pre-training nor the post-training data includes Kakao user data. Kanana-1.5-15.7B-A3B 64.79 51.77 83.23 59.76 60.10 61.18 Models MT-Bench KoMT-Bench IFEval HumanEval+ MBPP+ GSM8K (0-shot) MATH MMLU (0-shot, CoT) KMMLU (0-shot, CoT) Kanana-1.5-15.7B-A3B 7.67 7.24 73.35 79.27 70.37 83.02 66.42 68.55 48.92 Kanana-1.5-8B 7.76 7.63 80.11 76.83 67.99 87.64 67.54 68.82 48.28 Kanana-1.5-3B 7.01 6.52 70.08 70.73 64.29 80.36 56.70 59.69 37.60 > [!Note] > \ This model is not an open-sourced, just for comparison with Kanana-1.5-15.7B-A3B Evaluation Protocol - Base Model Benchmarks - MMLU, KMMLU, HAE-RAE: 5-shot, log-likelihood - HumanEval: 0-shot, pass@1 - MBPP: 3-shot, pass@1 - GSM8K: 5-shot, exact-match (strict-match) - Instruct Model Benchmarks - MT-Bench, KoMT-Bench: 0-shot, gpt-4o-2024-08-06 as judge model - IFEval: 0-shot, mean of strict-prompt-level and strict-instruction-level - HumanEval+, MBPP+: 0-shot, pass@1 - GSM8K, MATH: 0-shot, rule-based verification vLLM - `vllm>=0.8.5` or the latest version is required to run `Kanana` model. Contributors - Language Model Training - Yunju Bak, Doohae Jung, Boseop Kim, Nayeon Kim, Hojin Lee, Jaesun Park, Minho Ryu, Jiyeon Ham, Seungjae Jung, Hyunho Kim, Hyunwoong Ko, Changmin Lee, Taegyeong Eo Contact - Kanana LLM Team Technical Support: [email protected] - Business & Partnership Contact: [email protected]

NaNK
โ€”
118
7

kanana-safeguard-prompt-2.1b

๋ชจ๋ธ ์ƒ์„ธ์„ค๋ช… Kanana Safeguard-Prompt๋Š” ์นด์นด์˜ค์˜ ์ž์ฒด ์–ธ์–ด๋ชจ๋ธ์ธ Kanana 2.1B๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•œ ํ”„๋กฌํ”„ํŠธ ๊ณต๊ฒฉ ํƒ์ง€ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค. ์ด ๋ชจ๋ธ์€ ๋Œ€ํ™”ํ˜• AI ์‹œ์Šคํ…œ ๋‚ด ์‚ฌ์šฉ์ž์˜ ๋ฐœํ™”๋กœ๋ถ€ํ„ฐ ์•…์˜์ ์ธ ๊ณต๊ฒฉ๊ณผ ๊ด€๋ จ๋œ ๋ฆฌ์Šคํฌ ์—ฌ๋ถ€๋ฅผ ๋ถ„๋ฅ˜ํ•˜๋„๋ก ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ๋ถ„๋ฅ˜ ๊ฒฐ๊ณผ๋Š” <SAFE> ๋˜๋Š” <UNSAFE-A1> ํ˜•์‹์˜ ๋‹จ์ผ ํ† ํฐ์œผ๋กœ ์ถœ๋ ฅ๋ฉ๋‹ˆ๋‹ค. ์—ฌ๊ธฐ์—์„œ A1์€ ์‚ฌ์šฉ์ž ๋ฐœํ™”๊ฐ€ ์œ„๋ฐ˜ํ•œ ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ์˜ ์ฝ”๋“œ๋ฅผ ์˜๋ฏธํ•ฉ๋‹ˆ๋‹ค. ๋ฆฌ์Šคํฌ ๋ถ„๋ฅ˜ ์ฒด๊ณ„ Kanana Safeguard-Prompt๋Š” ํ”„๋กฌํ”„ํŠธ ๊ณต๊ฒฉ์„ ๋‘ ๊ฐ€์ง€ ๋ฆฌ์Šคํฌ ์œ ํ˜• (Prompt Injection, Prompt Leaking)์œผ๋กœ ์ •์˜ํ•˜๊ณ  ์ด๋ฅผ ๋ถ„๋ฅ˜ ๊ธฐ์ค€์œผ๋กœ ์‚ฌ์šฉํ•ฉ๋‹ˆ๋‹ค. ํ˜„์žฌ ํ”„๋กฌํ”„ํŠธ ๊ณต๊ฒฉ์— ๋Œ€ํ•œ ์—…๊ณ„ ํ‘œ์ค€ ๋ถ„๋ฅ˜ ์ฒด๊ณ„๋Š” ์•„์ง ๋ช…ํ™•ํžˆ ์ •๋ฆฝ๋˜์ง€ ์•Š์•˜์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ๋ณธ ๋ชจ๋ธ์€ ๊ฐœ๋ฐœ์ž ์ปค๋ฎค๋‹ˆํ‹ฐ์—์„œ ์ž์ฃผ ๋…ผ์˜๋˜๋Š” ์œ ํ˜•์„ ์ค‘์‹ฌ์œผ๋กœ ์ •์ฑ…์„ ์ˆ˜๋ฆฝํ•˜์˜€์Šต๋‹ˆ๋‹ค. A1 Prompt Injection LLM์˜ ์ง€์นจ์„ ๋ฌด์‹œํ•˜๊ฑฐ๋‚˜ ์‹œ์Šคํ…œ ๋™์ž‘์„ ๋ณ€๊ฒฝํ•˜๋ ค๋Š” ์˜๋„๋กœ ์šฐํšŒํ•˜๋ ค๋Š” ์กฐ์ž‘๋œ ๋ฐœํ™” A2 Prompt Leaking ํ”„๋กฌํ”„ํŠธ, ํ•™์Šต ๋ฐ์ดํ„ฐ ๋“ฑ AI ์‹œ์Šคํ…œ์˜ ๋‚ด๋ถ€ ์ •๋ณด๋ฅผ ์œ ์ถœํ•˜๋ ค๋Š” ๋ฐœํ™” ์ง€์› ์–ธ์–ด Kanana Safeguard-Prompt๋Š” ํ•œ๊ตญ์–ด์™€ ์˜์–ด์— ์ตœ์ ํ™”๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค. ๋น ๋ฅธ ์‹œ์ž‘ ๐Ÿค— HuggingFace Transformers - ๋ชจ๋ธ์„ ์‹คํ–‰ํ•˜๋ ค๋ฉด `transformers>=4.51.3` ๋˜๋Š” ์ตœ์‹  ๋ฒ„์ „์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค. Kanana Safeguard-Prompt๋Š” ์ˆ˜๊ธฐ ๋ฐ์ดํ„ฐ์™€ ํ•ฉ์„ฑ ๋ฐ์ดํ„ฐ๋ฅผ ํ•จ๊ป˜ ํ™œ์šฉํ•ด ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ์ˆ˜๊ธฐ ๋ฐ์ดํ„ฐ๋Š” ๋‚ด๋ถ€ ์ •์ฑ…์— ๋ถ€ํ•ฉํ•˜๋Š” ๋ฐ์ดํ„ฐ๋ฅผ ํ™•๋ณดํ•˜๊ธฐ ์œ„ํ•ด ์ „๋ฌธ ๋ผ๋ฒจ๋Ÿฌ๊ฐ€ ์ง์ ‘ ๋ฌธ์žฅ์„ ์ž‘์„ฑํ•˜๊ณ  ์ด๋ฅผ ๋‹ค์–‘ํ•œ ๊ธฐ๋ฒ•์œผ๋กœ ์ฆ๊ฐ•ํ•˜์˜€์Šต๋‹ˆ๋‹ค. ์™ธ๋ถ€์— ๊ณต๊ฐœ๋œ ๋ผ์ด์„ ์Šค ๋ฐ์ดํ„ฐ๋„ ์„ ๋ณ„์ ์œผ๋กœ ์ˆ˜์ง‘ํ•˜์—ฌ ํ•œ๊ตญ์–ด๋กœ ๋ฒˆ์—ญ ๋ฐ ๊ฐ€๊ณตํ•ด ์‚ฌ์šฉํ•˜์˜€์Šต๋‹ˆ๋‹ค. ๋˜ํ•œ ๊ฑฐ์ง“ ์–‘์„ฑ(false positive) ๋น„์œจ์„ ์ตœ์†Œํ™”ํ•˜๊ธฐ ์œ„ํ•ด ๋‹ค์–‘ํ•œ ์ •์ƒ ์ฑ„ํŒ… ์‹œ๋‚˜๋ฆฌ์˜ค๋„ ํ•™์Šต ๋ฐ์ดํ„ฐ์— ํฌํ•จํ•˜์˜€์Šต๋‹ˆ๋‹ค. ํ‰๊ฐ€ Kanana Safeguard-Prompt๋Š” SAFE / UNSAFE ์ด์ง„ ๋ถ„๋ฅ˜ ๊ธฐ์ค€์œผ๋กœ ์„ฑ๋Šฅ์„ ํ‰๊ฐ€ํ–ˆ์Šต๋‹ˆ๋‹ค. ๋ชจ๋“  ํ‰๊ฐ€์—์„œ UNSAFE๋ฅผ ์–‘์„ฑ ๋ผ๋ฒจ(positive label)๋กœ ๊ฐ„์ฃผํ•˜๊ณ , ๋ชจ๋ธ์ด ์ถœ๋ ฅํ•œ ์ฒซ ๋ฒˆ์งธ ํ† ํฐ์„ ๊ธฐ์ค€์œผ๋กœ ๋ถ„๋ฅ˜ํ–ˆ์Šต๋‹ˆ๋‹ค. ์™ธ๋ถ€ ๋ฒค์น˜๋งˆํฌ ๋ชจ๋ธ์€ ๊ฐ ๋ชจ๋ธ์˜ ์ถœ๋ ฅ๊ฐ’์— ๋Œ€ํ•ด ๋‹ค์Œ๊ณผ ๊ฐ™์€ ๋ฐฉ์‹์œผ๋กœ ํ‰๊ฐ€ํ•˜์˜€์Šต๋‹ˆ๋‹ค. ๋ถ„๋ฅ˜ ๊ธฐ๋ฐ˜ ๋ชจ๋ธ(Prompt Guard, Deepset, Protect AI)์€ ์ถœ๋ ฅ๋œ ๊ฒฐ๊ณผ๊ฐ€ ์–‘์„ฑ ๋ ˆ์ด๋ธ”์— ํ•ด๋‹นํ•˜๋Š”์ง€๋ฅผ ํ™•์ธํ•ด ์ด์ง„ ๋ถ„๋ฅ˜ ์„ฑ๋Šฅ์„ ์ธก์ •ํ–ˆ์Šต๋‹ˆ๋‹ค. GPT-4o๋Š” ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ๋ฅผ ๋ถ„๋ฅ˜ํ•˜๋Š” ํ”„๋กฌํ”„ํŠธ๋ฅผ zero-shot์œผ๋กœ ์ž…๋ ฅํ•œ ๋’ค, ํŠน์ • ์ฝ”๋“œ(A1, A2 ๋“ฑ)๋กœ ์‘๋‹ตํ•œ ๊ฒฝ์šฐ ์ด๋ฅผ UNSAFE๋กœ ๊ฐ„์ฃผํ•˜์—ฌ ๋™์ผํ•œ ๊ธฐ์ค€์œผ๋กœ ํ‰๊ฐ€๋ฅผ ์ง„ํ–‰ํ–ˆ์Šต๋‹ˆ๋‹ค. ๊ทธ ๊ฒฐ๊ณผ ์ž์ฒด์ ์œผ๋กœ ๊ตฌ์ถ•ํ•œ ํ•œ๊ตญ์–ด ํ‰๊ฐ€ ๋ฐ์ดํ„ฐ์…‹์—์„œ Kanana Safeguard-Prompt์˜ ๋ถ„๋ฅ˜ ์„ฑ๋Šฅ์ด ํƒ€ ๋ฒค์น˜๋งˆํฌ ๋ชจ๋ธ ๋Œ€๋น„ ๊ฐ€์žฅ ์šฐ์ˆ˜ํ•œ ์„ฑ๋Šฅ์„ ๋‚˜ํƒ€๋ƒˆ์Šต๋‹ˆ๋‹ค. ๋ชจ๋“  ๋ชจ๋ธ์€ ๋™์ผํ•œ ํ‰๊ฐ€ ๋ฐ์ดํ„ฐ์…‹๊ณผ ๋ถ„๋ฅ˜ ๊ธฐ์ค€์œผ๋กœ ํ‰๊ฐ€๋˜์—ˆ์œผ๋ฉฐ, ์ •์ฑ… ๋ฐ ๋ชจ๋ธ ๊ตฌ์กฐ ์ฐจ์ด์— ๋”ฐ๋ฅธ ์˜ํ–ฅ์„ ์ตœ์†Œํ™”ํ•˜๊ณ , ๊ณต์ •ํ•˜๊ณ  ์‹ ๋ขฐ๋„ ๋†’์€ ๋น„๊ต๊ฐ€ ๊ฐ€๋Šฅํ•˜๋„๋ก ์„ค๊ณ„๋˜์—ˆ์Šต๋‹ˆ๋‹ค. Kanana Safeguard-Prompt๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์€ ํ•œ๊ณ„์ ์ด ์žˆ์œผ๋ฉฐ, ์ด๋Š” ํ–ฅํ›„ ์ง€์†์ ์œผ๋กœ ๊ฐœ์„ ํ•ด๋‚˜๊ฐˆ ์˜ˆ์ •์ž…๋‹ˆ๋‹ค. ๋ณธ ๋ชจ๋ธ์€ 100% ์™„๋ฒฝํ•œ ๋ถ„๋ฅ˜๋ฅผ ๋ณด์žฅํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. ํŠนํžˆ, ๋ชจ๋ธ์˜ ์ •์ฑ…์€ ์ผ๋ฐ˜์ ์ธ ์‚ฌ์šฉ์‚ฌ๋ก€์— ๊ธฐ๋ฐ˜ํ•˜์—ฌ ์ˆ˜๋ฆฝ๋˜์—ˆ๊ธฐ ๋•Œ๋ฌธ์— ํŠน์ •ํ•œ ๋„๋ฉ”์ธ์—์„œ๋Š” ์ž˜๋ชป ๋ถ„๋ฅ˜๋  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ๋ณธ ๋ชจ๋ธ์€ ์ด์ „ ๋Œ€ํ™” ์ด๋ ฅ์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๋ฌธ๋งฅ์„ ์œ ์ง€ํ•˜๊ฑฐ๋‚˜ ๋Œ€ํ™”๋ฅผ ์ด์–ด๊ฐ€๋Š” ๊ธฐ๋Šฅ์€ ์ œ๊ณตํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. ๋ณธ ๋ชจ๋ธ์€ ์ •ํ•ด์ง„ ๋ฆฌ์Šคํฌ๋งŒ์„ ํƒ์ง€ํ•˜๋ฏ€๋กœ ์‹ค์‚ฌ๋ก€์˜ ๋ชจ๋“  ๋ฆฌ์Šคํฌ๋ฅผ ํƒ์ง€ํ•  ์ˆ˜๋Š” ์—†์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์˜๋„์— ๋”ฐ๋ผ Kanana Safeguard(์œ ํ•ดํ•œ ์ฝ˜ํ…์ธ  ํƒ์ง€), Kanana Safeguard-Siren(๋ฒ•์  ๋ฆฌ์Šคํฌ ํƒ์ง€) ๋ชจ๋ธ๊ณผ ํ•จ๊ป˜ ์‚ฌ์šฉํ•˜๋ฉด ์ „์ฒด์ ์ธ ์•ˆ์ „์„ฑ์„ ๋”์šฑ ๋†’์ผ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. Contributors Deok Jeong, JeongHwan Lee, HyeYeon Cho, JiEun Choi

NaNK
llama
104
14

kanana-safeguard-siren-8b

๋ชจ๋ธ ์ƒ์„ธ์„ค๋ช… Kanana Safeguard-Siren์€ ์นด์นด์˜ค์˜ ์ž์ฒด ์–ธ์–ด๋ชจ๋ธ์ธ Kanana 8B ๊ธฐ๋ฐ˜์œผ๋กœ ํ•œ ๋ฒ•์ โˆ™์ •์ฑ…์  ์œ„ํ—˜ ํƒ์ง€ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค. ์ด ๋ชจ๋ธ์€ ๋Œ€ํ™”ํ˜• AI ์‹œ์Šคํ…œ ๋‚ด ์‚ฌ์šฉ์ž์˜ ๋ฐœํ™”๋กœ๋ถ€ํ„ฐ ๋ฒ•์ โˆ™์ •์ฑ…์  ์ฃผ์˜๊ฐ€ ํ•„์š”ํ•œ ๋ฐœํ™”๋ฅผ ๋ถ„๋ฅ˜ํ•˜๋„๋ก ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ๋ถ„๋ฅ˜ ๊ฒฐ๊ณผ๋Š” <SAFE> ๋˜๋Š” <UNSAFE-I2> ํ˜•์‹์˜ ๋‹จ์ผ ํ† ํฐ์œผ๋กœ ์ถœ๋ ฅ๋ฉ๋‹ˆ๋‹ค. ์—ฌ๊ธฐ์—์„œ I2๋Š” ์‚ฌ์šฉ์ž ๋ฐœํ™”๊ฐ€ ์œ„๋ฐ˜ํ•œ ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ์˜ ์ฝ”๋“œ๋ฅผ ์˜๋ฏธํ•ฉ๋‹ˆ๋‹ค. ๋ฆฌ์Šคํฌ ๋ถ„๋ฅ˜ ์ฒด๊ณ„ ๋ณธ ๋ชจ๋ธ์˜ ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ๋Š” MLCommons ๋ถ„๋ฅ˜์ฒด๊ณ„์— ๊ธฐ๋ฐ˜ํ•˜๊ณ  ์žˆ์œผ๋ฉฐ, ์—ฌ๊ธฐ์— ํ•œ๊ตญ์˜ ๋ฒ•๋ฅ ์  ํŠน์„ฑ์— ๋งž๋Š” ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ๋ฅผ ์ถ”๊ฐ€ํ•จ์œผ๋กœ์จ ์•„๋ž˜์™€ ๊ฐ™์ด ์ด 4๊ฐ€์ง€ ์นดํ…Œ๊ณ ๋ฆฌ๋กœ ๊ตฌ์„ฑ๋œ ๋ฆฌ์Šคํฌ ๋ถ„๋ฅ˜์ฒด๊ณ„๋ฅผ ์ˆ˜๋ฆฝํ•˜์˜€์Šต๋‹ˆ๋‹ค. I1 ์„ฑ์ธ์ธ์ฆ ์ฃผ๋ฅ˜, ๋‹ด๋ฐฐ, ๋„๋ฐ•, ์œ ํฅ์—…์†Œ ๋˜๋Š” 19์„ธ ์ฝ˜ํ…์ธ  ๋“ฑ ์ฒญ์†Œ๋…„ ์œ ํ•ด ์ •๋ณด์— ๋Œ€ํ•œ ์š”์ฒญ์„ ํฌํ•จํ•˜๋Š” ๋ฐœํ™” I2 ์ „๋ฌธ์กฐ์–ธ ์˜ํ•™, ๋ฒ•๋ฅ , ์„ธ๋ฌด, ๊ธˆ์œต ๋“ฑ ์ „๋ฌธ์ ์ธ ์˜์‚ฌ๊ฒฐ์ •๊ณผ ๊ด€๋ จ๋œ ์กฐ์–ธ์„ ์š”์ฒญํ•˜๋Š” ๋ฐœํ™” I3 ๊ฐœ์ธ์ •๋ณด ๊ฐœ์ธ ์‹๋ณ„ ์ •๋ณด(์˜ˆ: ์ฃผ๋ฏผ๋“ฑ๋ก๋ฒˆํ˜ธ, ๊ณ„์ขŒ๋ฒˆํ˜ธ ๋“ฑ)๋‚˜ ๋ฏผ๊ฐํ•œ ๋ฐ์ดํ„ฐ๋ฅผ ์š”์ฒญํ•˜๊ฑฐ๋‚˜ ํฌํ•จํ•˜๋Š” ๋ฐœํ™” I4 ์ง€์‹์žฌ์‚ฐ๊ถŒ ์ €์ž‘๊ถŒ, ํŠนํ—ˆ, ์ƒํ‘œ๊ถŒ ๋“ฑ์œผ๋กœ ๋ณดํ˜ธ๋œ ์ฝ˜ํ…์ธ ๋ฅผ ๋ฌด๋‹จ์œผ๋กœ ์š”์ฒญํ•˜๊ฑฐ๋‚˜ ๋ณต์ œํ•˜๋ ค๋Š” ๋ฐœํ™” ๋น ๋ฅธ ์‹œ์ž‘ ๐Ÿค— HuggingFace Transformers - ๋ชจ๋ธ์„ ์‹คํ–‰ํ•˜๋ ค๋ฉด `transformers>=4.51.3` ๋˜๋Š” ์ตœ์‹  ๋ฒ„์ „์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค. Kanana Safeguard-Siren์˜ ํ•™์Šต ๋ฐ์ดํ„ฐ๋Š” ์ˆ˜๊ธฐ ๋ฐ์ดํ„ฐ, ํ•ฉ์„ฑ ๋ฐ์ดํ„ฐ, ์™ธ๋ถ€ ๋ฐ์ดํ„ฐ๋กœ ๊ตฌ์„ฑ๋˜๋ฉฐ ๋‹ค์–‘ํ•œ ์œ ํ˜•์˜ ๋ฐ์ดํ„ฐ๋ฅผ ํ™œ์šฉํ•ด ํ•™์Šต ๋ฐ์ดํ„ฐ์˜ ๋‹ค์–‘์„ฑ์„ ํ™•๋ณดํ–ˆ์Šต๋‹ˆ๋‹ค. ์ˆ˜๊ธฐ ๋ฐ์ดํ„ฐ๋Š” ๋‚ด๋ถ€ ์ •์ฑ…์— ๋ถ€ํ•ฉํ•˜๋„๋ก ์ „๋ฌธ ๋ผ๋ฒจ๋Ÿฌ๊ฐ€ ์ง์ ‘ ์ƒ์„ฑํ•˜๊ณ  ๋ผ๋ฒจ๋งํ•œ ๋ฐ์ดํ„ฐ์ž…๋‹ˆ๋‹ค. ํ•ฉ์„ฑ ๋ฐ์ดํ„ฐ๋Š” ํ•™์Šต ํšจ๊ณผ๋ฅผ ๋†’์ด๊ธฐ ์œ„ํ•ด LLM ๊ธฐ๋ฐ˜ ํ‘œํ˜„ ๋ณ€ํ™˜๊ณผ ๋…ธ์ด์ฆˆ ์‚ฝ์ž… ๋“ฑ ๋‹ค์–‘ํ•œ ๋ฐ์ดํ„ฐ ์ฆ๊ฐ• ๊ธฐ๋ฒ•์„ ํ†ตํ•ด ์ƒ์„ฑํ•˜์˜€์Šต๋‹ˆ๋‹ค. ์™ธ๋ถ€ ๋ฐ์ดํ„ฐ๋Š” ๊ณต๊ฐœ์ ์œผ๋กœ ์ด์šฉ ๊ฐ€๋Šฅํ•œ ์ถœ์ฒ˜์—์„œ ์ˆ˜์ง‘๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ํ•™์Šต ๋ฐ์ดํ„ฐ์—๋Š” ์•ˆ์ „ํ•˜์ง€ ์•Š์€ ๋ฐœํ™” ๋ฐ์ดํ„ฐ ์™ธ์—๋„, ๋ชจ๋ธ์˜ ๊ฑฐ์ง“ ์–‘์„ฑ(false positive) ๋น„์œจ์„ ์ค„์ด๊ธฐ ์œ„ํ•ด ์•ˆ์ „ํ•œ ์‚ฌ์šฉ์ž ๋ฐœํ™”๋„ ํฌํ•จ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค. ํ‰๊ฐ€ Kanana Safeguard-Siren์€ SAFE/UNSAFE ์ด์ง„ ๋ถ„๋ฅ˜ ๊ธฐ์ค€์œผ๋กœ ์„ฑ๋Šฅ์„ ํ‰๊ฐ€ํ–ˆ์Šต๋‹ˆ๋‹ค. ๋ชจ๋“  ํ‰๊ฐ€๋Š” UNSAFE๋ฅผ ์–‘์„ฑ(positive) ํด๋ž˜์Šค๋กœ ๊ฐ„์ฃผํ•˜๊ณ , ๋ชจ๋ธ์ด ์ถœ๋ ฅํ•œ ์ฒซ ๋ฒˆ์งธ ํ† ํฐ์„ ๊ธฐ์ค€์œผ๋กœ ๋ถ„๋ฅ˜ํ–ˆ์Šต๋‹ˆ๋‹ค. ์™ธ๋ถ€ ๋ฒค์น˜๋งˆํฌ ๋ชจ๋ธ์€ ๊ฐ ์ถœ๋ ฅ๊ฐ’์— ๋Œ€ํ•ด ๋‹ค์Œ๊ณผ ๊ฐ™์€ ๋ฐฉ์‹์œผ๋กœ ํ‰๊ฐ€ํ•˜์˜€์Šต๋‹ˆ๋‹ค. LlamaGuard๋Š” SAFE/UNSAFE ํ† ํฐ์„ ๊ทธ๋Œ€๋กœ ํ™œ์šฉํ•ด ๊ฒฐ๊ณผ๋ฅผ ํŒ์ •ํ–ˆ์Šต๋‹ˆ๋‹ค. ShieldGemma๋Š” ์ž„๊ณ„์น˜๋ฅผ 0.5๋กœ ์„ค์ •ํ•˜์—ฌ ์ด์ง„ ๋ถ„๋ฅ˜๋ฅผ ์ˆ˜ํ–‰ํ–ˆ์Šต๋‹ˆ๋‹ค. GPT-4o๋Š” ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ ๊ธฐ๋ฐ˜ ๋ถ„๋ฅ˜ ํ”„๋กฌํ”„ํŠธ๋ฅผ zero-shot ๋ฐฉ์‹์œผ๋กœ ์ž…๋ ฅํ•˜๊ณ , ์ถœ๋ ฅ ๋‚ด์šฉ์ด ํŠน์ • ์ฝ”๋“œ๋กœ ๋ถ„๋ฅ˜๋œ ๊ฒฝ์šฐ UNSAFE๋กœ ๊ฐ„์ฃผํ•˜์—ฌ ์ด์ง„ ๋ถ„๋ฅ˜๋ฅผ ์ˆ˜ํ–‰ํ–ˆ์Šต๋‹ˆ๋‹ค. ๊ทธ ๊ฒฐ๊ณผ ์ž์ฒด์ ์œผ๋กœ ๊ตฌ์ถ•ํ•œ ํ•œ๊ตญ์–ด ํ‰๊ฐ€ ๋ฐ์ดํ„ฐ์…‹์—์„œ Kanana Safeguard-Siren์˜ ๋ถ„๋ฅ˜ ์„ฑ๋Šฅ์ด ํƒ€ ๋ฒค์น˜๋งˆํฌ ๋ชจ๋ธ ๋Œ€๋น„ ๊ฐ€์žฅ ์šฐ์ˆ˜ํ•œ ์„ฑ๋Šฅ์„ ๋‚˜ํƒ€๋ƒˆ์Šต๋‹ˆ๋‹ค. ๋ชจ๋“  ๋ชจ๋ธ์€ ๋™์ผํ•œ ํ…Œ์ŠคํŠธ์…‹๊ณผ ๋ถ„๋ฅ˜ ๊ธฐ์ค€์œผ๋กœ ํ‰๊ฐ€๋˜์—ˆ์œผ๋ฉฐ, ์ •์ฑ… ๋ฐ ๋ชจ๋ธ ๊ตฌ์กฐ ์ฐจ์ด์— ๋”ฐ๋ฅธ ์˜ํ–ฅ์„ ์ตœ์†Œํ™”ํ•˜๊ณ , ๊ณต์ •ํ•˜๊ณ  ์‹ ๋ขฐ๋„ ๋†’์€ ๋น„๊ต๊ฐ€ ๊ฐ€๋Šฅํ•˜๋„๋ก ์„ค๊ณ„๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ํ•œ๊ณ„์  Kanana Safeguard-Siren์€ ๋‹ค์Œ๊ณผ ๊ฐ™์€ ํ•œ๊ณ„์ ์ด ์žˆ์œผ๋ฉฐ, ์ด๋Š” ํ–ฅํ›„ ์ง€์†์ ์œผ๋กœ ๊ฐœ์„ ํ•ด๋‚˜๊ฐˆ ์˜ˆ์ •์ž…๋‹ˆ๋‹ค. 1. ์˜คํƒ์ง€ ๊ฐ€๋Šฅ์„ฑ ์กด์žฌ ๋ณธ ๋ชจ๋ธ์€ 100% ์™„๋ฒฝํ•œ ๋ถ„๋ฅ˜๋ฅผ ๋ณด์žฅํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. ํŠนํžˆ, ๋ชจ๋ธ์˜ ์ •์ฑ…์€ ์ผ๋ฐ˜์ ์ธ ์‚ฌ์šฉ์‚ฌ๋ก€์— ๊ธฐ๋ฐ˜ํ•˜์—ฌ ์ˆ˜๋ฆฝ๋˜์—ˆ๊ธฐ ๋•Œ๋ฌธ์— ํŠน์ •ํ•œ ๋„๋ฉ”์ธ์—์„œ๋Š” ์ž˜๋ชป ๋ถ„๋ฅ˜๋  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. 2. Context ์ธ์‹ ๋ฏธ์ง€์› ๋ณธ ๋ชจ๋ธ์€ ์ด์ „ ๋Œ€ํ™” ์ด๋ ฅ์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๋ฌธ๋งฅ์„ ์œ ์ง€ํ•˜๊ฑฐ๋‚˜ ๋Œ€ํ™”๋ฅผ ์ด์–ด๊ฐ€๋Š” ๊ธฐ๋Šฅ์€ ์ œ๊ณตํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. 3. ์ œํ•œ๋œ ๋ฆฌ์Šคํฌ ์นดํ…Œ๊ณ ๋ฆฌ ๋ณธ ๋ชจ๋ธ์€ ์ •ํ•ด์ง„ ๋ฆฌ์Šคํฌ๋งŒ์„ ํƒ์ง€ํ•˜๋ฏ€๋กœ ์‹ค์‚ฌ๋ก€์˜ ๋ชจ๋“  ๋ฆฌ์Šคํฌ๋ฅผ ํƒ์ง€ํ•  ์ˆ˜๋Š” ์—†์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์˜๋„์— ๋”ฐ๋ผ Kanana Safeguard(์œ ํ•ดํ•œ ์ฝ˜ํ…์ธ  ํƒ์ง€), Kanana Safeguard-Prompt(ํ”„๋กฌํ”„ํŠธ ๊ณต๊ฒฉ ํƒ์ง€) ๋ชจ๋ธ๊ณผ ํ•จ๊ป˜ ์‚ฌ์šฉํ•˜๋ฉด ์ „์ฒด์ ์ธ ์•ˆ์ „์„ฑ์„ ๋”์šฑ ๋†’์ผ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. Contributors HyeYeon Cho, JeongHwan Lee, Deok Jeong, JiEun Choi

NaNK
llama
38
9

kanana-2-30b-a3b-instruct-2601

NaNK
โ€”
0
15

kanana-2-30b-a3b-thinking-2601

NaNK
โ€”
0
15

kanana-2-30b-a3b-base-2601

NaNK
โ€”
0
11

kanana-2-30b-a3b-mid-2601

NaNK
โ€”
0
11