A family of bilingual JA/EN LLMs. https://shisa.ai/posts/shisa-v2.1/
AI & ML interests
None defined yet.
-
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing
Paper • 2406.08464 • Published • 71 -
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Paper • 2406.20094 • Published • 104 -
argilla/magpie-ultra-v1.0
Viewer • Updated • 3.22M • 405 • 50 -
simplescaling/s1K-1.1
Viewer • Updated • 1k • 2.38k • 144
JA/EN Bilingual LLMs
A family of bilingual JA/EN LLMs
Comparing Efficiency and Quality of various formats
-
cyberagent/Mistral-Nemo-Japanese-Instruct-2408
Text Generation • 12B • Updated • 1.18k • 47 -
shisa-ai/Mistral-Nemo-Japanese-Instruct-FP8-Dynamic
12B • Updated • 2 -
shisa-ai/Mistral-Nemo-Japanese-Instruct-2408-SQ-GPTQ-W8A8-INT8
12B • Updated • 6 -
shisa-ai/Mistral-Nemo-Japanese-Instruct-2408-GPTQ-W4A16-gs32
12B • Updated • 3
A family of bilingual JA/EN LLMs. https://shisa.ai/posts/shisa-v2.1/
A family of bilingual JA/EN LLMs
-
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing
Paper • 2406.08464 • Published • 71 -
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Paper • 2406.20094 • Published • 104 -
argilla/magpie-ultra-v1.0
Viewer • Updated • 3.22M • 405 • 50 -
simplescaling/s1K-1.1
Viewer • Updated • 1k • 2.38k • 144
Comparing Efficiency and Quality of various formats
-
cyberagent/Mistral-Nemo-Japanese-Instruct-2408
Text Generation • 12B • Updated • 1.18k • 47 -
shisa-ai/Mistral-Nemo-Japanese-Instruct-FP8-Dynamic
12B • Updated • 2 -
shisa-ai/Mistral-Nemo-Japanese-Instruct-2408-SQ-GPTQ-W8A8-INT8
12B • Updated • 6 -
shisa-ai/Mistral-Nemo-Japanese-Instruct-2408-GPTQ-W4A16-gs32
12B • Updated • 3
JA/EN Bilingual LLMs