denial07/Qwen2-72B-Instruct-kor-dpo

The denial07/Qwen2-72B-Instruct-kor-dpo model is an instruction-tuned large language model with 72.7 billion parameters and a 131,072-token context length, based on the Qwen2-72B-Instruct architecture. This version is specifically improved for Korean language performance, demonstrating enhanced capabilities across various Korean benchmarks. It is optimized for general instruction-following tasks in Korean, including reasoning, math, writing, and coding.

Warm
Public
72.7B
FP8
131072
License: tongyi-qianwen
Hugging Face