kaist-ai/mistral-orpo-capybara-7k

kaist-ai/mistral-orpo-capybara-7k is a 7 billion parameter language model developed by KAIST AI, fine-tuned from Mistral-7B-v0.1 using the Odds Ratio Preference Optimization (ORPO) method. This model is specifically optimized for multi-turn conversational tasks, leveraging a distilled Capybara dataset. It demonstrates strong performance in alignment benchmarks like MT-Bench and AlpacaEval, making it suitable for dialogue-based applications.

Warm
Public
7B
FP8
8192
License: mit
Hugging Face