xxxxxccc/mediaDescr_2epoch_Mistral-Nemo-Base-2407_model

The xxxxxccc/mediaDescr_2epoch_Mistral-Nemo-Base-2407_model is a 12 billion parameter Mistral-based language model developed by xxxxxccc, fine-tuned from unsloth/Mistral-Nemo-Base-2407-bnb-4bit. This model was trained 2x faster using Unsloth and Huggingface's TRL library, offering a 32768 token context length. It is optimized for efficient training and deployment of Mistral-architecture models.

Warm
Public
12B
FP8
32768
License: apache-2.0
Hugging Face