xxxxxccc/mediaDescr_2epoch_Mistral-Nemo-Base-2407_model
The xxxxxccc/mediaDescr_2epoch_Mistral-Nemo-Base-2407_model is a 12 billion parameter Mistral-based language model developed by xxxxxccc, fine-tuned from unsloth/Mistral-Nemo-Base-2407-bnb-4bit. This model was trained 2x faster using Unsloth and Huggingface's TRL library, offering a 32768 token context length. It is optimized for efficient training and deployment of Mistral-architecture models.