Gensyn/Qwen2.5-0.5B-Instruct
Gensyn/Qwen2.5-0.5B-Instruct is an unmodified 0.49 billion parameter instruction-tuned causal language model from the Qwen2.5 family, featuring a 32,768 token context length. Developed by Qwen, this model utilizes a transformer architecture with RoPE, SwiGLU, and RMSNorm. It is specifically intended for local fine-tuning via peer-to-peer reinforcement learning within the Gensyn RL Swarm system. Its primary use case is as a base model for distributed RL training, after which it can be deployed in general workflows.