microsoft/phi-2-pytdml

Microsoft's Phi-2 is a 2.7 billion parameter Transformer-based causal language model, trained on a diverse dataset including synthetic NLP texts and filtered web data. It demonstrates near state-of-the-art performance among models under 13 billion parameters in common sense, language understanding, and logical reasoning benchmarks. This model is optimized for DirectML performance with fused operators and is primarily intended for research into safety challenges, as well as QA, chat, and code generation tasks.

Warm
Public
3B
BF16
2048
License: mit
Hugging Face