In April this year, Microsoft announced Phi-3 family of small language models (SLMs). The Phi-3-mini with a 3.8 billion parameter language model is trained on 3.3 trillion tokens and it beats Mixtral 8x7B and GPT-3.5. Microsoft’s recently announced Copilot+ PCs which use large language models (LLMs) running in Azure Cloud in concert with several of […]
The post Apple’s On-Device model performs better than Microsoft’s Phi-3-mini, Server model comparable to GPT-4 Turbo appeared first on MSPoweruser.