James Ding
Feb 26, 2025 15:38
Microsoft unveils new Phi SLMs, together with the multimodal Phi-4, educated on NVIDIA GPUs, enhancing AI capabilities with environment friendly useful resource utilization.
Microsoft has introduced the most recent additions to its Phi household of small language fashions (SLMs), that includes the brand new Phi-4-multimodal and Phi-4-mini fashions, each educated utilizing NVIDIA GPUs. This growth marks a big step within the evolution of language fashions, specializing in effectivity and flexibility, in keeping with NVIDIA.
Developments in Small Language Fashions
SLMs have emerged as a sensible answer to the challenges posed by giant language fashions (LLMs), which, regardless of their capabilities, require substantial computational sources. SLMs are designed to function effectively inside constrained environments, making them appropriate for deployment on gadgets with restricted reminiscence and computational energy.
Microsoft’s new Phi-4-multimodal mannequin is especially noteworthy for its skill to course of a number of kinds of information, together with textual content, audio, and pictures. This functionality opens up new potentialities for purposes resembling automated speech recognition, translation, and visible reasoning. The mannequin’s coaching concerned 512 NVIDIA A100-80GB GPUs over 21 days, underscoring the intensive computational efforts required to realize its capabilities.
Phi-4-multimodal and Phi-4-mini
The Phi-4-multimodal mannequin boasts 5.6 billion parameters and has demonstrated superior efficiency in automated speech recognition, rating first on the Huggingface OpenASR leaderboard with a phrase error charge of 6.14%. This achievement highlights the mannequin’s potential in enhancing speech recognition applied sciences.
Alongside Phi-4-multimodal, Microsoft additionally launched Phi-4-mini, a text-only mannequin optimized for chat purposes. With 3.8 billion parameters, Phi-4-mini is designed to deal with long-form content material effectively, providing a context window of 128K tokens. Its coaching concerned 1024 NVIDIA A100 80GB GPUs over 14 days, reflecting the mannequin’s concentrate on high-quality academic information and code.
Deployment and Accessibility
Each fashions can be found on Microsoft’s Azure AI Foundry, offering a platform for designing, customizing, and managing AI purposes. Customers may discover these fashions by the NVIDIA API Catalog, which gives a sandbox setting for testing and integrating these fashions into varied purposes.
NVIDIA’s collaboration with Microsoft extends past simply coaching these fashions. The partnership contains optimizing software program and fashions like Phi to advertise AI transparency and assist open-source tasks. This collaboration goals to advance AI expertise throughout industries, from healthcare to life sciences.
For extra detailed data, go to the NVIDIA weblog.
Picture supply: Shutterstock