In the early morning of May 22,Microsoft The Build 2024 Developer Conference was held in Seattle, USA. According to Satya Nadella, CEO and Chairman of Microsoft, OpeanAI The latest flagship model developed GPT-4o, now available in Azure AI Studio and as an API, this multimodal model integrates text, image, and audio processing capabilities to enable new generative and conversational AI experiences.
Additionally, Phi-3-vision, a new multimodal model from the Phi-3 AI Small Language Model (SLM) developed by Microsoft, is now available in Azure. The Phi-3 model is powerful, cost-effective, and optimized for personal devices. Phi-3-vision provides the ability to input images and text and receive text responses. For example, users can ask questions about a chart, or ask open-ended questions about a specific image.
Developers can experiment with these cutting-edge models in Azure AI Playground, and can start building and customizing with models in Azure AI Studio.