Fast and Flexible Inference for Open-Source AI Models at Scale with Azure

Microsoft Events presents a detailed session from Ignite 2025 focused on deploying and scaling open-source AI models using Azure Container Apps and AKS. The speakers showcase robust strategies for running custom and OSS models efficiently.

Fast and Flexible Inference for Open-Source AI Models at Scale with Azure

Session Overview

Presented at Microsoft Ignite 2025, this session explores how to deploy and operationalize open-source AI models at scale on Azure. The speakers, Mehrdad Abdolghafari, Cary Chai, and Sachi Desai, demonstrate how to leverage hybrid architectures and cloud technologies for high-performance and cost-efficient AI workloads.


Key Topics Covered


Technical Insights


Resources


Speaker Highlights

Real-World Example


Chapters Summary