Skip to main content

DISCOVER, CONNECT, GROW

Microsoft Reactor

Join Microsoft Reactor and engage with startups and developers live

Ready to get started with AI?  Microsoft Reactor provides events, training, and community resources to help startups, entrepreneurs and developers build their next business on AI technology. Join us!

DISCOVER, CONNECT, GROW

Microsoft Reactor

Join Microsoft Reactor and engage with startups and developers live

Ready to get started with AI?  Microsoft Reactor provides events, training, and community resources to help startups, entrepreneurs and developers build their next business on AI technology. Join us!

Go back

Build a multi-LLM chat application with Azure Container Apps

11 April, 2024 | 3:30 PM - 3:45 PM (UTC) Coordinated Universal Time

  • Format:
  • alt##LivestreamLivestream

Topic: Coding, Languages, and Frameworks

Language: English

In this demo, explore how to leverage GPU workload profiles in ACA to run your own model backend, and easily switch, compare, and speed up your inference times. You will also explore how to leverage LlamaIndex (https://github.com/run-llama/llama_index) to ingest data on-demand, and host models using Ollama (https://github.com/ollama/ollama). Then finally, decompose the application as a set of microservices written in Python, deployed on ACA.

  • Azure

Speakers

Related Events

The events below may be of interest to you as well. Be sure to visit our Reactor homepage to see all available events.

For questions please contact us at reactor@microsoft.com