top of page

The Future of AI Infrastructure on the Cloud - (AI Lover podcast)

Writer's picture: TensorOpsTensorOps

AI Lover Episode #5

Gad Benram⁠ and ⁠Charles Frye⁠ from ⁠Modal⁠ explore the strategic reasons behind companies choosing to host their own AI infrastructure versus relying on external cloud services. From controlling critical data to customizing AI applications.











Key topics include:

• 00:00 Introduction: Insights on AI Resources for Hosting AI Models

• 03:11 The Challenges of Existing Cloud Services

• 09:14 Introducing Modal: A Fast and Interactive Development Experience

• 15:13 Different Infrastructure Needs for Data Teams

• 19:42 Addressing Slowness in AI Services

• 26:20 Python and Notebooks for Data Scientists

• 33:35 Fast and Seamless Deployment with Modal

• 40:46 Future Directions and Closing Remarks




In this episode, Gad Benram and Charles Frye discuss the challenges of hosting AI models in production and the limitations of existing cloud services. They highlight the lack of resources and GPUs available for serving AI applications and the slow bootstrapping process. They introduce Modal, a serverless runtime for distributed applications built on top of cloud resources, as a solution to these challenges.

Modal offers fast deployment times, interactive development workflows, and support for large-scale models.


Does your AI project qualify for Google Cloud sponsorship?


Comments


Sign up to get updates when we release another amazing article

Thanks for subscribing!

bottom of page