Cloud Masters Episode #108
The cost impact of Large Language Models (LLMs) in production
We cover the ever-growing importance of Large Language Models (LLMs) in applications, how LLM costs can easily compound once in production, and breaking down the costs associated with using LLMs.
Cloud Masters Episode #108

With DoiT Spot Scaling, automate your AWS Spot Instances to save up to 90% on compute spend without compromising reliability.

Cloud Masters
Cloud Masters
The cost impact of Large Language Models (LLMs) in production
Loading
/
Cloud Masters
Cloud Masters
The cost impact of Large Language Models (LLMs) in production
Loading
/

Episode notes

About the guests

Gad Benram
Gad is the founder and CTO of TensorOps, which offers expert services for AI-driven applications as well AIOps and AI cost optimization
Gabriel Gonçalves
Gabriel is an ML Solutions Architect at TensorOps. He specializes in crafting intelligent solutions and architectures for Large Language Model applications.
Sascha Heyer
Sascha Heyer, a Senior Machine Learning Specialist at DoiT, stands out as a Google Developer Expert and Google Cloud Innovator. He has been crucial in helping over 306 companies grow in the field of Machine Learning. Sascha believes in keeping things simple, a mindset that has helped clarify complex tech concepts. Moreover, as an author, this expertise is showcased through engaging YouTube presentations and insightful Medium articles (https://medium.com/@saschaheyer), effectively demystifying complex tech topics for a broad audience.
Gad is the founder and CTO of TensorOps, which offers expert services for AI-driven applications as well AIOps and AI cost optimization
Gabriel is an ML Solutions Architect at TensorOps. He specializes in crafting intelligent solutions and architectures for Large Language Model applications.
Sascha Heyer, a Senior Machine Learning Specialist at DoiT, stands out as a Google Developer Expert and Google Cloud Innovator. He has been crucial in helping over 306 companies grow in the field of Machine Learning. Sascha believes in keeping things simple, a mindset that has helped clarify complex tech concepts. Moreover, as an author, this expertise is showcased through engaging YouTube presentations and insightful Medium articles (https://medium.com/@saschaheyer), effectively demystifying complex tech topics for a broad audience.

Related content

No longer a pipe dream — Gen AI and data pipelines
Exploring the impact that Gen AI will have on data pipelines and data engineering overall.
Observability of LLMs in Google Cloud
ML and AI specialists Eduardo Mota and Sascha Heyer join us to explore the complexities of observability of LLM-powered features. Packed with tons of real-life customer anecdotes and best practices, they discuss the challenges and strategies for monitoring Gen AI systems, emphasizing the importance of metrics in understanding system interactions, especially given Gen AI’s non-deterministic nature.
Gaining visibility over your LLMs with LLMStudio
Mentioned in the podcast, LLMStudio is designed to streamline interactions with large language models (LLMs). It focuses on prompt engineering, a critical skill for getting the most out of LLMs.

Connect With Us