This course will teach you how to deploy and manage large language models (LLMs) in production using AWS services like Amazon Bedrock. By the end of the course, you will know how to:
This course will teach you how to deploy and manage large language models (LLMs) in production using AWS services like Amazon Bedrock. By the end of the course, you will know how to:
Choose the right LLM architecture and model for your application using services.
Optimize cost, performance and scalability of LLMs on AWS using auto-scaling groups, spot instances and container orchestration
Monitor and log metrics from your LLM to detect issues and continuously improve quality
Build reliable and secure pipelines to train, deploy and update models using AWS services
Comply with regulations when deploying LLMs in production through techniques like differential privacy and controlled rollouts
This course is unique in its focus on real-world operationalization of large language models using AWS. You will work through hands-on labs to put concepts into practice as you learn. Whether you are a machine learning engineer, data scientist or technical leader, you will gain practical skills to run LLMs in production.
OpenCourser helps millions of learners each year. People visit us to learn workspace skills, ace their exams, and nurture their curiosity.
Our extensive catalog contains over 50,000 courses and twice as many books. Browse by search, by topic, or even by career interests. We'll match you to the right resources quickly.
Find this site helpful? Tell a friend about us.
We're supported by our community of learners. When you purchase or subscribe to courses and programs or purchase books, we may earn a commission from our partners.
Your purchases help us maintain our catalog and keep our servers humming without ads.
Thank you for supporting OpenCourser.