GenAI and LLMs on AWS

Software > Computer Software > Educational Software Duke University

Course Overview

This course will teach you how to deploy and manage large language models (LLMs) in production using AWS services like Amazon Bedrock. By the end of the course, you will know how to: Choose the right LLM architecture and model for your application using services. Optimize cost, performance and scalability of LLMs on AWS using auto-scaling groups, spot instances and container orchestration Monitor and log metrics from your LLM to detect issues and continuously improve quality Build reliable and secure pipelines to train, deploy and update models using AWS services Comply with regulations when deploying LLMs in production through techniques like differential privacy and controlled rollouts This course is unique in its focus on real-world operationalization of large language models using AWS. You will work through hands-on labs to put concepts into practice as you learn. Whether you are a machine learning engineer, data scientist or technical leader, you will gain practical skills to run LLMs in production.

Course FAQs

What are the prerequisites for 'GenAI and LLMs on AWS'?

Prerequisites for this continuing education class are set by Duke University. Most professional development online classes benefit from some prior knowledge. Please check the provider's page for specific requirements.

Will I receive a certificate for this CE class?

Yes, upon successful completion, Duke University typically offers a shareable certificate to showcase your new skills and fulfill your continuing education requirements.

How long does this online course take to complete?

Completion times for online continuing education courses vary. The provider's website will have the most accurate estimate of the time commitment needed.