In this course we'll learn how to design, develop, deploy and maintain ML applications at scale. While we will finetune an LLM for a text classification task, our focus will largely be around the model agnostic aspects of production ML. We'll cover the best practices for experimentation, as well as using mature CI/CD pipelines to launch, maintain and iterate (continual learning) on our application.
We will use Ray AI Runtime to implement core ML workloads at scale (data ingest/processing, training, hyperparameter tuning, serving, etc.) and to integrate with mature MLOps tooling for contextual workloads (experiment tracking, data validation, model monitoring, etc.).
By the end of this course, you'll know how to take an ML application from prototype to production and apply it to your own bespoke context, stack and scale.