Scale Your Large Training Jobs with Data and Model Parallelism

March 16, 2021

In the past, accessing a graphic processing unit (GPU) to accelerate your data processing or scientific simulation code was difficult. Today, startups can log on to their AWS console and choose from a range of GPU-based Amazon EC2 instances on demand. Whether you're looking for a do-it-yourself or a fully managed approach, we'll show you how to choose the right instance on AWS to meet your target performance goals. You'll learn how to maximize resource utilization to find performance bottlenecks, and how to reduce overall training and inference costs.

Previous Video
Confessions of AI/ML Startup Founders
Confessions of AI/ML Startup Founders

To help learn from those who’ve done it before, we’ve gathered AI/ML founders from some of the world’s top ...

Next Video
Scaling Your Startup: What to Expect When Building an ML Team
Scaling Your Startup: What to Expect When Building an ML Team

Discover how to set up a data lake and implement it into an ML experiment workflow, how to prepare an end-t...