Results for ""
In a statement, the company said Trainium will be available from 2021. In addition to delivering the most cost-effective ML training, Trainium offers the highest performance with the most teraflops (TFLOPS) of compute power for ML in the cloud and enables a broader set of ML applications.
The Trainium chip is specifically optimized for deep learning training workloads for applications including image classification, semantic search, translation, voice recognition, natural language processing and recommendation engines.
Last year, AWS launched AWS Inferentia, a custom chip that provides customers with high performance ML inference at the lowest cost in the cloud. While Inferentia addressed the cost of inference, which constitutes up to 90% of ML infrastructure costs, many development teams are also limited by fixed ML training budgets, putting a cap on the scope and frequency of training needed to improve their models and applications. AWS Trainium addresses this challenge by providing the highest performance and lowest cost for ML training in the cloud.
With both Trainium and Inferentia, customers will have an end-to-end flow of ML compute from scaling training workloads to deploying accelerated inference.
AWS Trainium shares the same AWS Neuron SDK as AWS Inferentia making it easy for developers using Inferentia to get started with Trainium. Because the Neuron SDK is integrated with popular ML frameworks including TensorFlow, PyTorch, and MXNet developers can easily migrate to AWS Trainium from GPU-based instances with minimal code changes. AWS Trainium will be available via Amazon EC2 instances and AWS Deep Learning AMIs as well as managed services including Amazon SageMaker, Amazon ECS, EKS, and AWS Batch.