AWS Deep Learning Containers
One stop shop for running AI/ML on AWS
Docs ยท Available Images ยท Tutorials
About¶
AWS Deep Learning Containers (DLCs) are pre-built Docker images for running AI/ML workloads on AWS. Each image is tested and patched for security vulnerabilities. For more details, visit our documentation.
๐ฅ What's New¶
๐ Release Highlights¶
- [2026/04/28] vLLM v0.20.0 โ EC2:
0.20-gpu-py312-ec2ยท SageMaker:0.20-gpu-py312ยท Introduces support for DeepSeek V4. - [2026/04/20] vLLM v0.19.1 โ EC2:
0.19-gpu-py312-ec2ยท SageMaker:0.19-gpu-py312ยท This upgrades Transformers to 5.5.4, enabling Gemma 4 support. - [2026/04/07] SGLang v0.5.10 โ EC2:
0.5.10-gpu-py312-ec2ยท SageMaker:0.5.10-gpu-py312 - [2026/04/07] vLLM v0.19.0 โ EC2:
0.19-gpu-py312-ec2ยท SageMaker:0.19-gpu-py312 - [2026/03/26] vLLM v0.18.0 โ EC2:
0.18-gpu-py312-ec2ยท SageMaker:0.18-gpu-py312
๐ข Support Updates¶
- [2026/04/28] We cannot guarantee security patching on Ubuntu-based vLLM and SGLang images due to the lack of Ubuntu Pro licensing. Customers may continue using these images at their own discretion and risk. We recommend migrating to our Amazon Linux-based images.
- [2026/02/10] Extended support for PyTorch 2.6 Inference containers until June 30, 2026
- PyTorch 2.6 Inference images will continue to receive security patches and updates through end of June 2026
- For complete framework support timelines, see our Support Policy
๐ Blog Posts¶
- Distributed Training on Amazon EKS - Configure and validate a distributed training cluster with DLCs on Amazon EKS.
- DLCs with Amazon SageMaker AI & MLflow - Use DLCs with SageMaker AI managed MLflow for experiment tracking and model management.
- LLM Serving on Amazon EKS with vLLM - Deploy and serve LLMs on Amazon EKS using vLLM DLCs.
- Fine-tuning Meta Llama 3.2 Vision - Fine-tune and deploy Llama 3.2 Vision for web automation using DLCs, Amazon EKS, and Amazon Bedrock.
- DLCs with Amazon Q Developer and MCP - Streamline deep learning environments with Amazon Q Developer and Model Context Protocol.
๐ Workshop¶
- LLM Deployment on Amazon EKS - Deploy and optimize LLMs on Amazon EKS using vLLM DLCs. See also: Sample Code
License¶
This project is licensed under the Apache-2.0 License.