The product integrates the latest AI accelerators such as NVIDIA AI compute and Trainium chips with high-speed networking, high-performance storage, databases, and energy-efficient infrastructure with AI services such as Amazon Bedrock and Amazon SageMaker to enable rapid development and deployment of large-scale AI applications.
As governments and large organizations look to scale their AI projects, many are turning to the concept of “AI factories” to meet their own sovereignty and compliance requirements. But building a high-performance AI factory requires a complex mix of management, database, storage, and security services, a challenge that most organizations prefer not to tackle on their own. To address this, AWS launched AWS AI Factories to provide enterprises and governments with purpose-built AI infrastructure deployed directly in their data centers. The product integrates the latest AI accelerators such as NVIDIA AI compute and Trainium chips with high-speed networking, high-performance storage, databases, and energy-efficient infrastructure with AI services such as Amazon Bedrock and Amazon SageMaker to enable rapid development and deployment of large-scale AI applications.
Regulated industries and public sector organizations face significant hurdles when deploying large-scale AI initiatives. Building an in-house AI capability requires coordinating procurement cycles, choosing the right AI model, and managing licenses from multiple AI providers, as well as significant capital investments in GPUs, data centers, and power. These challenges often result in multi-year schedules and operational complexities that take focus away from core business objectives. AWS AI Factories tackles these obstacles by providing dedicated AI infrastructure operated exclusively for customers, acting like a private AWS Region that ensures secure, low-latency access to compute, storage, database, and AI services. This approach allows organizations to leverage existing data center space and power while meeting strict security, data sovereignty, and regulatory requirements while accessing AWS AI infrastructure and managed services.
Over 15 years of collaboration between AWS and NVIDIA powers the deployment of advanced AI infrastructure. This integration enables customers to run large language models more efficiently, securely, and at scale within their data centers. AWS AI Factories provides seamless access to NVIDIA accelerated computing platforms, full-stack AI software, and GPU-accelerated applications to enhance performance and scalability. This infrastructure supports NVIDIA Grace Blackwell and Vera Rubin platforms, and in the future will also support NVIDIA NVLink Fusion interconnects for Trainium4 and Graviton chips. This combination accelerates time to market for AI solutions while optimizing data center performance.
AWS AI Factories is designed to support public sector AI adoption by meeting rigorous security standards across multiple classification levels. This infrastructure provides governments around the world with the reliability, control, and scalability they need to leverage AI technology for economic and technological growth. AWS is partnering with Saudi Arabia’s HUMAIN to build a dedicated “AI Zone” that will house up to 150,000 AI chips, dedicated infrastructure, and AI services in dedicated data centers. This initiative aims to address local and global AI computing demands while fostering a scalable ecosystem for AI innovation.
