AWS re:Invent 2025 Key Announcements

AWS re:Invent 2025 delivered one of the most transformative waves of innovation we’ve seen in years. These weren’t just incremental updates, but foundational leaps that reshape how we build, automate, and scale on the AWS cloud.  

From agent-driven development and next-generation AI infrastructure to major breakthroughs in data, security and serverless, AWS has clearly shifted into a new era of builders’ tooling. Whether you’re leading platforms, shipping products, running ML workloads or modernising operations, this year’s announcements redefine what’s possible and what will soon become the new normal.  

Below is a curated breakdown of the biggest releases you need to know and why they matter. 

Table of Contents

AI platforms, Agents & the Developer Experience 

AWS leaned hard into an agent-first developer experience this year. Amazon Bedrock AgentCore was introduced as a fully managed runtime for production AI agents, alongside new “frontier agents” such as Kiro (virtual developer), plus AWS Security Agent and AWS DevOps Agent designed to work through common SDLC and operational tasks. To support custom agent development, AWS expanded Strands Agents with TypeScript, CDK integration and edge deployment, making it easier than ever to build reliable, domain-specific agents. 

This is great because AgentCore and Frontier Agents finally take over all the repetitive coding and ops tasks we normally burn hours on, so we can ship features faster and give customers improvements in days instead of weeks. 

 

Models & Customisation: Nova, Bedrock & SageMaker AI

AWS made model choice and customisation a lot more practical this year. The new Nova 2 model family and Nova Forge, introduce “open training” style customisation using pre-trained checkpoints and curated datasets. Nova Act is specialised for UI and browser automation.

On top of that, Bedrock Reinforcement Fine-Tuning (RFT) enables alignment-style tuning for domain accuracy, while SageMaker AI now supports serverless customisation, cutting experimentation cycles from months to days. 

The takeaway is simple: it’s now easier to tailor AI to real business context without a heavy ML ops lift, so customers can roll out smarter, more relevant experiences faster.

Trainium3 UltraServers: Next-Generation AWS AI Silicon 

AWS unveiled Trainium3 UltraServers, a new 3nm AI accelerator platform enabling frontier-scale model training and high-throughput inference. With massive performance and efficiency improvements over Trainium2, these systems are positioned to power the next generation of cost-efficient LLM workloads both inside AWS and for early adopters like Anthropic. 

This will:

  • Lower cost to train and run big models, making advanced AI use cases more commercially viable

  • Higher performance per watt, which helps at scale when workloads run continuously

  • More headroom for ambitious features, unlocking AI capabilities that were previously hard to justify on budget

New NVIDIA GB300 UltraServers on EC2 

AWS introduced P6e-GB300 UltraServers, based on NVIDIA’s GB300 NVL72 architecture, enabling ultra-large model training and high-bandwidth reasoning at scale. These instances bring frontier GPU performance into the EC2 ecosystem with deep integration into Nitro, EKS, and AWS’s ML stack. 

These servers are amazing as they will:

  • Unlock extreme GPU scale on demand, without specialised hardware investment

  • Accelerate training and high-throughput inference, especially for large reasoning workloads

  • Simplify production deployment, thanks to tighter integration with EKS, networking and the broader AWS ML toolchain

AWS AI Factories for Sovereign & Regulated Workloads

AWS AI Factories allow organisations to deploy managed, hyperscale AI infrastructure inside their own data centers. These combine Trainium, NVIDIA GPUs, AWS networking, storage, databases, Bedrock, and SageMaker into a fully managed “private AWS region,” ideal for sovereign workloads and regulated industries. 

AI Factories are a game-changer for our regulated customers as they will:

  • Enable modern AI adoption while keeping data on-prem, supporting sovereignty and stricter regulatory needs

  • Bring AWS-managed AI infrastructure behind the firewall, reducing operational burden for customers

  • Accelerate time-to-value for regulated use cases, without compromising control, governance, or residency requirements

Amazon S3 Updates: Vectors, 50 TB Objects and Data Lake Enhancements 

Amazon S3 received major AI-era upgrades: S3 Vectors (high-scale vector search built into S3), 50 TB object support for massive datasets, 10× faster S3 Batch Operations, and enhancements to S3 Tables for Apache Iceberg like Intelligent-Tiering and cross-account replication.  

We love this, because S3 is basically becoming an AI-native database, which will:

This will:

  • Make vector search more native and scalable, reducing the need to stitch together extra services

  • Support much larger datasets more simply, especially for AI and analytics workloads

  • Speed up large-scale data operations, improving throughput for batch processing and lakehouse management

Lambda Upgrades: Durable Functions and Managed Instances

AWS Lambda now supports durable functions, enabling long-running workflows that persist for up to a year. Lambda Managed Instances give developers EC2-level control (hardware, networking, pricing) with the ergonomics of Lambda which is ideal for mission-critical pipelines and predictable workloads. 

This will:

  • Enable long-running, stateful workflows in serverless, without redesigning architectures

  • Give better cost and performance control, while keeping the simplicity of Lambda

  • Speed up delivery for event-driven and AI workflows, especially where orchestration and reliability matter most

Creation of a lambda function with Durable Function section highlighted. AWS ReInvent 2025

Networking & Multicloud: AWS Interconnect with Google Cloud 

AWS Interconnect – multicloud, built in partnership with Google Cloud, delivers private, high-bandwidth links between clouds using an open spec and coordinated operations, pushing multicloud from DIY networking to a fully managed experience. 

This will:

  • Finally make multicloud sane, eliminating the historical networking pain between AWS and GCP

  • Allow teams to combine services from both clouds without complex workarounds or fragile architectures

  • Enable customers to adopt true best-of-both-worlds solutions that maximise performance, resilience and choice

 

Security Enhancements: GuardDuty ETD, Security Hub Analytics and IAM Policy Autopilot

GuardDuty now provides Extended Threat Detection for EC2 and ECS, correlating host- and container-level signals into actionable, MITRE-mapped findings. Security Hub adds real-time risk analytics across AWS security services. IAM Policy Autopilot, an open-source MCP server, generates least-privilege IAM policies directly from your code. 

This will:

  • Reduce noisy alerts and cut through security signal overload

  • Automate least-privilege IAM, replacing manual policy tuning with intelligent defaults

  • Strengthen overall security posture while significantly reducing operational burden

  • Help teams avoid late-night fire drills by surfacing only the issues that actually matter

 

Customer Experience: Amazon Connect Agentic AI 

Amazon Connect now includes agentic AI for both self-service and human-assisted workflows. With Nova-powered speech models, Connect can automate tasks, generate documentation, analyse sentiment, and assist human agents with context-aware recommendations. 

This will:

  • Deliver quicker, more natural support for customers through agentic AI

  • Provide human agents with real-time guidance and next-best actions

  • Reduce wait times and boost service quality immediately

 

AI-Native Observability: CloudWatch and Related Services

CloudWatch gained AI-native observability for agents and LLMs, automatic application topology discovery, AI-generated incident reports, and expanded cross-account/Region telemetry via Logs Centralisation and Database Insights. 

This will:

  • Automate much of the detective work normally done by engineers

  • Spot issues earlier, identify root causes and analyse incidents faster

  • Improve system reliability and reduce time to recovery for customers

 

Dashboard view of a CloudWatch instance with GenAI observability enabled. AWS ReInvent 2025

Data Collaboration & Core Services: Clean Rooms, EKS, Route 53 Global Resolver, Partner Central 

AWS Clean Rooms introduced Clean Rooms ML synthetic data, a new capability that lets multiple parties collaborate on machine learning without sharing raw data. It generates noisy, de-identified synthetic datasets that preserve the statistical patterns of joint datasets (e.g., retailers + brands), enabling compliant ML training while reducing re-identification risk. Beyond data collaboration, AWS added new EKS capabilities for managed workload orchestration and cloud resource governance, launched Route 53 Global Resolver to unify DNS resolution across hybrid and multicloud environments, and integrated AWS Partner Central directly into the AWS Console for easier partner and marketplace management. 

This will:

  • Allow teams to collaborate on shared datasets without exposing sensitive information

  • Enable new joint modelling and analytics opportunities with partners

  • Unlock data-driven projects that were previously impossible due to privacy constraints

 

Updated AWS Support Experience

AWS reorganised support into Business Support+, Enterprise Support, and Unified Operations, with AI-powered assistance, faster response times, deeper security integration, and alignment with the new AI-driven operational tooling. 

This will:

  • Provide faster SLAs across incidents and operational requests

  • Use AI-powered troubleshooting to accelerate diagnostics and resolution

  • Improve uptime and reduce delays, delivering a smoother customer experience

 

Conclusion 

AWS re:Invent 2025 marks a turning point for builders on AWS, with breakthroughs across AI, compute, serverless, data, security, and operations that directly improve how teams design, deploy, and scale applications: 

  • AI becomes first-class with AgentCore, frontier agents, and expanded Bedrock/Nova models, giving developers production-ready AI building blocks instead of DIY glue and model customisation gets easier thanks to Nova Forge, RFT, and SageMaker AI serverless customisation, making domain-specific AI accessible without ML ops complexity. 
  • Lambda steps up with durable functions and Managed Instances, enabling long-running AI workflows and cost-optimised compute without losing serverless ergonomics. 
  • Multicloud networking with AWS Interconnect makes cross-cloud architectures reliable, high-bandwidth, and finally simple. 
  • Security shifts left and gets smarter with GuardDuty ETD, Security Hub analytics, and IAM Policy Autopilot, thus reducing noise and tightening least-privilege by default. 
  • CloudWatch’s AI-first observability and improved support tiers give teams faster insights, better incident response, and more intelligent operations. 

Across the board, these releases remove friction, automate complexity, and give builders more powerful primitives. Developing on AWS becomes faster, safer, more scalable, and far more AI-native than ever before, setting the stage for how modern applications will be built in the years ahead. 

The AWS re:Invent 2025 announcements open the door to a new era of AI-native, high-performance cloud engineering. If you’re ready to explore how agentic AI, Trainium3, S3 Vectors, multicloud networking or the latest serverless and security capabilities can accelerate your delivery, Cevo can help.

Whether you’re scaling platforms, modernising workloads, or building out new AI-powered capabilities, our consultants partner with your team to design solutions that are robust, secure and future-ready.

Move faster with confidence. Reach out to Cevo to start the conversation.

Enjoyed this blog?

Share it with your network!

Move faster with confidence