2026-02-12

2/12/2026, 12:00:00 AM ~ 2/13/2026, 12:00:00 AM (UTC) Recent Announcements Amazon Bedrock increases default quotas for Anthropic’s Claude Sonnet 4.5 model in AWS GovCloud (US) Amazon Bedrock has increased the default quotas for Anthropic’s Claude Sonnet 4.5 in AWS GovCloud (US-West) and AWS GovCloud (US-East) to 5,000,000 tokens per minute and 1,000 requests per minute, aligning with commercial AWS regions. This 25x increase enables customers to scale their AI workloads more effectively in regulated environments....

February 12, 2026

2025-12-10

12/10/2025, 12:00:00 AM ~ 12/11/2025, 12:00:00 AM (UTC) Recent Announcements Now generally available: Amazon EC2 C8gb instances Today, AWS announces the general availability of the new Amazon Elastic Block Storage (Amazon EBS) optimized Amazon Elastic Compute Cloud (Amazon EC2) C8gb instances. These instances are powered by AWS Graviton4 processors to deliver up to 30% better compute performance than AWS Graviton3 processors. At up to 150 Gbps of EBS bandwidth, these instances offer higher EBS performance compared to same-sized equivalent Graviton4-based instances....

December 10, 2025

2025-12-02

12/2/2025, 12:00:00 AM ~ 12/3/2025, 12:00:00 AM (UTC) Recent Announcements Announcing Amazon EC2 General purpose M8azn instances (Preview) Starting today, new general purpose high-frequency high-network Amazon Elastic Compute Cloud (Amazon EC2) M8azn instances are available for preview. These instances are powered by fifth generation AMD EPYC (formerly code named Turin) processors, offering the highest maximum CPU frequency, 5GHz in the cloud. The M8azn instances offer up to 2x compute performance versus previous generation M5zn instances....

December 2, 2025

2025-11-26

11/26/2025, 12:00:00 AM ~ 11/27/2025, 12:00:00 AM (UTC) Recent Announcements SageMaker HyperPod now supports Managed tiered KV cache and intelligent routing Amazon SageMaker HyperPod now supports Managed Tiered KV Cache and Intelligent Routing for large language model (LLM) inference, enabling customers to optimize inference performance for long-context prompts and multi-turn conversations. Customers deploying production LLM applications need fast response times while processing lengthy documents or maintaining conversation context, but traditional inference approaches require recalculating attention mechanisms for all previous tokens with each new token generation, creating computational overhead and escalating costs....

November 26, 2025

2024-11-19

11/19/2024, 12:00:00 AM ~ 11/20/2024, 12:00:00 AM (UTC) Recent Announcements AWS AppSync GraphQL APIs now support data plane logging to AWS CloudTrail Today, AWS AppSync announced support for logging GraphQL data plane operations (query, mutation, and subscription operations and connect requests to your real-time WebSocket endpoint) using AWS CloudTrail, enabling customers to have greater visibility into GraphQL API activity in their AWS account for best practices in security and operational troubleshooting....

November 19, 2024