All news with #aws tag
Thu, December 4, 2025
Amazon Bedrock Adds OpenAI-Compatible Responses API
🚀 Amazon Bedrock now exposes an OpenAI-compatible Responses API on new service endpoints, enabling asynchronous inference for long-running workloads, streaming and non-streaming modes, and automatic stateful conversation reconstruction so developers no longer must resend full histories. The endpoints provide Chat Completions with reasoning-effort support for models served by Mantle, Amazon’s distributed inference engine. Integration requires only a base URL change for OpenAI SDK–compatible code, and support starts today for OpenAI’s GPT OSS 20B and 120B models, with additional models coming soon.
Thu, December 4, 2025
AWS previews EC2 M9g instances powered by Graviton5
🚀 Amazon Web Services today previewed new Amazon EC2 M9g instances powered by AWS Graviton5 processors. AWS says M9g delivers up to 25% better compute performance and increased networking and EBS bandwidth versus Graviton4-based M8g, with up to 30–35% faster performance for databases, web applications, and machine learning. Built on the AWS Nitro System, M9g targets application servers, microservices, gaming servers, midsize data stores, and caching fleets, and is available in preview through a request process.
Wed, December 3, 2025
Amazon SageMaker HyperPod Adds Checkpointless Training
🚀 Amazon SageMaker HyperPod now supports checkpointless training, a foundational capability that eliminates the need for checkpoint-based, job-level restarts for distributed model training. Checkpointless training preserves forward training state across the cluster, automatically swaps out failed nodes, and uses peer-to-peer state transfer to resume progress, reducing recovery time from hours to minutes. The feature can deliver up to 95% training goodput at very large scale, is available in all Regions where HyperPod runs, and can be enabled with zero code changes for popular recipes or with minimal PyTorch modifications for custom models.
Wed, December 3, 2025
TypeScript Preview and Updates for Strands Agents on AWS
🚀 AWS has announced TypeScript support in preview for the Strands Agents SDK, giving developers a choice between Python and TypeScript for building model-driven AI agents. The TypeScript implementation provides idiomatic, type-safe APIs with async/await and modern JavaScript/TypeScript patterns, and is designed to run in browsers, client applications, and server runtimes such as AWS Lambda and Bedrock AgentCore. AWS also introduced three SDK updates: edge device support is now GA, Strands steering is available experimentally, and Strands evaluations is in preview to help validate agent behavior.
Wed, December 3, 2025
AWS SageMaker AI adds serverless model customization
🚀 Amazon SageMaker AI now offers a serverless model customization capability that lets developers quickly fine-tune popular models using supervised learning, reinforcement learning, and direct preference optimization. The fully managed, end-to-end workflow simplifies data preparation, synthetic data generation, training, evaluation, and deployment through an easy-to-use interface. Supported base models include Amazon Nova, Llama, Qwen, DeepSeek, and GPT-OSS. The AI agent-guided workflow is in preview with regional availability and a waitlist.
Wed, December 3, 2025
Amazon SageMaker HyperPod Adds Elastic Training at Scale
⚡ Amazon SageMaker HyperPod now supports elastic training, automatically scaling distributed training jobs to absorb idle accelerators and contract when higher‑priority workloads require resources. This eliminates the manual cycle of halting jobs, reconfiguring parameters, and restarting distributed training, which previously demanded specialized engineering time. Organizations can start training with minimal resources and grow opportunistically, improving cluster utilization and reducing costs. Elastic training can be enabled with zero code changes for public models like Llama and GPT OSS, and requires only lightweight configuration updates for custom architectures.
Wed, December 3, 2025
Amazon Bedrock Adds Reinforcement Fine‑Tuning for Models
🔧 Amazon Bedrock now supports reinforcement fine-tuning, enabling developers to improve model accuracy without deep ML expertise or large labeled datasets. The service automates the reinforcement fine-tuning workflow and trains models by learning from feedback on multiple candidate responses, improving model judgment about what makes a good reply. AWS reports an average 66% accuracy gain over base models, allowing teams to deploy smaller, faster, and more cost-effective variants while maintaining quality. At launch the feature supports Amazon Nova 2 Lite, and it can be accessed via the Bedrock console or APIs.
Tue, December 2, 2025
Fortinet and AWS at re:Invent: Expanding Cloud Security
🔒 Fortinet announced expanded integrations with AWS at re:Invent, including Fortinet Managed IPS Rules for AWS Network Firewall, FortiSASE on AWS Marketplace, and participation in the AWS European Sovereign Cloud. These offerings combine AI-driven FortiGuard threat intelligence, simplified procurement and Euro-denominated options for EU customers. The goal is to reduce operational burden, accelerate compliance with standards like PCI-DSS and HIPAA, and enable rapid deployment and scaling across hybrid and multi-cloud environments.
Tue, December 2, 2025
AWS Preview: EC2 M8azn Instances with 5GHz AMD CPUs
🚀 Starting today, AWS is previewing new general-purpose high-frequency Amazon EC2 M8azn instances powered by fifth-generation AMD EPYC processors that deliver up to 5 GHz maximum CPU frequency. These instances offer up to 2× the compute performance of M5zn and about 24% higher performance than M8a, and are built on the AWS Nitro System for secure, high-performance cloud delivery. They target workloads such as gaming, HFT, HPC, CI/CD, and simulation modeling; customers can request preview access.
Tue, December 2, 2025
AWS launches Apache Spark Upgrade Agent for Amazon EMR
🛠️ AWS announced the Apache Spark upgrade agent, a capability that automates and accelerates Spark version upgrades for Amazon EMR on EC2 and EMR Serverless. The agent performs automated code analysis across PySpark and Scala, identifies API and behavioral changes for Spark 2.4→3.5, and suggests precise code transformations. Engineers can invoke the agent from SageMaker Unified Studio, the Kiro CLI, or any MCP-compatible IDE, interact via natural-language prompts, review proposed edits, and approve implementations. Functional correctness is validated through data quality checks to help maintain processing accuracy during migration.
Tue, December 2, 2025
Amazon FSx for NetApp ONTAP Adds S3 Access Points Support
📂 You can now attach Amazon S3 Access Points to Amazon FSx for NetApp ONTAP file systems so applications can access file data as if it were stored in S3. This lets a broad range of AI, ML, and analytics services—including Amazon Bedrock, SageMaker, and Glue—use your FSx file data without copying it. Create and attach access points via the Amazon FSx console, AWS CLI, or SDK; support for existing file systems will arrive in a forthcoming maintenance window. The capability is available in select AWS Regions.
Tue, December 2, 2025
Amazon RDS for SQL Server Adds Developer Edition Support
🆕 Amazon RDS for SQL Server now supports SQL Server 2022 Developer Edition, enabling teams to run a feature-complete, free edition of SQL Server in non-production RDS instances. The Developer Edition includes all Enterprise features for building, testing, and demonstrating applications while reducing licensing costs for development and test environments. Core RDS capabilities — automated backups, automated software updates, monitoring, and encryption — are supported on Developer Edition. The license is strictly limited to development and testing and may not be used in production or for commercial end-user scenarios.
Tue, December 2, 2025
Amazon RDS for SQL Server: Optimize CPU on M7i/R7i
🔧Amazon RDS for SQL Server introduces an Optimize CPU option with support for M7i and R7i instance families, lowering prices by up to 55% compared with equivalent sixth‑generation instances. Optimize CPU disables SMT on instances with two or more physical CPU cores to halve vCPU counts and associated third‑party licensing charges while preserving the same number of physical cores and near‑equivalent performance. The biggest savings appear on 2Xlarge and larger sizes and on Multi‑AZ deployments; memory‑ or I/O‑intensive workloads can be further tuned to reduce costs.
Tue, December 2, 2025
Amazon EC2 P6e-GB300 UltraServers Now Generally Available
🚀 AWS has announced general availability of Amazon EC2 P6e-GB300 UltraServers powered by the NVIDIA GB300 NVL72. The new UltraServers deliver 1.5× GPU memory and 1.5× FP4 compute (without sparsity) compared with P6e-GB200, enabling higher-context inference and improved throughput for large models. Ideal for reasoning, Agentic AI, and production inference; contact your AWS sales representative to get started.
Tue, December 2, 2025
AWS Announces Memory-Optimized EC2 X8aedz Instances
🚀 AWS has introduced Amazon EC2 X8aedz, a new memory-optimized instance family powered by 5th Gen AMD EPYC processors (Turin) that deliver up to 5 GHz maximum CPU frequency. X8aedz claims up to 2x higher compute performance and ~31% improved price-performance versus the prior X2iezn generation, combining high single-thread speed with a 32:1 memory-to-vCPU ratio and local NVMe storage. Instances come in eight sizes (2–96 vCPUs, 64–3,072 GiB), include two bare-metal variants, and offer up to 8 TB of local NVMe SSD. They are available now in US West (Oregon) and Asia Pacific (Tokyo) and can be purchased via On-Demand, Spot, or Savings Plans.
Tue, December 2, 2025
Amazon EC2 X8i memory-optimized instances (Preview)
🚀 Amazon Web Services today announced a preview of Amazon EC2 X8i, a next-generation memory-optimized instance family built on custom Intel Xeon 6 processors. X8i offers up to 6 TB of memory—1.5× the capacity of X2i—and up to 3.4× the memory bandwidth of the previous generation. AWS reports 35% higher overall performance compared with X2i, and X8i is SAP-certified with a 46% SAPS increase for mission-critical SAP deployments. The instances target in-memory databases, large-scale databases, analytics, and EDA workloads; customers can request preview access to evaluate performance and fit.
Tue, December 2, 2025
S3 Storage Lens: performance metrics and prefix analytics
📊 S3 Storage Lens now provides three new capabilities: application performance metrics, expanded prefix analytics across billions of prefixes, and direct export of metrics to managed S3 Tables. The performance metrics include access pattern, request origin, and object access count metrics to surface inefficient requests, cross‑Region access, and hot object reads. Enable these features in your advanced dashboard to make metrics queryable via AWS analytics services.
Tue, December 2, 2025
Startup Frenetik Launches Patented Deception Technology
🔐 Frenetik, a Maryland cybersecurity startup, emerged from stealth with a patented approach called Deception In-Use that continuously rotates real identities and resources across Microsoft Entra (M365), AWS, Google Cloud and on-prem environments. By routing critical change details through out-of-band channels accessible only to trusted parties, defenders retain accurate visibility while attackers operate on stale intelligence and are more likely to be funneled into decoys and honeypots.
Tue, December 2, 2025
Unit 42 and AWS Launch No-Cost Incident Response Retainer
🔒 Palo Alto Networks Unit 42 and Amazon Web Services have expanded their partnership to offer a no-cost Unit 42 Incident Response Retainer in AWS Marketplace for qualified customers. The retainer provides 250 hours of initial incident response, a 2-hour response SLA and 24/7/365 access to Unit 42’s incident response team. The offering is designed to accelerate containment, enable holistic investigations across cloud and enterprise environments, and reduce procurement overhead while providing preferred pricing for proactive services.
Tue, December 2, 2025
Amazon Nova Forge: Build Frontier Models with Nova
🚀 Amazon Web Services announced general availability of Nova Forge, a SageMaker AI service that enables organizations to build custom frontier models from Nova checkpoints across pre-, mid-, and post-training phases. Developers can blend proprietary data with Amazon-curated datasets, run Reinforcement Fine Tuning (RFT) with in-environment reward functions, and apply custom safety guardrails via a built-in responsible AI toolkit. Nova Forge includes early access to Nova 2 Pro and Nova 2 Omni and is available today in US East (N. Virginia).