Tag Banner

All news with #amazon bedrock tag

Fri, December 5, 2025

Pegasus 1.2 Available with Global Cross-Region Inference

📣 Amazon Bedrock now offers TwelveLabs Pegasus 1.2 via Global cross-Region inference, expanding availability by 23 new Regions in addition to the seven where it was already supported. You can also access the model in all EU Regions using Geographic cross-Region inference to meet data-residency requirements. Pegasus 1.2 is a video-first model for long-form video-to-text generation and temporal understanding, enabling lower latency and simplified architecture for video-intelligence applications.

read more →

Thu, December 4, 2025

Amazon Bedrock Adds OpenAI-Compatible Responses API

🚀 Amazon Bedrock now exposes an OpenAI-compatible Responses API on new service endpoints, enabling asynchronous inference for long-running workloads, streaming and non-streaming modes, and automatic stateful conversation reconstruction so developers no longer must resend full histories. The endpoints provide Chat Completions with reasoning-effort support for models served by Mantle, Amazon’s distributed inference engine. Integration requires only a base URL change for OpenAI SDK–compatible code, and support starts today for OpenAI’s GPT OSS 20B and 120B models, with additional models coming soon.

read more →

Wed, December 3, 2025

Amazon Bedrock Adds Reinforcement Fine‑Tuning for Models

🔧 Amazon Bedrock now supports reinforcement fine-tuning, enabling developers to improve model accuracy without deep ML expertise or large labeled datasets. The service automates the reinforcement fine-tuning workflow and trains models by learning from feedback on multiple candidate responses, improving model judgment about what makes a good reply. AWS reports an average 66% accuracy gain over base models, allowing teams to deploy smaller, faster, and more cost-effective variants while maintaining quality. At launch the feature supports Amazon Nova 2 Lite, and it can be accessed via the Bedrock console or APIs.

read more →

Tue, December 2, 2025

AWS announces Amazon Nova 2 models in Amazon Bedrock

🤖AWS has introduced Amazon Nova 2, a next-generation family of foundation models now available in Amazon Bedrock. The release includes Nova 2 Lite, optimized for fast, cost-effective reasoning for everyday workloads, and Nova 2 Pro (Preview), designed for complex, multistep tasks. Both models support step-by-step reasoning, three thinking intensity levels, built-in tools such as code interpreter and web grounding, remote MCP tool support, and a one-million-token context window. Nova 2 Lite supports supervised fine-tuning on Bedrock and SageMaker; full fine-tuning is available on SageMaker. Nova 2 Pro is available in preview for Amazon Nova Forge customers with global cross-region inference.

read more →

Tue, December 2, 2025

Amazon Announces Nova 2 Sonic for Real‑Time Voice AI

🎙️ Amazon announced Amazon Nova 2 Sonic, a speech-to-speech model for natural, real-time conversational AI available via Amazon Bedrock. The model delivers streaming speech understanding robust to background noise and diverse speaking styles, expressive polyglot voices, turn-taking controllability, asynchronous tool calling, and a one‑million token context window. Developers can integrate Nova 2 Sonic with Amazon Connect, leading telephony providers, open-source frameworks, and Bedrock’s bidirectional streaming API; it’s initially available in select AWS Regions.

read more →

Tue, December 2, 2025

Mistral Large 3 and Ministral 3 Now on Amazon Bedrock

🚀 Amazon Bedrock now offers Mistral Large 3 and the Ministral 3 family alongside additional Mistral AI checkpoints, giving customers early access to open-weight multimodal models. Mistral Large 3 employs a granular Mixture-of-Experts architecture with 41B active and 675B total parameters and supports a 256K context window for long-form comprehension and agentic workflows. The Ministral 3 series (14B, 8B, 3B) plus Voxtral and Magistral small models let developers choose scales optimized for production assistants, RAG systems, single-GPU edge deployment, or low-resource environments.

read more →

Tue, December 2, 2025

Amazon Bedrock AgentCore Adds Policy and Evaluations

🛡️ Amazon Web Services' AgentCore introduces preview features — Policy and Evaluations — to help teams scale agents from prototypes into production. Policy intercepts real-time tool calls via AgentCore Gateway and converts natural-language rules into Cedar for auditability and compliance without custom code. Evaluations offers 13 built-in evaluators plus custom model-based scoring, with all quality metrics surfaced in an Amazon CloudWatch dashboard to simplify continuous testing and monitoring.

read more →

Tue, December 2, 2025

Amazon Bedrock Adds 18 Fully Managed Open Models Today

🚀 Amazon Bedrock expanded its model catalog with 18 new fully managed open-weight models, the largest single addition to date. The offering includes Gemma 3, Mistral Large 3, NVIDIA Nemotron Nano 2, OpenAI gpt-oss variants and other vendor models. Through a unified API, developers can evaluate, switch, and adopt these models in production without rewriting applications or changing infrastructure. Models are available in supported AWS Regions.

read more →

Tue, December 2, 2025

AWS AI Factories: Dedicated High-Performance AI Infrastructure

🚀 AWS AI Factories are now available to deploy high-performance AWS AI infrastructure inside customer data centers, combining AWS Trainium, NVIDIA GPUs, low-latency networking, and optimized storage. The service integrates Amazon Bedrock and Amazon SageMaker to provide immediate access to foundation models without separate provider contracts. AWS manages procurement, setup, and operations while customers supply space and power, enabling isolated, sovereign deployments that accelerate AI initiatives.

read more →

Tue, December 2, 2025

Bedrock AgentCore Runtime Adds Bi-Directional Streaming

🔁 Amazon Bedrock AgentCore Runtime now supports bi-directional streaming, enabling real-time, continuous conversations where agents listen and respond simultaneously and handle interruptions or context shifts mid-turn. This removes stop-start friction in voice and text agents and preserves context across exchanges. Built into AgentCore Runtime, the capability reduces months of engineering work required to implement streaming infrastructure, letting developers focus on agent experiences rather than plumbing. Available in nine AWS Regions with consumption-based pricing.

read more →

Tue, December 2, 2025

Amazon API Gateway Adds MCP Proxy for Agent Integration

🤖 Amazon API Gateway now supports the Model Context Protocol (MCP) via a proxy, enabling organizations to expose existing REST APIs to AI agents and MCP clients without modifying their applications. Integrated with Amazon Bedrock AgentCore's Gateway, the feature performs protocol translation, indexes APIs for semantic tool discovery, and eliminates the need to host additional intermediary infrastructure. It also enforces dual authentication to verify agent identities for inbound requests while managing secure outbound connections to REST endpoints. The capability is available in nine AWS Regions and follows Amazon Bedrock AgentCore pricing.

read more →

Sun, November 30, 2025

Amazon Connect adds Bedrock knowledge base integration

📘 Amazon Connect now supports connecting existing Amazon Bedrock Knowledge Bases directly to AI agents and allows multiple knowledge bases per agent. You can attach Bedrock KBs in a few clicks with no additional setup or data duplication, and leverage Bedrock connectors such as Adobe Experience Manager, Confluence, SharePoint, and OneDrive. With multiple KBs per agent, AI agents can query several sources in parallel for more comprehensive responses. This capability is available in all AWS Regions where both services are offered.

read more →

Sun, November 30, 2025

Amazon SageMaker Catalog Adds Automated Data Classification

🤖 Amazon SageMaker Catalog now provides automated data classification that suggests business glossary terms during dataset publishing to reduce manual tagging and improve metadata consistency. The capability leverages Amazon Bedrock language models to analyze table metadata and schema and recommend relevant business and sensitive-data terms from organizational glossaries. Data producers receive AI-generated suggestions they can accept or modify before publishing, helping standardize vocabulary and improve data discoverability. The feature is available in multiple AWS regions and can be managed via SageMaker Unified Studio, the AWS CLI, or SDKs.

read more →

Sun, November 30, 2025

AWS AI League 2026 Championship Expands Challenges

🤖 AWS has launched the AWS AI League 2026 Championship, expanding its flagship AI tournament with new challenge tracks and a doubled prize pool of $50,000 to drive builder innovation. The program pairs a brief orientation with two competition tracks: a Model Customization track using Amazon SageMaker AI to fine-tune foundation models for domain-specific tasks, and an Agentic AI track using Amazon Bedrock AgentCore to build planning and execution agents. Enterprises can apply to host internal tournaments and receive AWS credits to run team competitions, while individual developers can compete at AWS Summits to test skills and build with AWS AI services.

read more →

Sun, November 30, 2025

AWS Bedrock Knowledge Bases Adds Multimodal Retrieval

🔍 AWS has announced general availability of multimodal retrieval in Amazon Bedrock Knowledge Bases, enabling unified search across text, images, audio, and video. The managed Retrieval Augmented Generation (RAG) workflow provides developers full control over ingestion, parsing, chunking, embedding (including Amazon Nova multimodal), and vector storage. Users can submit text or image queries and receive relevant text, image, audio, and video segments back, which can be combined with the LLM of their choice to generate richer, lower-latency responses. Region availability varies by feature set and is documented by AWS.

read more →

Wed, November 26, 2025

AWS API MCP Server Now Available on AWS Marketplace

🔐 The AWS API MCP Server is now available on AWS Marketplace, enabling customers to deploy the Model Context Protocol (MCP) server to Amazon Bedrock AgentCore as a managed offering. The marketplace deployment provides built-in authentication (SigV4 or JWT), session isolation, and simplified container management while enabling configurable IAM roles and network settings to meet enterprise security requirements. Customers also benefit from AgentCore's logging and monitoring capabilities and regional availability where Bedrock AgentCore is supported.

read more →

Wed, November 26, 2025

Amazon Bedrock Reserved Tier for Predictable Performance

🔒 Amazon Bedrock now offers a Reserved service tier that provides prioritized compute and guaranteed input/output tokens-per-minute capacity for inference workloads. Customers can reserve asymmetric input and output capacities to match workload patterns, and excess traffic overflows automatically to the pay-as-you-go Standard tier to keep operations running. The tier targets 99.5% model response uptime and is available today for Anthropic Claude Sonnet 4.5, with 1- or 3-month reservations billed monthly at a fixed price per 1K tokens-per-minute.

read more →

Mon, November 24, 2025

Anthropic Claude Opus 4.5 Now Available in Amazon Bedrock

🚀 Anthropic's Claude Opus 4.5 is now available through Amazon Bedrock, giving Bedrock customers access to a high-performance foundation model at roughly one-third the prior cost. Opus 4.5 advances professional software engineering, agentic workflows, multilingual coding, and complex visual interpretation while supporting production-grade agent deployments. Bedrock adds two API features — tool search and tool use examples — plus a beta effort parameter to balance reasoning, tool calls, latency, and cost. The model is offered via global cross-region inference in multiple AWS regions.

read more →

Fri, November 21, 2025

Bedrock Guardrails: Natural-Language Test Generation

🧪 Amazon Web Services has added natural-language test Q&A generation to Automated Reasoning checks in Amazon Bedrock Guardrails. The capability generates up to N test Q&As from input documents to accelerate creating and validating formal verification policies. Automated Reasoning checks apply formal methods to detect correct model outputs and report up to 99% accuracy in identifying correct responses and reducing hallucinations. The feature is available in multiple US and EU Regions and accessible via the Bedrock console and Python SDK.

read more →

Fri, November 21, 2025

Amazon Bedrock Data Automation: Synchronous Image Processing

🚀 Amazon Bedrock Data Automation (BDA) now supports synchronous image processing, enabling low-latency extraction of structured insights from visual content. Synchronous APIs complement existing asynchronous workflows, removing the need for polling or callbacks and simplifying application architecture. BDA supports Standard Output for common analyses and Custom Output via Blueprints for industry-specific field extraction.

read more →