All news with #amazon bedrock tag
Wed, October 15, 2025
Simplified Amazon Bedrock Model Access and Governance Controls
🔐 Amazon Bedrock now automatically enables serverless foundation models in each AWS Region, removing the prior per-model enablement step and retiring the Model Access page and PutFoundationModelEntitlement IAM permission. Access is managed through standard AWS controls—IAM and Service Control Policies (SCPs)—so account- and organization-level governance remains intact. Existing model restrictions enforced by IAM or SCPs continue to apply, and previously enabled models are unaffected. Administrators should transition to scoped IAM/SCP policies and patterns such as wildcards and NotResource denies to maintain least-privilege control.
Wed, October 15, 2025
Anthropic Claude Haiku 4.5 Now Available in Bedrock
🚀 Claude Haiku 4.5 is now available in Amazon Bedrock, offering near-frontier performance comparable to Claude Sonnet 4 while reducing cost and improving inference speed. The model targets latency-sensitive and budget-conscious deployments, excelling at coding, computer use, agent tasks, and vision-enabled workflows. Haiku 4.5 supports global cross-region inference and is positioned for scaled production use; consult Bedrock documentation, the console, and pricing pages for region and billing details.
Wed, October 15, 2025
Amazon Bedrock automatically enables serverless models
🔓 Amazon Bedrock now automatically enables access to all serverless foundation models by default in all commercial AWS regions. This removes the prior manual activation step and lets users immediately use models via the Amazon Bedrock console, AWS SDK, and features such as Agents, Flows, and Prompt Management. Anthropic models remain enabled but require a one-time usage form before first use; completing the form via the console or API and submitting it from an AWS organization management account will enable Anthropic across member accounts. Administrators continue to control access through IAM policies and Service Control Policies (SCPs).
Wed, October 15, 2025
Amazon Bedrock expands DeepSeek, OpenAI, Qwen models
🚀 Amazon Bedrock has expanded regional access to several foundation models, adding DeepSeek-V3.1, OpenAI open-weight models (20B, 120B), and multiple Qwen3 variants. The update makes DeepSeek-V3.1 and Qwen3 Coder-480B available in US East (Ohio) and Asia Pacific (Jakarta), and brings OpenAI open-weight and additional Qwen models to US East (Ohio), Europe (Frankfurt), and Asia Pacific (Jakarta). Customers can deploy these models locally to meet data residency needs, reduce latency, and enable faster AI-powered experiences.
Tue, October 14, 2025
AgentCore Identity: Secure Identity for AI Agents at Scale
🔐 Amazon Bedrock AgentCore Identity centralizes and secures identities and credentials for AI agents, integrating with existing identity providers such as Amazon Cognito to avoid user migration and rework of authentication flows. It provides a token vault encrypted with AWS KMS, native AWS Secrets Manager support, and orchestrates OAuth 2.0 flows (2LO and 3LO). Declarative SDK annotations and built-in error handling simplify credential injection and refresh workflows, helping teams deploy agentic workloads securely at scale.
Mon, October 13, 2025
Amazon Bedrock AgentCore GA: Secure Agent Platform
🔐 Amazon Bedrock AgentCore is now generally available, offering an agentic platform to build, deploy, and operate agents securely at scale without infrastructure management. It adds VPC support, extended eight-hour runtimes, session isolation, and initial Agent-to-Agent (A2A) protocol support. Identity, Gateway, and Observability features provide secure authorization, tool discovery, and CloudWatch/OTEL monitoring.
Fri, October 3, 2025
Amazon OpenSearch Service Adds Batch AI Inference Support
🧠 You can now run asynchronous batch AI inference inside Amazon OpenSearch Ingestion pipelines to enrich and ingest very large datasets for Amazon OpenSearch Service domains. The same AI connectors previously used for real-time calls to Amazon Bedrock, Amazon SageMaker, and third parties now support high-throughput, offline jobs. Batch inference is intended for offline enrichment scenarios—generating up to billions of vector embeddings—with improved performance and cost efficiency versus streaming inference. The feature is available in regions that support OpenSearch Ingestion on domains running 2.17+.
Thu, October 2, 2025
Cohere Embed v4 Multimodal Embeddings on Amazon Bedrock
🚀 Amazon Bedrock now supports Cohere Embed v4, a multimodal embedding model that generates high-quality embeddings for text, images, and complex business documents. The model natively processes tables, charts, diagrams, code snippets, and handwritten notes, reducing the need for extensive preprocessing and data cleanup. It supports over 100 languages and includes industry fine-tuning for finance, healthcare, and manufacturing. Cohere Embed v4 is available for on-demand inference in select AWS Regions; access is requested via the Bedrock console.
Wed, October 1, 2025
Amazon Bedrock Data Automation Adds Enhanced Transcription
🔊 Amazon Bedrock Data Automation (BDA) now offers enhanced transcription with speaker diarization and channel identification, letting developers separate and process individual speakers or channels in audio files. It also provides a guided, natural language blueprint workflow for extracting custom audio insights. These capabilities simplify reading and analysis of multi-party recordings—customer calls, telehealth visits, webinars, public-safety recordings, and meetings—and support subtitle creation, compliance monitoring, and productivity analysis. BDA is available in seven AWS Regions.
Mon, September 29, 2025
Secure Network Architectures for Generative AI on AWS
🔐 This post explains how to design defense-in-depth network architectures for generative AI workloads using AWS services. It outlines common external threats — including layer 4 and layer 7 DDoS, web request floods, application-specific exploits, and malicious bots — and maps mitigations to AWS capabilities. The guidance recommends private connectivity via Amazon Bedrock and AWS PrivateLink, edge protections with AWS WAF and AWS Shield, subnet-level controls using AWS Network Firewall, and continuous detection and response with GuardDuty, Inspector, and CloudWatch.
Mon, September 29, 2025
Anthropic Claude Sonnet 4.5 Now Available in Bedrock
🚀 Anthropic’s Claude Sonnet 4.5 is now available through Amazon Bedrock, providing managed API access to the company’s most capable model. The model leads SWE-bench Verified benchmarks with improved instruction following, stronger code-refactoring judgment, and enhanced production-ready code generation. Bedrock adds automated context editing and a memory tool to extend usable context and boost accuracy for long-running agents across global regions.
Mon, September 29, 2025
Amazon Bedrock Available in Thailand, Malaysia, and Taipei
🚀 Amazon has launched Amazon Bedrock in the Asia Pacific (Thailand), Asia Pacific (Malaysia), and Asia Pacific (Taipei) regions, enabling local customers to build and scale generative AI applications using a range of foundation models and developer tools. The managed service supports deploying agents and productionizing models to shorten the path from experimentation to real-world deployment. Customers can expect improved latency, regional data residency options, and integration with AWS operational and security services.
Mon, September 29, 2025
Amazon Bedrock Now Available in Israel (Tel Aviv) Region
🚀 Beginning today, Amazon Bedrock is available in the Israel (Tel Aviv) region, enabling customers to build and scale generative AI applications with local infrastructure. The managed service connects organizations to a variety of foundation models (FMs) and provides tools to deploy and operate agents, reducing time-to-production. Local availability can lower latency, support regional compliance needs, and help move projects from experimentation to real-world deployment.
Mon, September 29, 2025
Amazon Bedrock Launches in Middle East (UAE) Region
🚀 Amazon Bedrock is now available in the Middle East (UAE) Region, enabling customers to build, experiment with, and scale generative AI applications using a broad selection of foundation models (FMs) and integrated developer tools. The managed service provides capabilities to deploy and operate agents and production workloads with built-in controls for security and operational management. Customers in the region can begin using Bedrock today and should consult the documentation for supported models, APIs, and recommended practices.
Thu, September 25, 2025
Adapting Enterprise Risk Management for Generative AI
🛡️ This post explains how to adapt enterprise risk management frameworks to safely scale cloud-based generative AI, combining governance foundations with practical controls. It emphasizes the cloud as the foundational infrastructure and identifies differences from on‑premises models that change risk profiles and vendor relationships. The guidance maps traditional ERMF elements to AI-specific controls across fairness, explainability, privacy/security, safety, controllability, veracity/robustness, governance, and transparency, and references tools such as Amazon Bedrock Guardrails, SageMaker Clarify, and the ISO/IEC 42001 standard to operationalize those controls.
Thu, September 25, 2025
AgentCore Supports VPC, PrivateLink, CloudFormation
🔒 Amazon Web Services announced that Amazon Bedrock AgentCore Runtime, AgentCore Browser, and AgentCore Code Interpreter now support VPC connectivity, AWS PrivateLink, CloudFormation, and resource tagging. These additions let developers deploy AI agents that access private resources such as databases and internal APIs without internet exposure. CloudFormation integration enables infrastructure-as-code provisioning, while tagging provides cost allocation and access-control organization. AgentCore is in preview in US East (N. Virginia), US West (Oregon), Asia Pacific (Sydney), and Europe (Frankfurt).
Thu, September 18, 2025
Source-of-Truth Authorization for RAG Knowledge Bases
🔒 This post presents an architecture to enforce strong, source-of-truth authorization for Retrieval-Augmented Generation (RAG) knowledge bases using Amazon S3 Access Grants with Amazon Bedrock. It explains why vector DB metadata filtering is insufficient—permission changes can be delayed and complex identity memberships are hard to represent—and recommends validating permissions at the data source before returning chunks to an LLM. The blog includes a practical Python walkthrough for exchanging identity tokens, retrieving caller grant scopes, filtering returned chunks, and logging withheld items to reduce the risk of sensitive data leaking into LLM prompts.
Thu, September 18, 2025
Stability AI Image Services Now Available in Amazon Bedrock
🖼️ Amazon Bedrock now includes Stability AI Image Services, a suite of nine specialized image-editing tools available via the Bedrock API. The offering splits into Edit tools (Remove Background, Erase Object, Search and Replace, Search and Recolor, Inpaint) and Control tools (Structure, Sketch, Style Guide, Style Transfer). It is currently supported in US West (Oregon), US East (N. Virginia), and US East (Ohio), and is intended to accelerate professional creative workflows with granular edit control.
Thu, September 18, 2025
OpenAI Open-Weight Models Now in Eight More AWS Regions
🚀 AWS has expanded availability of OpenAI open weight models on Amazon Bedrock to eight additional regions. The update adds US East (N. Virginia), Asia Pacific (Tokyo), Europe (Stockholm), Asia Pacific (Mumbai), Europe (Ireland), South America (São Paulo), Europe (London), and Europe (Milan) to the previously supported US West (Oregon). This broader regional coverage reduces network latency, helps meet data residency preferences, and makes it easier for customers to deploy AI-powered applications closer to their users. Customers can access the models through the Amazon Bedrock console and supporting documentation to get started.
Thu, September 18, 2025
DeepSeek-V3.1 Available as Fully Managed in Bedrock
🔍 DeepSeek-V3.1 is now available as a fully managed foundation model in Amazon Bedrock, offering an open-weight option designed for enterprise deployment. The model supports a selectable 'thinking' mode for step-by-step analysis and a faster non-thinking mode for quicker replies, with improved multilingual accuracy and reduced hallucinations. Enhanced tool-calling, transparent reasoning, and strong coding and analytical performance make it well suited for building AI agents, automating workflows, and tackling complex technical tasks. DeepSeek-V3.1 is available in US West (Oregon), Asia Pacific (Tokyo, Mumbai), and Europe (London, Stockholm).