< ciso
brief />
Tag Banner

All news with #rag security tag

26 articles

Cloudflare AI Search: A Search Primitive for Agents

πŸ” Cloudflare introduced AI Search, a plug-and-play search primitive that provides a unified retrieval layer for agents, support bots, and coding assistants. It pairs hybrid semantic (vector) and BM25 keyword matching, managed storage, and built-in vector indexes so developers can create instances at runtime without provisioning separate infrastructure. The service integrates with Workers, the Agents SDK, and Wrangler, supports metadata boosts and cross-instance queries, and can optionally rerank results with a cross-encoder.
read more β†’

Introducing QueryData: Near-100% Accurate Data Agents

πŸ” QueryData launches in preview, offering near-100% accuracy translating natural language into database queries across AlloyDB, Cloud SQL (MySQL and PostgreSQL) and Spanner. Built on Google Cloud’s Gemini LLM and augmented by rich database context, it uses schema ontologies, query blueprints and ambiguity detection to generate precise queries. Deterministic security is enforced via Parameterized Secure Views (PSVs), and integration is supported through a unified QueryData API, the MCP Toolbox for Databases, and context-engineering tools including an Evalbench framework.
read more β†’

Amazon S3 Vectors Adds Availability in 17 Regions Globally

πŸš€ Amazon expanded S3 Vectors into 17 additional AWS Regions β€” now available in 31 Regions worldwide. S3 Vectors is the first cloud object storage with native vector support, built for AI agents, inference, Retrieval-Augmented Generation (RAG), and semantic search at billion-vector scale. It supports up to two billion vectors per index, elastic scaling to 10,000 vector indexes per bucket, low-latency queries (frequent queries as fast as 100 ms; infrequent under one second), and native integration with Amazon Bedrock Knowledge Bases to help reduce RAG costs.
read more β†’

Spanner's Multi-Model Advantage for Agentic AI in Production

πŸ”Spanner positions itself as a unified, globally consistent database designed for agentic AI by combining relational, key-value, graph, vector and full-text search capabilities in one platform. The post argues this interoperable multi-model approach reduces data silos, removes brittle synchronization logic, and improves governance, availability, and development velocity. Google highlights features such as GQL graph support, a Cassandra native endpoint for lift-and-shift, and ScaNN-based ANN vector search. The customer example of MakeMyTrip illustrates significant operational simplification and faster AI feature delivery.
read more β†’

Five Priorities CISOs Must Address at RSAC 2026 Summit

πŸ€–RSA Conference 2026 reframes AI from a single track to the event itself, with roughly 40% of sessions AI-weighted and artificial intelligence woven across identity, cloud, threat intelligence and human-focused tracks. CISOs face a dual mandate: accelerate AI adoption to remain competitive while protecting the enterprise from new attack surfaces such as RAG pipelines, vector databases, prompt injection and model inversion. Key priorities at RSAC include securing the AI stack, defining AI governance and compliance (including preparation for the EU AI Act), managing non‑human identities, mitigating shadow AI and AI-assisted coding risks, and preparing SOCs for autonomous remediation.
read more β†’

Transforming Developers into AI Architects with Google Cloud

🧭 This post launches Google Cloud's "Data Strategy = AI Strategy" series and reframes the database as the central context engine for production AI. It argues that by using fully PostgreSQL-compatible services such as AlloyDB and Cloud SQL, teams can eliminate latency and improve retrieval accuracy while reducing infrastructure friction. The article emphasizes three enterprise pillars β€” speed, scale, and security β€” and describes hands-on labs that cover batch embeddings, real-time inference with Gemini 3 Flash, and row-level security for zero-trust agents.
read more β†’

Private Connectivity for RAG AI Applications on Google Cloud

πŸ”’ This Google Cloud blog outlines a reference architecture to deliver private-IP only connectivity for retrieval-augmented generation (RAG) applications that must not transit the public internet. It describes a multi-project topologyβ€”routing project, Shared VPC host, and service projects for Data Ingestion, Serving, and Frontendβ€”and maps required services such as Cloud Interconnect/Cloud VPN, Network Connectivity Center, Private Service Connect, Cloud Router, Cloud Armor, and VPC Service Controls. The post also details RAG population and inference flows to show end-to-end private traffic paths and highlights management and routing orchestration for hybrid and VPC spokes.
read more β†’

Using the Neo4j Gemini CLI Extension on Google Cloud

πŸ”— Gemini CLI's Neo4j extension connects graph databases to Gemini's reasoning via the Model Context Protocol (MCP). The extension bundles four MCP servers to manage Neo4j Aura, translate natural language into Cypher, support interactive data modeling and visualization, and use Neo4j as long-term memory for agentic flows. Developers can provision databases, run Cypher queries, and persist knowledge from the terminal to accelerate GraphRAG workflows.
read more β†’

Amazon Neptune Analytics Expands to Seven Regions Globally

πŸ”” Amazon Neptune Analytics is now available in seven additional AWS Regions: Middle East (Bahrain), Middle East (UAE), Israel (Tel Aviv), Africa (Cape Town), Canada (Calgary), Asia Pacific (Malaysia), and Europe (Zurich). Neptune is a serverless graph database that automatically scales graph workloads, reduces operational overhead, and improves AI accuracy and explainability by modeling connected data. It also provides fully managed GraphRAG with Amazon Bedrock Knowledge Bases and integrates with the Strands AI Agents SDK and popular agentic memory tools to accelerate graph-powered AI development.
read more β†’

Model Security Misses the Point: Secure AI Workflows

πŸ›‘οΈAs AI copilots and assistants are embedded into daily work, recent incidents show the primary risk lies in surrounding workflows rather than in the models themselves. Malicious Chrome extensions that exfiltrated ChatGPT and DeepSeek chats and prompt injections that tricked an AI coding assistant into executing malware exploited integration contexts, not model internals. The piece advises mapping AI usage, applying least-privilege, enforcing middleware guardrails to scan outputs, and using dynamic SaaS platforms like Reco to detect and control risky workflows.
read more β†’

Data Leakage in AI: Addressing Risks in LLM Systems

πŸ” This article explains how sensitive data commonly leaks from AI systems β€” from RAG retrievals and agentic tool chains to user-initiated oversharing β€” and why LLMs cannot enforce document-level permissions. It recommends a layered, defense-in-depth approach: automatic identification and classification, data minimization at ingress, sanitization, redaction, and strict access controls that follow data through the pipeline. The authors also stress threat modeling and vendor due diligence to limit regulatory, competitive, and reputational harm.
read more β†’

Google Patches Zero-Click Gemini Enterprise Vulnerability

πŸ”’ Google has patched a zero-click vulnerability in Gemini Enterprise and Vertex AI Search that could have allowed attackers to exfiltrate corporate data via hidden instructions embedded in shared Workspace content. Discovered by Noma Security in June 2025 and dubbed "GeminiJack," the flaw exploited Retrieval-Augmented Generation (RAG) retrieval to execute indirect prompt injection without any user interaction. Google updated how the systems interact, separated Vertex AI Search from Gemini Enterprise, and changed retrieval and indexing workflows to mitigate the issue.
read more β†’

Amazon S3 Vectors GA: Scalable, Cost‑Optimized Vector Store

πŸš€ Amazon S3 Vectors is now generally available, delivering native, purpose-built vector storage and query capabilities in cloud object storage. It supports up to two billion vectors per index, 10,000 indexes per vector bucket, and offers up to 90% lower costs to upload, store, and query vectors. S3 Vectors integrates with Amazon Bedrock, SageMaker Unified Studio, and OpenSearch Service, supports SSE-S3 and optional SSE-KMS encryption with per-index keys, and provides tagging for ABAC and cost allocation.
read more β†’

Amazon Connect adds Bedrock knowledge base integration

πŸ“˜ Amazon Connect now supports connecting existing Amazon Bedrock Knowledge Bases directly to AI agents and allows multiple knowledge bases per agent. You can attach Bedrock KBs in a few clicks with no additional setup or data duplication, and leverage Bedrock connectors such as Adobe Experience Manager, Confluence, SharePoint, and OneDrive. With multiple KBs per agent, AI agents can query several sources in parallel for more comprehensive responses. This capability is available in all AWS Regions where both services are offered.
read more β†’

AWS Bedrock Knowledge Bases Adds Multimodal Retrieval

πŸ” AWS has announced general availability of multimodal retrieval in Amazon Bedrock Knowledge Bases, enabling unified search across text, images, audio, and video. The managed Retrieval Augmented Generation (RAG) workflow provides developers full control over ingestion, parsing, chunking, embedding (including Amazon Nova multimodal), and vector storage. Users can submit text or image queries and receive relevant text, image, audio, and video segments back, which can be combined with the LLM of their choice to generate richer, lower-latency responses. Region availability varies by feature set and is documented by AWS.
read more β†’

How BigQuery Brought Vector Search to Analytics at Scale

πŸ” In early 2024 Google introduced native vector search in BigQuery, embedding semantic search directly into the data warehouse to remove the need for separate vector databases. Users can create indexes with a simple CREATE VECTOR INDEX statement and run semantic queries via the VECTOR_SEARCH function or through Python integrations like LangChain. BigQuery provides serverless scaling, asynchronous index refreshes, model rebuilds with no downtime, partitioned indexes, and ScaNN-based TreeAH for improved price/performance, while retaining row- and column-level security and a pay-as-you-go pricing model.
read more β†’

Amazon Web Grounding for Nova Models Now Generally Available

🌐 Web Grounding is now generally available as a built-in tool for Nova models, usable today with Nova Premier via the Amazon Bedrock tool use API. It retrieves and incorporates publicly available information with citations to support responses, enabling a turnkey RAG solution that reduces hallucinations and improves accuracy. Cross-region inference makes the tool available in US East (N. Virginia), US East (Ohio), and US West (Oregon). Support for additional Nova models will follow.
read more β†’

Amazon Nova Multimodal Embeddings β€” Unified Cross-Modal

πŸš€ Amazon announces general availability of Amazon Nova Multimodal Embeddings, a unified embedding model designed for agentic RAG and semantic search across text, documents, images, video, and audio. The model handles inputs up to 8K tokens and video/audio segments up to 30 seconds, with segmentation for larger files and selectable embedding dimensions. Both synchronous and asynchronous APIs are supported to balance latency and throughput, and Nova is available in Amazon Bedrock in US East (N. Virginia).
read more β†’

Amazon ElastiCache Adds Vector Search with Valkey 8.2

πŸš€ Amazon ElastiCache now offers vector search generally available with Valkey 8.2, enabling indexing, searching, and updating billions of high-dimensional embeddings from providers such as Amazon Bedrock, Amazon SageMaker, Anthropic, and OpenAI with microsecond latency and up to 99% recall. Key use cases include semantic caching for LLMs, multi-turn conversational agents, and RAG-enabled agentic systems to reduce latency and cost. Vector search runs on node-based clusters in all AWS Regions at no additional cost, and existing Valkey or Redis OSS clusters can be upgraded to Valkey 8.2 with no downtime.
read more β†’

INDOT Used Google AI to Save 360 Hours and Meet Deadline

πŸš€ Indiana Department of Transportation built a week-long pilot on Google Cloud to meet a 30-day executive order, using a Retrieval-Augmented Generation workflow that combined rapid ETL, Vertex AI Search indexing, and Gemini. The system scraped and parsed decades of internal policies and manuals, produced draft reports across nine divisions with 98% fidelity, and saved an estimated 360 hours of manual effort, enabling INDOT to submit on time.
read more β†’