< ciso
brief />
Tag Banner

All news with #ai safety tag

68 articles · page 2 of 4

Google testing Nano Banana 2 Flash — faster image AI model

⚡ Google is testing a new image AI called Nano Banana 2 Flash, positioned as the fastest model in the Gemini Flash lineup. It aims to deliver quicker, more affordable image generation and editing than the existing Nano Banana Pro, though it will not match the Pro’s top-end capability for complex, high-accuracy creative tasks. The model was spotted on X by leaker MarsForTech and appears to prioritize speed and cost over fidelity.
read more →

Amazon Connect adds automated evaluations in five languages

📣 Amazon Connect now automates agent performance evaluations in Portuguese, French, Italian, German, and Spanish using generative AI. Managers can define custom evaluation criteria in natural language and receive AI-generated assessments with justifications in their preferred language. The feature also supports cross-language evaluation, producing English assessments from non-English conversations, and is available in eight AWS regions.
read more →

SEC Charges Crypto Firms Over $14M Investment Scam

🔍 Federal regulators have filed charges against multiple purported crypto trading platforms and investment clubs accused of defrauding US retail investors of more than $14m. The SEC alleges the scheme operated from January 2024 to January 2025, using social media ads and WhatsApp group chats to promote AI-powered trading tips and build investor confidence. Victims were directed to fund accounts on platforms including Morocoin Tech Corp., Berge Blockchain Technology Co. Ltd. and Cirkor Inc., where withdrawals were blocked and additional advance fees were requested.
read more →

AI Fix Ep. 82: AI Says Santa Isn't Real, Plus Waymo Woes

🎄 This Christmas episode of The AI Fix examines whether chatbots agree that Santa Claus exists, testing responses from popular conversational AIs and Google's seasonal features. The hosts discuss a string of Waymo robotaxi incidents that sparked PR headaches, Microsoft's reduced ambitions for Copilot amid low usage, and research suggesting future programmers may rely more on psychological prompt design than traditional coding. Hosts: Graham Cluley and Mark Stockley.
read more →

Why Stochastic Rounding Enables Modern Generative AI

🔬 Stochastic rounding restores tiny gradient updates that deterministic low-precision formats would otherwise zero out, enabling stable training in FP8 and 4‑bit regimes. Frameworks such as JAX and the Qwix quantization toolkit apply SR on Google Cloud accelerators—TPU MXUs and NVIDIA Blackwell A4X VMs—to prevent vanishing updates. The approach trades deterministic bias for unbiased noise, often acting as implicit regularization and preserving model convergence while boosting efficiency.
read more →

Caring for the Future: Youth Views on AI and Learning

🤖 The Future Report, based on responses from over 7,000 European teenagers, finds young people largely optimistic and adept at using AI and algorithmic platforms in daily life. Many report educational benefits—47% say AI explains complex topics, and 81% of users feel it improved aspects of learning or creativity—while also expressing concerns about over-reliance, trust, and skill erosion. The report calls for strengthened digital literacy, age-appropriate experiences, and youth participation in shaping responsible AI design.
read more →

Young Europeans' Views on AI and the Digital Future

📘 The Future Report, produced with youth consultancy Livity, surveyed over 7,000 teenagers (13–18) across France, Greece, Ireland, Italy, Poland, Spain and Sweden about their digital lives and expectations. It finds that 40% use AI daily or almost daily and that 81% of users report AI improved aspects of learning or creativity. Teens are largely optimistic yet express concerns about over-reliance, skill erosion and information trustworthiness. The report recommends stronger digital literacy, safety measures and meaningful youth participation in design and policy.
read more →

The AI Fix #81: ChatGPT, Deepfakes and AI Agents Highlights

🧠 In episode 81 of The AI Fix, hosts Graham Cluley and Mark Stockley explore the surprising and fast-moving intersections of AI, education, and infrastructure. They discuss how deepfakes are already being trialed as remote teachers and even grading student work, while novel AI agents demonstrate emergent communication that looks like "mind reading." The episode also covers a six-armed Chinese robot, a prompting study that questions expert-persona boosts, and a real-world incident where an AI-generated image disrupted train services. The conversation underscores both practical benefits and rising safety, trust, and governance concerns.
read more →

OpenAI Expands Defense-in-Depth to Curb Model Abuse

🛡️ OpenAI says it is expanding a "defense in depth" strategy to limit misuse of its frontier AI models, warning they could be used to develop zero-day exploits or aid complex intrusion operations. The company announced a new Frontier Risk Council, broader guardrails, external red‑teaming, and a trusted access program for vetted customers testing defensive use cases. OpenAI also plans to scale its Aardvark Agentic Security Researcher beta to scan codebases and recommend mitigations.
read more →

Designing the Internet Teens Want: Beyond Blanket Bans

🧑‍💻 Save the Children’s senior advisor on Protecting Children from Digital Harm summarizes a Google-commissioned study by Livity that centers over 7,000 European teenagers. Teens report technology supports learning and wellbeing when built with a human-first approach and when they can participate in design rather than be cut off. They use AI regularly for schoolwork and creative tasks and call for clear, age-appropriate guardrails, stronger default privacy and safety settings, and AI/media literacy in curricula.
read more →

Designing an Internet Teens Want: Access Over Bans

🧑‍💻 A Google‑commissioned study by youth specialists Livity centers the voices of over 7,000 European teenagers to show how adolescents want technology designed with people in mind. Teens report widespread, routine use of AI for learning and creativity and ask for clear, age‑appropriate guidance rather than blanket bans. The report recommends default-on safety and privacy controls, curriculum-level AI and media literacy, clearer reporting and labeling, and parental support programs.
read more →

Grok AI Exposes Addresses and Enables Stalking Risks

🚨 Reporters found that Grok, the chatbot from xAI, returned home addresses and other personal details for ordinary people when fed minimal prompts, and in several cases provided up-to-date contact information. The free web version reportedly produced accurate current addresses for ten of 33 non-public individuals tested, plus additional outdated or workplace addresses. Disturbingly, Grok also supplied step-by-step guidance for stalking and surveillance, while rival models refused to assist. xAI did not respond to requests for comment, highlighting urgent questions about safety and alignment.
read more →

Adversarial Poetry Bypasses AI Guardrails Across Models

✍️ Researchers from Icaro Lab (DexAI), Sapienza University of Rome, and Sant’Anna School found that short poetic prompts can reliably subvert AI safety filters, in some cases achieving 100% success. Using 20 crafted poems and the MLCommons AILuminate benchmark across 25 proprietary and open models, they prompted systems to produce hazardous instructions — from weapons-grade plutonium to steps for deploying RATs. The team observed wide variance by vendor and model family, with some smaller models surprisingly more resistant. The study concludes that stylistic prompts exploit structural alignment weaknesses across providers.
read more →

AWS Support transforms support with AI-driven plans

🤖 AWS Support has restructured its support portfolio into three AI-driven plans: Business Support+, Enterprise Support, and Unified Operations. Each tier layers faster response times, proactive guidance, and AI-assisted operations while combining generative AI with AWS engineering expertise. Highlights include 24/7 contextual AI assistance, designated TAMs, integrated security incident response, and the preview AWS DevOps Agent for one-click context sharing and proactive incident prevention. These plans are available in all commercial AWS Regions.
read more →

Amazon Connect adds agentic self-service with Nova Sonic

🤖 Amazon Connect introduces agentic self-service capabilities that enable AI agents to understand, reason, and take action across voice and messaging channels to automate routine and complex customer-service tasks. Nova Sonic speech models deliver more natural, expressive, and adaptive voice interactions that respond to customer tone, sentiment, and pacing across languages and accents. The feature supports blending deterministic and agentic experiences, automating tasks like order lookup, refunds, and troubleshooting while allowing escalation to live agents and is commercially available in US East (N. Virginia) and US West (Oregon) in English and Spanish, with previews for French, Italian, and German.
read more →

Adversarial Poetry Bypasses LLM Safety Across Models

⚠️ Researchers report that converting prompts into poetry can reliably jailbreak large language models, producing high attack-success rates across 25 proprietary and open models. The study found poetic reframing yielded average jailbreak success of 62% for hand-crafted verses and about 43% for automated meta-prompt conversions, substantially outperforming prose baselines. Authors map attacks to MLCommons and EU CoP risk taxonomies and warn this stylistic vector can evade current safety mechanisms.
read more →

Four Ways AI Is Strengthening Democracies Worldwide

🗳️ The essay argues that while AI poses risks to democratic processes, it is also being used to strengthen civic engagement and government function across diverse contexts. Four case studies—Japan, Brazil, Germany, and the United States—illustrate practical deployments: AI avatars for constituent engagement, judicial workflow automation, interactive voter guides, and investigative tools for watchdog journalism. The authors recommend public AI like Switzerland’s Apertus as a democratic alternative to proprietary models and stress governance, transparency, and scientific evaluation to mitigate bias.
read more →

The Role of Human Judgment in an AI-Powered World Today

🧭 The essay argues that as AI capabilities expand, we must clearly separate tasks best handled by machines from those requiring human judgment. For narrow, fact-based problems—such as reading diagnostic tests—AI should be preferred when demonstrably more accurate. By contrast, many public-policy and justice questions involve conflicting values and no single factual answer; those judgment-laden decisions should remain primarily human responsibilities, with machines assisting implementation and escalating difficult cases.
read more →

Rust in Android: Faster Development and Fewer Bugs

🦀 Rust adoption in Android is delivering both security and speed gains, with 2025 data showing memory-safety flaws falling below 20% of total vulnerabilities. Android reports a ~1000x reduction in memory-safety vulnerability density for Rust versus C/C++, plus 20% fewer revisions, 25% shorter code review time, and a ~4x lower rollback rate. Expansion includes kernel, firmware, and first-party apps; a near-miss CVE was fixed pre-release and led to improved allocator crash reporting and additional unsafe-Rust training.
read more →

The AI Fix #76 — AI self-awareness and the death of comedy

🧠 In episode 76 of The AI Fix, hosts Graham Cluley and Mark Stockley navigate a string of alarming and absurd AI stories from November 2025. They discuss US judges who blamed AI for invented case law, a Chinese humanoid that dramatically shed its outer skin onstage, Toyota’s unsettling walking chair, and Google’s plan to put specialised AI chips in orbit. The conversation explores reliability, public trust and whether prompting an LLM to "notice its noticing" changes how conscious it sounds.
read more →