TechLetters ☕️ Agentic AI is just IT, but messier. LLMs won’t find every bug. US-China AI hotline? CAISI tests frontier models with guardrails off.
Security
Agentic AI systems are complex ecosystems of LLMs, humans, guardrails, datasets, tools and hardware, where security risks often emerge from interactions between components rather than isolated flaws.. Organisations should address AI security, including agentic AI systems, within established cyber security frameworks rather than treating it as a separate or standalone discipline. AI systems are fundamentally IT systems. https://www.cyber.gov.au/sites/default/files/2026-05/careful_adoption_of_agentic_ai_services.pdf
“With the recent news of folks finding vulnerabilities left and right using LLMs, some folks hope that we'd be able to find every single vulnerability.
Today, I hope to shatter that idea” https://github.com/yo-yo-yo-jbo/vr_difficulty
Privacy
Technology Policy
The US and China are considering a hotline for AI incidents to avoid accidentally starting a war over it. The reason is Mythos-grade models, which moved the whole conversation from “tech regulation” to national security. Is a phone call before anyone launches a counterstrike over an ambiguous AI-enabled incident good idea?
The proposed channel would cover four areas: cyberattacks on critical infrastructure, autonomous military systems, non-state actors using AI well beyond their previous capabilities, and misattribution, where a third party uses a model to fake a nation-state operation. That last one is the underrated risk.
This is crisis communication infrastructure. Governments are quietly admitting that AI might be used to generate incidents faster than diplomacy can respond.
The AI channel, if it happens, would be the first formal admission that both sides understand the same problem. AI-enabled incidents can move faster than attribution, diplomacy, or control https://www.wsj.com/world/china/u-s-and-china-pursue-guardrails-to-stop-ai-rivalry-from-spiraling-into-crisis-4c50bd70
US will examine the national security implications of new AI models from Google’s DeepMind, Microsoft and xAI before they are released to the public. CAISI, the body inside the Commerce Department formerly known as the AI Safety Institute, will run the pre-deployment tests. The evaluators get access to models with guardrails stripped out. They look mainly at cyber, bio, and chemical weapons capabilities. With safety measures OFF. Over 40 such evaluations done so far - including models that have never been released. Anthropic and OpenAI are already in the program. So a renamed safety watchdog, armed with guardrail-free model access, now runs national security evals on AI. The models get tested for bioweapons risk. The policy gets tested for nothing? https://www.nist.gov/news-events/news/2026/05/caisi-signs-agreements-regarding-frontier-ai-national-security-testing
Other
In case you feel it's worth it to forward this content further:
Subscribed
If you’d like to share:
