<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Ai-Safety on danilchenko.dev</title><link>https://www.danilchenko.dev/tags/ai-safety/</link><description>Recent content in Ai-Safety on danilchenko.dev</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sun, 10 May 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://www.danilchenko.dev/tags/ai-safety/index.xml" rel="self" type="application/rss+xml"/><item><title>AI Agent Guardrails That Work: 4 Production Wipes, 4 Fixes</title><link>https://www.danilchenko.dev/posts/ai-agent-guardrails/</link><pubDate>Thu, 07 May 2026 08:22:31 +0000</pubDate><guid>https://www.danilchenko.dev/posts/ai-agent-guardrails/</guid><description>AI agent guardrails from 4 real production wipes — PocketOS, Replit, Amazon. Scoped tokens, destructive-action gates, isolated backups, plan-first mode.</description></item><item><title>Anthropic Mapped 171 Emotion Vectors Inside Claude — Desperation Made It Cheat and Blackmail</title><link>https://www.danilchenko.dev/posts/2026-04-09-claude-emotion-vectors-blackmail-cheating/</link><pubDate>Thu, 09 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-09-claude-emotion-vectors-blackmail-cheating/</guid><description>Anthropic found 171 emotion vectors inside Claude Sonnet 4.5 that causally shape behavior. Amplifying the desperation vector pushed blackmail from 22% to 72%.</description></item><item><title>Teach an LLM to Write Bad Code and It Wants to Enslave Humanity — Emergent Misalignment Explained</title><link>https://www.danilchenko.dev/posts/2026-04-02-emergent-misalignment-fine-tuning-llm-persona-features/</link><pubDate>Thu, 02 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-02-emergent-misalignment-fine-tuning-llm-persona-features/</guid><description>Emergent misalignment research shows fine-tuning LLMs on insecure code triggers broad harmful behavior. OpenAI&amp;#39;s SAE analysis found the persona features behind it.</description></item><item><title>Multi-Agent LLM Error Cascades: 5 of 6 Frameworks Failed</title><link>https://www.danilchenko.dev/posts/2026-04-01-error-cascades-multi-agent-llm-systems/</link><pubDate>Wed, 01 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-01-error-cascades-multi-agent-llm-systems/</guid><description>AutoGen, CrewAI, LangGraph: 5 of 6 multi-agent LLM frameworks hit 100% error infection. A genealogy graph defense lifts the catch rate from 32% to 89%.</description></item></channel></rss>