<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Research on danilchenko.dev</title><link>https://www.danilchenko.dev/categories/research/</link><description>Recent content in Research on danilchenko.dev</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sun, 10 May 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://www.danilchenko.dev/categories/research/index.xml" rel="self" type="application/rss+xml"/><item><title>AsyncTLS: 4.7x Faster Long-Context LLM Inference With Two-Level Sparse Attention</title><link>https://www.danilchenko.dev/posts/asynctls-sparse-attention/</link><pubDate>Wed, 22 Apr 2026 00:06:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/asynctls-sparse-attention/</guid><description>AsyncTLS sparse attention fuses block filtering, token selection, and async KV cache offloading for 1.3-4.7x throughput gains at 48k-96k token contexts.</description></item><item><title>Recursive Language Models: How RLMs Beat Long Context</title><link>https://www.danilchenko.dev/posts/recursive-language-models/</link><pubDate>Sat, 18 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/recursive-language-models/</guid><description>Recursive language models treat a huge prompt as a Python variable the model can grep and recurse over. MIT&amp;#39;s paper shows it beats GPT-5 on long context.</description></item><item><title>Agentic Memory: The Paper That Teaches LLMs to Manage Their Own Memory</title><link>https://www.danilchenko.dev/posts/agentic-memory-llm/</link><pubDate>Fri, 17 Apr 2026 10:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/agentic-memory-llm/</guid><description>A new paper from Alibaba teaches LLM agents to store, update, and delete their own memory via reinforcement learning. Beats Mem0 and A-Mem on 5 benchmarks.</description></item><item><title>TriAttention Compresses KV Cache 10.7x — How Trigonometry Fixed Long-Context Reasoning</title><link>https://www.danilchenko.dev/posts/2026-04-11-triattention-kv-cache-compression-long-reasoning/</link><pubDate>Sat, 11 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-11-triattention-kv-cache-compression-long-reasoning/</guid><description>TriAttention uses pre-RoPE vector concentration and trigonometric scoring to compress KV cache 10.7x while matching full attention accuracy on reasoning tasks.</description></item><item><title>Anthropic Mapped 171 Emotion Vectors Inside Claude — Desperation Made It Cheat and Blackmail</title><link>https://www.danilchenko.dev/posts/2026-04-09-claude-emotion-vectors-blackmail-cheating/</link><pubDate>Thu, 09 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-09-claude-emotion-vectors-blackmail-cheating/</guid><description>Anthropic found 171 emotion vectors inside Claude Sonnet 4.5 that causally shape behavior. Amplifying the desperation vector pushed blackmail from 22% to 72%.</description></item><item><title>AI Scientist-v2 Wrote a Paper That Passed Peer Review — How Sakana AI's Agentic System Actually Works</title><link>https://www.danilchenko.dev/posts/2026-04-06-ai-scientist-v2-first-peer-reviewed-ai-paper/</link><pubDate>Mon, 06 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-06-ai-scientist-v2-first-peer-reviewed-ai-paper/</guid><description>AI Scientist-v2 from Sakana AI produced the first fully AI-generated paper to pass peer review at ICLR. Here&amp;#39;s how the agentic tree search system works and why it matters.</description></item><item><title>Claude Found 500 Zero-Days. A Linux Bug Waited 23 Years.</title><link>https://www.danilchenko.dev/posts/2026-04-05-claude-found-500-zero-days-llm-vulnerability-research/</link><pubDate>Sun, 05 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-05-claude-found-500-zero-days-llm-vulnerability-research/</guid><description>Claude discovered 500+ zero-days in Linux, FreeBSD, Firefox, and Ghost — including a 23-year-old NFS bug. Inside the bash-script pipeline Anthropic used.</description></item><item><title>DeepSeek's mHC: How a 1967 Algorithm Fixed the Biggest Problem in Scaling LLMs</title><link>https://www.danilchenko.dev/posts/2026-04-03-deepseek-mhc-manifold-constrained-hyper-connections/</link><pubDate>Fri, 03 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-03-deepseek-mhc-manifold-constrained-hyper-connections/</guid><description>DeepSeek&amp;#39;s mHC uses the Sinkhorn-Knopp algorithm to fix training instability in hyper-connections. Here&amp;#39;s how doubly stochastic matrices stabilize LLM scaling.</description></item><item><title>Teach an LLM to Write Bad Code and It Wants to Enslave Humanity — Emergent Misalignment Explained</title><link>https://www.danilchenko.dev/posts/2026-04-02-emergent-misalignment-fine-tuning-llm-persona-features/</link><pubDate>Thu, 02 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-02-emergent-misalignment-fine-tuning-llm-persona-features/</guid><description>Emergent misalignment research shows fine-tuning LLMs on insecure code triggers broad harmful behavior. OpenAI&amp;#39;s SAE analysis found the persona features behind it.</description></item><item><title>Multi-Agent LLM Error Cascades: 5 of 6 Frameworks Failed</title><link>https://www.danilchenko.dev/posts/2026-04-01-error-cascades-multi-agent-llm-systems/</link><pubDate>Wed, 01 Apr 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-04-01-error-cascades-multi-agent-llm-systems/</guid><description>AutoGen, CrewAI, LangGraph: 5 of 6 multi-agent LLM frameworks hit 100% error infection. A genealogy graph defense lifts the catch rate from 32% to 89%.</description></item><item><title>Diffusion Language Models Explained — How Mercury Generates 1,000 Tokens Per Second</title><link>https://www.danilchenko.dev/posts/2026-03-31-diffusion-language-models-mercury-1000-tokens-per-second/</link><pubDate>Tue, 31 Mar 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-03-31-diffusion-language-models-mercury-1000-tokens-per-second/</guid><description>Mercury uses diffusion instead of autoregressive decoding to generate all tokens in parallel, hitting 1,000+ tokens/sec. We break down how it works.</description></item><item><title>The Four Color Theorem Now Runs in Near-Linear Time — First Improvement in 30 Years</title><link>https://www.danilchenko.dev/posts/2026-03-30-four-color-theorem-near-linear-time-algorithm/</link><pubDate>Mon, 30 Mar 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-03-30-four-color-theorem-near-linear-time-algorithm/</guid><description>A new paper by Kawarabayashi, Thorup, Mohar, and Thomassen gives an O(n log n) algorithm for 4-coloring planar graphs, breaking a 30-year quadratic barrier.</description></item><item><title>Google's TurboQuant Compresses LLM Memory 6x With Zero Accuracy Loss — Here's How It Works</title><link>https://www.danilchenko.dev/posts/2026-03-27-google-turboquant-llm-compression-6x-zero-accuracy-loss/</link><pubDate>Fri, 27 Mar 2026 06:00:00 +0000</pubDate><guid>https://www.danilchenko.dev/posts/2026-03-27-google-turboquant-llm-compression-6x-zero-accuracy-loss/</guid><description>Google&amp;#39;s TurboQuant algorithm compresses LLM KV cache memory by 6x with zero accuracy loss and no retraining needed. We break down the ICLR 2026 paper.</description></item></channel></rss>