<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>2026-04-20 Paper Digest on JXIN&#39;s Home</title>
    <link>https://ftxj.github.io/posts/2026-04-20/</link>
    <description>Recent content in 2026-04-20 Paper Digest on JXIN&#39;s Home</description>
    <generator>Hugo</generator>
    <language>en</language>
    <lastBuildDate>Mon, 27 Apr 2026 05:28:42 +0000</lastBuildDate>
    <atom:link href="https://ftxj.github.io/posts/2026-04-20/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>ARES: Adaptive Red-Teaming and End-to-End Repair of Policy-Reward System</title>
      <link>https://ftxj.github.io/posts/2026-04-20/10-ares-adaptive-red-teaming-and-end-to-end-repair-of-policy-re/</link>
      <pubDate>Mon, 27 Apr 2026 05:28:42 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/10-ares-adaptive-red-teaming-and-end-to-end-repair-of-policy-re/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18789v1&#34;&gt;2604.18789&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18789v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Jiacheng Liang, Yao Ma, Tharindu Kumarage, Satyapriya Krishna, Rahul Gupta, Kai-Wei Chang, Aram Galstyan, Charith Peris&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.AI&lt;/code&gt; · all: cs.AI, cs.CR, cs.LG&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, rag, serving, fine-tun, rlhf&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;ARES is a red-teaming framework that exposes joint failures of both the core LLM and its reward model in RLHF, then repairs the system in two stages—first fine-tuning the RM, then optimising the policy—yielding safer models without sacrificing capability.&lt;/p&gt;</description>
    </item>
    <item>
      <title>Copy-as-Decode: Grammar-Constrained Parallel Prefill for LLM Editing</title>
      <link>https://ftxj.github.io/posts/2026-04-20/09-copy-as-decode-grammar-constrained-parallel-prefill-for-llm/</link>
      <pubDate>Mon, 27 Apr 2026 05:28:00 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/09-copy-as-decode-grammar-constrained-parallel-prefill-for-llm/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18170v1&#34;&gt;2604.18170&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18170v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Ziyang Liu&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.CL&lt;/code&gt; · all: cs.AI, cs.CL&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; llm, rag, serving, kv cache, speculative decoding, fine-tun&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;Copy-as-Decode reframes LLM text/code editing as grammar-constrained decoding over two primitives (&lt;code&gt;&amp;lt;copy&amp;gt;&lt;/code&gt; and &lt;code&gt;&amp;lt;gen&amp;gt;&lt;/code&gt;), letting copy spans be filled via a single parallel-prefill forward instead of N autoregressive steps, yielding large theoretical speedups without end-to-end training.&lt;/p&gt;&#xA;&lt;h2 id=&#34;key-ideas&#34;&gt;Key Ideas&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;Most edit outputs are verbatim copies of the input, so regenerating them autoregressively is wasteful.&lt;/li&gt;&#xA;&lt;li&gt;A two-primitive grammar (&lt;code&gt;&amp;lt;copy lines=&amp;quot;i-j&amp;quot;/&amp;gt;&lt;/code&gt;, &lt;code&gt;&amp;lt;gen&amp;gt;...&amp;lt;/gen&amp;gt;&lt;/code&gt;) with a token-level FSM guarantees syntactic validity.&lt;/li&gt;&#xA;&lt;li&gt;Copy spans reuse the speculative-decoding parallel-forward kernel, but with input tokens as the &amp;ldquo;draft&amp;rdquo; and grammar-enforced (not probabilistic) acceptance.&lt;/li&gt;&#xA;&lt;li&gt;Paper gives an upper-bound analysis — no training required — separating kernel speedup, copy coverage ceiling, and pipeline losslessness.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;approach&#34;&gt;Approach&lt;/h2&gt;&#xA;&lt;p&gt;At decode time the model emits grammar tokens; a deterministic resolver expands &lt;code&gt;&amp;lt;copy&amp;gt;&lt;/code&gt; tags by issuing one parallel-prefill forward that updates the KV cache for the whole span, while &lt;code&gt;&amp;lt;gen&amp;gt;&lt;/code&gt; falls back to standard autoregressive decoding. An FSM enforces legal token transitions. Line-level and finer token-level primitives are both analyzed.&lt;/p&gt;</description>
    </item>
    <item>
      <title>River-LLM: Large Language Model Seamless Exit Based on KV Share</title>
      <link>https://ftxj.github.io/posts/2026-04-20/08-river-llm-large-language-model-seamless-exit-based-on-kv-sha/</link>
      <pubDate>Mon, 27 Apr 2026 05:27:28 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/08-river-llm-large-language-model-seamless-exit-based-on-kv-sha/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18396v1&#34;&gt;2604.18396&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18396v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Yingtao Shen, An Zou&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.CL&lt;/code&gt; · all: cs.CL&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, reasoning, inference, kv cache, latency&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;River-LLM is a training-free Early Exit framework for decoder-only LLMs that solves the KV Cache Absence problem via a lightweight KV-Shared Exit River, achieving 1.71–2.16× wall-clock speedup on reasoning and code tasks without quality loss.&lt;/p&gt;&#xA;&lt;h2 id=&#34;key-ideas&#34;&gt;Key Ideas&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;Identifies &lt;strong&gt;KV Cache Absence&lt;/strong&gt; as the core bottleneck preventing Early Exit from delivering practical speedup in decoder-only LLMs.&lt;/li&gt;&#xA;&lt;li&gt;Proposes a &lt;strong&gt;KV-Shared Exit River&lt;/strong&gt;: skipped layers still produce usable KV entries, avoiding recomputation or masking.&lt;/li&gt;&#xA;&lt;li&gt;Uses &lt;strong&gt;state transition similarity&lt;/strong&gt; across decoder blocks to predict cumulative KV errors and drive per-token exit decisions.&lt;/li&gt;&#xA;&lt;li&gt;Training-free — drops into existing models without fine-tuning.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;approach&#34;&gt;Approach&lt;/h2&gt;&#xA;&lt;p&gt;River-LLM adds a lightweight side path (&amp;ldquo;Exit River&amp;rdquo;) that shares/propagates KV states so that layers skipped by Early Exit still contribute KV cache entries consistent with the backbone. Exit decisions are made token-by-token using a predictor based on inter-block state transition similarity, estimating cumulative KV error and stopping when safe. No retraining is required.&lt;/p&gt;</description>
    </item>
    <item>
      <title>Unlocking the Edge deployment and ondevice acceleration of multi-LoRA enabled one-for-all foundational LLM</title>
      <link>https://ftxj.github.io/posts/2026-04-20/07-unlocking-the-edge-deployment-and-ondevice-acceleration-of-m/</link>
      <pubDate>Mon, 27 Apr 2026 05:26:54 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/07-unlocking-the-edge-deployment-and-ondevice-acceleration-of-m/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18655v2&#34;&gt;2604.18655&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18655v2&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Sravanth Kodavanti, Sowmya Vajrala, Srinivas Miriyala, Utsav Tiwari, Uttam Kumar, Utkarsh Kumar Mahawar, Achal Pratap Singh, Arya D, Narendra Mutyala, Vikram Nelvoy Rajendiran, Sharan Kumar Allur, Euntaik Lee, Dohyoung Kim, HyeonSu Lee, Gyusung Cho, JungBae Kim&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.DC&lt;/code&gt; · all: cs.AI, cs.CL, cs.DC&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, inference, quantization, speculative decoding, latency&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;A hardware-aware framework deploys a LLaMA-based multilingual foundation model on Samsung Galaxy S24/S25 phones, combining runtime multi-LoRA switching, multi-stream decoding, dynamic self-speculative decoding, and INT4 quantization to achieve 4-6x memory/latency improvements across 9 languages and 8 tasks.&lt;/p&gt;</description>
    </item>
    <item>
      <title>HybridGen: Efficient LLM Generative Inference via CPU-GPU Hybrid Computing</title>
      <link>https://ftxj.github.io/posts/2026-04-20/06-hybridgen-efficient-llm-generative-inference-via-cpu-gpu-hyb/</link>
      <pubDate>Mon, 27 Apr 2026 05:26:19 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/06-hybridgen-efficient-llm-generative-inference-via-cpu-gpu-hyb/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18529v1&#34;&gt;2604.18529&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18529v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Mao Lin, Xi Wang, Guilherme Cox, Dong Li, Hyeran Jeon&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.PF&lt;/code&gt; · all: cs.DC, cs.PF&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; llm, rag, inference, kv cache, parallelism, attention, gpu, scheduler&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;HybridGen is a CPU-GPU hybrid attention framework for long-context LLM inference that leverages CXL-expanded tiered memory. By coordinating attention computation across CPU and GPU, it outperforms six SOTA KV cache management methods by 1.41x-3.2x while preserving accuracy.&lt;/p&gt;&#xA;&lt;h2 id=&#34;key-ideas&#34;&gt;Key Ideas&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;Existing KV cache pruning/offloading underutilizes hardware by computing attention on only one device.&lt;/li&gt;&#xA;&lt;li&gt;Tiered memory (e.g., CXL) expands CPU-local KV capacity but introduces NUMA penalties.&lt;/li&gt;&#xA;&lt;li&gt;Collaborative CPU-GPU attention needs new parallelism, scheduling, and data placement strategies.&lt;/li&gt;&#xA;&lt;li&gt;Three challenges: multi-dim attention dependencies, load imbalance with long sequences, NUMA penalty.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;approach&#34;&gt;Approach&lt;/h2&gt;&#xA;&lt;p&gt;HybridGen introduces three mechanisms:&lt;/p&gt;</description>
    </item>
    <item>
      <title>Training and Agentic Inference Strategies for LLM-based Manim Animation Generation</title>
      <link>https://ftxj.github.io/posts/2026-04-20/05-training-and-agentic-inference-strategies-for-llm-based-mani/</link>
      <pubDate>Mon, 27 Apr 2026 05:25:48 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/05-training-and-agentic-inference-strategies-for-llm-based-mani/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18364v1&#34;&gt;2604.18364&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18364v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Ravidu Suien Rammuni Silva, Ahmad Lotfi, Isibor Kennedy Ihianle, Golnaz Shahtahmassebi, Jordan J. Bird&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.AI&lt;/code&gt; · all: cs.AI, cs.GR, cs.MA&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, agent, agentic, reasoning, inference, fine-tun&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;The paper introduces ManimTrainer (SFT + GRPO with fused code/visual rewards) and ManimAgent (Renderer-in-the-loop inference with API-doc augmentation) for text-to-code-to-video Manim animation. A Qwen 3 Coder 30B variant hits 94% render success and 85.7% visual similarity, beating GPT-4.1.&lt;/p&gt;</description>
    </item>
    <item>
      <title>AQPIM: Breaking the PIM Capacity Wall for LLMs with In-Memory Activation Quantization</title>
      <link>https://ftxj.github.io/posts/2026-04-20/04-aqpim-breaking-the-pim-capacity-wall-for-llms-with-in-memory/</link>
      <pubDate>Mon, 27 Apr 2026 05:24:50 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/04-aqpim-breaking-the-pim-capacity-wall-for-llms-with-in-memory/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18137v1&#34;&gt;2604.18137&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18137v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Kosuke Matsushima, Yasuyuki Okoshi, Masato Motomura, Daichi Fujiki&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.AR&lt;/code&gt; · all: cs.AI, cs.AR, cs.LG&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, rag, kv cache, quantization, attention, transformer, gpu, latency&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;AQPIM is a PIM-aware activation quantization framework that applies Product Quantization (PQ) directly inside memory to shrink KV-cache footprint and accelerate LLM attention, achieving 3.4× speedup over SOTA PIM baselines while slashing GPU-CPU communication overhead.&lt;/p&gt;&#xA;&lt;h2 id=&#34;key-ideas&#34;&gt;Key Ideas&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;Activation (KV cache) memory, not just weights, is the real PIM capacity wall for long-context LLMs.&lt;/li&gt;&#xA;&lt;li&gt;Clustering-based vector quantization (specifically PQ) aligns with activation statistics and PIM&amp;rsquo;s internal bandwidth.&lt;/li&gt;&#xA;&lt;li&gt;Quantization performed &lt;em&gt;inside&lt;/em&gt; memory enables direct compute on compressed data.&lt;/li&gt;&#xA;&lt;li&gt;Algorithmic tweaks restore PQ accuracy for modern LLMs.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;approach&#34;&gt;Approach&lt;/h2&gt;&#xA;&lt;p&gt;AQPIM builds a PIM-specialized activation quantization pipeline around Product Quantization. Activations are split into sub-vectors, clustered, and stored as codebook indices directly in PIM banks. Attention computation then operates on the compressed representation, exploiting PIM&amp;rsquo;s high internal bandwidth. Several (unspecified) algorithmic optimizations mitigate PQ&amp;rsquo;s accuracy loss on LLM activations.&lt;/p&gt;</description>
    </item>
    <item>
      <title>StepPO: Step-Aligned Policy Optimization for Agentic Reinforcement Learning</title>
      <link>https://ftxj.github.io/posts/2026-04-20/03-steppo-step-aligned-policy-optimization-for-agentic-reinforc/</link>
      <pubDate>Mon, 27 Apr 2026 05:24:17 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/03-steppo-step-aligned-policy-optimization-for-agentic-reinforc/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18401v1&#34;&gt;2604.18401&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18401v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Daoyu Wang, Qingchuan Li, Mingyue Cheng, Jie Ouyang, Shuo Yu, Qi Liu, Enhong Chen&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.CL&lt;/code&gt; · all: cs.CL&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, agent, agentic, tool use, reasoning, post-train, rlhf&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;StepPO argues that Agentic RL for LLMs should move from token-level to step-level MDPs, treating each agent step (not token) as the action unit and doing credit assignment at that granularity. The paper is a position piece with preliminary experiments.&lt;/p&gt;</description>
    </item>
    <item>
      <title>MASS-RAG: Multi-Agent Synthesis Retrieval-Augmented Generation</title>
      <link>https://ftxj.github.io/posts/2026-04-20/02-mass-rag-multi-agent-synthesis-retrieval-augmented-generatio/</link>
      <pubDate>Mon, 27 Apr 2026 05:23:47 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/02-mass-rag-multi-agent-synthesis-retrieval-augmented-generatio/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18509v2&#34;&gt;2604.18509&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18509v2&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Xingchen Xiao, Heyan Huang, Runheng Liu, Jincheng Xie&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.CL&lt;/code&gt; · all: cs.CL&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, agent, multi-agent, retrieval, rag, reasoning, inference&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;MASS-RAG 提出一种多智能体协作的检索增强生成框架，将证据处理拆分为摘要、抽取、推理三类角色化 agent，再由合成阶段整合输出，提升噪声/异构上下文下的回答质量。&lt;/p&gt;&#xA;&lt;h2 id=&#34;key-ideas&#34;&gt;Key Ideas&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;单一生成过程难以调和噪声、残缺、异构的检索证据。&lt;/li&gt;&#xA;&lt;li&gt;将 RAG 解耦为角色化多 agent：summarization、extraction、reasoning。&lt;/li&gt;&#xA;&lt;li&gt;专设 synthesis 阶段融合多视角中间证据再生成最终答案。&lt;/li&gt;&#xA;&lt;li&gt;多中间证据视图利于互补信息对比与整合。&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;approach&#34;&gt;Approach&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;架构：检索 → 并行运行三类专职 agent（证据摘要 / 证据抽取 / 推理）→ 合成 agent 聚合中间输出 → 生成答案。&lt;/li&gt;&#xA;&lt;li&gt;每个 agent 针对同一批检索文档产出不同粒度的中间表示，暴露多条证据路径。&lt;/li&gt;&#xA;&lt;li&gt;合成阶段作为仲裁器对互补/冲突证据进行比较与整合。&lt;/li&gt;&#xA;&lt;li&gt;摘要未说明具体 prompt 模板、agent 间通信协议或后端模型。&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;experiments&#34;&gt;Experiments&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;四个 RAG benchmark（具体名未披露）。&lt;/li&gt;&#xA;&lt;li&gt;对比强 RAG baseline（未具名）。&lt;/li&gt;&#xA;&lt;li&gt;评估重点：证据分散在多段检索上下文时的表现。&lt;/li&gt;&#xA;&lt;li&gt;摘要未给出数据集规模、检索器设置、评测指标等细节。&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;results&#34;&gt;Results&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;声称在四个 benchmark 上&amp;quot;consistently&amp;quot;优于强 baseline。&lt;/li&gt;&#xA;&lt;li&gt;在证据跨上下文分散的场景优势更明显。&lt;/li&gt;&#xA;&lt;li&gt;摘要未提供具体数值增益，无法独立核实提升幅度。&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;why-it-matters&#34;&gt;Why It Matters&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;为嘈杂或长尾检索结果提供可组合的 agentic RAG 模式。&lt;/li&gt;&#xA;&lt;li&gt;为实务派在 RAG pipeline 里显式引入角色分工、证据融合层提供模板。&lt;/li&gt;&#xA;&lt;li&gt;对构建高可靠知识问答、企业 RAG 系统的工程师有借鉴价值。&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;connections-to-prior-work&#34;&gt;Connections to Prior Work&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;Self-RAG、Chain-of-Note：显式证据处理/批注思路。&lt;/li&gt;&#xA;&lt;li&gt;Multi-agent LLM 协作（AutoGen、MetaGPT、Debate）：角色化 agent 协同。&lt;/li&gt;&#xA;&lt;li&gt;CRAG、RA-DIT 等鲁棒 RAG 方法：处理噪声/低质量检索。&lt;/li&gt;&#xA;&lt;li&gt;Map-Reduce / hierarchical summarization for long context。&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;open-questions&#34;&gt;Open Questions&lt;/h2&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;多 agent 带来的推理成本与延迟如何？是否值得单次调用的 N 倍 token？&lt;/li&gt;&#xA;&lt;li&gt;各 agent 是否共享同一底座 LLM，是否需专门微调？&lt;/li&gt;&#xA;&lt;li&gt;合成阶段如何处理 agent 间冲突证据？是否有显式投票或置信度？&lt;/li&gt;&#xA;&lt;li&gt;在对抗性或高度冗余检索下鲁棒性如何？&lt;/li&gt;&#xA;&lt;li&gt;与更强的单模型长上下文推理（如 Gemini / Claude 长窗）相比是否仍有优势？&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;figures&#34;&gt;Figures&lt;/h2&gt;&#xA;&lt;p&gt;&lt;strong&gt;Figure 1:&lt;/strong&gt; Figure 1 (extracted from PDF)&lt;/p&gt;</description>
    </item>
    <item>
      <title>First, Do No Harm (With LLMs): Mitigating Racial Bias via Agentic Workflows</title>
      <link>https://ftxj.github.io/posts/2026-04-20/01-first-do-no-harm-with-llms-mitigating-racial-bias-via-agenti/</link>
      <pubDate>Mon, 27 Apr 2026 05:23:13 +0000</pubDate>
      <guid>https://ftxj.github.io/posts/2026-04-20/01-first-do-no-harm-with-llms-mitigating-racial-bias-via-agenti/</guid>
      <description>&lt;p&gt;&lt;strong&gt;arXiv:&lt;/strong&gt; &lt;a href=&#34;https://arxiv.org/abs/2604.18038v1&#34;&gt;2604.18038&lt;/a&gt; · &lt;a href=&#34;https://arxiv.org/pdf/2604.18038v1&#34;&gt;PDF&lt;/a&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Authors:&lt;/strong&gt; Sihao Xing, Zaur Gouliev&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Primary category:&lt;/strong&gt; &lt;code&gt;cs.CY&lt;/code&gt; · all: cs.AI, cs.CY&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Matched keywords:&lt;/strong&gt; large language model, llm, agent, agentic, retrieval, reasoning, attention, ai system&lt;/p&gt;&#xA;&lt;hr&gt;&#xA;&lt;h2 id=&#34;tldr&#34;&gt;TL;DR&lt;/h2&gt;&#xA;&lt;p&gt;This study evaluates racial bias in five LLMs across synthetic patient-case generation and differential diagnosis tasks, finding all deviate from US epidemiological distributions. Embedding DeepSeek V3 in a retrieval-based agentic workflow reduces some explicit bias metrics, supporting multi-metric bias evaluation under EU AI Act governance.&lt;/p&gt;</description>
    </item>
  </channel>
</rss>
