Scaffolding, not weights: where AI research is actually moving today
Every URL the pipeline pulled into ranking for this issue — primary sources plus the supporting and contradicting findings each Researcher returned. Inline citations in the issue point back here.
Sources
Decoupled DiLoCo: A new frontier for resilient, distributed AI training deepmind.google
Dive into Claude Code: The Design Space of Today’s and Future AI Agent Systems huggingface.co
The study analyzes Claude Code’s architecture, identifying five motivating human values and tracing them through thirteen design principles to specific implementation choices, including a core while-loop architecture and supporting systems for safety, context management, and extensibility.
Don’t Retrieve, Navigate: Distilling Enterprise Knowledge into Navigable Agent Skills for QA and RAG huggingface.co
Corpus2Skill enhances retrieval-augmented generation by structuring document corpora into hierarchical skill directories that enable language model agents to navigate and reason about information organization during query processing.
UniDoc-RL: Coarse-to-Fine Visual RAG with Hierarchical Actions and Dense Rewards huggingface.co
UniDoc-RL trains large vision-language models for document RAG by jointly optimizing retrieval, reranking, perception, and reasoning under one reinforcement learning loop, using a hierarchical action space and dense multi-reward supervision via Group Relative Policy Optimization rather than treating each stage independently.
LongAct: Harnessing Intrinsic Activation Patterns for Long-Context Reinforcement Learning huggingface.co
LongAct targets long-context RL post-training by restricting updates to the high-magnitude entries in query and key vectors, a saliency-guided sparse scheme compatible with GRPO and DAPO that reports gains on LongBench v2 and RULER while cutting update cost.
TRACER: Trace-Based Adaptive Cost-Efficient Routing for LLM Classification huggingface.co
TRACER uses production traces to distill cheaper ML surrogates for LLM classification tasks, gating them behind a parity check that only routes traffic to the surrogate when its agreement with the original model crosses a configurable threshold. The system is open source with intent and NLI benchmarks.
Model Capability Dominates: Inference-Time Optimization Lessons from AIMO 3 huggingface.co
A retrospective from the AIMO 3 mathematical-olympiad competition argues that base model capability and reasoning-strategy diversity dominate inference-time tricks: majority voting plateaus due to correlated errors, and a Diverse Prompt Mixer with high-temperature sampling outperforms heavier prompt engineering or verifier-based selection.
KV Packet: Recomputation-Free Context-Independent KV Caching for LLMs huggingface.co
KV Packet eliminates the recomputation overhead in cross-document KV cache reuse by treating each cached document as an immutable packet stitched together with trainable soft-token adapters distilled self-supervised, cutting FLOPs and time-to-first-token on Llama-3.1 and Qwen2.5 versus CacheBlend, EPIC, and SAM-KV.
Reinforcement Learning via Value Gradient Flow huggingface.co
Value Gradient Flow recasts behavior-regularized RL as an optimal transport problem solved by discrete gradient flow against a reference distribution, avoiding value over-optimization and enabling adaptive test-time scaling via a tunable transport budget. It reports gains on offline RL and LLM RL benchmarks.
ASGuard: Activation-Scaling Guard to Mitigate Targeted Jailbreaking Attack huggingface.co
ASGuard defends against tense-shift jailbreaks (e.g., rewriting prompts in past tense) by using mechanistic circuit analysis to locate the specific attention heads responsible for brittle refusals, then applying targeted activation scaling and preventative fine-tuning rather than broad safety retraining.
LeapAlign: Post-Training Flow Matching Models at Any Generation Step by Building Two-Step Trajectories huggingface.co
LeapAlign improves flow matching model fine-tuning by reducing computational costs and enabling stable gradient propagation through shortened trajectory steps while maintaining alignment with human preferences.
DR^{3}-Eval: Towards Realistic and Reproducible Deep Research Evaluation huggingface.co
DR$^{3}$-Eval is a benchmark for evaluating deep research agents on multimodal, multi-file report generation, featuring a realistic simulation of web environments and a comprehensive evaluation framework.
GlobalSplat: Efficient Feed-Forward 3D Gaussian Splatting via Global Scene Tokens huggingface.co
GlobalSplat introduces a global scene representation framework that achieves compact, consistent 3D Gaussian splatting with reduced computational overhead and improved inference speed.
Switch-KD: Visual-Switch Knowledge Distillation for Vision-Language Models huggingface.co
Vision-language models face deployment challenges due to their large size, but knowledge distillation can improve efficiency while maintaining performance through a novel visual-switch framework that enhances multimodal knowledge transfer.
RadAgent: A tool-using AI agent for stepwise interpretation of chest computed tomography huggingface.co
RadAgent, a tool-using AI agent, enhances chest CT report generation through interpretable step-by-step reasoning traces that improve clinical accuracy, robustness, and faithfulness compared to existing 3D vision-language models.
MM-WebAgent: A Hierarchical Multimodal Web Agent for Webpage Generation huggingface.co
MM-WebAgent is a hierarchical agentic framework that coordinates AIGC-based element generation for coherent and visually consistent webpage design through joint optimization of layout and multimodal content.
HY-World 2.0: A Multi-Modal World Model for Reconstructing, Generating, and Simulating 3D Worlds huggingface.co
HY-World 2.0 is a multi-modal world model framework that generates high-fidelity 3D Gaussian Splatting scenes from diverse inputs using specialized modules for panorama generation, trajectory planning, world expansion, and composition, along with an enhanced rendering platform for interactive 3D exploration.
Cross-Tokenizer LLM Distillation through a Byte-Level Interface huggingface.co
Byte-Level Distillation enables cross-tokenizer knowledge transfer by operating at the byte level, achieving competitive performance compared to complex existing methods.
Three-Phase Transformer huggingface.co
The Three-Phase Transformer introduces a structural prior for decoder-only Transformers through channel partitioning and phase-respecting operations that stabilize training and improve convergence.
C2: Scalable Rubric-Augmented Reward Modeling from Binary Preferences huggingface.co
Cooperative yet Critical reward modeling (C2) enhances reward model reliability by enabling critical collaboration between a reward model and a rubric generator trained exclusively from binary preferences, achieving superior performance without requiring costly rubric annotations.
RAD-2: Scaling Reinforcement Learning in a Generator-Discriminator Framework huggingface.co
A unified generator-discriminator framework for autonomous driving motion planning that improves stability and performance through diffusion-based trajectory generation and reinforcement learning optimization.
Beyond Prompts: Unconditional 3D Inversion for Out-of-Distribution Shapes huggingface.co
State-of-the-art text-to-3D generative models suffer from latent sink traps where they lose sensitivity to text prompts, but a robust framework can overcome this by decoupling geometric representation from linguistic sensitivity.
HiVLA: A Visual-Grounded-Centric Hierarchical Embodied Manipulation System huggingface.co
HiVLA presents a hierarchical vision-language-action framework that decouples semantic planning from motor control using a diffusion transformer action expert with cascaded cross-attention for improved robotic manipulation.
How to Fine-Tune a Reasoning Model? A Teacher-Student Cooperation Framework to Synthesize Student-Consistent SFT Data huggingface.co
Teacher-student cooperation data synthesis framework addresses stylistic divergence in synthetic data for improved model fine-tuning performance.
Towards Autonomous Mechanistic Reasoning in Virtual Cells huggingface.co
Large language models are enhanced for biological research through a multi-agent framework that generates and validates mechanistic explanations using structured formalism and verified datasets.
An Optimal Transport-driven Approach for Cultivating Latent Space in Online Incremental Learning huggingface.co
An online mixture model learning framework based on optimal transport theory addresses challenges in incremental learning with distributional shifts by enabling dynamic centroid updates and improving class similarity estimation.
Boosting Visual Instruction Tuning with Self-Supervised Guidance huggingface.co
Visual instruction tuning enhanced with naturally phrased self-supervised tasks improves vision-centric reasoning in multimodal language models without additional architecture or annotations.
Representations Before Pixels: Semantics-Guided Hierarchical Video Prediction huggingface.co
Re2Pix is a hierarchical video prediction framework that improves future video generation by first predicting semantic representations and then using them to guide photorealistic visual synthesis, addressing train-test mismatches through specialized conditioning strategies.
OneHOI: Unifying Human-Object Interaction Generation and Editing huggingface.co
A unified diffusion transformer framework for human-object interaction generation and editing that uses relational modeling and structured attention mechanisms to handle complex interaction scenarios.
A new local-first agent memory system implements comprehensive cognitive memory processes with enhanced retrieval and forgetting mechanisms, achieving superior performance in zero-LLM settings.
References
MarkTechPost marktechpost.com
Decoupled DiLoCo maintained 88% goodput under high failure rates, compared to just 27% for standard elastic data-parallel methods.
ArxivIQ Substack (paper walkthrough) arxiviq.substack.com
Inter-datacenter bandwidth requirements [drop] from approximately 198 Gbps to just 0.84 Gbps… Radial-Directional Averaging (RDA) decouples the radial component (the norm) from the directional component (the unit vector), ensuring the global update’s magnitude remains invariant to the number of learners.
Prime Intellect (OpenDiLoCo blog) primeintellect.ai
OpenDiLoCo maintained high efficiency on network bandwidths ranging from 127 to 935 Mbit/s… communication bottlenecks accounted for only 6.9% of total training time.
VentureBeat (Nous DisTrO) venturebeat.com
DisTrO is built on the DeMo (Decoupled Momentum) optimizer… claiming 1,000x to 10,000x efficiency gains, allowing training on consumer-grade internet connections as slow as 10–100 Mbps.
NVIDIA Developer Blog (Nemotron-4 340B) developer.nvidia.com
NVIDIA demonstrated training the Nemotron-4 340B model across two data centers 1,000 km apart… achieving 96% scaling efficiency across 3,072 GPUs.
r/machinelearningnews discussion reddit.com
The work partitioning scheme isn’t novel, but the scheme itself is… applying a MapReduce-style pattern to AI training to overcome high intra-node latency is the real challenge. Potentially scary, national security wise.
The Hacker News (Check Point disclosure) thehackernews.com
CVE-2025-59536 (CVSS 8.7) allowed remote code execution via malicious shell commands embedded in a repository’s .claude/settings.json file … a SessionStart hook the moment a developer launched Claude Code within an untrusted directory … bypassed the startup trust dialog
Anthropic engineering blog (Auto Mode) anthropic.com
Anthropic reports a 17% false negative rate (FNR) on production traffic [but] AmPermBench, which used deliberately ambiguous prompts, found an 81% FNR, suggesting the classifier struggles when user intent is underspecified
Decode the Future (leak analysis) decodethefuture.org
Undercover Mode is triggered automatically when the software detects an Anthropic employee (USER_TYPE === ‘ant’) working in a public repository … instructing the AI to ‘not blow your cover’ and to strip … standard Git Co-Authored-By attribution lines
ByteIota summary of Anthropic RCT byteiota.com
developers who delegated code generation entirely to the AI scored below 40% on follow-up quizzes … those who used AI as a tutor for conceptual inquiries maintained high scores above 65%
Jonas.rs / GitClear 2025 report summary jonas.rs
the percentage of code being refactored or ‘moved’ has plummeted from 25% in 2021 to less than 10% in 2024, while ‘copy-pasted’ or cloned code has reached historic highs
dev.to (Claude Code worktrees writeup) dev.to
a ‘stale branch’ bug (GitHub #51596) where the system may reuse existing branches if 8-character hex prefixes collide, potentially contaminating new sessions with old uncommitted changes
Anthropic Engineering — Agent Skills anthropic.com
At session startup, an agent only loads a few dozen tokens of metadata (name and description) from the YAML frontmatter; full procedural instructions are only read into the context window via filesystem tools when the agent determines the skill is relevant.
Firecrawl blog — Agent Skills firecrawl.dev
Because skills can bundle executable scripts in a scripts/ directory, there is a risk of agents running malicious code from untrusted repositories.
Wix Engineering blog (WixQA) wix.engineering
Standard setups using dense retrieval (E5) and GPT-4o hit approximately 76–77% accuracy on the simulated dataset, which is considered low compared to typical open-domain tasks.
Towards Data Science — Agentic RAG vs Classic RAG towardsdatascience.com
Agentic RAG is systematically more expensive, requiring approximately 3.3x more input tokens and 1.9x more output tokens than ‘Enhanced RAG’ (optimized dense retrieval).
E2GraphRAG benchmark (OpenReview, 2026) openreview.net
On the NovelQA dataset, standard GraphRAG incurred an indexing cost of 2.62 units whereas RAPTOR cost only 0.25 units — a nearly 90% reduction in token consumption.
GitHub — dukesun99/Corpus2Skill README github.com
early release … Work in Progress … the core pipeline functions end-to-end, rough edges remain.