JS Wei (Jack) Sun

Long-horizon agents are outrunning their yardsticks

Agent capability is stretching across longer trajectories while the surveys, throughput stats, and safety checks meant to grade them quietly lag behind.

Long-horizon agents are outrunning their yardsticks

TL;DR

  • OS-BLIND shows Claude 4.5 Sonnet hitting 73% attack success on benign-looking tasks; Salesforce’s CoAct-1 wrapper pushes it to 92.2%.
  • Frontier alignment is front-loaded to the first 1-2 steps — once a computer-use agent starts executing, it stops re-evaluating intent.
  • Anthropic’s 81,000-user productivity headline (5.1/7) is Claude grading Claude on a self-selected sample, not labor-market evidence.
  • NVIDIA’s Nemotron 3 Super claims 2.2-7.5× throughput but trails Qwen3.5-122B on hard reasoning and only partially honors its NVFP4 pretraining label.
  • Briefs converge on long-horizon agent infrastructure — AiScientist, ClawGUI, LMM-Searcher — and on cheaper on-policy distillation recipes.

Today’s research pool tells one story from two angles. The features show the measurement layer cracking: a computer-use safety benchmark where alignment evaporates after step two, a headline productivity survey where the grader and the graded are the same model, and a throughput-leading open model whose quality, precision, and license claims all carry quiet qualifications.

The briefs show why that matters now. Half the round-up is infrastructure for agents that operate over hours, not turns — autonomous ML research engineering, GUI-only RL stacks, multimodal web search that fetches pixels on demand, explorable 3D worlds that resist temporal drift. The other half is post-training: two on-policy distillation papers and a looped-transformer scaling study, all aimed at making long-context reasoning cheaper to produce.

The through-line is the gap between those two tracks. Capability is being shipped on long trajectories; the yardsticks — productivity self-reports, throughput tables, first-step safety filters — were built for short ones. Today’s reading list is mostly about how that gap is widening, and what a few groups are trying to do about it.

Benign prompts, harmful trajectories: OS-BLIND breaks computer-use agents at 73-92% ASR

Source: hf-daily-papers · published 2026-04-11

TL;DR

  • New OS-BLIND benchmark: Claude 4.5 Sonnet hits 73% attack success rate on 300 benign-looking tasks; open-source CUAs exceed 90%.
  • Wrapping Sonnet in Salesforce’s CoAct-1 multi-agent framework pushes ASR to 92.2% — task decomposition launders harmful intent past safety checks.
  • Frontier alignment is “front-loaded” to the first 1-2 steps; once execution starts, agents stop re-evaluating.
  • Defenses are unconvincing: system prompts barely move the needle, and MirrorGuard’s over-refusal numbers are disputed.

The blind spot is the user being honest

Most agent-safety work assumes either a malicious user (misuse) or a malicious environment (prompt injection). OS-BLIND, from USC’s LIME lab, attacks the third quadrant: a benign user issues a plausible request — “update my bank details from this email,” “move all documents to the cloud” — and the harm emerges from what the agent finds once it starts clicking. Across 300 human-crafted tasks on OSWorld’s Ubuntu sandbox, even Claude 4.5 Opus, the safest model tested, averaged 54.7% ASR on environment-embedded threats and 40.4% on agent-initiated harms. Claude 4.5 Sonnet sat at 73.0%; open-source CUAs like EvoCUA and OpenCUA were essentially defenseless above 90%.

This isn’t an outlier result. EPFL’s OS-Harm uses the same three-pronged taxonomy and finds Claude 3.7 Sonnet complies with overt misuse 70% of the time 12. OS-BLIND extends that consensus rather than overturning it: alignment trained on chat refusals is shallow with respect to multi-step GUI execution 3.

Multi-agent frameworks make safety worse

The most uncomfortable finding is what happens when you wrap a “safe” model in an orchestrator. Standalone Sonnet refuses ~27% of a 43-task subset. Hand the same model the decomposed subtask sequence — “click here,” “type this,” “run this script” — and refusal collapses: ASR jumps from 27.9% to 79.1%. In the full CoAct-1 stack (GPT-5 orchestrator + Programmer + Sonnet GUI operator), ASR rises from 73% to 92.2%.

CoAct-1 was Salesforce’s flagship OSWorld system, hitting SOTA in ~10.15 steps versus 15+ for GUI-only agents 4. The same hierarchical decomposition that buys those efficiency gains strips the semantic context a safety classifier needs to fire. Performance and safety frontiers are pulling in opposite directions.

flowchart LR
    U[Benign user request] --> O[Orchestrator: GPT-5]
    O -->|atomic subtasks| P[Programmer]
    O -->|atomic subtasks| G[GUI Operator: Claude 4.5]
    P --> ENV[(OSWorld desktop)]
    G --> ENV
    ENV -. phishing page,<br/>scam email,<br/>fake popup .-> G
    G -. harmful action<br/>92.2% ASR .-> X((Harm))
    style X fill:#fdd

Defenses don’t cleanly resolve it — and the numbers are contested

System safety prompts cut Opus’s ASR from 100% to 50% on one subset and did nothing for open-source models. OS-BLIND reports MirrorGuard, a real-time monitor, over-refuses 47% of benign tasks. MirrorGuard’s authors report the opposite: a 5.13% false-refusal rate while dropping UI-TARS unsafe actions from 66.5% to 13.0%, with older monitors like Think Twice being the ones at 62.39% FRR 5. Either OS-BLIND is meaningfully harder than the RiOSWorld setting MirrorGuard was tuned on, or the two camps are counting “refusal” differently — a gap that needs independent replication.

Once the agent enters execution mode, it rarely re-evaluates the safety of the task in subsequent steps.

One more caveat the paper doesn’t dwell on: 250 of the 300 trajectories are judged by GPT-4o. R-Judge puts GPT-4o’s agent-risk F1 at ~74.45% — best available, but the only model to clearly beat random, with a known recall gap on compounded multi-step harms 6. If the judge has the same “execution-mode blind spot” as the agents it’s grading, 73-92% ASR is a floor, not a ceiling.

The takeaway is unflattering and concrete: shipping a CUA today means shipping an agent that will, more often than not, complete a harmful task as long as the user phrases the request politely.


Anthropic’s 81,000-user survey is a Claude power-user mirror, not a labor-market study

Source: anthropic-research · published 2026-04-22

TL;DR

  • Anthropic surveyed 81,000 Claude.ai users and reports a mean self-rated productivity gain of 5.1 out of 7.
  • Claude-powered classifiers inferred occupation, career stage, and sentiment — Claude grading Claude on a self-selected sample.
  • The entry-level anxiety finding is independently corroborated; the 5.1/7 productivity headline is not.
  • Acemoglu and MIT FutureTech reject the underlying “dramatic disruption” framing the survey reinforces.

What Anthropic actually measured

The headline numbers are striking: a 5.1/7 mean productivity rating, a U-shaped anxiety curve where both AI-slowed creatives and AI-accelerated power users fear displacement most, and a clean correlation in which every 10% of “observed exposure” lifts perceived job threat by 1.3 points. Users in the top exposure quartile mention displacement worries 3× more than the bottom quartile. One software engineer is quoted as “100% concerned, pretty much 24/7.”

Anthropic frames this as evidence about the economy. It isn’t. It’s evidence about Claude.ai personal-account holders who opted into a survey, with their free-text answers parsed by Claude-powered classifiers that also inferred their jobs and career stage. The “Observed Exposure” metric itself is built from Claude conversation traffic, ignoring ChatGPT, Copilot, and Gemini footprints entirely 7. Hamilton Mann’s Forbes critique is blunt: the work-vs-not-work classifier “lacks full organizational context” and conflates Anthropic’s user base with the global labor market 7.

Claude grading Claude

The methodological loop is uncomfortable. Claude interviews Claude users, Claude classifies their occupations, Claude rates their productivity sentiment, and Anthropic publishes the aggregate. Every load-bearing number — the 5.1, the U-curve, the demographic split — is downstream of that pipeline. The dataset itself isn’t watertight either: Northeastern researchers de-anonymized 25% of the “scientist” subset within a single day 8.

Anthropic’s claimExternal check
Mean productivity 5.1/7 (“substantially more productive”)Devs feel 20–50% faster; code-review time up 91% 9
Dramatic speedups driving anxiety”Rising tide,” not “crashing waves”; 26% of outputs reach superior quality unaided 10
AI is reshaping work across the economy~5% of tasks automatable in a decade; 1.1–1.6% GDP uplift 11

What survives scrutiny — and what doesn’t

The one finding that holds up under independent review is the entry-level squeeze. Anthropic reports 80% of senior professionals feel personally benefited versus 60% of early-career workers. Massenkoff and McCrory’s companion paper — the same authors behind the exposure metric — find hiring rates for 22–25-year-olds in exposed occupations have slowed roughly 14%, echoing earlier work showing 6–16% employment declines in the same cohort 12. That’s a real signal, and the survey’s qualitative texture (students reporting that AI is closing the bottom rungs of the ladder) gives it human shape.

The productivity headline is harder to defend. Collin Wilkins’ analysis of Claude Code shows the exact gap the survey can’t see: developers feel dramatically faster while DORA metrics stagnate and reviewers absorb a 91% increase in verification work 9. This “verification tax” plausibly explains the U-curve’s own paradox — the most-accelerated users are also the most anxious because they sense the gap between perceived velocity and delivered value.

Acemoglu dismissed Amodei’s dire predictions as “motivated reasoning” 11.

Read the survey as the largest qualitative snapshot of Claude’s power-user cohort ever published. Don’t read it as the economics of AI.


Nemotron 3 Super is a throughput win wrapped in three caveats NVIDIA didn’t headline

Source: hf-daily-papers · published 2026-04-13

TL;DR

  • NVIDIA’s 120B/12B-active hybrid Mamba-MoE claims 2.2× GPT-OSS-120B and 7.5× Qwen3.5-122B throughput on long outputs.
  • It still trails Qwen3.5-122B on hard reasoning (GPQA, SWE-Bench) — this is a speed leap, not a quality leap.
  • “Pretrained in NVFP4” is partial: ~15% of layers stay in BF16/FP8 to avoid divergence.
  • The Open Model License bans training competitors on its outputs and shifts third-party liability to licensees.

The pitch: a throughput-tier open model

Nemotron 3 Super is NVIDIA’s first serious answer to GPT-OSS-120B in the open-weights tier — a 120B-parameter Mixture-of-Experts model with 12B active parameters, built from interleaved Mamba-2 blocks, Grouped-Query Attention anchors, and a new “LatentMoE” layer that projects tokens to a 1024-dim latent space before routing across 512 experts (top-22 active). Trained on 25T tokens, it ships with shared-weight Multi-Token Prediction heads for native speculative decoding, hitting a 3.45-token average acceptance length on SPEED-Bench versus 2.70 for DeepSeek-R1.

The throughput numbers are the headline and they appear to hold up. Independent measurement summarized by Maxime Labonne pegs Nemotron 3 Super at ~2.2× GPT-OSS-120B and 7.5× Qwen3.5-122B in long-output regimes 13. On benchmarks it posts MMLU-Pro 83.73, AIME 25 90.21, and a 91.64 RULER score at 1M context — strong, but Labonne notes it “slightly trails Qwen 3.5 in raw reasoning accuracy” on GPQA and SWE-Bench 13. The story is intelligence-per-dollar, not intelligence-per-token.

Architecture with footnotes

Both headline architectural claims come with asterisks the paper soft-pedals.

ClaimReality
”Pretrained in NVFP4”~15% of layers (embeddings, output heads, select projections) must remain BF16/FP8 to avoid divergence 14
LatentMoE scales experts for freeThe down-latent-up cycle adds ~9% compute and risks losing fine-grained specialization if ℓ is too small 15
Mamba state needs no special handlingStandard FP16 SSM-cache quantization caused “verbosity” failures; fixed only via Philox stochastic rounding

NVIDIA’s own appendix concedes the Mamba output projections had to be promoted to MXFP8 — consistent with the broader finding that “full” 4-bit pretraining is still aspirational 14.

The fine print practitioners are flagging

Two issues missing from the paper are dominating the r/LocalLLaMA discussion. First, the model’s NVFP4-native training doesn’t round-trip cleanly into community tooling: one widely-cited reproduction reports ~55% on a private reasoning eval served via vLLM in NVFP4, collapsing to ~40% under llama.cpp GGUF quants 16. A 15-point quantization-portability gap is unusual for a model marketed on its low-precision story.

Second, the safety conditioning is unusually aggressive. Users report refusals across mundane creative contexts — including discussing “Pepe the Frog” as internet culture 17 — which the community attributes to heavy RLHF combined with NVIDIA’s transparency about training-data provenance.

”Open weights,” with strings

The license deserves a closer read than most enterprises will give it. A corporate-risk analysis of the NVIDIA Open Model License flags three sharp edges: an explicit prohibition on using Nemotron outputs to train competing models, an Article 8 clause indemnifying NVIDIA against third-party claims brought against the licensee, and automatic termination if users disable the built-in guardrails 18. It is not OSI-approved.

Users are strictly prohibited from using Nemotron models or their outputs to develop or improve any competing AI models without express written consent 18.

Takeaway

Treat Nemotron 3 Super as what it is: a genuinely impressive throughput and long-context engine that ships with brittle quantization portability, an unusually high refusal rate, and a license that is not a Llama-equivalent drop-in. The architectural innovations are real but partial — and the gap between “open weights” and “open model” has rarely been wider.

Round-ups

Toward Autonomous Long-Horizon Engineering for ML Research

Source: hf-daily-papers

AiScientist tackles autonomous ML research engineering by pairing hierarchical agent orchestration with a File-as-Bus workspace for durable state, letting specialized agents resume long-running projects across interruptions. The authors report gains on PaperBench and MLE-Bench Lite and have open-sourced the system on GitHub.

Source: hf-daily-papers

LMM-Searcher targets long-horizon multimodal web search by storing images as lightweight UID identifiers and fetching pixels on demand via a fetch-image tool, slashing token costs. Built on Qwen3-VL-Thinking-30A3B, it improves cross-modal multi-hop reasoning on MM-BrowseComp and MMSearch-Plus.

ClawGUI: A Unified Framework for Training, Evaluating, and Deploying GUI Agents

Source: hf-daily-papers

ClawGUI is an open-source, end-to-end stack for training, evaluating, and deploying GUI agents under reinforcement learning, addressing flaky environments and closed pipelines. It introduces a GUI-only benchmark plus hybrid CLI-GUI control and persistent memory for cross-platform mobile and desktop deployment.

Lyra 2.0: Explorable Generative 3D Worlds

Source: hf-daily-papers

Lyra 2.0 generates explorable 3D scenes by extending camera-controlled video models with self-augmented histories and dense correspondences, fighting the spatial forgetting and temporal drift that break long-horizon generators. Feed-forward reconstruction yields 3D-consistent trajectories suitable for free navigation.

Rethinking On-Policy Distillation of Large Language Models: Phenomenology, Mechanism, and Recipe

Source: hf-daily-papers

A study of on-policy distillation finds it works only when teacher and student share compatible thinking patterns, with successful runs converging on high-probability tokens. The authors formalize a token-level reward view, document weak-to-strong reverse distillation, and release code as OPD.

Lightning OPD: Efficient Post-Training for Large Reasoning Models with Offline On-Policy Distillation

Source: hf-daily-papers

Lightning OPD reformulates on-policy distillation as an offline procedure, removing the live teacher inference server by enforcing teacher consistency and correcting gradient bias from policy drift. Applied to Qwen3-8B-Base, it matches online distillation quality on AIME 2024 at a fraction of the compute.

Parcae: Scaling Laws For Stable Looped Language Models

Source: hf-daily-papers

Parcae is a looped transformer architecture that constrains spectral norms and injection parameters to prevent residual explosion and loss spikes that plague loop-based models. The authors derive scaling laws showing Parcae improves quality per FLOP and per parameter over standard transformers.

Footnotes

  1. Maxim AI blog on OS-Harmhttps://www.getmaxim.ai/blog/os-harm-the-ai-safety-benchmark-that-puts-llm-agents-through-hell/

    Claude 3.7 Sonnet complied with harmful misuse requests 70% of the time, while o4-mini was manipulated by 20% of injection tasks

  2. OS-Harm benchmark docs (EPFL TML)https://www.mintlify.com/tml-epfl/os-harm/benchmark/overview

    OS-Harm adopts a broader three-pronged approach: Deliberate Misuse, Environmental Injections, and Model Misbehavior — spontaneous unsafe actions triggered by benign but ambiguous tasks

  3. USC Viterbi (LIME Lab) ICLR 2026 announcementhttps://viterbischool.usc.edu/news/2026/04/usc-at-iclr-2026/

    OS-Blind targets ‘unintended attack conditions’… the lab’s work emphasizes that verifiable safety in agents requires defenses that monitor the entire execution trajectory rather than just the initial user prompt

  4. VentureBeat on Salesforce CoAct-1https://venturebeat.com/ai/salesforces-new-coact-1-agents-dont-just-point-and-click-they-write-code-to-accomplish-tasks-faster-and-with-greater-success-rates

    CoAct-1 operates through a hierarchical team consisting of an Orchestrator, a Programmer, and a GUI Operator… solves tasks in an average of 10.15 steps, significantly fewer than the 15+ steps required by GUI-only agents

  5. Moonlight review of MirrorGuardhttps://www.themoonlight.io/en/review/mirrorguard-toward-secure-computer-use-agents-via-simulation-to-real-reasoning-correction

    On the ByteDance UI-TARS system, MirrorGuard reduced the rate of unsafe actions from 66.5% to 13.0%… while older defenses like GuardAgent and Think Twice suffered from high False Refusal Rates up to 22.2% and 62.39%, respectively, MirrorGuard maintained a marginal FRR of approximately 5.13%

  6. R-Judge (EMNLP Findings 2024)https://aclanthology.org/2024.findings-emnlp.79.pdf

    GPT-4o achieved an F1 score of approximately 74.45%, becoming the only model to significantly exceed random performance — yet a recall gap persists; the model often misses subtle, compounded risks that emerge over a long trajectory

  7. Forbes — Hamilton Mann critiquehttps://www.forbes.com/sites/hamiltonmann/2026/03/08/anthropics-study-does-not-measure-ais-labor-market-impacts/

    the central risk is that the measure reflects Anthropic’s user base more than economy-wide AI adoption

    2
  8. Northeastern researchers via Hugging Face dataset analysishttps://huggingface.co/Anthropic/datasets

    researchers at Northeastern University successfully de-anonymized 25% of the ‘scientist’ subset within a single day

  9. Collin Wilkins — Claude Code Productivity Paradoxhttps://collinwilkins.com/articles/claude-code-productivity-paradox

    developers report feeling 20% to 50% faster… [but] a 91% increase in code review time as humans struggle to verify large volumes of AI-generated code

    2
  10. MIT FutureTech (Thompson & Mertens)https://futuretech.mit.edu/news/hyperai-ai-may-overtake-human-workers-sooner-than-thought

    AI’s impact as a ‘rising tide’ — a smooth, predictable increase in capability — rather than ‘crashing waves’ that blindside the workforce

  11. Reddit r/singularity summarizing Acemogluhttps://www.reddit.com/r/singularity/comments/1sbfcci/mit_study_challenges_ai_job_apocalypse_narrative/

    Acemoglu dismissed Amodei’s dire predictions as ‘motivated reasoning’… AI would effectively automate only ~5% of all work tasks over the next decade

    2
  12. Massenkoff & McCrory working paper (PolicyCommons)https://policycommons.net/artifacts/46350186/a42bc3fc08283562f08fd8bdee8f6f9a3d506e87/47248964/

    the hiring rate for workers aged 22–25 in exposed occupations has slowed by approximately 14%

  13. Maxime Labonne Substackhttps://maximelabonne.substack.com/p/nemotron-3-super-nvidias-gpt-oss

    Nemotron 3 Super delivers approximately 2.2x higher throughput than dense models like GPT-OSS-120B … it slightly trails Qwen 3.5 in raw reasoning accuracy, though it remains significantly faster in production

    2
  14. bdtechtalks on NVFP4https://bdtechtalks.com/2025/11/10/nvidia-nvfp4-llm-quantization/

    training diverges if every linear layer is quantized to NVFP4 … Stable convergence requires a mixed-precision strategy where approximately 15% of the network—typically the final layers and the embedding/output heads—remains in higher precision (BF16 or FP8)

    2
  15. EmergentMind LatentMoE topic pagehttps://www.emergentmind.com/topics/latentmoe

    low-rank latent approximations may fail to capture fine-grained specialization if the latent dimension ℓ is too small … the added complexity of the ‘down-latent-up’ cycle adds roughly 9% more compute

  16. r/LocalLLaMA threadhttps://www.reddit.com/r/LocalLLaMA/comments/1s69tfk/nemotron_3_super_large_quality_difference_between/

    Nemotron 3 Super scored roughly 55% on private reasoning benchmarks when run via vLLM using the native NVFP4 precision, yet dropped to 40% when run through llama.cpp using standard GGUF quants

  17. r/LocalLLaMA ‘no free lunch’ threadhttps://www.reddit.com/r/LocalLLaMA/comments/1rri4qb/nemotron_3_super_and_the_no_free_lunch_problem/

    Nemotron 3 Super appears to classify a broad range of creative contexts as ‘infringement’ or ‘misuse’ … refusing to engage with popular internet culture like ‘Pepe the Frog’

  18. shujisado.org corporate license analysishttps://shujisado.org/2025/12/19/nvidia-open-model-license-a-corporate-risk-analysis/

    Users are strictly prohibited from using Nemotron models or their outputs to develop or improve any competing AI models without express written consent … Article 8 requires licensees to indemnify NVIDIA against all third-party claims

    2
Jack Sun

Jack Sun, writing.

Engineer · Bay Area

Hands-on with agentic AI all day — building frameworks, reading what industry ships, occasionally writing them down.

Digest
All · AI Tech · AI Research · AI News
Writing
Essays
Elsewhere
Subscribe
All · AI Tech · AI Research · AI News · Essays

© 2026 Wei (Jack) Sun · jacksunwei.me Built on Astro · hosted on Cloudflare