Paper review: Not quite the AlphaGo moment yet

ASI-ARCH presents a fully autonomous LLM-driven pipeline that reports discovering 106 "state-of-the-art" linear-attention architectures and frames this as an "AlphaGo-like" leap.


Summary

ASI-ARCH presents a fully autonomous LLM-driven pipeline that reports discovering 106 “state-of-the-art” linear-attention architectures and frames this as an “AlphaGo-like” leap with a scaling law for discovery oai_citation:0‡arXiv.
Table 1 shows only 1-3-point gains over DeltaNet and Mamba2 at 340 M parameters and provides no confidence intervals or efficiency data oai_citation:1‡arXiv.
Key modules (novelty filter, LLM-as-Judge, cognition base) still depend on a hand-curated set of ≈100 prior papers and carefully engineered prompts, so the process is not fully “human-free” oai_citation:2‡GitHub.
Overall, the study is an interesting automation prototype, but the evidence falls short of an AlphaGo-scale breakthrough.

Claimed contributions vs. documented evidence

Metrics & evaluation design

Pipeline robustness

External perspective

Verdict & recommendations

  1. Interesting engineering - LLM-centric automation pipeline is worth replicating.
  2. Claims overstated - Results are narrow, mid-scale, and lack statistical depth.
  3. Future validation
    • Add Transformer, Performer, and SSM baselines at ≥1 B parameters.
    • Report variance, significance, and compute/energy per model.
    • Benchmark inference speed vs. sequence length.
    • Ablate each agent (Judge, cognition base, etc.) to measure contribution.

© Tobias Klein 2025 · All rights reserved
LinkedIn: https://www.linkedin.com/in/deep-learning-mastery/