Paper review: Not quite the AlphaGo moment yet

ASI‑ARCH presents a fully autonomous LLM‑driven pipeline that reports discovering 106 "state‑of‑the‑art" linear‑attention architectures and frames this as an "AlphaGo‑like" leap.


Summary

ASI‑ARCH presents a fully autonomous LLM‑driven pipeline that reports discovering 106 “state‑of‑the‑art” linear‑attention architectures and frames this as an “AlphaGo‑like” leap with a scaling law for discovery  oai_citation:0‡arXiv.
Table 1 shows only 1–3‑point gains over DeltaNet and Mamba2 at 340 M parameters and provides no confidence intervals or efficiency data  oai_citation:1‡arXiv.
Key modules (novelty filter, LLM‑as‑Judge, cognition base) still depend on a hand‑curated set of ≈100 prior papers and carefully engineered prompts, so the process is not fully “human‑free”  oai_citation:2‡GitHub.
Overall, the study is an interesting automation prototype, but the evidence falls short of an AlphaGo‑scale breakthrough.

Claimed contributions vs. documented evidence

Metrics & evaluation design

Pipeline robustness

External perspective

Verdict & recommendations

  1. Interesting engineering – LLM‑centric automation pipeline is worth replicating.
  2. Claims overstated – Results are narrow, mid‑scale, and lack statistical depth.
  3. Future validation
    • Add Transformer, Performer, and SSM baselines at ≥1 B parameters.
    • Report variance, significance, and compute/energy per model.
    • Benchmark inference speed vs. sequence length.
    • Ablate each agent (Judge, cognition base, etc.) to measure contribution.