ArtoriaZero

A decoder-only transformer that plays chess through pure pattern recognition — no search, no MCTS.

Implementation inspired by “Grandmaster-Level Chess Without Search” (Ruoss et al., 2024) — not the original paper's model

No Search

Pure neural network policy — no MCTS, no Alpha-Beta, no search tree.

Transformer Architecture

LLaMA-style decoder with RMSNorm, SwiGLU, and bidirectional attention.

Behavioral Cloning

Trained to imitate strong players directly from millions of chess games.

Dual Head

Policy head for move prediction + Value head for position evaluation.

Multiple Scales

Small (19M), Mid (100M), and Large (500M) parameter variants.

Instant Inference

Single forward pass per move — no thinking time, no depth limits.

Model Variants

Variantd_modelLayersHeadsParams
Small25688~19M
Mid512168~100M
Large10244032~500M