§ feed · cluster
Zyphra Releases ZAYA1-8B: A Reasoning MoE Trained on AMD Hardware That Punches Far Above Its Weight Class
Zyphra released ZAYA1-8B, a small MoE model with 760M active parameters trained on AMD hardware, outperforming larger models on math and coding benchmarks, available under Apache 2.0.
§ sources1 publication · timeline below