All Reviews (12)
Qwen3.6-35B-A3B
35B MoE model with only 3B active params that beats models 10× its inference size
GLM-5.1
The first open-source model to beat GPT-5.4 and Claude Opus on real-world coding
Meta Llama 4
Open-weight multimodal MoE models with 10M context — free to run
LFM2.5-VL
450M vision-language model that runs in under 250ms on edge hardware
GLM-5.1
#1 on SWE-Bench Pro — Zhipu's open 754B MoE beats GPT-5 on coding
Bonsai-8B
First commercially usable 1-bit LLM: 8B capabilities in 1.15 GB of RAM
Darwin-4B-David
4.5B merged model beats Gemma-4-31B on GPQA — no training needed
OmniVoice
Zero-shot TTS for 600+ languages — voice cloning at 40x real-time speed
Kimi K2.5
Open-weight multimodal model with 100-agent swarm mode and 256K context
GLM-5.1
#1 on SWE-Bench Pro — 744B MoE model that runs autonomously for 8 hours
PrismML (1-Bit Bonsai)
Commercially viable 1-bit LLMs that run on almost any hardware
Qwen3.6-Plus
The agentic coding model beating Claude Opus 4.5 — free on OpenRouter