Back to reviews
MacMind

MacMind

A working backprop transformer built in HyperCard on a 1989 Mac SE/30 with 4 MB RAM

MacMind is a complete single-layer transformer — attention, positional encoding, backpropagation, and weight updates — implemented entirely in HyperTalk, the scripting language built into Apple HyperCard, running on a Mac SE/30 with an 8 MHz processor and 4 MB of RAM. It trains to learn the bit-reversal permutation fundamental to the Fast Fourier Transform, and in doing so, the attention mechanism independently discovers the Cooley-Tukey butterfly routing pattern — not because it was designed in, but because the gradient descent finds it. Every operation is visible and editable in HyperCard's stack interface. Weights persist between sessions in card fields. The project is a deliberate demonstration that the mathematical operations underlying modern AI — matrix multiplication, softmax, cross-entropy, backprop — are substrate-independent: they work identically on hardware from 1989 as on an H100 cluster today, just much slower. The HN thread was warmly received as a genuine educational artifact: seeing attention, positional encoding, and gradient descent laid bare in HyperTalk's English-like syntax strips away 35 years of abstraction and reveals what transformers actually are. For educators, students, and curious engineers, MacMind is an unusually effective explanation tool.

Panel Reviews

The Builder

The Builder

Developer Perspective

Ship

Every engineer who works on LLMs should read this code. HyperTalk's readable syntax forces you to confront what's actually happening in a forward pass — there's no PyTorch autograd magic to hide behind. The fact that attention discovers the FFT butterfly on its own is a genuinely beautiful result worth the price of admission alone.

The Skeptic

The Skeptic

Reality Check

Skip

This is a teaching toy, not a tool — calling it 'ship' in a practical sense is misleading. The SE/30 trains a trivial task in an hour that PyTorch does in milliseconds. The intellectual point is valid but if you're looking for something to put in a workflow, look elsewhere.

The Futurist

The Futurist

Big Picture

Ship

The timing is significant: as AI systems become increasingly opaque and proprietary, projects like MacMind go in the opposite direction — maximally transparent, maximally accessible. Demystification at this level has real cultural value. The next generation of AI researchers may be inspired by seeing a transformer in HyperTalk before they see one in PyTorch.

The Creator

The Creator

Content & Design

Ship

As someone who uses AI tools daily without fully understanding them, MacMind made me genuinely understand what attention is doing for the first time. Clicking through the HyperCard stack and watching weights update in real time is a better explainer than any Medium article. This belongs in every AI literacy curriculum.

Community Sentiment

Overall189 mentions
82% positive14% neutral4% negative
Hacker News34 mentions
88%10%2%

Attention discovering the FFT butterfly pattern independently

Reddit60 mentions
82%14%4%

Nostalgia for HyperCard and educational value for beginners

Twitter/X95 mentions
80%16%4%

Substrate independence of neural network math