MacMind
A working backprop transformer built in HyperCard on a 1989 Mac SE/30 with 4 MB RAM
MacMind is a complete single-layer transformer — attention, positional encoding, backpropagation, and weight updates — implemented entirely in HyperTalk, the scripting language built into Apple HyperCard, running on a Mac SE/30 with an 8 MHz processor and 4 MB of RAM. It trains to learn the bit-reversal permutation fundamental to the Fast Fourier Transform, and in doing so, the attention mechanism independently discovers the Cooley-Tukey butterfly routing pattern — not because it was designed in, but because the gradient descent finds it. Every operation is visible and editable in HyperCard's stack interface. Weights persist between sessions in card fields. The project is a deliberate demonstration that the mathematical operations underlying modern AI — matrix multiplication, softmax, cross-entropy, backprop — are substrate-independent: they work identically on hardware from 1989 as on an H100 cluster today, just much slower. The HN thread was warmly received as a genuine educational artifact: seeing attention, positional encoding, and gradient descent laid bare in HyperTalk's English-like syntax strips away 35 years of abstraction and reveals what transformers actually are. For educators, students, and curious engineers, MacMind is an unusually effective explanation tool.
Panel Reviews
The Builder
Developer Perspective
“Every engineer who works on LLMs should read this code. HyperTalk's readable syntax forces you to confront what's actually happening in a forward pass — there's no PyTorch autograd magic to hide behind. The fact that attention discovers the FFT butterfly on its own is a genuinely beautiful result worth the price of admission alone.”
The Skeptic
Reality Check
“This is a teaching toy, not a tool — calling it 'ship' in a practical sense is misleading. The SE/30 trains a trivial task in an hour that PyTorch does in milliseconds. The intellectual point is valid but if you're looking for something to put in a workflow, look elsewhere.”
The Futurist
Big Picture
“The timing is significant: as AI systems become increasingly opaque and proprietary, projects like MacMind go in the opposite direction — maximally transparent, maximally accessible. Demystification at this level has real cultural value. The next generation of AI researchers may be inspired by seeing a transformer in HyperTalk before they see one in PyTorch.”
The Creator
Content & Design
“As someone who uses AI tools daily without fully understanding them, MacMind made me genuinely understand what attention is doing for the first time. Clicking through the HyperCard stack and watching weights update in real time is a better explainer than any Medium article. This belongs in every AI literacy curriculum.”
Community Sentiment
“Attention discovering the FFT butterfly pattern independently”
“Nostalgia for HyperCard and educational value for beginners”
“Substrate independence of neural network math”