Do you want to understand AI at its most fundamental level?
This program takes you back to the 1950s, where artificial intelligence was born β not inside a computer, but in analog hardware. The Mark I Perceptron was a physical machine made of wires, switches, and resistors that could learn. AI didn't start as software. It started as a circuit.
This simulator lets you operate that machine. Flip toggle switches, turn weight dials, and watch a neural network learn in real time. But it goes further than simulation: it gives you everything you need to build the real thing. Train any network, and the program generates a complete circuit schematic, an interactive breadboard layout, and a bill of materials with real part numbers you can order today. You can solder together a working perceptron on your desk.
The included Operator's Manual is styled as a declassified 1950s government document and covers the real history, the math, and step-by-step build instructions. Video tutorials walk you through it all.
Every line of this program was written by AI. And embedded inside it β in the Operator's Manual β is a complete copy of itself, written in plain English: the prompt that was used to generate the entire codebase. Feed that prompt back into an AI, and it rebuilds the program from scratch. This application carries an embryonic copy of itself within itself.
An AI program about AI, written by AI, that contains the instructions to recreate itself.
Keyboard Shortcuts: When you click on a knob, use the up-down, left-right arrows.
"The Navy revealed the embryo of an electronic computer today that it expects will be able to walk, talk, see, write, reproduce itself, and be conscious of its existence." β New York Times, July 1958
A topological look at the hardware configuration.
In 1958, psychologist Frank Rosenblatt at Cornell built the Mark I Perceptron β a room-sized machine that could learn to recognize patterns without being explicitly programmed. It was the first artificial neural network, and it terrified and amazed the world in equal measure.
This simulator recreates that historic machine with a faithful vintage industrial aesthetic: toggle switches that click, rotary dials that turn, analog meters that swing, and backlit buttons that glow. It's not just a simulation β it's a time machine.
ποΈ Flip switches. Turn dials. Watch a machine learn. ποΈ
| If you're interested in... | This project offers... |
|---|---|
| AI/ML fundamentals | Hands-on understanding of how neural networks actually learn β no black boxes |
| History of computing | A window into the 1958 breakthrough that started the AI revolution |
| Beautiful UI design | 15 custom-drawn controls with metallic textures, glow effects, and vintage aesthetics |
| AI-assisted development | A case study in building complex apps through natural language prompts |
Every control is custom-rendered with GDI+ graphics:
- Toggle switches with LED indicators
- Rotary knobs with precise dial control (-30 to +30)
- Analog meter showing real-time output
- Mechanical push buttons with backlit glow effects
- Brushed metal plates with engraved text and corner screws
This isn't a toy β it implements Rosenblatt's actual learning algorithm:
OUTPUT = Ξ£(Switch Γ Weight) + Bias
Train it to recognize "T" shapes vs "J" shapes. Watch the weights converge. Feel the math working.
Turn the Math Dial to switch between seven different learning algorithms:
- 1958 (1 o'clock): Original Rosenblatt perceptron (1-to-1 connectivity)
- 1958+ (2 o'clock): Fully connected, sum rule
- 1958m (3 o'clock): Fully connected, average rule
- 1958/+ (4 o'clock): Input-normalized, sum rule
- 1958/m (5 o'clock): Input-normalized, average rule
- 1960 (9 o'clock): Widrow-Hoff / LMS / Delta rule
- 1986 (10 o'clock): Multi-layer perceptron with ReLU + backpropagation
- Experience the evolution of neural network math from Rosenblatt to Rumelhart!
A complete operator's manual styled like a declassified 1950s government document:
- Original 1958 New York Times press release
- Step-by-step operating procedures
- The algorithm explained
- "DECLASSIFIED" watermark for authenticity
- Chapter X: Build It Yourself β Complete hardware build guide with:
- Professional circuit schematic with op-amp symbols
- Interactive breadboard layout with hover tooltips
- Bill of Materials with real part numbers and prices
- Export BOM to CSV for ordering
Click BRAIN to see "THE BRAIN" β an interactive network topology view:
- Input nodes light up based on switch states β click them to toggle switches!
- Connection weights shown on each line (red for active +1 inputs)
- Click hidden nodes then use arrow keys to adjust weights directly
- Watch signals flow through the network as you adjust dials
Toggle the TELETYPE switch to open a 1950s-style teletype printer window:
- Striped dot-matrix printer paper with alternating cream/blue bars
- Perforated tractor-feed holes along both edges
- Embedded Special Elite typewriter font for authentic output
- Real-time debug logging to
%AppData%\Mark1\Log.txt - Adjustable font size with +/- buttons
Adjust the grid from 1Γ1 to 10Γ10 β the entire UI dynamically scales. At 10Γ10, you have 100 switches and 100 dials working in concert.
At grid size 1, keep turning the knob down to enter linear mode:
- Add nodes one at a time (2, 3, 4... up to 25)
- Nodes wrap at 5 per row β create any configuration you want
- Perfect for experimenting without the constraints of square grids
The circuit has been professionally audited and verified:
- β Schematic validated against electronics principles
- β Breadboard verified to match schematic exactly
- β KiCad export tested with complete netlist
- β Buildable and will function correctly when built
- See CIRCUIT_AUDIT.md for full validation report
The app includes complete hardware build documentation with PROFESSIONAL-GRADE tools:
- Circuit Schematic: Professional schematic with gap-free connections, calculated resistor values
- KiCad Export: Export to .kicad_sch format for professional EDA editing
- SPICE Simulation: Verify circuit operation in KiCad/ngspice
- Interactive Breadboard: Electrically-accurate breadboard layout with:
- Hole-based placement: Components at specific (row, column) coordinates
- Electrical connectivity modeling: Holes represent actual breadboard connections
- Power rails in dense clusters (groups of 5 with 1-space gaps)
- Scalable view: Shows 4 inputs in detail, notes pattern for additional inputs
- Row/column numbering (A-Z, 5/10/15...) for precise placement
- Thick, visible wiring (4-5px) with color-coding
- Power supply selector: Choose off-board, on-board battery, USB 5V, or bench supply
- Hover tooltips showing component specs and exact hole positions
- Bill of Materials: Complete BOM with real manufacturer part numbers, prices, and suppliers
- Export to CSV: One-click export for ordering from DigiKey/Mouser
- Total kit cost: ~$15-25 for basic 4-input perceptron
This program contains its own DNA.
Embedded in the application is prompt.txt β the complete English-language instructions used to generate this entire codebase. Feed it into Claude Code, and you can recreate the application from scratch.
I predict, one day, we won't share code β we will share prompts.
And so I am starting the trend today.
This is executable English: software that reproduces itself through natural language.
Download the latest release from the Releases page and run PerceptronSimulator.exe.
git clone https://github.com/srives/Perceptron.git
cd Perceptron
dotnet build
dotnet run- Install Claude Code
- Copy
resources/perceptrons.pngtoc:\temp\ - Feed
resources/complete_prompt.txtinto Claude Code - Watch it build the entire application from English
- Create a pattern β Click switches to form a shape (try a "T")
- Press Learn+ β Tell the machine this pattern should output POSITIVE
- Create another pattern β Clear and make a "J" shape
- Press Learn- β Tell the machine this pattern should output NEGATIVE
- Test it β The meter and LED now correctly classify your patterns!
The machine learns by adjusting dial weights. After a few examples, it generalizes to patterns it's never seen.
This project was built entirely through conversation with Claude Code AI as an homage to the pioneers of neural networks. No code was written by hand β every line emerged from natural language prompts.
What took Frank Rosenblatt years of research and a room full of hardware, I recreated in a weekend through conversation with an AI. The irony isn't lost on me: using a descendant of the perceptron to rebuild its ancestor.
This program was written using Claude Code. It is an AI program that teaches the foundations of AI, written with AI, for AI and the ability to remake itself with built-in AI prompts. | Efficiency Multiplier | 8-12Γ |
The full simulator with toggle switches, rotary weight knobs, analog meter, and vintage control panel.
After training, the perceptron has learned weights and the output LED lights up for recognized patterns.
A declassified 1950s-style manual with redacted chapters, operating procedures, and build instructions.
Professional schematic of the analog summing circuit with LM358 op-amp, SPDT switches, and calculated resistor values. Exportable to KiCad.
Interactive, electrically-accurate breadboard view with hole-based component placement, color-coded wiring, and hover tooltips.
- Frank Rosenblatt (1928-1971) β For inventing the perceptron and starting the AI revolution
- Welch Labs β For the excellent educational video that inspired this project
- Anthropic's Claude β For being an exceptional coding partner
- Steve Rives, AI Engineer
MIT License β Use it, learn from it, share it.
"The perceptron is the embryo of an electronic computer that will be able to walk, talk, see, write, reproduce itself, and be conscious of its existence."
β Dr. Frank Rosenblatt, 1958
He wasn't entirely wrong. It just took 67 years.
β Star this repo if you believe the future of software is natural language β





