TL;DR: On a Friday afternoon at Stanford in 1959, Bernard Widrow and his grad student Ted Hoff came within one substitution of inventing modern backpropagation — 27 years early. They missed because their neurons used a binary step function whose derivative killed the gradient. Swap the step for a sigmoid and you have deep learning. Nobody made that swap until 1986. Ted Hoff, meanwhile, left Stanford and built the microprocessor. The man who almost invented backprop built the hardware that now runs it. Every breakthrough is one substitution away from the previous dead end. Build at the substitution →
What Is the Widrow-Hoff LMS Algorithm?
The Bronx Science piece told the story of how Frank Rosenblatt's perceptron rose, fell, and eventually returned to eat the world. Most histories stop there. They go from Rosenblatt's 1957 demo to Minsky's 1969 book to the 1986 backpropagation paper, treating the intervening 17 years as wilderness.
The wilderness wasn't empty.
In 1959, at Stanford — not at MIT, not at Cornell, not in Rosenblatt's lab — two engineers sat down on a Friday afternoon and almost invented the algorithm that powers every modern AI system.
They missed by one substitution.
That substitution — a single change to one component of their system — would not be made for another 27 years. The story of that near-miss is the most instructive "what if" in the history of computing. It's also the story of how the guy who came closest to backpropagation walked out of Stanford and built the hardware that would eventually run it.
This is that story.
The Two Men
Bernard Widrow was a 30-year-old assistant professor of electrical engineering at Stanford in 1959. He had finished his PhD at MIT four years earlier and was already building a reputation in adaptive signal processing — the mathematics of systems that adjust themselves based on feedback. Widrow cared about control theory, noise cancellation, and the emerging question of whether circuits could learn.
Marcian Edward Hoff — everybody called him Ted — was Widrow's first graduate student. He had arrived at Stanford from RPI in 1958 with a freshly minted bachelor's degree and an interest in solid-state electronics. He was 21 years old.
Widrow had just read about Rosenblatt's perceptron in The New York Times. The article was breathless — electronic brains, machines that would walk and talk and reproduce. Widrow read past the hype and saw the actual mathematics. The perceptron was using a heuristic learning rule: nudge the weights in the direction of the error, with a fixed step size. It worked, but it was ad hoc. There was no underlying optimization principle.
Widrow thought: what if we derived the learning rule from calculus?
The difference looks small. It wasn't.
The Insight
Rosenblatt's learning rule was: if the network got the answer wrong, push each weight in the direction that would have given a better answer. It worked because it was, in spirit, a discrete approximation to gradient descent. But it was a recipe, not a derivation.
Widrow and Hoff went to the whiteboard and did the derivation.
They defined an error function: the squared difference between the network's output and the desired output. They took the derivative of this error function with respect to each weight. They used the chain rule. They got a formula for how each weight should change to minimize the error most efficiently.
The result, which they would later call the LMS algorithm — for Least Mean Squares — is elegant:
$$
w_{i}(t+1) = w_{i}(t) + \eta \cdot (d - y) \cdot x_{i}
$$
Where $w_i$ is the weight for input $i$, $\eta$ is the learning rate, $d$ is the desired output, $y$ is the actual output, and $x_i$ is the input. The term $(d - y)$ is the error. The rule says: move each weight proportionally to the error times its input.
Compare this to the modern delta rule used in backpropagation:
$$
\Delta w_{ij} = \eta \cdot \delta_{j} \cdot x_{i}
$$
They are essentially the same equation. Widrow would later remark, with justified pride: "You don't have to square anything or compute the actual error. The power of that compared to earlier methods is just fantastic."
In 1959, on that Friday afternoon, Widrow and Hoff had written down a gradient-descent learning rule that — with one substitution — is the learning rule that trains GPT-5, Claude, and every other frontier model of 2026.
They built a hardware implementation and called it ADALINE — Adaptive Linear Neuron. It was a physical device, a box full of transistors and potentiometers representing adjustable weights, that could be trained to recognize patterns. It worked beautifully. By 1961, Widrow had deployed a variant to do adaptive noise cancellation for telephone lines — one of the first real commercial deployments of a learning system.
And then they tried to stack them.
The Wall
Single-layer ADALINE could do a lot. But it had the same mathematical limits as Rosenblatt's perceptron — no single-layer network can compute XOR, as Minsky would famously prove in 1969. The obvious next move was to stack multiple layers: feed ADALINE's output into another ADALINE, creating a network with depth.
Widrow and Hoff tried. They couldn't get it to work.
The LMS algorithm was beautiful for a single layer. Extended to multiple layers, it collapsed. The gradient calculated at the output wouldn't propagate backward through the network. Somewhere between the output and the earlier layers, the signal died.
They didn't know why.
Neither did anyone else. This is the part of the story that's worth sitting with. The mathematical framework was correct. The optimization target was correct. The learning rule for a single layer was exactly right. The only thing standing between 1959 and modern deep learning was a specific, identifiable component — and nobody could see it.
The component was the activation function.
The Wrong Switch
Every artificial neuron in 1959 used a binary step activation function. You summed the weighted inputs. If the sum was above a threshold, the neuron output 1. Otherwise it output 0. This was the canonical model, the McCulloch-Pitts neuron from 1943, the thing every perceptron and every ADALINE was built from.
The step function is a brick wall. Look at its shape:
Step Activation Function
output
───────────── ──────
│1.0 │
│ │
│ │
│ │
│ │
│0.0 ───────────────────┘
─────────────────────────────── input
▲
threshold
The derivative is zero everywhere the function is flat,
and undefined at the single jump point.
Calculus cannot flow through it.
For a single layer, the step function's lack of a derivative was fine — LMS didn't need it. LMS used the raw pre-activation sum for its gradient calculation and the step function only at the output for the final decision. The gradient didn't have to pass through the activation.
For multiple layers, this strategy failed. In a multi-layer network, the output of one layer is the input to the next, and the gradient has to flow backward through every activation function along the way. A step function's derivative is zero everywhere it's flat — which is everywhere except the single threshold point, where it's undefined. Multiply any signal by zero and it becomes zero. The gradient died at the first layer boundary.
The fix was obscenely simple. Replace the step function with a smooth curve that looks similar but has a non-zero derivative everywhere. The natural candidate is the sigmoid:
$$
\sigma(x) = \frac{1}{1 + e^{-x}}
$$
It's S-shaped. It goes from near-0 for very negative inputs to near-1 for very positive inputs, smoothly transitioning through the middle. It looks like a step function that somebody softened with a rolling pin.
Sigmoid Activation Function
output
───────────────────────────
│1.0 ╱─────────
│ ╱╱
│ ╱
│ ╱
│ ╱
│0.0 ───
─────────────────────────── input
The derivative is non-zero everywhere.
Calculus can flow through it.
Multi-layer training becomes possible.
Its derivative is beautiful: $\sigma'(x) = \sigma(x)(1 - \sigma(x))$. Never zero. Always differentiable. Gradients can flow through it in both directions.
That's it. That's the substitution. Swap the step function for the sigmoid and the LMS algorithm generalizes to multiple layers — which is essentially what backpropagation is.
Nobody made that swap until 1986.
The 27-Year Gap
An ASCII view of the gap, with the substrate (hardware + networking + data) catching up in parallel:
1959 ────●── LMS algorithm @ Stanford (single layer works)
1960 ────│
1962 ────├── Engelbart publishes "Augmenting Human Intellect"
1965 ────│
1969 ────●── Minsky & Papert "Perceptrons" → AI winter begins
1971 ────●── Ted Hoff's 4004 microprocessor ships
1974 ────●── Werbos PhD thesis: essentially backprop (community ignores it)
1977 ────●── Apple II
1981 ────●── IBM PC
1982 ────●── Hopfield network (neural nets quietly revive)
1984 ────●── Apple Macintosh
1985 ────●── Boltzmann machine
1986 ────●── Rumelhart / Hinton / Williams: backprop + sigmoid ← THE SUBSTITUTION
│
│ ...the substrate finally catches up...
│
2012 ────●── AlexNet: GPUs + ImageNet
2017 ────●── Attention is All You Need: transformers
2022 ────●── ChatGPT
2025 ────●── Taskade Genesis: the execution layer
Twenty-seven years. Think about what existed in the world during those 27 years.
| Year | In the world | In neural networks |
|---|---|---|
| 1959 | LMS algorithm exists | Single-layer ADALINE works |
| 1962 | Engelbart publishes "Augmenting Human Intellect" | — |
| 1969 | Moon landing / ARPANET / Minsky's Perceptrons book | AI Winter begins |
| 1971 | Ted Hoff ships the Intel 4004 microprocessor | — |
| 1974 | Paul Werbos describes backprop in his Harvard PhD thesis | Largely ignored by the neural network community |
| 1977 | Apple II ships | — |
| 1981 | IBM PC ships | — |
| 1982 | Hopfield network published | Neural networks quietly revive |
| 1984 | Apple Macintosh ships | — |
| 1985 | Boltzmann machine | — |
| 1986 | Rumelhart, Hinton, Williams publish backpropagation | The substitution is finally made |
The infrastructure arrived. Personal computers. Networking. Moore's Law. Even the theoretical precursors — Werbos had published essentially the backprop algorithm in 1974. The community didn't pick it up. The AI winter had frozen out the people who would have.
The 1986 paper by David Rumelhart, Geoffrey Hinton, and Ronald Williams — "Learning representations by back-propagating errors" — cited Widrow and Hoff's LMS algorithm as the foundation. They extended it to multi-layer networks by doing exactly what Widrow and Hoff could not do in 1959: differentiating through the activation function. They used the sigmoid. The gradient flowed. Deep learning was born.
The Stanford Fork
The real twist of the story is what happened to Ted Hoff.
After finishing his PhD with Widrow in 1962, Hoff stayed at Stanford as a research associate. He kept working on adaptive systems, neural networks, and the edges of machine learning. By 1968 he had done good work but wasn't chasing the grand AI prize. When a new Silicon Valley semiconductor startup called Intel offered him a job, he took it. He was employee number 12.
In 1969, Intel took on a contract from a Japanese calculator company called Busicom. Busicom wanted Intel to design a set of custom chips for a new line of calculators — twelve different chips, each handling a specific function.
Ted Hoff looked at the spec and said: what if we built one general-purpose programmable chip instead?
The result was the Intel 4004, shipped in November 1971. It contained 2,300 transistors on a single die and was the first commercial microprocessor. Hoff's core insight was that computation could be decoupled from application: build a chip that executes instructions, and any application becomes a matter of writing the right instructions.
The 4004 led to the 8008, the 8080, the 8086, the Pentium, and every CPU you have ever touched. It led to the personal computer. It led to the smartphone. It led to the GPU. It led to the data centers that now train the transformers that run modern AI.
Ted Hoff's 4004 is the ancestor of the hardware that runs the algorithm Ted Hoff almost invented.
The man who came closest to backpropagation in 1959 left to build the microprocessor in 1971, and the microprocessor is what eventually made backpropagation useful in 2012 when Ilya Sutskever and the AlexNet team used a few thousand Nvidia GPUs to train a network that Widrow and Hoff could have trained on a whiteboard — if only they'd changed the activation function.
The irony is not accidental. It's constitutive. AI history is made of these almost-discoveries, and the people who almost made them often go on to build the infrastructure that enables the eventual discovery to matter.
What Widrow Said Later
Bernard Widrow is still alive as of 2026 — he's 97 and emeritus at Stanford. He has given a number of interviews and lectures reflecting on the period. The through-line is instructive.
He doesn't treat the 1959-to-1986 gap as a tragedy. He treats it as a structural feature of how research progresses. The mathematics was there. The insight about gradient-based learning was there. The hardware wasn't. The culture wasn't. The specific piece of mathematical furniture — the smooth activation — wasn't obvious until the community had been forced, by the 1970s wilderness, to think more carefully about what was blocking progress.
Widrow has also pointed out, correctly, that modern frontier models are essentially massive stacks of the same thing he was building in 1959. GPT-3 alone contains roughly 10 million artificial neurons across 96 layers. Each of those neurons is, at heart, a 1959 ADALINE — with a different activation function, trained with a multi-layer extension of his algorithm, on billions of times more data. GPT-4 is an order of magnitude larger. GPT-5 is larger still.
You would need 10 million ADALINE boxes, each with over 10,000 adjustable weights, to reproduce GPT-3 in 1960s hardware. The dream was correct. The implementation was one substitution and three decades of hardware progress away.
The Lesson for Builders
I am writing this post as the founder of an AI company that is, in its small way, trying to make a 58-year-old vision finally ship. The parallel to Widrow and Hoff is not lost on me.
The lesson from 1959 is not be patient, the world will come around. The world did not come around to Widrow and Hoff — it ignored them and had to rediscover their work from scratch decades later. The lesson is about the structure of stuck problems.
The "one missing substitution" pattern
| Field | What was right | What one component was wrong | Years until substituted |
|---|---|---|---|
| Neural networks (1959) | Gradient-based learning, error surface, weight update | Step activation function | 27 (sigmoid, 1986) |
| Rosenblatt's perceptron (1957) | Learning from data | Single-layer architecture | 29 (multi-layer + backprop, 1986) |
| Expert systems (1980s) | Symbolic reasoning | No common-sense layer | Still open (LLMs are the partial substitution) |
| Chatbots (2022) | Conversational LLMs | No persistent memory / workspace | ~3 years (execution-layer workspaces, 2025) |
Every entry is the same shape: the framework is correct, the substrate is the blocker, and the substitution is a single component swap — not a reinvention. The trick is seeing which single component is holding the system back while it still looks like the framework is wrong.
Most stuck problems are not stuck because the framework is wrong. They are stuck because one specific component is wrong, and the wrongness is invisible because the rest of the system looks so right. The people inside the problem rarely see the blocker. The people who solve it usually do so by replacing exactly one piece and leaving everything else intact.
- Widrow and Hoff had the learning rule right. One activation function was wrong. That was the whole gap.
- Rosenblatt had the hype right. One theoretical result from Minsky got over-interpreted. That was the whole AI winter.
- Expert systems had the knowledge-engineering idea right. One thing — common sense — was missing. That was the whole second AI winter.
- Chatbots have the conversational interface right. One thing — persistent memory inside a shared workspace — is missing. That's the whole gap between demo and product.
The Execution Layer thesis is, in the end, a Widrow-Hoff argument. The infrastructure for AI-as-teammate has existed for years. Foundation models are good enough. Integration platforms are mature. UX patterns are understood. The missing substitution is the persistent, structured, multiplayer memory layer that lets an agent exist inside work rather than alongside it.
That's what Taskade Genesis is. Not a rebuild of AI. A substitution. The chat box was the step function; its derivative was zero; no gradient could flow from a session into ongoing work. The workspace is the sigmoid.
This is what it feels like, inside, to build at a moment when the ceiling is one substitution away. You are not chasing a grand breakthrough. You are looking for the component that's quietly killing the gradient for everyone else in the field — and you are replacing it, while leaving everything upstream and downstream intact.
That's the Widrow-Hoff move, executed in time rather than too early.
What If
The "what if" is irresistible. What if Widrow and Hoff had, in 1959, tried a sigmoid by accident? What if one of Widrow's students had been thinking about the shape of the activation function rather than the weight update rule?
You can construct a plausible counterfactual:
- 1960: Multi-layer ADALINE works. The XOR problem is solved before anyone proves it's unsolvable.
- 1965: Neural networks become a credible research program. Minsky's criticism never lands, because the criticism was aimed at single-layer networks.
- 1970: The first serious deep-learning systems are trained on what little hardware exists.
- 1975: The field absorbs Werbos's work naturally rather than rediscovering it.
- 1985: By the time personal computers arrive, there is a mature neural network industry ready to use them.
- 2000: ChatGPT-class systems. Twenty years early.
It's a plausible counterfactual and a useless one. The hardware was not going to catch up before the late 2000s no matter what the algorithm looked like. Trained on 1960s hardware, even a correctly-designed multi-layer sigmoid network would have taken weeks to do what a modern laptop does in seconds. The algorithm alone is not enough. The substrate has to be ready.
This is the more sober version of the lesson: the substitution unlocks the breakthrough, but the substrate determines when the unlock matters. Widrow and Hoff missed the substitution by one component. They also predated the substrate by two decades. Either gap alone would have delayed deep learning. Both together produced 27 years of dormancy.
The current era is the opposite. The substrate is overwhelming — thousands of H100s per training run, petabytes of data, trillions in capex. What we are missing is the substitutions that turn all this available compute into actually-shipped work. That is the opportunity. That is where this decade's Widrows and Hoffs are, right now, whiteboarding something that looks almost right and is blocked by one invisible piece.
Closing
Ted Hoff is still alive too — he's 88, retired in the Bay Area. He wrote the microprocessor and walked away from the algorithm. Most people know him for the first and not the second. Both were extraordinary.
The 4004 is in the Smithsonian. A copy of the original ADALINE is in the Computer History Museum in Mountain View, which you can visit on any day of the week, and which I recommend. You can stand in front of Widrow and Hoff's hardware and see how close they were.
Seventy years from now, historians will write about the 2020s the way we write about 1959. They will find the Friday afternoons where someone sketched almost the right architecture and missed by one substitution. They will also find the few projects that made the substitution. Those projects will be what survives.
- Find the substitution.
- Build the system.
- Ship before the other 27 years start.
Deeper Reading
- From Bronx Science to Taskade Genesis — The lineage Widrow and Hoff fit into
- Doug Engelbart's 1968 Demo Was Taskade — A different 58-year substitution story, on the human-augmentation track
- The Execution Layer: Why the Chatbot Era Is Over — Today's equivalent of the 1959 whiteboard moment
- The Genesis Equation: P × A mod Ω — What comes after the substitution
- How Do LLMs Actually Work? — What modern backpropagation actually does inside a transformer
- What Is Grokking in AI? — The latest phase transition we don't yet fully understand
- From VisiCalc to Spreadsheet-of-Thought — The end-user programming lineage
- Memory Reanimation Protocol — The missing substitution for AI agents
John Xie is the founder and CEO of Taskade. He went to Bronx Science, ran a hosting company out of the computer lab, and spent more of his twenties than he should admit re-deriving other people's almost-discoveries from scratch. He is a large fan of Bernard Widrow.
Build with Taskade Genesis: Create an AI App | Deploy AI Agents | Automate Workflows | Explore the Community
Frequently Asked Questions
Who were Bernard Widrow and Ted Hoff?
Bernard Widrow was a Stanford professor who pioneered adaptive signal processing and early neural networks in the late 1950s. Ted Hoff (Marcian Edward Hoff) was his graduate student who later left Stanford to join Intel, where he co-invented the Intel 4004 — the world's first commercial microprocessor — in 1971. Together in 1959 they developed the LMS (Least Mean Squares) algorithm, a gradient-based learning rule that came remarkably close to modern backpropagation but missed by one critical substitution.
What is the LMS algorithm?
The LMS (Least Mean Squares) algorithm, also called the Widrow-Hoff learning rule, is a gradient-descent method for adjusting the weights of a linear system to minimize the squared error between its output and a target. It was developed by Bernard Widrow and Ted Hoff at Stanford in 1959 and used to train ADALINE (Adaptive Linear Neuron), one of the earliest practical neural networks. The LMS algorithm is mathematically nearly identical to the delta rule used in modern backpropagation — it differs primarily in that LMS operates on linear neurons with step activation functions, while backpropagation operates on multi-layer networks with differentiable activation functions like the sigmoid.
Why didn't Widrow and Hoff invent backpropagation?
They came within one substitution of it. Their LMS algorithm used calculus to compute the gradient of error — exactly the principle behind backpropagation. But their neurons used a binary step activation function (output 1 or 0), and the derivative of a step function is zero almost everywhere and undefined at the jump. This killed the gradient — it couldn't flow backward through the layer boundary. The fix was to replace the step function with a smooth sigmoid curve, whose derivative is never zero. Nobody made this substitution until Rumelhart, Hinton, and Williams published backpropagation in 1986 — 27 years later.
What was ADALINE?
ADALINE (Adaptive Linear Neuron) was a single-layer neural network built by Widrow and Hoff at Stanford in 1960. It could be trained with the LMS algorithm to recognize patterns, filter noise, and perform adaptive control tasks. ADALINE was used in practical applications including adaptive noise cancellation for telephone lines — one of the first real-world deployments of a learning system. It was the direct successor to Rosenblatt's perceptron and the direct ancestor of modern deep learning.
How did Ted Hoff help invent the microprocessor?
After leaving Stanford, Ted Hoff joined Intel in 1968 as employee #12. In 1969, tasked with designing custom chips for a Japanese calculator company called Busicom, Hoff proposed instead building a single general-purpose programmable chip — what became the Intel 4004, released in 1971. The 4004 is considered the world's first commercial microprocessor. It contained 2,300 transistors and is the ancestor of every modern CPU. The irony: the same person who almost invented the algorithm that powers modern AI went on to build the hardware that eventually runs it.
What is the sigmoid activation function?
The sigmoid activation function is a smooth S-shaped curve that maps any real number to a value between 0 and 1. Its formula is sigma(x) = 1 / (1 + e to the negative x). Unlike the binary step function, the sigmoid is differentiable everywhere, and its derivative is never zero — which means gradients can flow through it during backpropagation. Replacing the step function with the sigmoid was the critical technical move that unlocked multi-layer neural network training in 1986. Modern networks often use variants like ReLU, GELU, or Swish, but the principle is the same: smoothness enables gradient flow.
Why did it take 27 years to make such a simple substitution?
Several reasons. First, the 1969 publication of Minsky and Papert's 'Perceptrons' book convinced most researchers that multi-layer neural networks were a dead end, cutting funding and talent pipelines. Second, the computational cost of training deeper networks was prohibitive on 1960s and 1970s hardware. Third, the specific insight that the step function was the blocker wasn't obvious — researchers were focused on bigger architectural questions, not activation function choice. Fourth, the backpropagation algorithm was actually discovered and rediscovered several times in different contexts (Linnainmaa 1970, Werbos 1974) but didn't reach the neural network community until Rumelhart, Hinton, and Williams published it in 1986.
What does the Widrow-Hoff story teach about AI progress?
It teaches that the distance between a dead end and a breakthrough is often a single substitution, not a fundamental rethink. The mathematical framework was essentially right in 1959. The learning rule was right. The optimization target was right. One component — the activation function — was wrong, and the consequence was 27 years of stalled progress. This pattern recurs throughout AI history: attention mechanisms (2014) unlocked the transformer (2017); RLHF (2017) unlocked useful chatbots (2022); persistent memory layers are currently unlocking agentic workflows. The infrastructure for a breakthrough is often already built. The breakthrough is finding the missing substitution.
How does this connect to Taskade Genesis?
Taskade Genesis is built on the same pattern the Widrow-Hoff story teaches: the technical infrastructure for AI-as-teammate has existed for years, but one critical component — persistent, structured, multiplayer memory inside an agentic workspace — was missing. The chat interface was the step function, blocking the flow. Replacing it with a workspace layer was the substitution. The rest of the execution layer fell out as consequence. Sometimes the right move isn't to rethink the system — it's to find the one substitution that unsticks it.
What other near-misses exist in AI history?
Many. In 1970, Seppo Linnainmaa published the essential backpropagation algorithm in Finnish as part of his master's thesis, but the paper didn't reach the neural network community. Paul Werbos described backpropagation for neural networks in his 1974 Harvard PhD thesis, but the work was largely ignored until the 1980s. Shun'ichi Amari published foundational work on neural network training in 1967 that anticipated many later developments. The history of AI is full of ideas that were technically correct decades before they became recognized, usually because the surrounding infrastructure — hardware, data, adjacent techniques — hadn't caught up. Rosenblatt's perceptron itself was one of these: correct in 1957, fully validated in the 2010s.




