Boston, Massachusetts. 1974. A researcher at MIT’s Artificial Intelligence Laboratory sits in his office and reads a letter from DARPA — the Defence Advanced Research Projects Agency that has been funding the lab’s work since the early 1960s. The letter is polite, precise, and devastating. The unconstrained research contract that has supported the lab’s most ambitious work — the contract that has allowed researchers to follow their intellectual instincts without reporting to specific deliverables or timelines — is being restructured. The new arrangement will require explicit justification of every project against specific defence applications.
The researcher folds the letter. He looks around at the lab that has grown up around him — the students, the programs, the machines, the culture of ambitious, creative, sometimes chaotic research that has produced some of the most remarkable results in the history of computing. He thinks about what the restructuring means.
It means the golden age is over.
Not immediately. Not completely. The lab will survive, adapt, continue to produce important work. But the era of unconstrained, patient, long-horizon funding that made the early MIT AI Lab possible — that made the Logic Theorist and GPS and LISP and a dozen other foundational contributions possible — that era is ending. The field that began with such confidence in the summer of 1956 has run into something harder and more resistant than anyone expected, and the people and institutions that funded it are beginning to lose patience.
This is what the first AI winter looked like from the inside: not a dramatic collapse, but a slow cooling, a gradual withdrawal of warmth and resources, a landscape becoming incrementally less hospitable until some things that had thrived in the warmth could no longer survive.
What the Winter Was: A Definition
Before telling the story of the first AI winter, it is worth being precise about what the term means — because “AI winter” has become a phrase that gets used loosely, and its loose use obscures some important distinctions.
An AI winter is not simply a period when AI research makes slow progress. AI research has always made slow progress on the hardest problems — that is the nature of hard problems. An AI winter is a period when the institutional infrastructure supporting AI research — the funding, the talent pipeline, the institutional home — contracts significantly, typically in response to a gap between the promises that had attracted support and the results that had been delivered.
The first AI winter was, in this specific sense, a funding and credibility crisis. It was not primarily a scientific crisis — the scientific questions were as interesting in 1974 as they had been in 1964. It was a crisis of institutional support, triggered by the accumulating evidence that the field’s most confident predictions were not being fulfilled on the timelines that had justified the investment.
The winter lasted, depending on how you measure it, from roughly 1974 to roughly 1980. Different parts of the field experienced it differently — British AI research was hit hard and early by the Lighthill Report, American AI research experienced a more gradual contraction driven primarily by DARPA’s restructuring and the failure of machine translation. Both recovered, eventually, as the expert systems approach demonstrated that AI could produce commercially valuable results in specific domains.
What makes the first AI winter historically significant is not its severity — the second winter was in some ways more damaging to commercial AI — but its timing and its causes. It was the first time the field experienced the consequences of systematic overpromising, and the lessons it forced — about the gap between laboratory demonstration and real-world capability, about the dangers of extrapolating from specific successes to general claims — shaped the culture of AI research for decades.
The Dream at Its Peak: 1965
To understand the crash, you need to understand how high the dream had flown.
By 1965, ten years after the Dartmouth Conference, the AI field was at the peak of its early confidence. The programs that had been built in the preceding decade were genuinely impressive, and the community of researchers who had built them had developed a coherent and compelling vision of what the next decade would bring.
The vision was not vague. It was specific. Herbert Simon had predicted in 1957 that within ten years a digital computer would be the world’s chess champion and would discover and prove an important new mathematical theorem. Marvin Minsky had been telling journalists for several years that the problem of artificial intelligence would be substantially solved within a generation. The DARPA programme officers who were funding AI research were receiving progress reports that described each year’s results as significant advances toward systems of human-level capability.
The institutional infrastructure supporting this vision was substantial. MIT’s Project MAC — the first large-scale time-sharing computer project, funded by DARPA — had given the AI Lab access to computing resources that had previously been unavailable. Stanford’s AI Laboratory, founded by McCarthy in 1963, was attracting talented researchers and producing important results. Carnegie Mellon’s computer science department, where Newell and Simon were based, was becoming a world centre for AI and cognitive science research.
The funding was generous and, for the most part, unconstrained. DARPA’s model — give money to good people, trust them to work on important problems, don’t demand specific near-term results — was exactly what basic research required. It had produced the Logic Theorist, GPS, LISP, and a dozen other foundational contributions. It seemed to be working.
The research community was confident, connected, and productive. The people working in AI knew each other, read each other’s work, argued with each other at conferences and in seminars. The field had a shared identity, a shared set of problems, a shared sense of being engaged in one of the most important scientific and engineering projects in human history.
This was the peak. From here, the descent began.
The First Cracks: Machine Translation Fails
The first serious crack in the AI dream appeared not in the academic AI community but in the applied world of machine translation — the attempt to build systems that could automatically translate text from one language to another.
Machine translation had attracted substantial government funding since the early 1950s, driven by a specific and urgent need: the US intelligence community wanted to be able to process large quantities of Soviet scientific and military documents without the expense and delay of human translation. The technical proposal seemed straightforward — if you could capture the rules of grammar and the correspondences between words in different languages, translation should be automatable.
The early demonstrations were encouraging. A famous 1954 demonstration at IBM, using a vocabulary of 250 words and six grammar rules, successfully translated a small set of Russian sentences into English. The demonstration generated considerable excitement and led to predictions that fully automatic high-quality machine translation would be available within three to five years.
By the early 1960s, ten years had passed, tens of millions of dollars had been spent, and the predictions had not been fulfilled. The systems that existed could produce rough, useful translations of some kinds of text but could not approach the quality of human translators for most purposes. The problem, it turned out, was not grammar rules and vocabulary correspondences. It was meaning — the fact that the same word in one language could correspond to many different words in another language depending on context, that the context itself could require understanding of the real world that no translation program possessed, that language use was embedded in cultural assumptions that could not be captured in any finite list of rules.
The US government commissioned an independent review. In 1966, the Automatic Language Processing Advisory Committee — ALPAC — published a report that was devastating in its clarity. Machine translation as currently practiced, ALPAC concluded, was slower, less accurate, and twice as expensive as human translation. There was no immediate or future prospect of useful machine translation. The research programme should be redirected.
The ALPAC Report cut US government funding for machine translation research immediately and almost completely. Research groups that had been built around the machine translation problem dissolved or redirected. The most prominently funded AI application of the 1950s and early 1960s had failed, publicly and expensively, and the failure damaged the credibility of AI research more broadly.
The failure was not random. It was a failure of the specific approach — the rule-based, linguistic approach that had dominated early machine translation research — in the face of the real complexity of language. The approach could handle what it had been designed to handle. It failed when confronted with the vast, implicit, contextual knowledge that real translation required.
The lessons of the ALPAC failure were available to anyone who cared to learn them. The specific capability demonstrated in laboratory conditions did not generalise to the complexity of real-world use. The rules that were sufficient for carefully chosen demonstration examples were insufficient for the full range of inputs the system would encounter in deployment. The gap between demonstration and deployment was large and had been underestimated.
These lessons were not fully absorbed. The machine translation failure was treated as a problem specific to that research programme — a consequence of the particular difficulty of language, not a symptom of a more general problem with the AI research approach. The broader AI research community continued with its confident predictions, its generous funding, its assumption that the remaining problems were tractable and the remaining distance was short.
The Combinatorial Wall: When the Problems Got Harder
While machine translation was failing in the applied world, a different and more fundamental problem was becoming apparent in the academic AI research community: the combinatorial explosion.
The combinatorial explosion is the tendency of search spaces to grow exponentially with problem complexity. In chess, each additional move considered doubles or triples the size of the search tree. In natural language, each additional word in a sentence multiplies the number of possible interpretations. In planning, each additional step in a plan multiplies the number of possible action sequences. In any problem that involves searching through a space of possibilities, the space grows faster than linear with the size of the problem.
The early AI programs had worked in simplified, constrained domains precisely because those domains were small enough to be searchable. The Logic Theorist proved theorems in propositional logic — a domain with a small number of inference rules and relatively short proofs. GPS solved puzzles with a small number of objects and operations. ELIZA engaged in superficial conversation that avoided the deep structure of language understanding. MacHack played chess at an amateur level by searching only a few moves ahead.
When researchers tried to extend these programs to harder problems — longer proofs, more complex planning tasks, more realistic language understanding — the search spaces exploded. A program that could solve simple puzzles in seconds could not solve harder puzzles in years, because the search space grew too fast for any available computational power to handle. The heuristics that pruned the search tree efficiently in simple cases were insufficient in complex ones.
This was not simply a problem of computing power, though the computers of the 1960s and early 1970s were severely limited. It was a structural problem with the approach. Rule-based, search-based systems encountered an exponential barrier when moved from simplified domains to realistic ones, and no amount of additional computing power or better heuristics could overcome the fundamental exponential growth.
By the early 1970s, this problem was becoming impossible to ignore. Programs that had seemed to be approaching general intelligence in simplified domains were failing to make the transition to more realistic settings. The progress that had seemed to be linear — each year bringing systems closer to general capability — was plateauing. The simplified domains were being exhausted; the realistic domains were resisting.
This was what the Lighthill Report identified and analysed. Lighthill’s diagnosis of the combinatorial explosion as the central obstacle to AI progress was substantially correct, even if his conclusion — that the obstacle was fatal rather than merely requiring different approaches — was too pessimistic.
The Lighthill Report in Context
We have discussed the Lighthill Report in detail in the Events article on that topic. Here the focus is on what it meant for the people who experienced it — on how it felt to be an AI researcher in Britain in 1973 and to watch your field’s credibility attacked in a major government review and your institution’s funding cut in response.
For the researchers at Edinburgh’s machine intelligence unit, the Lighthill Report was a personal catastrophe dressed in the language of policy analysis. Donald Michie, who had been building the unit for years, who had attracted talented researchers from around the world, who had produced genuine results in machine learning and robotics — Michie experienced the report as a fundamental injustice. He had been building something real. He believed in it. And it was being dismantled on the basis of an assessment that, in his view, misunderstood what his group was doing and why.
His response — appearing on the BBC to argue against the report, writing public responses, challenging Lighthill’s methodology — was the response of a person fighting for something he valued. The fight did not succeed. The funding was cut. The unit was diminished. But it revealed something important about what was at stake: not just research programmes and careers, but a vision of what AI could be and what it could mean.
For American researchers, the experience was different — less concentrated, less sudden, spread over several years rather than triggered by a single report. DARPA’s gradual restructuring of its AI funding — the shift from unconstrained basic research to contracts requiring specific defence applications — changed the environment in which research was done rather than eliminating it. The shift was real but it was not a verdict. It was a request for greater accountability, delivered in language that made the request feel more like a demand.
The response of the American AI community to this pressure was revealing. Some researchers reframed their work in terms of defence applications — arguing that basic AI research was relevant to military planning, to intelligence analysis, to autonomous systems. Some retreated to narrower, more defensible claims — focusing on specific capabilities that could be demonstrated concretely rather than on the general intelligence that had been the field’s founding ambition. Some continued to make the ambitious claims that had characterised the field, finding that the funding environment could still support such claims if they were packaged appropriately.
The diversity of responses reflected genuine disagreement about what had gone wrong and what needed to change. Some researchers believed that the winter was primarily a communication problem — that the field had made claims it should not have made, and that more careful communication would restore credibility. Others believed it was a more fundamental problem — that the research approaches themselves needed to change, that the ambition needed to be recalibrated toward domains where genuine progress was achievable.
The Perceptrons Effect: Neural Networks Before the Winter
One of the most important stories of the first AI winter is the role of Minsky and Papert’s Perceptrons book — published in 1969, before the winter had fully arrived — in damaging the neural network research programme that would eventually provide the path forward.
The Perceptrons book, as we have discussed in previous articles, was technically correct in its analysis of the limitations of single-layer perceptrons. What was damaging was the way its conclusions were read — as demonstrating that neural networks in general were a dead end, rather than that a specific simple architecture had specific limitations.
Neural network research had been developing throughout the 1960s alongside the symbolic AI mainstream. Frank Rosenblatt’s perceptron work had attracted attention and funding. Several researchers were exploring learning systems, adaptive networks, systems that could improve through experience rather than through explicit programming. This work was not the dominant paradigm — the symbolic AI tradition held that position — but it was active and potentially productive.
The Perceptrons book changed this. After 1969, neural network research became substantially harder to fund and substantially less attractive to graduate students choosing research areas. The book’s authority — Minsky was the most prominent figure in American AI, and Papert’s mathematical credentials were impeccable — gave its conclusions a weight that deterred further exploration.
This had a specific effect on the winter: it meant that the approach that would eventually prove most productive — learning-based, neural network approaches — was suppressed precisely in the period when its development was most needed. If the neural network research of the late 1960s had continued without the Perceptrons disruption, backpropagation might have been developed and demonstrated in the early 1970s rather than the mid-1980s. The deep learning revolution might have happened fifteen years earlier.
This counterfactual cannot be stated with certainty. The development of deep learning required not just the backpropagation algorithm but also large datasets and substantial computing power, which were not available in the early 1970s. The algorithm alone might not have been sufficient to change the field’s direction.
But the suppression of neural network research in the late 1960s and early 1970s was a genuine cost, and it is a cost that can be partly attributed to the Perceptrons book’s influence. Understanding this cost is important for understanding the full picture of the first AI winter — not just the external pressure from funding agencies, but the internal intellectual dynamics that shaped what the field worked on and what it did not.
The Human Geography of the Winter: Who Left, Who Stayed, Who Came Back
The first AI winter was not experienced uniformly by everyone in the field. Its effects depended on where you worked, what you were working on, how deeply your institution was affected by funding cuts, and what alternatives were available.
Researchers at the major American AI labs — MIT, Stanford, Carnegie Mellon — experienced the winter as a change in atmosphere and constraints more than as a catastrophic loss of institutional support. DARPA continued to fund AI research; it restructured the funding to require more explicit connection to defence applications. This was constraining, sometimes frustrating, but not fatal. The labs survived, adapted, and continued to produce important work.
Researchers in the UK, particularly at Edinburgh, experienced something more severe. The Lighthill Report provided a specific, public justification for cutting funding in ways that the more diffuse American funding environment did not. Edinburgh’s machine intelligence unit lost a significant fraction of its funding and several of its key researchers. The concentration of AI research capability that Michie had built was permanently reduced.
Graduate students and early-career researchers faced the most difficult decisions. The pipeline of AI talent that had been building through the 1960s encountered a narrowed set of opportunities in the early 1970s. Some redirected their research toward related fields — cognitive science, computational linguistics, software engineering — that offered better employment prospects. Some left academia for industry, finding that the skills they had developed in AI research were applicable in computing more broadly. Some persisted in AI research, sustaining themselves on whatever funding remained available.
The people who left were not, in general, people who had given up on AI as a project. They were people who had made pragmatic decisions about where they could do useful work and build sustainable careers. Many of them brought their AI backgrounds with them into other fields, producing work that cross-pollinated ideas between AI and adjacent disciplines in productive ways.
And many of them came back. When the expert systems boom of the 1980s restored the commercial AI market, researchers who had redirected their careers during the winter found pathways back into the field. When the neural network revival of the late 1980s and early 1990s created new intellectual opportunities, researchers who had maintained their interest in learning-based approaches were ready to contribute.
The AI field has a strong tendency to recapture talented people who have temporarily moved elsewhere. The problems are compelling enough, the intellectual excitement sufficient, that people who have been trained in AI and who maintain any connection to it tend to return when the institutional environment becomes more hospitable.
What the Programs Could Not Do: The Technical Dimension of the Crisis
The winter was ultimately grounded in a technical reality: the programs that had been built were not doing what they had been claimed to do. Understanding the specific technical failures — not just that the programs fell short, but why and in what specific ways — is essential to understanding what the winter was really about.
The most fundamental failure was the failure to generalise from simplified to realistic domains. The programs that worked in laboratory conditions worked because the laboratory conditions had been carefully designed to match the programs’ capabilities. The blocks world that SHRDLU inhabited was not a simplification of the real world — it was a different kind of world, one that had been constructed specifically to make language understanding tractable by eliminating the open-endedness, the ambiguity, and the implicit world knowledge that real language required.
When programs were moved from simplified to realistic domains, several things happened simultaneously. The search spaces exploded, making exhaustive or near-exhaustive search impractical. The knowledge representations that had been adequate for simplified domains proved inadequate for realistic ones — the relevant concepts were more numerous, more contextually dependent, more interrelated. The heuristics that had worked in simplified domains failed to scale to realistic complexity.
This failure of generalisation was not accidental. It was a consequence of the fundamental approach — the assumption that intelligence was explicit rule-following in a well-defined domain. Real-world intelligence operated in domains that were not well-defined, that involved implicit knowledge, that required continuous calibration against the reality of the situation rather than the application of pre-specified rules.
The programs also failed to learn. Every program that was built had its knowledge encoded by human programmers. It could not acquire new knowledge from experience, could not update its strategies based on feedback, could not improve its performance through practice. Human intelligence — even at the level of simple skill acquisition — involves continuous learning and adaptation. The programs that had been built were static in a way that real intelligence was not.
The absence of learning was particularly damaging for the machine translation failures. Translation required knowledge of context, usage, idiomatic expression, the cultural assumptions embedded in language — knowledge that was constantly evolving as language evolved and that could not be fully captured in any fixed set of rules. A translation system that could not learn from its mistakes, that could not update its knowledge as language use changed, was structurally unable to approach the quality of a human translator who was continuously learning.
The Institutional Response: What Funding Agencies Did
The funding agencies that had supported AI research through its optimistic years responded to the winter in ways that shaped the field’s subsequent development for decades.
DARPA’s restructuring was the most consequential American institutional response. The shift from unconstrained basic research funding to more directed funding with explicit defence applications was not a rejection of AI research but a demand for greater accountability. DARPA wanted to see specific results — demonstrations that AI technology was relevant to specific defence problems — rather than the general progress toward human-level intelligence that the field’s founders had been promising.
This demand produced a specific kind of research. AI researchers who wanted DARPA funding learned to frame their work in terms of applications — speech recognition for military communications, computer vision for reconnaissance, planning systems for logistics and strategy. Some of this framing was genuine: these were real applications where AI technology had genuine relevance. Some of it was strategic: researchers packaged work that they would have done anyway in language that connected it to defence needs.
The Science Research Council in Britain had a more straightforwardly negative response, cutting AI funding across multiple institutions in response to the Lighthill Report. The British institutional response was less sophisticated than DARPA’s — rather than redirecting AI research toward applications, it simply reduced support for the field. The consequences for British AI research were correspondingly more severe.
The differences between the British and American institutional responses produced differences in how the two countries’ AI communities experienced and recovered from the winter. British AI took longer to recover its pre-winter strength, partly because the institutional support had been more completely withdrawn. American AI contracted but maintained a substantial institutional base from which recovery was possible when the expert systems boom restored commercial interest.
The institutional lesson — drawn explicitly or implicitly by researchers who experienced the winter — was that AI research needed to maintain a connection to demonstrable results in specific applications. The era of patient funding for research on the general problem of machine intelligence, without specific near-term deliverables, was over. Future funding would require more specific justification, more concrete results, more demonstration that the technology was heading somewhere useful.
This lesson shaped the expert systems era that followed: an AI research programme that was explicitly focused on specific applications, that could demonstrate concrete results in specific domains, that made modest rather than extravagant claims about its capabilities. The expert systems approach was, in part, a strategic response to the institutional pressures of the winter.
The Philosophical Dimension: What the Winter Revealed About Intelligence
Beyond the institutional and technical dimensions, the first AI winter had a philosophical dimension that is easy to overlook but important to understand.
The programs that failed to generalise were not failing because of minor technical limitations that better algorithms or more computing power would overcome. They were failing because they were based on a conception of intelligence that was, in important respects, wrong.
The conception was this: intelligence is rule-following in a well-defined domain. Give a program the right rules, implemented in a sufficiently powerful computing system, and it will exhibit intelligent behaviour. The intelligence is in the rules.
This conception was not entirely wrong. Rule-following is an important component of many kinds of intelligent behaviour. Mathematical proof, logical inference, many kinds of calculation — these can be done by explicit rule-following, and doing them by rule-following produces behaviour that is, in those specific domains, indistinguishable from intelligent behaviour.
But the conception was wrong about what intelligence fundamentally was. Intelligence — real-world intelligence, the kind that allowed a person to navigate an unfamiliar situation, to understand a new kind of problem, to communicate effectively in a new context — was not primarily rule-following. It was something else: a flexible, adaptive engagement with an open-ended world, drawing on vast reserves of implicit knowledge that could not be fully made explicit, always contextually sensitive in ways that defied complete formalisation.
The failures of the first AI winter were, in this philosophical sense, demonstrations that the wrong conception of intelligence had been guiding the field. The programs failed not because they implemented the right conception badly, but because they implemented the wrong conception, even if reasonably well.
This philosophical lesson was not fully drawn from the first winter. The expert systems era that followed made a strategic retreat from the grand claims of general intelligence to the more defensible claim of domain-specific expertise — but it maintained the same fundamental conception: intelligence as explicit rule-following. It was only with the machine learning revolution that the field shifted to a genuinely different conception: intelligence as learned pattern recognition, emerging from exposure to data rather than from explicit rule specification.
The Voices That Said It Was Coming
In retrospect, the first AI winter was predictable. The specific predictions that failed to materialise — the decade-scale claims about chess champions and proven mathematical theorems — were clearly overconfident given the actual state of the technology. The gap between laboratory demonstrations and real-world capability was visible, if not widely acknowledged. The difficulty of the remaining problems — common sense reasoning, real-world perception, natural language understanding — was apparent to anyone who thought carefully about what those problems required.
There were voices, before and during the winter, that said something like this. Dreyfus’s 1965 RAND report — later expanded into “What Computers Can’t Do” — argued that the symbolic AI approach was philosophically misguided and would not achieve its goals. He was dismissed, sometimes contemptuously, but he was substantially right.
Within the AI community itself, there were researchers who were more cautious in their public statements than Minsky or Simon — who qualified their predictions, who acknowledged the difficulty of the remaining problems, who tried to maintain a distinction between what had been demonstrated and what had been promised. These researchers were less visible than the optimists, partly because caution is less newsworthy than confidence, and partly because the culture of AI in the 1960s selected for boldness.
The absence of a strong culture of honest self-assessment was one of the contributing factors to the winter. If the field had been more willing to acknowledge the gap between its demonstrations and its promises — to say publicly what it was sometimes saying privately — the winter might have been less severe, because the gap between expectations and reality would have been smaller.
This is one of the lasting lessons of the first AI winter: the importance of honest communication about what has been achieved, what has not been achieved, and how large the remaining distance is. The overconfidence of the 1960s was not just strategically costly — in the sense of damaging credibility — but intellectually costly, in that it discouraged honest assessment of what the field was actually learning from its attempts to build intelligent systems.
The Recovery: How the Dream Survived
The first AI winter did not end AI. It did not permanently damage the intellectual project or prove that the goal was unachievable. It was a crisis of institutional support and credibility that required a period of retrenchment, recalibration, and rebuilding before the field could advance again.
The recovery came from an unexpected direction. The expert systems approach — which had been developing quietly during the winter years — demonstrated in the late 1970s and early 1980s that AI could produce commercially valuable results in specific domains. XCON saving DEC forty million dollars per year was not the general machine intelligence that the field had originally promised. But it was something real, something demonstrably valuable, something that corporations and investors could understand and support.
The expert systems recovery had its own problems, as we have discussed. It led to a second wave of overconfidence and a second winter. But it demonstrated that the field could survive a crisis of credibility by recalibrating its ambitions and demonstrating genuine value in specific domains.
The deeper recovery came from the neural network revival of the mid-to-late 1980s — from researchers who had maintained their faith in learning-based approaches through the winters, who had kept working on the ideas that the field’s mainstream had dismissed, and who eventually produced results that forced a reconsideration of the whole field’s direction.
The first AI winter was, in this sense, necessary. Not in the sense that it was optimal — the field would have been better off without the credibility damage, without the talent loss, without the years of reduced funding. But necessary in the sense that the crisis it produced forced the field to confront, in a way that the optimism of the 1960s had prevented, the real difficulty of the problems it was working on and the real limitations of the approaches it had been using.
The Pattern and Its Meaning
The first AI winter was not the last. The second winter followed the expert systems boom in the late 1980s. Subsequent waves of AI progress have been accompanied by cycles of elevated expectations that outrun the evidence and then contract when the evidence catches up.
Understanding this pattern — the recurring cycle of enthusiasm, overpromising, disappointment, and contraction — requires understanding what drives it. The pattern is not simply a failure of honesty or a failure of realism. It is driven by forces that are more structural.
Scientific and technical communities that are making genuine progress are embedded in funding environments that reward confident claims. Researchers who want to attract funding, students, institutional support, and public attention are rewarded for boldness and penalised for caution. The social dynamics of science — the competition for resources, for recognition, for the most talented students — select for the kinds of claims that the funding environment rewards.
These dynamics do not require dishonesty. Researchers who make overconfident claims are often genuinely confident — their confidence is a real expression of their belief in the work, filtered through social dynamics that amplify confidence and attenuate doubt. The problem is not lying. It is a systematic bias in how scientific confidence is expressed and communicated.
The correction for this bias is not simply to ask researchers to be more honest. It is to create institutional structures that reward honesty as much as confidence — peer review processes that scrutinise claims carefully, funding mechanisms that reward demonstrations over predictions, norms of scientific communication that treat acknowledged uncertainty as a sign of intellectual maturity rather than weakness.
These structures exist, in various forms, in mature scientific fields. They are less well-developed in AI, a field that has grown very rapidly and that has often been more oriented toward engineering outcomes than toward the careful epistemic standards of basic science.
The first AI winter was a painful lesson in why these structures matter. It was the first demonstration that a field can outrun its evidence in ways that damage its credibility and its institutional support, that the consequences of this outrunning are real and lasting, and that the correction requires more than the individual honesty of individual researchers — it requires collective institutional commitments to honest assessment of what has and has not been achieved.
The Survivors and What They Built
The researchers who survived the first AI winter and continued to work through it built the foundations of everything that came after.
Those who stayed in symbolic AI — in knowledge representation, in planning, in formal reasoning — developed ideas that are still relevant. Non-monotonic reasoning, temporal logic, situation calculus, constraint satisfaction — these are contributions of the winter years that appear, in modified form, in the AI systems of today. The limitations of the symbolic approach do not negate these contributions; they contextualise them as contributions to specific aspects of intelligence rather than the foundation of all intelligence.
Those who sustained the connectionist research tradition — Hinton, LeCun, and their colleagues — built the intellectual infrastructure that would eventually produce the deep learning revolution. The ideas they developed during the winter — backpropagation, convolutional networks, the theory of distributed representations — were not just recovered from pre-winter neural network research. They were significantly advanced, refined, and extended. The winter years were not empty years for connectionist research. They were productive years during which important ideas were developed and refined, even without the validation of mainstream recognition.
Those who built the expert systems approach — Feigenbaum, Buchanan, and their colleagues — demonstrated that AI could produce commercially valuable results, even if the approach had fundamental limitations. The commercial applications they built, and the methodologies they developed for knowledge engineering and system validation, contributed to subsequent AI development in ways that are easy to overlook from the perspective of the deep learning era.
The winter survivors were not people who had lost faith. They were people who had maintained their conviction that the project was worth pursuing — who continued to work, with reduced resources and reduced recognition, on problems they believed were important. Their persistence was not merely admirable on a personal level. It was functionally essential to the field’s survival and eventual recovery.
The Dream That Would Not Die
The first AI winter killed careers. It killed research programmes. It killed companies, and conferences, and the institutional momentum that had been building since Dartmouth. It killed the optimism that had animated the field’s founding generation.
It did not kill the dream.
The dream — the conviction that machines could be made to think, that intelligence was a natural phenomenon that could be understood and built, that the project was worth pursuing — survived the winter in the minds and the work of the researchers who had not given up. It survived in the papers published during the lean years, in the graduate students who chose AI despite the funding difficulties, in the informal conversations and the maintained connections that sustained the intellectual community even when the institutional infrastructure was contracting.
It survived because it was right. The goal — machine intelligence — was real and achievable. The path was longer and more winding than the founders had imagined, and the terrain was rougher. But the destination was real.
The first AI winter was a collision between an idea and reality. The idea was that intelligence was achievable by the approaches that had been tried. The reality was that it was not achievable by those approaches on those timelines. The collision was painful. And it was necessary for the field to learn what the dream actually required — what the path to machine intelligence genuinely demanded — rather than what the optimistic projections of the early years had promised.
That learning was hard. It was painful. It cost years and careers and institutional credibility that took decades to rebuild. But it was also genuine learning — the kind of learning that only comes from trying something and discovering that it does not work, and then asking, with honesty and with rigour, why not.
The answer to that question — discovered slowly, painfully, over decades of work by generations of researchers — is the story of the rest of this series.
Further Reading
- “Machines Who Think” by Pamela McCorduck (1979) — The first comprehensive history of AI, written while the first winter was still fresh. Vivid, honest, and essential for understanding the human dimensions of the period.
- “Artificial Intelligence: A General Survey” by James Lighthill (1973) — Read the original report that triggered the British winter. Short, clear, and historically essential.
- “The Sciences of the Artificial” by Herbert Simon (1969) — Simon’s most accessible and philosophical work, written at the beginning of the winter and capturing the pre-winter optimism at its most thoughtful.
- “What Computers Can’t Do” by Hubert Dreyfus (1972) — The most sustained critique of the AI approach that the winter would partially vindicate. Essential reading alongside the optimist literature.
- “Knowledge and Power” by Philip Agre (1997) — A later work that analyses the cultural dimensions of AI research and the institutional pressures that shaped it. Provides important context for understanding the winter.
Next in the Articles series: A11 — Expert Systems: AI Learns to Be a Specialist — How the field that had been humbled by the first winter found a commercially viable path forward through the expert systems approach: going narrow, going deep, and going to work. The full story of AI’s first commercial era, from DENDRAL and MYCIN to XCON and the billion-dollar industry that grew around knowledge engineering.
Minds & Machines: The Story of AI is published weekly. If the story of the first AI winter — the crash, the survivors, and the dream that refused to die — resonates with the current moment in AI, share it with someone who would find the historical perspective valuable.