Maynard, Massachusetts. 1982. The Digital Equipment Corporation is struggling with a problem that sits at the heart of its business model. DEC makes minicomputers — powerful machines customised to the specific requirements of each customer. A university might need a machine configured for scientific computation. A bank might need one configured for transaction processing. A manufacturing plant might need one configured for process control. The configurations involve hundreds of components — processors, memory boards, storage devices, terminal controllers, cables, power supplies — and the rules governing which components can be combined with which others fill several thick manuals.

Human engineers called “configurers” manage this complexity. They are expensive, they take years to train, and even after years of training they make mistakes — errors that result in machines that are shipped to customers with incompatible components, machines that do not work until a technician visits the site to correct the configuration. Each mistake costs DEC thousands of dollars in rework and customer goodwill.

There is a program — R1, developed at Carnegie Mellon University over the past three years — that can configure VAX computer systems automatically. DEC has been using it in pilot deployments, and the results have been extraordinary: configurations produced faster, with fewer errors, at a fraction of the cost of human configurers. DEC is now deploying R1 across its entire configuration operation.

By 1986, R1 — renamed XCON — will be handling virtually all of DEC’s VAX configurations. The company will estimate that it saves DEC more than forty million dollars per year. It will become the most successful commercial AI application of its era — the demonstration that AI was not just a research curiosity but a technology with real economic value.

And it will do it not by being generally intelligent, not by reasoning from first principles, not by understanding the world in any deep sense. It will do it by encoding the specific, detailed, domain-specific knowledge of DEC’s expert configurers in a form that a computer program could apply.

This was expert systems. This was AI getting a job.


The Philosophy of Going Narrow

The expert systems era represented a fundamental shift in the ambition of AI — a shift that was simultaneously a retreat from the grand vision of the field’s founders and a genuine intellectual advance in the understanding of what AI could practically achieve.

The grand vision, articulated at Dartmouth in 1956 and pursued through the 1960s, was general intelligence: machines that could do anything an intelligent person could do, across any domain, with the flexibility and adaptability of human cognition. This vision was never abandoned entirely, but by the late 1970s it was clear that it was not going to be achieved in any near-term timeframe by any approaches then available.

The expert systems approach accepted this reality and drew the logical conclusion: if general intelligence was out of reach, perhaps domain-specific intelligence was achievable. If you could not build a machine that knew everything, perhaps you could build a machine that knew everything about a specific domain — chemical analysis, medical diagnosis, computer configuration — and could apply that knowledge with the reliability and speed that computers provided.

This was not a small ambition. The knowledge of a genuine expert — a specialist physician, an experienced engineer, a master configurator — was deep, detailed, and hard won. It took years to acquire, it was maintained only through constant updating and practice, and it was often tacit — embedded in patterns of recognition and judgment that experts could not always articulate. Making that knowledge explicit enough to be encoded in a computer program was a significant intellectual challenge.

But it was a tractable challenge in a way that general intelligence was not. You could interview experts. You could ask them to describe their reasoning process. You could observe them solving problems, ask them to think aloud, identify the rules and patterns and heuristics they were applying. You could then encode those rules in a knowledge base and implement an inference engine that applied them.

This process — extracting knowledge from experts and encoding it in a computational form — was called knowledge engineering, and it became the defining practice of the expert systems era. Knowledge engineers were the translators between human expertise and computational implementation, the people who could talk to physicians and engineers and financial analysts and extract from their expertise a structured, formal representation that a computer program could use.

The approach was modestly ambitious in its goals but surprisingly effective in its results. Expert systems that encoded specific domain knowledge could perform at the level of human experts in their domains — sometimes exceeding human performance in consistency and speed, while falling short in flexibility and the ability to handle cases outside their training domain.


The Intellectual Roots: From DENDRAL to MYCIN

The expert systems approach did not appear from nowhere in 1980. Its intellectual roots were in the work that had been done through the 1960s and 1970s by researchers who had recognised, earlier than most, that domain-specific knowledge was the key to practical AI.

The most important precursor was DENDRAL, developed at Stanford University starting in the mid-1960s by Edward Feigenbaum, Joshua Lederberg, Bruce Buchanan, and their colleagues. DENDRAL was a program designed to determine the molecular structure of organic compounds from mass spectrometer data — a task that required specialised chemical knowledge and pattern recognition that expert chemists had developed through years of training.

DENDRAL’s approach was distinctive. Rather than trying to implement a general problem-solving method that could be applied across domains, the developers worked closely with expert chemists to understand the specific knowledge and reasoning patterns that those experts used to interpret mass spectrometer data. They encoded that knowledge explicitly — rules describing which spectral patterns corresponded to which structural features, heuristics for ruling out impossible structures, procedures for evaluating candidate structures against the observed data.

The result was a program that performed well on its specific task — matching or exceeding the performance of expert chemists on the identification problems it was designed to handle. DENDRAL was not trying to be intelligent in general. It was trying to apply the specific knowledge of expert chemists to a specific class of problems, and it succeeded.

Feigenbaum drew several important lessons from the DENDRAL project. The most important was what he called the “knowledge principle” — the hypothesis that the power of an intelligent program was primarily determined by the amount of knowledge it had about its domain, rather than by the sophistication of its reasoning methods. A program with rich, detailed domain knowledge and simple reasoning would outperform a program with sparse domain knowledge and sophisticated reasoning. Knowledge was the key variable.

This hypothesis was not obvious, and it cut against the prevailing tendency in AI research to focus on general reasoning methods — heuristic search, theorem proving, general problem-solving — and to treat domain knowledge as a secondary concern. Feigenbaum was saying something different: that the search for general reasoning methods was misguided, that the right way to build intelligent programs was to start with the domain knowledge and build the reasoning around it.

The knowledge principle was the theoretical foundation of expert systems. It was also, as subsequent experience would reveal, a half-truth: knowledge was indeed essential, but the form in which knowledge was represented and the methods by which it was reasoned about turned out to matter enormously. The expert systems era would discover, painfully, the limitations of simple rule-based representation and inference.

MYCIN, developed at Stanford in the early 1970s by Edward Shortliffe and colleagues, was the application of the DENDRAL approach to medical diagnosis — specifically to the diagnosis of bacterial infections and the recommendation of antibiotic treatments. MYCIN’s development ran from 1972 to 1980, entirely within the first AI winter, and its relative success in the face of the broader field’s difficulties was a sign of things to come.

MYCIN’s knowledge base consisted of approximately five hundred rules — if-then production rules encoding the reasoning patterns of expert physicians. A typical rule might say: IF the infection is of type gram-positive AND the patient has a penicillin allergy AND the severity is high, THEN consider cephalosporin antibiotics with probability 0.7. The rules encoded not just the knowledge but the uncertainty — the probability attached to each conclusion reflected the fact that medical diagnosis was rarely certain.

MYCIN also implemented a backward-chaining inference mechanism — an inference engine that would reason backward from a goal (identify the probable infection) through chains of rules to the observable evidence (the patient’s symptoms and test results). This backward chaining was efficient for the diagnosis task because it focused the reasoning on the goal rather than forward-chaining through all possible conclusions from all available evidence.

The program was evaluated by having it diagnose a series of cases and comparing its recommendations to those of expert physicians who had not seen MYCIN’s diagnoses. The evaluations showed that MYCIN performed at or near the level of human specialists on the class of cases it was designed to handle. This was a genuine achievement, and it demonstrated that the expert systems approach was not just a research curiosity but a practical approach to implementing specialised intelligence.


XCON: The Program That Proved the Business Case

If MYCIN was the demonstration that expert systems could match human experts in a specific domain, XCON was the demonstration that expert systems could produce real economic value — that the technology had a business case, not just a research case.

The story of XCON — originally called R1 — began in 1978 when Carnegie Mellon graduate student John McDermott began working on the problem of configuring DEC’s VAX computer systems. McDermott was a researcher in the emerging field of production system programming, and he was interested in applying the technology to practical problems. The DEC configuration problem was practically important — DEC was losing significant money to configuration errors — and technically tractable — the rules governing component compatibility could, in principle, be encoded in a production system.

The development of R1 required extensive knowledge engineering — months of interviews with DEC’s expert configurers, careful documentation of the rules they applied, analysis of the cases where configurations went wrong and why. McDermott and his colleagues extracted knowledge that the expert configurers themselves could not always articulate — patterns of judgment that were implicit in their expertise and that had to be made explicit to be encoded in the program.

The resulting program contained several thousand production rules — if-then rules of the form “IF the system requires a floating-point processor AND the processor being configured is a VAX-11/780, THEN add the appropriate floating-point accelerator to the component list.” The rules covered the full range of compatibility and constraint satisfaction problems involved in VAX configuration, from high-level architectural decisions to low-level cable length constraints.

DEC deployed R1 in 1980 and began tracking its performance carefully. The results were consistently impressive. The program configured systems faster than human experts — a configuration that took a human configurer several hours could be completed in minutes. It made fewer errors — not zero errors, because its knowledge was incomplete, but significantly fewer than human configurers. And it was consistently available — it did not get tired, did not have sick days, did not retire and take its expertise with it.

By 1982, DEC had expanded R1’s deployment and was reporting annual savings that would grow, by the mid-1980s, to more than forty million dollars per year. The program — renamed XCON, for eXpert CONfigurer — became the most successful commercial AI application of its era and the benchmark against which other expert systems were evaluated.

XCON’s success had effects that went far beyond DEC. It demonstrated, conclusively, that AI technology could produce commercially valuable results — that the investment in AI research was not just an investment in academic curiosity but an investment in technology that could save real money for real businesses. This demonstration attracted the attention of investors, corporations, and venture capitalists who had not previously taken AI seriously.

The commercial expert systems industry that grew from XCON’s success was substantial. Companies including Teknowledge, IntelliCorp, and dozens of others were founded to develop and sell expert systems technology. IBM, Digital Equipment Corporation, Texas Instruments, and other established technology companies established AI research groups and began developing expert systems products. The AI industry, which had barely existed as a commercial entity in the mid-1970s, was worth hundreds of millions of dollars by the mid-1980s and would grow to billions before the second AI winter ended the boom.


The Technology: How Expert Systems Worked

Understanding the expert systems era requires understanding how expert systems actually worked — what they were doing when they appeared to reason like experts, and what was genuinely impressive about that and what was not.

Expert systems were built around two main components: a knowledge base and an inference engine.

The knowledge base was a collection of domain-specific knowledge, typically encoded as production rules — if-then statements that captured the reasoning patterns of human experts. A rule in a medical expert system might say “IF the patient has a fever AND the patient has a stiff neck AND the patient has sensitivity to light, THEN the patient may have meningitis.” A rule in a configuration expert system might say “IF the CPU requires a specific type of memory controller AND the available memory controllers are of types A, B, and C AND the system has an existing memory controller of type A, THEN do not add another memory controller.”

The knowledge base was the heart of the expert system — it encoded the expertise that made the system capable. A larger, more accurate, more complete knowledge base produced a more capable system. The process of building the knowledge base — the knowledge engineering process — was the most difficult and most labour-intensive part of building an expert system.

The inference engine was the reasoning mechanism that applied the knowledge base to specific cases. The two main inference strategies were forward chaining and backward chaining. Forward chaining began with the available evidence and applied rules to derive conclusions, then applied more rules to those conclusions, continuing until no more rules could be applied. Backward chaining began with a goal and worked backward, identifying which rules could establish the goal and what evidence those rules required.

Expert systems also typically included an explanation facility — a component that could explain why it had reached a particular conclusion, by tracing through the chain of rules that had been applied. This transparency was one of the most practically important features of expert systems: physicians and engineers who used them wanted to understand why the system had recommended a particular course of action, not just what it had recommended. The ability to say “I recommended antibiotic X because the patient has symptom Y and test result Z, which according to rule 47 indicates a gram-positive infection that is likely to respond to antibiotic X” was crucial for user acceptance.

The transparency of expert systems — the fact that you could inspect the rules and follow the reasoning — was both a practical advantage and a theoretical distinction from later AI approaches. Neural networks, which became dominant in the subsequent era, were opaque: their knowledge was encoded in numerical weights that could not be directly inspected, and their reasoning processes could not be explained in terms that humans could follow. The contrast between the interpretable expert system and the opaque neural network would become a central issue in AI ethics and AI safety in subsequent decades.


The Knowledge Engineering Challenge

The process of building a knowledge base — extracting expertise from human experts and encoding it in a formal representation that a computer could use — turned out to be considerably harder than the early expert systems proponents had anticipated.

The difficulty had several sources. Human expertise was not always explicit. Experts who had been practising their craft for decades often could not articulate the rules they were applying — the knowledge was embedded in patterns of recognition that operated below the level of conscious access. When asked “how do you know this patient has bacterial meningitis?” a physician might say “I just know — you develop a feel for it after years of practice.” This tacit knowledge was precisely what the expert system needed, and precisely what was hardest to extract.

Expertise was also contextual in ways that were difficult to capture in rules. A rule that was almost always correct might have exceptions that an experienced practitioner recognised implicitly but could not always state explicitly. The boundaries between cases where a rule applied and cases where it did not were often fuzzy, dependent on subtle features of the situation that were hard to formalise.

Knowledge was also inconsistent. Different experts had different knowledge, and their rules sometimes conflicted. Reconciling expert disagreements — deciding which version of a rule was correct, or how to handle cases where the rules gave contradictory recommendations — required judgments that went beyond what any individual expert could provide.

And knowledge was dynamic. Medical knowledge, for instance, changed as new research was published, new treatments became available, and clinical experience accumulated. An expert system’s knowledge base became outdated as the domain evolved, and updating it was expensive and time-consuming. The knowledge engineer had to be brought back in, the experts had to be interviewed again, the rules had to be revised. This maintenance overhead was not adequately anticipated in the early enthusiasm for expert systems.

The knowledge acquisition bottleneck — the difficulty and expense of extracting expert knowledge and encoding it in a machine-readable form — became the central practical problem of the expert systems era. It limited the scale and scope of expert systems that could be built, inflated the cost of their development, and made their maintenance difficult. It also, ultimately, contributed to the second AI winter: the cost of building and maintaining large knowledge bases proved to be higher than the commercial value they generated for many applications, and the market contracted when this became apparent.


The Breadth of the Deployment: AI Goes Everywhere

Despite the difficulties, expert systems were developed and deployed across an extraordinary range of domains in the 1980s. The breadth of the deployment is itself historically significant — it demonstrated both how widely the approach could be applied and how varied its results were in different domains.

Medicine. MYCIN and its successors were the paradigmatic medical expert systems, but dozens of others were developed for different specialties. Systems for interpreting electrocardiograms, for diagnosing genetic disorders, for planning radiation therapy for cancer, for guiding the selection of drug dosages — all were developed and, in some cases, deployed in clinical settings. The medical domain was attractive for expert systems because the knowledge was well-structured, the rules of inference were relatively explicit, and the potential value of improving the consistency and availability of expert-level diagnosis was enormous.

Financial services. Banks and financial institutions developed expert systems for credit scoring, loan evaluation, fraud detection, and investment advice. These systems encoded the judgment of experienced credit officers and financial analysts in forms that could be applied consistently and quickly to large numbers of cases. The financial domain was well-suited to expert systems because many financial decisions were indeed rule-based — following explicit regulations, applying established criteria — and consistency and speed were commercially valuable.

Engineering and manufacturing. XCON was the paradigmatic engineering expert system, but similar systems were developed for many other engineering domains: diagnosing faults in complex machinery, planning manufacturing processes, designing electrical circuits, configuring complex systems of many kinds. The engineering domain was attractive because engineering knowledge was often highly formalised — engineers documented their knowledge in standards, specifications, and design rules — and because errors in engineering systems were expensive.

Legal and regulatory. Expert systems were developed to assist with legal research, to apply regulatory rules to specific cases, to advise on tax strategy. The legal domain was in some ways ideal for expert systems: law consisted largely of explicit rules, and the application of those rules to specific cases was precisely the kind of task that rule-based expert systems did well. The limitations appeared at the edges — where the law was ambiguous, where competing principles had to be balanced, where the spirit and the letter of the law diverged.

Geology and resource exploration. PROSPECTOR, developed at SRI in the late 1970s, was a geological expert system designed to assess the likelihood of mineral deposits in a given area. The system encoded the knowledge of expert geologists about the geological indicators of different types of ore deposits, and it was used in the assessment of sites for mineral exploration. In 1981, PROSPECTOR correctly identified a molybdenum deposit in Washington State that had not been previously identified — a result that attracted significant attention and demonstrated the potential value of expert systems in resource exploration.

The breadth of deployment was genuinely impressive, and it established AI as a commercially important technology for the first time. But it also exposed, across many domains, the fundamental limitations of the expert systems approach.


The Limitations Emerge: What Expert Systems Could Not Do

As expert systems were deployed at scale across many domains, their limitations became increasingly apparent. The limitations were not random or unpredictable — they flowed directly from the fundamental assumptions of the approach.

Brittleness at domain boundaries. Expert systems performed well within their intended domains and failed catastrophically at the edges. A medical expert system designed for bacterial infections performed well on bacterial infection cases and produced nonsense on cases that fell outside its design — cases of viral infection, of unusual presentations, of combinations of conditions that the system’s designers had not anticipated. Human experts, encountering an unusual case, recognised that it was unusual, sought additional information, consulted colleagues, applied general medical knowledge. Expert systems had no such fallback. They either applied their rules, regardless of whether the case fit the rules’ assumptions, or they produced no recommendation — neither of which was satisfactory.

The knowledge acquisition bottleneck. As discussed above, the extraction of expert knowledge was slow, expensive, and often incomplete. Expert systems could only know what their knowledge engineers had successfully extracted, and the tacit, contextual, and inconsistent nature of human expertise made complete extraction essentially impossible. The gap between what was in the knowledge base and what a human expert would know in practice was a permanent feature of expert systems, not a problem that would be solved by more careful knowledge engineering.

Lack of learning. Expert systems could not learn from experience. Their knowledge bases were static — fixed at the time of development and updated only through expensive manual revision. Human experts, by contrast, learned continuously — updating their knowledge with every case they encountered, integrating new research findings, refining their judgment based on experience. An expert system deployed in 1985 knew exactly what it had known in 1980 unless its developers had explicitly updated it. This inability to learn meant that expert systems became increasingly outdated over time, and their maintenance was a continuing cost that many organisations underestimated.

Combinatorial explosion in complex domains. In highly complex domains with many interacting variables, the number of rules required to capture all relevant expertise grew combinatorially. XCON eventually contained more than ten thousand rules — a knowledge base so large that maintenance became extremely difficult and the interactions between rules were hard to manage. Rules that were individually correct could interact in unexpected ways to produce incorrect conclusions. Debugging a large rule-based system required understanding not just individual rules but the complex web of their interactions, a task that became increasingly intractable as the knowledge base grew.

No deep understanding. Expert systems encoded surface knowledge — the rules that experts applied — without any understanding of the principles that generated those rules. A medical expert system knew that patients with symptom A and test result B probably had condition C, but it did not know why that was true, what the biological mechanism was, why the correlation existed. This lack of deep understanding meant that expert systems could not reason by analogy, could not recognise new patterns based on underlying principles, could not adapt to genuinely novel situations. Human experts could; expert systems could not.


The Commercial Boom: The Late 1980s at Peak Expert Systems

The period from approximately 1984 to 1988 was the commercial peak of the expert systems era — a period of rapid growth, significant investment, and genuine commercial success that had no precedent in the history of AI.

The market for expert systems software and services grew from effectively zero in the late 1970s to several hundred million dollars annually by the mid-1980s and was projected to reach billions by the early 1990s. Companies including Intellicorp, Teknowledge, Inference Corporation, and Carnegie Group were founded specifically to develop and sell expert systems technology. IBM, DEC, Texas Instruments, and other established technology companies launched AI divisions and invested heavily in expert systems development.

The growth was driven partly by genuine commercial value — systems like XCON demonstrated that the technology could produce real returns — and partly by the hype that always accompanies new technology with genuine promise. Corporations that had no specific application in mind invested in expert systems because they did not want to be left behind. Universities expanded their AI programmes to meet demand from students who saw AI as the hot field to be in. Venture capital flowed into AI startups.

The Japanese Fifth Generation Computer Project, which we will discuss in a later article, intensified the investment by creating the impression of a national competition: if Japan was investing billions in AI, the United States could not afford to lag behind. DARPA, responding to this perceived competitive threat, launched the Strategic Computing Initiative in 1983, which channelled hundreds of millions of dollars into AI research with an explicit military focus. The combination of corporate investment, venture capital, and government funding made the mid-1980s the most generously funded period in AI history to that point.

The quality of the applications being developed varied enormously. Some expert systems were genuine successes — XCON, the medical diagnosis systems that performed at expert level, the geological exploration systems that identified valuable mineral deposits. Others were oversold, underperforming systems that had been rushed to market on the strength of the enthusiasm rather than the results. The gap between the best expert systems and the average was large, and the average was often not impressive enough to justify the investment.


LISP Machines: The Hardware That Built an Industry

One of the distinctive features of the expert systems era was the development of specialised hardware designed specifically to run LISP — the programming language of AI research. These LISP machines were some of the most sophisticated computers of their era and, for a brief period, a booming business.

LISP, as McCarthy had designed it, was computationally expensive to run on conventional computer hardware. Its dynamic memory management, its support for recursion, and its symbolic processing requirements all required significant overhead on machines designed primarily for numerical computation. By the late 1970s, the demand for AI computing was sufficient to justify the development of hardware specifically designed for LISP — machines with specialised instruction sets, hardware support for garbage collection, and architectures optimised for symbolic processing.

The major LISP machine manufacturers — Symbolics, LISP Machine Inc. (LMI), and Xerox PARC, which produced the Interlisp-D machines — produced hardware that was, for AI applications, dramatically faster than the general-purpose minicomputers and workstations of the era. A Symbolics 3600, running in 1984, could execute LISP code at speeds that were an order of magnitude faster than a DEC VAX running the same programs.

The LISP machine companies attracted significant investment and grew rapidly through the early 1980s. Symbolics, the most successful, had hundreds of employees and revenue in the tens of millions of dollars. The machines they sold were expensive — a Symbolics workstation cost more than a new car — but they were the best available tool for AI development, and organisations that were serious about building expert systems needed them.

The LISP machine era ended abruptly in the late 1980s when the rapidly improving performance of general-purpose workstations — Sun Microsystems, Apollo, and eventually high-end personal computers — made the specialised machines economically uncompetitive. A Sun workstation running Common LISP could do, by 1987 or 1988, much of what a Symbolics machine had been needed for in 1984, at a fraction of the cost. The LISP machine manufacturers were caught in the classic innovator’s dilemma: their specialised hardware was overtaken by the general-purpose hardware that their own success had helped drive to greater capability.

The collapse of the LISP machine market was one of the early signals that the expert systems boom was approaching its limits. The companies that had built their businesses around the assumption of continuing AI investment were about to face a market that was contracting rather than growing.


The Seeds of the Second Winter

Even at the peak of the expert systems boom, the seeds of its eventual collapse were visible to those who looked carefully.

The fundamental problem was the knowledge acquisition bottleneck. Expert systems were expensive to build because knowledge engineering was expensive. They were expensive to maintain because updating knowledge bases was expensive. And the commercial value they produced was often not sufficient to justify these costs, particularly for applications that were less clearly structured than XCON or less urgently needed than MYCIN’s diagnostic capabilities.

Many organisations had invested in expert systems during the boom years not because they had specific applications that clearly benefited from the technology but because they did not want to be left behind. These investments produced systems that worked adequately in demonstration but were never deployed at production scale, systems that solved problems that were not actually important, systems that were overtaken by events before they could be deployed.

The gap between the demonstrations and the deployments was large. The most impressive expert systems demonstrations — XCON saving DEC forty million dollars a year, PROSPECTOR identifying a mineral deposit, MYCIN diagnosing bacterial infections at expert level — were achievements of years of careful knowledge engineering applied to well-chosen problems by skilled researchers. The average expert systems deployment — rushed to market by a startup, applied to a problem that was less well-structured than the showcase cases — was not nearly as impressive.

As the boom peaked and began to decline, the mismatch between expectations and results became increasingly apparent. Companies that had invested millions in expert systems technology began to question the returns. The growth projections that had justified the investments were not being met. The promised productivity improvements were not materialising at the scale that had been claimed. The maintenance costs were higher than anticipated, the expert systems were becoming outdated faster than they could be updated, and the range of problems they could handle was more limited than the demonstrations had suggested.

By 1987, the signs of trouble were clear. The LISP machine market was collapsing. Several expert systems companies were struggling. The growth projections for the AI industry were being revised downward. The second AI winter was beginning.


What Expert Systems Contributed: The Permanent Legacy

Despite the collapse, the expert systems era left a permanent and important legacy — contributions to AI and to practical computing that outlasted the specific technology and the specific era.

Knowledge representation. The expert systems era produced the most extensive practical experience with knowledge representation that AI had yet had. The limitations that were discovered — the difficulty of capturing tacit knowledge, the brittleness at domain boundaries, the maintenance problem — were genuinely important lessons. They shaped the development of more sophisticated knowledge representation frameworks: ontologies, description logics, semantic networks, and the formal knowledge representation languages that are still in use in AI systems today.

The economics of applied AI. XCON and a handful of other successful expert systems demonstrated that AI could produce measurable commercial value. This demonstration was important even after the specific technology became obsolete, because it established in the corporate imagination the idea that AI was a commercially relevant technology. When the next wave of AI capability arrived — machine learning in the 2000s, deep learning in the 2010s — companies and investors were already primed to think about commercial applications. The expert systems era established the template.

The knowledge engineering methodology. The practice of knowledge engineering — the careful extraction and formalisation of domain expertise — developed a rigorous methodology during the expert systems era that has applications beyond rule-based AI. The techniques developed for interviewing experts, documenting their reasoning, validating knowledge bases against expert performance, and maintaining knowledge bases through the evolution of a domain have proved useful in subsequent AI development, including in the curation of training data for machine learning systems.

Verification and validation. Expert systems, because their knowledge was explicit and their reasoning was transparent, could be systematically verified and validated. The methodology for testing expert systems — comparing their recommendations to expert consensus, identifying cases where they failed, tracing the reasoning that led to failures, correcting the underlying rules — was developed during the expert systems era and has influenced the testing and evaluation of AI systems ever since.

Domain-specific AI. Perhaps the most durable legacy of the expert systems era is the demonstration that domain-specific intelligence was achievable and commercially valuable, even in the absence of general intelligence. This lesson — that you did not need to solve the general problem to build useful, valuable AI systems — has shaped every subsequent wave of AI development. The natural language processing systems, the computer vision systems, the recommendation systems, the fraud detection systems, and all the other domain-specific AI applications that followed were all, in a broad sense, applications of the expert systems lesson: go narrow, be specific, focus on where you can add genuine value.


The Medical Expert System Question: What Should Have Been Done

One of the most interesting and most troubling aspects of the expert systems era is the story of medical expert systems — systems that were technically successful by the criteria of the time, that demonstrated genuine expert-level performance, and that were never widely deployed in clinical practice.

MYCIN is the paradigmatic case. It worked. Its recommendations were evaluated by expert physicians and found to be at or above the level of specialists in the domain it was designed for. The technology existed in 1980 to deploy it in clinical settings and potentially improve the quality of antibiotic prescribing. It was not deployed, primarily for reasons that had nothing to do with its technical performance.

The barriers to clinical deployment were institutional and liability-related. Physicians were not comfortable with the idea of a computer program making clinical recommendations, even in an advisory capacity. Hospital administrators were concerned about liability — if the program made an incorrect recommendation and a patient was harmed, who was responsible? Regulators had no framework for evaluating or approving medical AI systems. The clinical institutions, the legal system, and the regulatory environment were not ready for AI in clinical practice, even when the technology was ready.

This story has a melancholy resonance with the present. Decades later, the regulatory, liability, and institutional barriers to deploying AI in clinical practice remain significant obstacles, even though the technology has advanced enormously. The question of how to evaluate, approve, and deploy AI in clinical settings — how to assign responsibility when AI makes an incorrect recommendation, how to ensure that AI systems are used as supplements rather than substitutes for clinical judgment — is still being worked out.

The medical expert systems story is a reminder that technology deployment is not just a technical problem. It is also an institutional, regulatory, and social problem, and the non-technical barriers can be as significant as the technical ones.


The Transition: From Expert Systems to Machine Learning

The transition from the expert systems era to the machine learning era was not a clean break — it was a gradual shift in the balance of attention and investment, driven by the accumulating evidence of expert systems’ limitations and the growing evidence of machine learning’s potential.

The key insight that drove the transition was simple but profound: the knowledge acquisition bottleneck could be bypassed if you could learn the knowledge from data rather than extracting it from humans. A machine learning system trained on a large dataset of examples — medical cases with known diagnoses, financial transactions with known outcomes of fraud or legitimacy, chemical compounds with known properties — could potentially acquire the same knowledge that a knowledge engineer would spend months extracting, without the expensive and imperfect process of human extraction.

This insight was not new — Samuel’s checkers program had demonstrated it in 1952, and the neural network researchers who kept working through the AI winters had been pursuing it consistently. What was new in the late 1980s and 1990s was the combination of growing datasets, improving algorithms, and increasing computing power that made machine learning practically effective in a wider range of domains.

The transition was gradual and uneven. Expert systems were not replaced overnight — many continued to be maintained and used for years, and some are still in use today. But the investment in new expert systems development declined steadily through the late 1980s and early 1990s as the investment in machine learning increased. The researchers who had built careers in knowledge engineering shifted their attention toward machine learning. The companies that had been built around expert systems technology failed or pivoted. The journals that had published expert systems research shifted their attention toward learning-based approaches.

By the mid-1990s, the transition was effectively complete: machine learning had become the dominant approach to practical AI, and expert systems were a legacy technology maintained for existing deployments rather than a frontier being actively developed.

The expert systems era had demonstrated that AI could be commercially valuable. It had demonstrated that domain-specific intelligence was achievable. And it had demonstrated, through the knowledge acquisition bottleneck, exactly the problem that machine learning would eventually solve. The two eras were connected by the problem that one identified and the other addressed.


The Broader Lesson: What the Expert Systems Era Taught

The expert systems era taught several lessons that have shaped AI development ever since, including in ways that are directly relevant to the current moment.

The most important lesson was about the relationship between impressive demonstrations and practical deployment. Expert systems that worked brilliantly on carefully chosen demonstration cases often failed to scale to production use. The gap between the best expert systems and the average was large, and the gap between demonstrations and full-scale deployment was larger still. This lesson — that impressive demonstrations do not automatically translate into practical value, that scaling from demonstration to deployment involves challenges that are often underestimated — is one that every subsequent wave of AI has had to relearn.

The second lesson was about maintenance and brittleness. AI systems that work at deployment time need to continue working as the world changes, as the knowledge they encode becomes outdated, as they encounter cases outside their design scope. Expert systems were expensive to maintain and brittle in practice. This lesson has not been fully solved in subsequent AI approaches — large language models also become outdated as the world changes, also fail on cases outside their training distribution — but it has been more clearly recognised as a critical practical challenge.

The third lesson was about the economics of AI. Commercial success in AI required not just technically capable systems but economically viable development and deployment processes. The knowledge acquisition bottleneck made expert systems economically unviable for many applications. The lesson — that AI systems need to be not just capable but cost-effective — is fundamental and enduring.

The expert systems era was not a failure. It was a generation of commercial AI that achieved real results, demonstrated real value, and left a real legacy. It was also a demonstration of the limits of one approach — rule-based, explicitly encoded, static knowledge — and an indication of what a better approach would need to provide. It prepared the field, intellectually and institutionally, for what came next.


Further Reading

  • “The Rise of the Expert Company” by Edward Feigenbaum, Pamela McCorduck, and H. Penny Nii (1988) — Written at the peak of the expert systems era, this book captures the excitement and the ambition of the period, with case studies of successful applications. An essential primary source.
  • “Building Expert Systems” by Hayes-Roth, Waterman, and Lenat (1983) — The definitive technical guide to expert systems development during the boom era. Comprehensive and rigorous.
  • “The Art of Artificial Intelligence: Themes and Case Studies of Knowledge Engineering” by Feigenbaum and McCorduck — A broader treatment of the knowledge engineering approach and its foundations.
  • “Computers and Thought” edited by Feigenbaum and Feldman (1963) — The anthology that captured the early optimism about AI before the expert systems era, providing essential context for understanding how the approach developed.
  • “The Handbook of Artificial Intelligence” edited by Barr and Feigenbaum (1981–1982) — A comprehensive reference to AI as it stood at the beginning of the expert systems boom. Valuable for understanding the intellectual context.

Next in the Events series: E8 — Japan’s Fifth Generation Project, 1982: The Billion-Dollar Gamble — The most audacious AI project in history: Japan’s announcement that it would build a new generation of computers based on artificial intelligence, the global panic it triggered in the United States and United Kingdom, and the spectacular, quiet failure that nobody wanted to acknowledge. How national ambition, geopolitical anxiety, and genuine scientific enthusiasm combined to produce one of the great misadventures of the technology era.


Minds & Machines: The Story of AI is published weekly. If the story of expert systems — the real achievements, the real limitations, and the gap between them — resonates with AI conversations you are having today, share this with someone who would find the parallel instructive.