New York City. May 11, 1997. The forty-fourth floor of the Equitable Center on Seventh Avenue. A room set up with a chess board, two chairs, cameras, and the kind of hushed, tense atmosphere that settles over rooms where something important is about to happen.

On one side of the board sits Garry Kasparov, thirty-four years old, the reigning World Chess Champion since 1985, widely regarded as the greatest chess player in human history. He has won every major chess tournament he has entered for the past twelve years. He plays chess with a ferocity and a creativity that has made him famous beyond the chess world, a celebrity in Russia and increasingly in the West, a man who carries the weight of representing human intellectual supremacy against what he and the world understand to be a challenge from the machines.

On the other side of the board sits nobody. The chair is empty. The moves for the other side will be made by a human operator following the instructions of a computer: Deep Blue, built by IBM, filling several refrigerator-sized cabinets elsewhere in the building, capable of examining two hundred million chess positions per second.

It is Game 6 of a rematch between Kasparov and Deep Blue. The score is tied at two and a half points each. The winner of this game will win the match.

Kasparov, playing white, opens with 1. e4. Deep Blue, playing black, responds with 1… c6 — the Caro-Kann Defence, a solid, classical opening. The game begins.

Forty-five moves later, Kasparov resigns. He knocks over his king. He walks off the stage.

The machine has won. History has changed.


Before the Match: The Road to 1997

The 1997 match between Deep Blue and Kasparov was not the first encounter between man and machine at the chess board, and it was not the first time a computer had defeated a world-class human player. But it was the first time a computer defeated the reigning world champion in a formal match under standard chess conditions — and it was this specific, symbolic threshold that made it one of the most culturally significant events in the history of artificial intelligence.

To understand the significance, you need to understand how chess had been positioned in the AI story.

Chess had been a touchstone for AI research since the field’s founding. Shannon’s 1950 paper “Programming a Computer for Playing Chess” had established chess as the paradigm problem for AI — complex enough to require something resembling intelligence, well-defined enough to be tractable as a computational problem. Simon’s famous prediction in 1957 that a computer would be the world chess champion within ten years had made chess the most cited example of AI’s coming capabilities and its most visible failure when the prediction was not met.

The failure of the ten-year timeline had made chess almost embarrassing for AI — a problem the field had confidently claimed it was about to solve, that proved stubbornly resistant for three decades. Chess programs improved steadily through the 1960s, 1970s, and 1980s, but the gap between the best programs and the best human players remained substantial. Programs that could defeat strong amateurs could not compete with grandmasters; programs that could compete with grandmasters could not approach the level of the world’s elite players.

By the late 1980s, the gap was narrowing. Deep Thought, developed at Carnegie Mellon by Feng-hsiung Hsu, Murray Campbell, and their collaborators, had achieved a rating of approximately 2550 — in the range of the world’s strongest grandmasters, though still clearly below the level of the absolute elite. IBM, recognising the potential for a public demonstration of computing capability, recruited Hsu and Campbell and gave them the resources to build a much stronger system: Deep Blue.

Deep Blue was not a fundamentally different system from Deep Thought — it used the same basic approach, the same architecture, the same evaluation function paradigm. What IBM provided was brute force at a previously impossible scale: specialised chess chips that could examine 200 million positions per second, running in parallel across a massively powerful hardware platform. The system searched game trees to depths of six to eight moves — sometimes deeper in critical positions — examining all possible sequences of moves and evaluating the resulting positions with a sophisticated evaluation function developed in collaboration with grandmaster consultants.

The first match between Kasparov and Deep Blue took place in February 1996. Deep Blue won Game 1 — the first time a world chess champion had been defeated by a computer under tournament conditions. Kasparov recovered, adapting his play to exploit the computer’s weaknesses, and won the match 4-2. He was shaken by the first game but confident after the match that he understood the computer’s limitations well enough to manage it.

IBM spent the year between the matches upgrading Deep Blue significantly — increasing its speed, improving its evaluation function, incorporating new opening book knowledge. When the rematch began in May 1997, the system Kasparov faced was substantially stronger than the one he had defeated in 1996.


The Players: Kasparov and the Machine

Garry Kimovich Kasparov was born in 1963 in Baku, Azerbaijan, then part of the Soviet Union. He learned chess from his father at the age of five, became a Soviet youth champion, and by his early twenties was recognised as one of the most gifted chess players in history. His 1985 victory over Anatoly Karpov in the World Championship match — a victory achieved at twenty-two, making him the youngest world champion in history — announced the arrival of a player of exceptional originality and ferocity.

Kasparov’s chess was characterised by dynamic, aggressive play, a willingness to take risks in return for complex positions where his superior calculation and imagination could be decisive, and an astonishing ability to generate creative ideas in seemingly impossible positions. He was not the most technically precise player of his era — there were players with deeper memorisation of opening theory and more accurate calculation in simple positions. But in complex, unclear positions — positions where the evaluation was uncertain and the paths forward multiple — Kasparov was without peer.

He also had a personality that matched his chess: intense, competitive, emotionally engaged, unwilling to accept defeat gracefully. His match against Deep Blue was not, for Kasparov, simply a chess competition. It was a contest with existential stakes — a question of whether human creativity and intelligence could prevail against machine calculation, a question of whether there was something distinctively human about the kind of understanding that great chess required.

Deep Blue, by contrast, was defined entirely by its computational characteristics. It examined approximately two hundred million positions per second — orders of magnitude more than any human player. It had perfect recall of its opening book, which contained millions of grandmaster games. It evaluated positions using a carefully tuned function that incorporated dozens of factors — material balance, piece activity, king safety, pawn structure, control of key squares — with weights that had been calibrated by grandmaster consultants.

What Deep Blue did not have was anything resembling chess understanding in the sense that a human grandmaster would recognise. It did not see the board as a narrative — as a story of pieces developing toward an objective, of plans being formed and executed, of one player’s initiative being seized or lost. It saw the board as a state to be evaluated and a tree of possible future states to be searched. The evaluation was based on features that grandmasters had identified as important, but the program did not understand why those features were important or how they connected to the broader themes of a chess game.

Kasparov understood this, and had structured his 1996 strategy partly around exploiting the computer’s lack of understanding. He played positions that were strategically rich but tactically quiet — positions where human understanding of long-term plans was more valuable than tactical calculation depth. The strategy had worked in 1996. It would prove less effective in 1997.


The 1997 Match: Game by Game

Game 1 was played on May 3, 1997. Deep Blue played 1. e4, and Kasparov responded with the Ruy Lopez — one of the oldest and most deeply analysed openings in chess. The game developed along standard theoretical lines until move 26, when Deep Blue played a move that surprised Kasparov: Rxf6, sacrificing a rook for a bishop and knight in a position that most grandmasters would have evaluated as offering insufficient compensation.

Kasparov assumed the sacrifice was wrong — that the computer had made a miscalculation. He accepted the rook and played on, expecting to win. But the position proved more complex than he had anticipated, and Deep Blue’s compensation turned out to be adequate. Kasparov eventually had to accept a draw.

The draw itself was not significant. But Kasparov’s assessment — that the computer had made a mistaken sacrifice — was wrong, and his wrongness was significant. He had misjudged Deep Blue’s capabilities. The computer’s positional sacrifice, which looked wrong to Kasparov, was actually a sound strategic idea that his analysis had failed to refute. He went into the rest of the match with a slightly miscalibrated sense of what he was facing.

Game 2 was played on May 4, 1997, and it produced the most dramatic and most discussed single chess game in the match.

Deep Blue, playing white, opened with 1. d4. Kasparov responded with the Nimzo-Indian Defence — a sophisticated opening that he had played many times. The game developed normally until move 36, when Deep Blue played Bd6 — a bishop move that Kasparov later said he found deeply shocking.

The move was not obviously correct. In fact, it looked potentially dubious — the bishop was moving to a square where it appeared to have limited activity. But the deeper calculation showed that the move was part of a plan that would eventually improve Deep Blue’s position significantly. Kasparov, unable to find a refutation, played on, but he was psychologically unsettled.

What disturbed Kasparov about Bd6 was not its tactical depth — he was accustomed to deep tactical calculations. What disturbed him was its strategic character. The move did not look like what he expected from a computer. It looked like the kind of deep, quiet, strategically motivated move that a human grandmaster with profound positional understanding might play — not the sharp, tactically oriented moves that he had expected from a machine relying primarily on calculation.

Kasparov continued to play but his confidence was visibly shaken. He made an uncharacteristic positional error on move 45 that allowed Deep Blue to gain a decisive advantage, and he resigned on move 50.

The defeat was significant in itself, but Kasparov’s reaction to it was more significant still. He accused IBM of cheating. He claimed that the Bd6 move was too sophisticated for a machine to have found by calculation — that it must have been suggested to the machine by a human grandmaster with access to the position in real time. IBM denied the accusation and provided logs of Deep Blue’s calculation. Independent analysis suggested that the move was within Deep Blue’s calculational reach — that it had found the move through search, not through human input.

Kasparov’s accusation was not accepted by the chess community or by the broader public, and it damaged his credibility in the subsequent controversy. But the accusation was itself revealing: it showed how profoundly the Game 2 loss had disrupted Kasparov’s mental model of what Deep Blue was doing. He had expected a machine that calculated tactically and struggled strategically. Game 2 showed him a machine that appeared to understand strategy as well as tactics. The experience was, for Kasparov, deeply disorienting.

Game 3 was drawn after fifty hours of play. Kasparov played solidly and found enough resources to hold the draw, but the game felt like a defensive achievement rather than an offensive effort.

Game 4 was also drawn, after a complicated middlegame in which both players had chances. Kasparov’s play was more confident in this game, suggesting some recovery of equilibrium.

Game 5 was a quick draw — a quiet game that ended after only forty-five minutes. Kasparov seemed to be playing for psychological recovery rather than for a win.

Game 6, the decisive game, was played on May 11, 1997. Kasparov, needing a win, played aggressively as white. But his opening preparation failed him — he went wrong in a theoretical position and found himself in a difficult endgame. Deep Blue’s endgame technique was flawless. On move 19, after Deep Blue played a precise exchange that simplified to a winning position, Kasparov — rather than continuing to play a lost position — resigned.

The resignation was abrupt and came in a position where many analysts felt he could have played on for a significant number of additional moves, even if the position was lost. Kasparov later explained that he had lost the will to continue — that the psychological weight of the match had become too great. Whether this was the right decision is debated among chess analysts, but it was a very human decision — one that no computer would have made, because computers do not have a will to continue or lose.

The final score was Deep Blue 3.5, Kasparov 2.5. The machine had won.


Kasparov’s Accusations: The Controversy That Would Not Die

Immediately after the match, Kasparov’s accusations of cheating became the dominant public narrative surrounding the result. He alleged that IBM had violated the rules of the match by allowing human grandmasters to modify Deep Blue’s opening book between games, effectively giving the computer human assistance in real time.

The specific accusation focused on Game 2 and the Bd6 move. Kasparov believed the move was too strategically sophisticated for a machine to find without human guidance — that it bore the hallmarks of human positional understanding rather than machine calculation.

IBM denied the accusations and provided extensive documentation of Deep Blue’s decision-making process. The match rules had specified that IBM was permitted to modify the program between games, which was standard practice in computer chess competitions — adjustments to the opening book and evaluation function were considered part of normal system maintenance. IBM maintained that all modifications had been within the scope of what the rules permitted.

The chess community was divided. Some grandmasters found Kasparov’s accusation plausible — they too found Bd6 surprisingly human-seeming, and they were aware of the competitive pressures on IBM that might have motivated improper assistance. Others found the accusation unfounded — subsequent analysis suggested that Bd6 was within Deep Blue’s calculational reach, and the accusation seemed to reflect Kasparov’s psychological state after the loss more than any genuine evidence of wrongdoing.

The controversy was never fully resolved in a formal sense. IBM did not provide complete access to Deep Blue’s logs, citing proprietary interests, which gave ammunition to those who suspected something had been concealed. Kasparov did not retract his accusation. The match was never replayed, as Kasparov requested, partly because IBM had no incentive to replay — they had won, and a replay created the risk of losing the result they already had.

The controversy had several long-term effects. It complicated the cultural significance of the match by casting doubt on whether the result was fully legitimate. It damaged Kasparov’s public image, making him appear a poor loser willing to make unsubstantiated accusations. And it prevented the kind of clean, universally accepted verdict that both AI and chess communities might have benefited from.

From a historical perspective, the accusations were almost certainly wrong about IBM cheating. Deep Blue’s play throughout the match was consistent with a very strong program playing at the limit of its capabilities — there were games where it made moves that were clearly suboptimal, games where its positional understanding was limited in ways that human grandmasters exploited, and games where its calculation was impressive but not inexplicable. The match as a whole looked like what it was supposed to be: the best chess-playing computer in history defeating the best human chess player in history in a six-game match.

Whether the specific move Bd6 in Game 2 was within Deep Blue’s independent calculational reach is a technical question that has been studied in detail by chess computer experts. The consensus view is that it was — that Deep Blue could and did find the move through search, without human assistance. But the complete logs were never published, and the consensus is based on reconstruction rather than definitive evidence.


What Deep Blue Actually Was: Inside the Machine

Understanding what Deep Blue actually was — what it was doing when it defeated Kasparov — is essential for understanding what the victory meant and what it did not mean.

Deep Blue was, at its core, a very large, very fast, very well-implemented version of the approach that Shannon had described in 1950: search the game tree to some depth, evaluate the leaf positions, choose the move that leads to the best evaluated position. The specific implementation was enormously sophisticated — the search used a number of advanced pruning techniques to focus computation on the most important parts of the tree, the evaluation function incorporated dozens of carefully tuned factors, the opening book contained millions of grandmaster games, and the hardware was purpose-built for speed. But the fundamental approach was Shannon’s.

What Deep Blue did not have was chess understanding in the sense that human grandmasters had it. It did not see the chess position as a narrative with themes and plans. It did not recognise positional patterns as instances of types — the way a grandmaster would say “this is a hedgehog structure, and in hedgehog structures the key plan is to break in the center with …d5.” It evaluated positions based on quantified features, not on qualitative understanding.

This distinction — between the deep pattern-based understanding of a human grandmaster and the quantified-feature evaluation of a chess program — was what Kasparov had tried to exploit in both matches. He attempted to create positions where human understanding of complex plans and strategic themes was more valuable than the computer’s ability to calculate specific tactical lines. In 1996, this strategy worked. In 1997, Deep Blue’s evaluation function was sophisticated enough to handle these strategic complexities better than Kasparov had anticipated.

Deep Blue’s evaluation function had been developed with significant input from grandmaster consultants, particularly Joel Benjamin, who served as the primary chess consultant for the 1997 match. The consultants helped identify the positional features that mattered, how they should be weighted, and how the function should trade off different considerations. In this sense, Deep Blue’s “strategic understanding” was actually the strategic understanding of grandmasters, encoded in the evaluation function — a form of expert systems thinking applied to chess evaluation.

The relationship between Deep Blue’s evaluation function and genuine chess understanding is philosophically interesting. The evaluation function was a compression of human chess wisdom into a mathematical formula. It captured the insights of grandmasters about what made positions good or bad, but it captured them in a form that could be computed mechanically without any ongoing engagement with the reasons those insights were correct. It was, in a sense, a very sophisticated form of the expert systems approach applied to chess: human expertise encoded in explicit rules (the evaluation function factors) applied by a mechanical inference engine (the search algorithm).


The Cultural Earthquake: How the World Responded

The reactions to Deep Blue’s victory revealed a great deal about how humans understood the relationship between intelligence and computing — and how the relationship between the two was being renegotiated in real time by events like this match.

The reactions were varied and often contradictory.

Some responded with grief — a genuine sense of loss at the demonstration that machines could excel at a domain that had been considered distinctively human. Chess had been held up, for decades, as a paradigm of intellectual activity requiring creativity, insight, and the kind of deep understanding that was quintessentially human. If a machine could defeat the world’s best chess player, what did that mean for the special status of human intelligence?

Some responded with relief — at proof that the AI project was working, that the decades of research and investment were producing real results, that machine intelligence was not just a promise but an achievable reality. For those who had been working in AI through the winters and the disappointments, Deep Blue’s victory was validation.

Some responded with scepticism — arguing that chess was not really a test of intelligence, that a machine which searched two hundred million positions per second was not thinking in any meaningful sense, that the victory said nothing about whether machines could match human intelligence in the domains that actually mattered.

Some responded with wonder — at the specific beauty of the competition, at the drama of watching the world’s greatest chess player face a machine that did not care about winning or losing, that had no ego and no history and no fear.

The range of reactions reflected the ambiguity of what had actually happened. Deep Blue had defeated Kasparov at chess. But what this meant for the broader question of machine intelligence was genuinely unclear, and the clarity of the outcome at chess did not translate into clarity about the deeper questions.


The Kasparov Question: Was It Cheating, or Was He Simply Beaten?

Returning to the most contested aspect of the match: was Kasparov right that something improper had occurred, or was he simply beaten by a better opponent?

The evidence, considered in full, suggests that Deep Blue won legitimately and that Kasparov’s accusations reflected psychological distress rather than genuine evidence of wrongdoing.

The specific accusation — that human grandmasters had provided real-time assistance during the match — is difficult to reconcile with what is known about how computer chess systems work and how the match was organised. The match rules specified procedures for modifying the program between games; what IBM did was within those rules. The accusation that Bd6 was too sophisticated for Deep Blue rests on the assumption that the move was beyond the search capabilities of the system — an assumption that subsequent analysis has not confirmed.

But the broader accusation — that IBM played the match to win at the expense of sportsmanship, that they maximised every advantage available to them, that the spirit of the competition as a genuine test of machine capability was compromised by commercial pressures — has more substance.

IBM did not disallow access to Deep Blue’s logs after the match. They were not required to by the match rules, but their refusal to provide complete transparency contributed to the atmosphere of suspicion. They disbanded the Deep Blue team and dismantled the computer after the match, preventing the kind of post-match analysis that might have definitively answered the questions Kasparov raised. These decisions looked, from the outside, like an organisation protecting a PR victory rather than a scientific result.

The commercial context of the match was also significant. IBM’s primary interest in Deep Blue was not scientific but commercial — the match was a marketing exercise, designed to associate IBM’s brand with the most impressive demonstration of computing capability that the company could produce. The Deep Blue project was funded as a marketing investment, not as a research programme. This context shaped how IBM managed the match and its aftermath in ways that prioritised maintaining the PR value of the result over providing the transparency that genuine scientific progress required.

Whether or not IBM cheated in the technical sense — whether or not any rule violation occurred — the manner in which they managed the match and its aftermath was not consistent with the ideals of open scientific inquiry. The match was a demonstration, not an experiment. The result, while real, was surrounded by circumstances that made it harder to draw the clear conclusions that a cleaner competition might have supported.


The Philosophical Significance: What Chess Reveals About Intelligence

The philosophical questions raised by Deep Blue’s victory were more interesting than the question of cheating, and they have been discussed and debated ever since.

The most fundamental question: did Deep Blue think?

The answer depends on what you mean by “think.” If thinking means producing intelligent-seeming outputs through information processing, then Deep Blue was thinking — it was processing information about the chess position and producing outputs (chess moves) that were highly effective within the domain of chess. If thinking means something more specific — if it requires conscious experience, subjective understanding, genuine comprehension of what the moves mean — then Deep Blue was not thinking in that sense.

The chess match threw this ambiguity into unusually sharp relief because chess had long been held to be the paradigm case of intellectual activity requiring the distinctively human kind of thinking. The assumption was that playing chess well required insight, creativity, understanding — not just calculation. Deep Blue played chess well without insight, creativity, or understanding in any philosophically interesting sense. What it had was extremely effective calculation.

This raised a question that the chess community found uncomfortable: if chess could be played at the world champion level through calculation alone, what did that mean for the claim that playing chess well required insight? Either the claim was wrong — chess could indeed be played excellently without anything like human insight — or Deep Blue was doing something that deserved to be called insight, despite the fact that it was implemented through calculation.

The second possibility is more interesting than it sounds. Perhaps insight, at the level of a chess grandmaster, just is very sophisticated pattern recognition and calculation — patterns learned through experience, calculations performed in the domain of chess-relevant information. If so, then the distinction between human chess understanding and machine chess calculation is a distinction about implementation, not about what is fundamentally happening. Both are implementations of effective chess decision-making; one is implemented in biological neural tissue, the other in silicon. The chess performance is the same.

This is a version of the functionalist position in philosophy of mind: that mental states are defined by their functional role — their relationship to inputs, outputs, and other mental states — rather than by their physical implementation. On a functionalist view, if Deep Blue was playing chess as well as a grandmaster, it was playing chess with something that deserved to be called chess intelligence, regardless of what was happening in the silicon underneath.

Most philosophers of mind find strict functionalism too simple — there are good arguments that functional equivalence in outputs does not guarantee mental equivalence, that there is something about subjective experience that is not captured by functional description alone. But the chess match was at least a challenge to the naive view that human chess performance was obviously different in kind from machine chess performance.


The AI Community’s Response: What the Match Meant for the Field

Within the AI research community, the reaction to Deep Blue’s victory was complex and somewhat divided.

For those working in the mainstream of AI — in machine learning, in statistical approaches, in the kinds of practical AI that was becoming commercially relevant in the late 1990s — the chess match was impressive but somewhat beside the point. Deep Blue was a domain-specific system, extraordinarily powerful in chess and completely useless for anything else. Its victory demonstrated the power of search and evaluation in a well-defined game, not the emergence of general machine intelligence. The techniques it used — specialised hardware, game tree search, hand-crafted evaluation functions — were not the techniques that were going to produce general AI.

For those who had been working on game-playing AI specifically, the victory was a milestone worth celebrating — the culmination of a research programme that had begun with Shannon’s 1950 paper and had been progressing steadily for nearly half a century. The victory demonstrated that the approach could scale to world champion level, that the combination of hardware power and algorithmic sophistication could overcome the gap between program and grandmaster that had seemed so persistent.

For the public, the match was a cultural event of genuine significance — a moment when the question of machine intelligence moved from abstraction to demonstration in a way that was visible, understandable, and emotionally engaging. Chess was a domain that people understood to require intelligence. A machine had beaten the best human. Whatever that meant philosophically, it meant something in the cultural imagination.

The cultural significance outlasted the specific technology. Deep Blue itself was quickly superseded — chess programs running on consumer hardware surpassed Deep Blue within a few years, and by the 2010s, the best chess programs were dramatically stronger than any human. The threshold that Deep Blue crossed in 1997 — defeating a world champion — was crossed so routinely by subsequent programs that it lost its symbolic weight. But the moment of crossing remained significant in memory.


The Aftermath: Kasparov’s Continued Fight

After the 1997 match, Kasparov continued to engage with the question of machine intelligence at chess and more broadly. He played matches against subsequent computer programs, using different strategies and continuing to explore the boundary between human and machine intelligence at chess.

In 2003, he played matches against Junior and Fritz — two strong commercial chess programs — with different rules than the Deep Blue matches. Both matches ended in draws, demonstrating that Kasparov could hold his own against strong programs even if he could not consistently win against them.

Kasparov also developed what he called “Advanced Chess” or “Centaur Chess” — a variant in which human players were assisted by computer programs, the human-computer team playing against other human-computer teams. In this format, Kasparov found that the combined intelligence of human and computer was stronger than either alone — that the human player’s strategic understanding and the computer’s tactical calculation were complementary rather than competing.

This observation — that human-machine collaboration could be stronger than either pure human or pure machine play — became one of Kasparov’s most interesting contributions to the AI conversation. He argued that the right response to increasingly capable AI was not to view humans and machines as competitors but as complementary partners — that the future of intelligence was collaborative rather than adversarial.

The observation has proven prescient in many domains beyond chess. The most effective uses of AI in medicine, in science, in engineering, in creative work, have generally been those that combine human judgment and domain understanding with machine processing power and pattern recognition — that use AI as a tool that amplifies human capability rather than replacing it.

Kasparov’s post-1997 intellectual evolution — from competitor against machines to advocate for human-machine collaboration — is itself one of the more interesting intellectual journeys in the story of AI and human intelligence.


The Legacy: What 1997 Actually Meant

The 1997 Deep Blue victory is sometimes described as the moment when machines first demonstrated that they could match or exceed human performance in a domain requiring intelligence. This is accurate but requires careful qualification.

What machines demonstrated in 1997 was that they could match or exceed human performance in chess through methods that were fundamentally different from how humans played chess. Deep Blue searched more positions in a second than Kasparov could search in his lifetime. It evaluated positions using dozens of quantified factors that grandmaster consultants had identified as relevant. It had perfect recall of millions of grandmaster games and had processed opening theory that no human could fully absorb.

The performance was real. The achievement was genuine. But it was achieved through means that had little in common with the deep pattern-based, experience-grounded understanding that characterised Kasparov’s chess mastery.

The deeper question — whether machines could replicate or equal the specific kind of intelligence that human chess masters exercised — was not answered by Deep Blue’s victory. Deep Blue showed that machines could win at chess. It did not show that machines could win at chess in the way humans won at chess — through the accumulation of deep positional understanding over years of experience, through the recognition of thematic patterns and strategic ideas, through the kind of creative, exploratory thinking that produced Kasparov’s most memorable games.

This distinction mattered because the question of AI was ultimately not about chess performance. It was about intelligence — about whether machines could develop the kinds of cognitive capabilities that made human intelligence so extraordinary and so consequential. Deep Blue’s victory was evidence that machines could produce human-level performance on specific tasks through methods that did not involve human-like intelligence. This was important to know. It was not the same as demonstrating human-like intelligence.

The 1997 match was, in this sense, both more and less significant than its cultural reception suggested. More significant, because it was a genuine demonstration that decades of AI research could produce systems that exceeded the best human performance in a domain that had been considered the paradigm of human intellectual activity. Less significant, because the method of that demonstration — computational brute force combined with encoded human expertise — was not the kind of artificial intelligence that the field was ultimately trying to build.


Deep Blue to AlphaZero: The Evolution Continues

The story of chess AI did not end with Deep Blue’s 1997 victory. It continued, and the continuation was more significant in some respects than the original breakthrough.

In the twenty years after 1997, chess programs improved continuously and substantially. Commercial programs like Fritz, Rybka, and Stockfish surpassed Deep Blue within a few years and continued to improve. By the 2010s, the best programs were playing chess at a level that made Deep Blue look modest — with better evaluation functions, better search algorithms, and better opening books than Deep Blue had.

The most significant development came in 2017, when DeepMind’s AlphaZero program defeated Stockfish — then the strongest chess program in the world — in a hundred-game match. AlphaZero won decisively.

What made AlphaZero historically significant was not its performance — by 2017, it was expected that a well-engineered AI system could outplay any chess program. What made it significant was its method. AlphaZero was trained entirely through self-play, starting from random play and improving through reinforcement learning, without any access to human games or human-designed evaluation functions. After approximately nine hours of training, it had become the strongest chess player — human or machine — that had ever existed.

AlphaZero’s chess was different from Deep Blue’s in a way that was immediately visible to grandmasters who analysed its games. Where Deep Blue’s play was characterised by tactical precision and straightforward exploitation of material advantages, AlphaZero’s play was characterised by dynamic, active piece placement, willingness to sacrifice material for long-term positional advantages, and a kind of creative, forward-looking strategic thinking that grandmasters described as more human-like than any previous computer chess.

This description — that AlphaZero played more “human-like” chess than its predecessors — is interesting and somewhat ironic. AlphaZero had never seen a human chess game. Its chess had emerged entirely from self-play, from millions of games against itself, with the only feedback being whether it won or lost. And yet its chess looked more like great human chess than Deep Blue’s did.

This suggests that some of what makes chess beautiful — the dynamic piece activity, the willingness to sacrifice for long-term advantage, the creative exploration of the position’s possibilities — is not a uniquely human way of playing chess. It is an effective way of playing chess that emerges when a system with sufficient capability explores the game deeply enough. Deep Blue found effective chess through calculation. AlphaZero found beautiful chess through learning. The difference was not just aesthetic — AlphaZero was significantly stronger.

The arc from Deep Blue to AlphaZero reflects the broader arc of AI development from the 1990s to the 2010s: from systems that encoded human expertise explicitly to systems that learned their expertise from experience. AlphaZero did not beat chess by knowing more chess than Deep Blue. It beat it by having learned chess more deeply, from a perspective unconstrained by human assumptions about how the game should be played.


The Question That Remains: Human and Machine Intelligence

The chess story, from Shannon’s 1950 paper through Deep Blue’s 1997 victory to AlphaZero’s 2017 dominance, illuminates a question that is still unresolved: what is the relationship between what machines are doing when they play chess at superhuman levels and what human grandmasters are doing when they play chess at their best?

The question matters because chess was supposed to be a test of intelligence — a domain where the specific qualities of human intellectual performance could be assessed and compared to machine performance. Deep Blue’s victory raised the possibility that machine performance could equal human performance through fundamentally different mechanisms. AlphaZero’s victory complicated the picture further by demonstrating that machine performance could be not just equivalent to but stylistically similar to the best human performance, while arising through a completely different learning process.

The most honest answer is that we do not know whether AlphaZero’s chess intelligence is the same kind of thing as human chess intelligence, similar to it, or completely different from it. We know that the performance is equivalent or better. We know that the learning process is different. We know that the resulting play is, to human eyes, recognisably chess-like in its style. Whether these observations add up to “similar intelligence” or “different intelligence producing similar outputs” is a question that the available evidence does not definitively answer.

This unresolved question is part of the enduring significance of the Deep Blue match. The 1997 match asked whether a machine could beat the world chess champion. The answer was yes. But the deeper question — what the machine was doing, and whether it was the same thing as what Kasparov was doing, and what either tells us about the nature of intelligence — remains open.

It is, in the end, the right question. And it is still being asked.


Further Reading

  • “Deep Thinking: Where Machine Intelligence Ends and Human Creativity Begins” by Garry Kasparov (2017) — Kasparov’s own retrospective on the match and on the broader question of human versus machine intelligence. Remarkably thoughtful and has evolved significantly from his 1997 position.
  • “Behind Deep Blue: Building the Computer That Defeated the World Chess Champion” by Feng-hsiung Hsu (2002) — The primary architect of Deep Blue’s account of how the machine was built. Honest and technically detailed.
  • “Kasparov versus Deep Blue: Computer Chess Comes of Age” by Monty Newborn (1997) — The most immediate scholarly account of the matches, written by a computer chess historian.
  • “Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm” by the AlphaZero team (2018) — The AlphaZero paper, showing how far chess AI came in the two decades after Deep Blue. Essential for understanding the full arc.
  • “Game Over: Kasparov and the Machine” (documentary, 2003) — A documentary film about the 1997 match that includes extensive interviews with Kasparov and examines the cheating accusation in detail.

Next in the Events series: E12 — The Netflix Prize, 2006: The Moment the Crowd Beat the Experts — How a million-dollar open competition to improve a recommendation algorithm accidentally accelerated machine learning by a decade, established collaborative open research as a model for the field, and produced a result so good that Netflix never actually deployed it. The strange, productive, ultimately instructive story of science by competition.


Minds & Machines: The Story of AI is published weekly. If the Deep Blue story — the machine that won, the man who lost, and the question that neither settled — opens up something about what intelligence means and what it requires, share it with someone who would find the question worth sitting with.