How IBM’s Watson hammered its ‘Jeopardy’ foes (2024)

How IBM’s Watson hammered its ‘Jeopardy’ foes (1)

byPeter Wayner

Contributing writer

analysis

Feb 16, 20115 mins

Business IntelligenceTechnology Industry

The historic “Jeopardy” matches between IBM’s Watson supercomputer and two representatives of the human race are almost over. Watson is winning. The best and brightest wetware is on the ropes.

The secret to Watson’s success is a collection of algorithms that can analyze natural language and guess whether the results yielded by those algorithms stand a good chance of being correct. This “self-knowledge” is important because Watson poses the question to all of the algorithms at once, then ranks the results. If confidence is high on any single result, the computer triggers a little servo that pushes the same button as the competitors from meatspace do.

IBM calls Watson “DeepQA,” a reference to DeepBlue, the computer that defeated human chess champion Garry Kasparov and a descendant of an earlier chess machine that borrowed the name Deep Thought from “The Hitchhiker’s Guide to the Galaxy.” It was built deep in the machine rooms at IBM’s Thomas Watson labs in Yorktown Heights, N.Y., a building from the “Mad Men” era designed by Eero Saarinen.

Behind Watson’s curtainWatson’s brain showcases several IBM technologies. The hardware is jammed into 10 refrigerator-sized racks filled with Power7 server blades. To be exact, there are 90 Power750 servers filled with four processors each — and each processor has 8 cores, for a total of 2,880 cores altogether.

The software is built on top of IBM’s UIMA architecture, which IBM created and then open-sourced. UIMA stands for “unstructured information management application,” and it offers a framework where modules can work independently on the problem at hand and then vote on the best answer. The Apache project hosts the current UIMA version, as well as a number of common modules, but not all of the modules used on the show.

Much of the power comes from IBM’s carefully curated collection of data. Jennifer Chu-Carroll, one of the scientists who has worked on the project since it began over four years ago, says that Watson excels, predictably enough, when the answer is a detail stored in its database.

“Watson is very good at things we consider facts,” Chu-Carroll explains. “The Beatles category is one we think of as ‘filling in the blank.’ They give you something and ask what’s missing. If we have the source, that’s fairly easy for the computer to do. Of course that requires understanding that type of question and actually having the source.”

She points out that the entire corpus of information is built out of largely natural language sources, including licensed Beatles lyrics. Watson is completely self-contained and does not have access to the Internet during the game. “We do have things from Project Gutenberg,” explains Chu-Carroll. “Things like Wikipedia. But primarily free text that is available to us: dictionaries, encyclopedias, newspaper articles, things that…cover ‘Jeopardy’ topics well.”

Unlike Google and its Books project, IBM chose to obey licensing rules. “If we don’t have a license, we don’t have it,” notes Chu-Carroll. Nor did IBM include any structure from CYC, a controversial artificial intelligence project designed to categorize much of human knowledge in a structured database. Most of the information used by Watson is plain text searched by the computer.

Beating the “Jeopardy” clockChu-Carroll says that search is among the most time-consuming operations for the algorithm. The team studied past “Jeopardy” games, timed Alex Trebek’s reading speed, and estimated that Watson would have three seconds to work on a problem before buzzing in. Some observers have noticed that Watson is often unprepared if the clue is short and read quickly. This is because Watson allocates the same amount of time regardless of the length of the question, something Chu-Carroll says may be fixed in the future.

During this time, Watson’s algorithms work to identify the most likely possible answers. It boils down the million or so book-length chunks of text into 100 or so likely answers, the top three of which were displayed for the entertainment of “Jeopardy” viewers.

The researchers analyzed the “Jeopardy” game and devoted an “entire team” to deciding what to wager on the Daily Double, the confidence level to reach before buzzing, and so on. Watson also tracks the amount of money it has and the amount its competitors have — details that change the odds and change the strategy. It knows enough to avoid “pulling a Clavin,” a reference to an episode of the 1980s television sitcom “Cheers,” where postman Cliff Clavin made it to the final round of “Jeopardy” only to lose a commanding lead by wagering too much.

What’s next for Watson’s technology? IBM plans to commercialize the software by approaching industries with huge collections of knowledge that must be sorted and searched by employees. The focus on using plain, relatively unfiltered, natural-language texts means that the system can be reprogrammed for a new domain just by choosing a different corpus of knowledge. Call center representatives, for instance, may have Watson search through the vast databases of product manuals and other texts to help find the best possible answer.

IBM frequently mentions creating a doctor’s assistant, a tool that will help suggest potential diagnoses to the doctor who will make a final decision and bear the brunt of any second guessing during a malpractice lawsuit. Humans, it seems, are still necessary for some things.

This article, “How IBM’s Watson hammered its ‘Jeopardy’ foes,” was originally published at InfoWorld.com. Get the first word on what the important tech news really means with the InfoWorld Tech Watch blog, and for the latest in business technology news, follow InfoWorld.com on Twitter.

Related content

  • newsActiveState's Python taps Intel MKL to speed data science and machine learning The MKL libraries for accelerating math operations debuted in Intel's own Python distribution, but now other Pythons are following suit By Serdar YegulalpMay 18, 20173 minsData ScienceMachine LearningOpen Source
  • newsCrateDB 2.0 Enterprise stresses security and monitoring—and open source The open source database for processing high-speed freeform data with SQL queries now has enterprise features, available as open source for faster developer uptakeBy Serdar YegulalpMay 16, 20173 minsNoSQL DatabasesTechnology IndustryDatabases
  • news analysisWaah! WannaCry shifts the blame game into high gear Every security crisis presents the opportunity to point fingers, but that's just wasted energy. The criminals are at fault—and we need to work together to stop themBy Fahmida RashidMay 16, 20177 minsSmall and Medium BusinessTechnology IndustryMalware
  • newsFaster machine learning is coming to the Linux kernel The addition of heterogenous memory management to the Linux kernel will unlock new ways to speed up GPUs, and potentially other kinds of machine learning hardwareBy Serdar YegulalpMay 15, 20173 minsTechnology IndustryMachine LearningOpen Source
  • Resources
  • Videos
How IBM’s Watson hammered its ‘Jeopardy’ foes (2024)

FAQs

How IBM’s Watson hammered its ‘Jeopardy’ foes? ›

Watson's main innovation centered on its ability to quickly execute hundreds of algorithms to simultaneously analyze a question from many directions, find and score potential answers, gather additional supporting evidence for each answer, and evaluate everything using natural language processing.

What is the significance of the contestant Watson on Jeopardy? ›

Three nights, two people, one machine and $1 million: the victory of IBM's Watson over two human contestants on Jeopardy was the first, and possibly only, time the machine impressed itself on the general public's consciousness.

Is Watson connected to the Internet when playing Jeopardy? ›

By February 2010, Watson could beat human Jeopardy! contestants on a regular basis. During the game, Watson had access to 200 million pages of structured and unstructured content consuming four terabytes of disk storage including the full text of the 2011 edition of Wikipedia, but was not connected to the Internet.

Does IBM's Watson's Jeopardy project represent a change in strategy? ›

Does IBM's Watson's Jeopardy Project Represent a Change in Strategy? The IBM's Watson jeopardy project does not represent a remarkable change in strategy. It apparently shifted its service strategy to a different domain. IBM primary business model was focused on selling computers and hardware.

How did Watson beat Jeopardy? ›

Watson's main innovation centered on its ability to quickly execute hundreds of algorithms to simultaneously analyze a question from many directions, find and score potential answers, gather additional supporting evidence for each answer, and evaluate everything using natural language processing.

What is the story behind IBM Watson? ›

IBM Watson IBM Research started working on the grand challenge of building a computer system that could compete with champions at the game of Jeopardy!. Just four years later in 2011, the open-domain question-answering system dubbed Watson beat the two highest ranked players in a nationally televised two-game Jeopardy! ...

Do Jeopardy contestants know the topics beforehand? ›

Do they tell you in advance what the categories are going to be on your show, so you can study up? I don't know where people get this idea, but I get asked this all the time. No, you don't see the categories until the second the viewers at home do.

Can Jeopardy contestants see the question? ›

Contestants can read each clue on the gameboard from their podiums as Alex reads them aloud. But occasionally a clue will be in a visual form, which means a special monitor is used so contestants can clearly decipher anything from a celebrity photo to a Broadway musical excerpt.

Do they read the questions live on Jeopardy? ›

During this time, contestants also phrase the question, which is pre-written during the wager. After the break, the Final Jeopardy! clue is revealed and read by the host. The contestants have 30 seconds to write their responses on the electronic display, while the show's "Think!" music plays.

What is the IBM Watson personality model? ›

Watson™ Personality Insights provides a reading of each consumer's psychology based on the Big Five personality model, which measures the traits of Openness, Conscientiousness, Extraversion, Agreeableness, and Neuroticism. It also provides data on consumer needs, decision-making values, and consumption preferences.

What is IBM Watson used for? ›

IBM Watson is a data analytics processor that uses natural language processing, a technology that analyzes human speech for meaning and syntax. IBM Watson performs analytics on vast repositories of data that it processes to answer human-posed questions, often in a fraction of a second.

What is Jeopardy the IBM challenge? ›

The IBM Challenge was the modern Jeopardy's first exhibition tournament to and take place during episodes of the syndicated version. Featuring the three-week, two-round format as the traditional tournaments on Jeopardy!, it was held in honor of Watson and offered a $1,000,000 top prize.

How does Watson work? ›

IBM Watson is a data analytics processor that uses natural language processing, a technology that analyzes human speech for meaning and syntax. IBM Watson performs analytics on vast repositories of data that it processes to answer human-posed questions, often in a fraction of a second.

What happened to Watson Health? ›

In 2022, IBM sold off parts of its Watson Health division to Francisco Partners, a private equity firm. The sale was seen as a sign that IBM was scaling back its ambitions in healthcare. However, IBM has said that it will continue to develop and support Watson for Oncology and other products.

What was the final Jeopardy question for Watson? ›

The category was "U.S. Cities" and the clue was: "Its largest airport is named for a World War II hero; its second largest for a World War II battle." The human competitors Ken Jennings and Brad Rutter both answered correctly with "Chicago" but IBM's supercomputer Watson said "Toronto."

What is the difference between Watson and Watsonx? ›

What is the difference between IBM Watson and watsonx? IBM Watson is a suite of AI-powered tools and services that includes natural language processing, machine learning, and speech-to-text capabilities. Watsonx, on the other hand, is a platform that provides AI-powered solutions for businesses and developers.

References

Top Articles
Latest Posts
Article information

Author: Eusebia Nader

Last Updated:

Views: 5785

Rating: 5 / 5 (80 voted)

Reviews: 87% of readers found this page helpful

Author information

Name: Eusebia Nader

Birthday: 1994-11-11

Address: Apt. 721 977 Ebert Meadows, Jereville, GA 73618-6603

Phone: +2316203969400

Job: International Farming Consultant

Hobby: Reading, Photography, Shooting, Singing, Magic, Kayaking, Mushroom hunting

Introduction: My name is Eusebia Nader, I am a encouraging, brainy, lively, nice, famous, healthy, clever person who loves writing and wants to share my knowledge and understanding with you.