Phillips Exeter Academy

Artificial

Artificial

Intelligence

A brief history of artificial intelligence

1796

Jonathan Swift’s satiric novel, Gulliver’s Travels, refers to the Engine, a large contraption used by scholars to generate new ideas, sentences and books.

1950

British mathematician Alan Turing publishes an academic paper addressing whether machines can think. He developed the Turing Test, a way to measure machine intelligence by assessing its ability to mimic human conversation and behavior. (The Loebner Prize competition is based on the Turing Test.)

1956

Dartmouth College mathematics professor John McCarthy coins the term “artificial intelligence” during the Dartmouth Summer Research Project on Artificial Intelligence, a conference exploring how machines could simulate human intelligence.

1958

Perceptron, the first artificial neural network, is developed by American psychologist Frank Rosenblatt. The program makes decisions in a way similar to the human brain. It can distinguish between punch cards marked on the left and right and is described by its creator as the first machine capable of having an original idea.

1960

Adaline (Adaptive Linear Neuron), a single-layer artificial neural network, is developed by Stanford University professor Bernard Widrow and his student Marcian Hoff. It’s an adaptive system for pattern recognition and the foundation for future advances in neural network and machine learning.

1997

Deep Blue, developed by IBM, is the first computer system to defeat a reigning world chess champion, Garry Kasparov. The computer’s underlying technology advances the ability of supercomputers to tackle complex calculations to perform tasks like uncovering patterns in databases.

2012

AlexNet, a deep learning neural network with eight layers, is a breakthrough in image recognition, identifying images of dogs and cars at a level similar to humans.

2017

Google Research develops Transformer, a neural network architecture that can train a computer to recognize the next word in a chain of words.

2019

OpenAI’s Generative Pretrained Transformer 2 (or GPT-2) demonstrates the power of natural language processing. GPT-2 is able to predict the next item in a sequence, perform tasks such as summarizing and translating text. GPT-3, introduced in 2020, is able to produce text often indistinguishable from human writing.

2021

DALL-E, a neural network that creates pictures from language prompts is introduced by Open AI.

2022

ChatGPT, Open AI’s chatbot, built on a large language model, introduces generative AI, which can create new content based on existing data. It can produce text, images, videos, audio and more.

2023

Google Labs releases Notebook LM, which summarizes up to 50 sources, including documents, videos and books.

2024

Using Google’s AI algorithms, Google Research and Harvard publish the first synaptic resolution of the human brain. Open AI releases Sora, an AI tool that creates videos from text, images and other video.