← 返回知识库

The History of AI: From Turing to Transformers

history· 2 分钟阅读

The history of artificial intelligence spans over seven decades, from theoretical foundations to the transformer revolution that powers today's AI systems.

The story begins with Alan Turing's 1950 paper "Computing Machinery and Intelligence," which posed the question "Can machines think?" and proposed the Turing Test as a measure of machine intelligence. This laid the philosophical groundwork for the entire field.

The term "Artificial Intelligence" was coined at the 1956 Dartmouth Conference, organized by John McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon. Early optimism was high — researchers predicted human-level AI within a generation.

The 1960s and 70s saw early successes in symbolic AI — programs that manipulated symbols and rules to solve problems. ELIZA (1966) simulated a psychotherapist through pattern matching. SHRDLU (1970) could understand natural language commands in a limited block world. Expert systems like MYCIN diagnosed bacterial infections using hand-coded rules.

However, progress stalled in periods known as "AI Winters." The first (1974-1980) came when early promises went unfulfilled and funding dried up. The second (1987-1993) followed the collapse of the expert systems market. These winters taught the field humility about the difficulty of general intelligence.

The machine learning revolution began in the 1990s and 2000s. Instead of hand-coding rules, researchers developed algorithms that learned from data. Support vector machines, random forests, and boosting methods achieved practical success in spam filtering, recommendation systems, and fraud detection.

The deep learning breakthrough came in 2012 when AlexNet, a deep convolutional neural network, dramatically won the ImageNet competition. This demonstrated that deep neural networks, trained on large datasets with GPU computing, could achieve superhuman performance on visual recognition tasks.

Key milestones followed rapidly: DeepMind's AlphaGo defeated world Go champion Lee Sedol in 2016. GPT-2 (2019) showed that language models could generate coherent text. GPT-3 (2020) demonstrated that scaling up models produced emergent capabilities.

The Transformer architecture, introduced in the 2017 paper "Attention Is All You Need," became the foundation for modern AI. Its self-attention mechanism enabled efficient processing of sequential data, leading directly to BERT, GPT, and all modern large language models.

The current era, beginning with ChatGPT's launch in November 2022, has brought AI into mainstream awareness. Generative AI tools for text, images, code, and video are transforming industries and sparking global conversations about the future of work, creativity, and human-AI collaboration.