AI History Project

2010s

2010
Read more

Self-Supervised Learning

Training methods where the data itself provides the supervision, eliminating manual labeling.

2010
Read more

Explainable AI (XAI)

Methods and techniques to interpret the hidden decisions mathematically of deep neural networks.

2012
Read more

AlexNet

The deep CNN that sparked the deep learning revolution by crushing the ImageNet benchmark.

2013
Read more

Word2Vec

Highly scalable models to produce dense vector representations of language.

2013
Read more

Variational Autoencoders (VAE)

A generative probabilistic model that compresses data into a latent space and reconstructs it.

2014
Read more

Generative Adversarial Networks

An architecture where two neural networks contest with each other to generate entirely new, realistic data.

2014
Read more

Attention Mechanism

A mechanism allowing models to automatically focus on relevant parts of the input sequentially.

2014
Read more

Neural Machine Translation (NMT)

End-to-end translation models replacing complex pipelined statistical translation methods.

2015
Read more

Batch Normalization

A mechanism to accelerate the training of deep neural networks by normalizing layer inputs.

2015
Read more

Residual Networks (ResNets)

An architecture using skip connections to enable the training of hundred-layer deep networks.

2015
Read more

Neural Style Transfer

An algorithm that takes the artistic style of one image and applies it to the content of another.

2015
Read more

Real-time Object Detection (YOLO)

You Only Look Once (YOLO) algorithms framing object detection as a single regression problem.

2016
Read more

Neural Architecture Search (NAS)

Using algorithms (like RL or evolutionary methods) to automatically design optimal neural networks.

2016
Read more

Federated Learning

Training a centralized AI model using decentralized data stored on millions of user devices.

2016
Read more

AlphaGo / AlphaZero

The AI system that conquered the ancient, mathematically intractable game of Go.

2017
Read more

Transformer Architecture

The defining neural network architecture built entirely heavily reliant on 'self-attention'.

2017
Read more

Reinforcement Learning from Human Feedback (RLHF)

A safety and fine-tuning mechanism for aligning AI behavior with human intent.

2017
Read more

Capsule Networks

A novel neural network design to better capture spatial hierarchies in images.

2018
Read more

BERT (Bidirectional Encoder Representations)

A landmark pre-trained open-source transformer model for understanding language context.

2018
Read more

GPT Series (GPT-1, GPT-2)

The first generation of the Generative Pre-trained Transformer family.

2020s

2020
Read more

GPT-3

A 175-billion parameter LLM that proved language models possessed 'emergent' zero-shot capabilities.

2020
Read more

Diffusion Models

Thermodynamics-inspired networks that generate data by reversing a noise-addition process.

2020
Read more

Retrieval-Augmented Generation (RAG)

Combining LLMs with external knowledge bases to prevent hallucinations.

2020
Read more

Vision Transformers (ViT)

Applying the Transformer NLP architecture directly to image sequences.

2020
Read more

AI Safety & Alignment

The formalized research field ensuring superintelligent systems follow human intent without disastrous side effects.

2021
Read more

DALL-E Series

The definitive text-to-image synthesis models that brought generative AI to the mainstream.

2021
Read more

Low-Rank Adaptation (LoRA)

A wildly efficient parameter-tuning method allowing individuals to fine-tune massive models on consumer hardware.

2022
Read more

ChatGPT

The conversational AI interface that reached 100 million users in two months.

2022
Read more

Constitutional AI

A mechanism for training helpful and harmless AI via a set of written principles.

2022
Read more

Chain-of-Thought Prompting

Forcing standard language models to elicit step-by-step logical reasoning before answering.

2022
Read more

Mixture of Experts (MoE)

A sparse network architecture selectively activating only specialized sub-networks per token.

2023
Read more

GPT-4 / GPT-4o

The apex multimodal foundational model, parsing text, vision, and audio natively.

2023
Read more

AI Agents & Tool Use

Equipping LLMs with the ability to dynamically call external APIs, calculators, and search browsers.

2023
Read more

Synthetic Data Generation

Using frontier AI models to perfectly generate complex training data for smaller models.

2023
Read more

State Space Models (Mamba)

A linear-time sequence architecture positioned as a highly efficient alternative to Transformers.

2023
Read more

Multimodal Foundation Models

Models trained jointly on text, images, video, and audio data streams.

2023
Read more

Reinforcement Learning from AI Feedback (RLAIF)

Replacing human trainers with AI trainers to scale model alignment securely.

2024
Read more

AI Regulation (EU AI Act)

The world's first comprehensive horizontal legal framework strictly governing AI deployment.

2024
Read more

Agentic Coding

AI systems that autonomously write, debug, test, deploy, and review complete codebases.

2024
Read more

Test-Time Compute Scaling

Scaling inference compute dynamically to vastly improve model reasoning.

20202020s

AI Safety & Alignment

View on Wikipedia

The formalized research field ensuring superintelligent systems follow human intent without disastrous side effects.

Why It Was Important

As timelines for AGI sharply decreased, organizations heavily funded mathematical alignment framing. Concepts like 'instrumental convergence' (the idea an AI will gain power as a side effect of achieving its goal) drove billions of dollars into interpretability research, 'red teaming', and containment strategies.

Who Invented It

Alignment Community (Anthropic, Redwood, Alignment Forum)

Philosophers, mathematicians, and engineers focused on existential risk.

Applications

  • Superalignment (SSI)
  • Model Red-teaming
  • Sleeper Agent defense