(ENG) Development of Artificial Intelligence

Table of Contents


1940s–1950s - Early Foundations

Discovery: Boolean logic and the Turing Test

Problem: How can we formalize reasoning and test machine intelligence?

Solution: Alan Turing introduced the concept of a machine capable of computation (Turing Machine) and proposed the Turing Test as a way to determine if a machine exhibits intelligent behavior.


1950s–1960s - Discovery of Neural Networks

Discovery: Perceptron (1958, Frank Rosenblatt)

Problem: How can we mimic the human brain’s ability to learn patterns?

Solution: The perceptron, a simple single-layer neural network, was developed to classify data linearly by adjusting weights using feedback. However, it was unable to solve non-linear problems, such as XOR.


1970s–1980s - AI’s First Winter(인공지능의 첫 겨울) & Backpropagation(역전파기법)

Discovery: Backpropagation (1986, Rumelhart, Hinton, Williams)

Problem: How can multi-layer neural networks be efficiently trained?

Solution: Backpropagation introduced a systematic way to compute gradients and update weights in deep networks using the chain rule, which is the cornerstone of deep learning.This allowed to solve non-linear problems.Applications suggested during that period included image recognition and character classification but were limited by computational power and data availability.


1980s–1990s - Sequence Processing and Memory

Discovery: Recurrent Neural Networks(RNN) (1986, Rumelhart & McClelland)

Problem: How can sequential data be modeled and retain context from prior inputs?

Solution: RNNs introduced a feedback loop where outputs from previous steps are fed back as inputs, allowing networks to maintain a “memory” over time.

Limitations: RNNs faced vanishing gradient issues, making them ineffective for long sequences -> LSTM was introduced to resolve this issue.


1997 - Long Short-Term Memory(LSTM) (Hochreiter & Schmidhuber)

Problem: How can we learn and retain long-term dependencies in sequences

Solution: LSTMs introduced gated cells to control the flow of information, solving vanishing gradient problems. Applications include speech recognition and language translation.


1980s–1990s - Visual Recognition and Spatial Data

Discovery: Convolutional Neural Networks (CNNs, 1989, LeCun)

Problem: How to efficiently process and recognize spatially structured data like images.

Solution: CNNs use convolutional layers to detect patterns such as edges and textures by learning spatial hierarchies. Pooling layers reduce dimensionality while preserving critical information.

Applications: Handwritten digit recognition (e.g., MNIST dataset) and later extended to more complex tasks like object detection.


2012 - Deep Learning Revolution

Discovery: Deep CNNs and AlexNet (2012, Krizhevsky, Sutskever, Hinton)

Problem: How to achieve state-of-the-art(SOTA) accuracy in image recognition.

Solution: AlexNet leveraged deeper architectures, ReLU activations, and GPUs for training, achieving a breakthrough in the ImageNet competition. (Previous SOTA classification : 74% -> Alexnet 84.69%)

Applications: Facial recognition, autonomous vehicles, and medical imaging.


2014 - Generative Adversarial Networks(GAN) (Goodfellow)

Problem: How can we generate new output data that resembles an input dataset?

Solution: GANs use a generator and discriminator in a competitive framework, enabling applications like image synthesis and style transfer.


2017 - Attention Mechanisms and Transformers (Vaswani et al.)

Problem: How can we handle long-term dependencies and parallelize(병렬화) sequence processing?

Solution: The attention mechanism learns relationships between all inputs simultaneously, eliminating sequential constraints. Transformers like BERT and GPT revolutionized natural language processing (NLP).

Applications: Machine translation, summarization, and chatbots.


2020s - Multi-Modal AI and Real-Time Applications

Discovery: Multi-Modal Models (e.g., CLIP, DALL-E)

Problem: How can we unify processing across different modalities like text, images, and audio?

Solution: Pretraining large-scale models that align multiple data types, enabling applications in creative AI and real-time analysis.


TL;DR

  • Backpropagation solved problem of non-linearity.
  • RNNs/LSTMs enabled handling sequential data.
  • CNNs revolutionized image processing.
  • Transformers led to breakthroughs in NLP and multi-modal AI.



Enjoy Reading This Article?

Here are some more articles you might like to read next:

  • (ENG) Are Sixteen Heads Really Better than One?
  • (작성중)(ENG) Attention Is All You Need Paper Review
  • (KOR) 소프트웨어 서비스 설계 방법론
  • (KOR) Retrieval-Augmented Machine Translation with Unstructured Knowledge 논문 리뷰
  • (KOR) Arc Browser 사용법