Supervised Learning
Common applications of supervised learning include credit scoring, handwriting recognition, speech recognition, information retrieval, financial analysis, and spam detection.
02 Unsupervised Learning
Common applications of unsupervised learning include anti-money laundering, customer segmentation, advertisement recommendations, and sales trend forecasting.
Reinforcement Learning
Common applications of reinforcement learning include autonomous driving, machine translation, healthcare, news customization, advertising marketing, and robot control.
Deep learning is a branch of machine learning that simulates the brain’s neural network structure to perform representation learning on data. Deep learning originated from the study of the mechanisms of the human brain. American neurophysiologists David Hubel and Torsten Wiesel, who won the Nobel Prize in Physiology or Medicine in 1981, discovered that the information processing of the human visual system is hierarchical, and human perception of high-level features is based on combinations of low-level features. For example, face recognition goes through the process of pixel intake through the pupil (shape judgment) to abstract the concept of a face—recognizing it as a face—where the feature representation becomes increasingly abstract and conceptual from low to high levels. This discovery implies that the brain is a deep architecture, and cognitive processes are also deep, while deep learning precisely forms more abstract high-level features by combining low-level features. The development of deep learning can be divided into three stages: perceptrons, neural networks, and deep learning.
In 1943, American psychologist Warren S. McCulloch and mathematician Walter Pitts proposed the concept of artificial neural networks and constructed a mathematical model of artificial neurons, known as the MCP model, thus pioneering the era of artificial neural network research. In 1949, Canadian psychologist Donald Hebb described the basic principles of synaptic plasticity, explaining the changes that occur in brain neurons during the learning process from a neuroscientific perspective. Hebb’s theory is the biological basis for artificial neural networks. In 1958, Rosenblatt invented the perceptron algorithm at the Cornell Aeronautical Laboratory, which is the world’s first neural network learning algorithm with a complete algorithm description. The perceptron algorithm is a simple configuration of a single-layer neural network that can distinguish basic shapes like triangles. However, limited by computer hardware, the perceptron algorithm could not be widely applied at that time. In 1969, Minsky and Seymour Papert proved that perceptrons could not solve simple linear inseparable problems like XOR, leading perceptron research to decline in the 1970s.
In 1959, Hubel and Wiesel discovered two types of cells in the primary visual cortex of cats while studying their visual nervous system: simple cells and complex cells, where simple cells perceive light information, and complex cells perceive motion information. Inspired by this, in 1980, Japanese computer scientist Kunihiko Fukushima proposed a network model called the “Neocognitron.” This network is divided into multiple layers, each consisting of a type of neuron. Within the network, these two types of neurons alternate, used respectively for extracting and combining pattern information. These two types of neurons later evolved into convolutional layers and pooling layers. However, the neurons in this network were artificially designed and could not adjust based on computational results, so it could only recognize a limited number of simple digits and lacked learning capabilities.
In 1982, American physicist John J. Hopfield proposed the Hopfield neural network model with limited memory capabilities based on statistical physics, pioneeringly demonstrating the stability of neural networks designed according to Hebb’s law. In the same year, Finnish computer scientist Teuvo Kohonen proposed the self-organizing map network, simulating the signal processing mechanisms of brain neurons, used for data analysis and exploration, with its first application in speech analysis. Kohonen’s key invention introduced a system model consisting of a competitive neural network that implements the winner-takes-all function and a subsystem that implements plasticity control. In 1987, American scientists Stephen Grossberg and Gail Carpenter proposed the adaptive resonance theory network, which achieves analogical learning by allowing known and unknown information to resonate. However, these neural networks had limitations in learning efficiency, required constant optimization of design, and had small network memory capacity, which restricted their practical application range.
In 1986, American psychologist David Rumelhart, computer scientist Ronald Williams, and Canadian cognitive psychologist and computer scientist Geoffrey E. Hinton jointly proposed the backpropagation algorithm (BP algorithm). The BP algorithm uses the chain rule of gradients to feedback the differences between output results and true values to the weights of each layer, allowing each layer’s function to be trained like a perceptron. The BP algorithm phase solved the problem of neural network adaptability and autonomous learning. In 1989, Yann LeCun, a French computer scientist at Bell Labs, successfully implemented the practical application of neural networks for the first time. He combined convolutional neural networks with the BP algorithm to propose the LeNet network. In the 1990s, the United States Postal Service used the LeNet network to automatically read postal codes on envelopes. However, neural networks based on the BP algorithm could only solve local optima, and this situation worsened as the number of network layers increased, which limited the development of neural networks.
In 2006, Hinton proposed the deep learning algorithm, which effectively reduced the training difficulty through unsupervised learning and layer-wise pre-training, thus solving the problem that BP neural networks found it difficult to achieve global optimality. In 2012, Hinton’s research team won the championship in the ImageNet image classification competition using deep learning, with an accuracy exceeding the second place by more than 10%, causing a huge sensation in the field of computer vision and triggering a wave of interest in deep learning. In 2013, MIT Technology Review listed deep learning as the top technological breakthrough of the year. Today, deep learning is widely used in search engines, speech recognition, automatic machine translation, natural language processing, autonomous driving, facial recognition, and is one of the hottest research directions in artificial intelligence.
“The work of the award winners has already produced tremendous benefits. In the field of physics, we apply artificial neural networks to a wide range of areas, such as developing new materials with specific properties,” said Ellen Moons, chair of the 2024 Nobel Prize in Physics Committee.
· END ·
This article is excerpted from “Algorithm Development in the Intelligent Era,” with modifications made during publication on WeChat public account. All images in the text are from the Royal Swedish Academy of Sciences.
