Decoding Deep Learning: Understanding Its Core Concepts

Decoding Deep Learning: Understanding Its Core Concepts

Non-Business Promotion Space
Decoding Deep Learning: Understanding Its Core Concepts

Cover Image Source:CAM (Cambridge Alumni Magazine) Issue 85

Written by | Jin Chunqi

In recent years, deep learning methods have achieved tremendous success in learning complex data forms, leading to new modeling and data processing methods across many scientific disciplines. It can address numerous data-intensive tasks, including but not limited to 3D computer vision, sequence modeling, causal reasoning, generative modeling, and efficient probabilistic reasoning, which are crucial for further scientific discoveries.

01

What is Deep Learning?

Along with deep learning, the concepts we most often hear are artificial intelligence, machine learning, and neural networks.

Artificial Intelligence (AI) is a broad concept that involves simulating, extending, and enhancing human intelligence using machines. For example, the robots in the TV series “Westworld” can think like humans and possess self-awareness.

Machine Learning (ML) is a branch of artificial intelligence aimed at allowing machines to learn how to judge and output corresponding results automatically by providing them with raw “learning materials.” When the system performs correctly, it is rewarded to reinforce such behavior, and when it makes mistakes, it is punished to weaken that behavior. Models are trained using data, which are then used for predictions.

Neural Networks generally refer to networks composed of brain neurons, synapses, cells, etc., used to generate consciousness and help organisms think and act. Later, artificial intelligence was inspired by neural networks to develop Artificial Neural Networks (ANNs).

Deep Learning (DL) originates from the study of artificial neural networks, referring to multi-layer artificial neural networks and the methods to train them. A layer of the neural network takes a large matrix of numbers as input, applies a non-linear activation method to weight them, and produces another data set as output. This process mimics the workings of biological neural brains, where suitable quantities of matrices are linked together to form a neural network “brain” that performs precise and complex processing, similar to how humans recognize objects and label images.

Deep learning is a new research direction in the field of machine learning, introduced to bring it closer to its original goal—artificial intelligence.

Unlike traditional machine learning, deep learning requires more samples, resulting in fewer manual annotations and higher accuracy. The early notion of “deep” referred to neural networks with more than one layer, but with the rapid development of deep learning, its connotation has exceeded the traditional multi-layer neural networks and even the scope of machine learning, rapidly evolving towards artificial intelligence. It can learn the intrinsic patterns and expressive hierarchies of sample data, and the information obtained is highly beneficial for interpreting data such as text, images, and sounds. Its ultimate goal is to enable machines to possess analytical learning capabilities akin to humans, capable of recognizing data like text, images, and sounds.

Deep learning has achieved significant results in search technologies, data mining, machine learning, machine translation, natural language processing, multimedia learning, speech, recommendation and personalization technologies, and other related fields.

Decoding Deep Learning: Understanding Its Core Concepts

Image Source: quantdare

The basic idea of learning using deep neural networks is quite simple. Based on a database of input-output pairs, the correct output is linked to the input data through machine learning. Deep learning employs multi-layer neural networks, where input data is fed into the first layer, and its output serves as input for the next layer, and so on. Each layer consists of the product of input and weight matrix, along with a non-linear function. The number of repetitions of the learning process is determined by the number of layers.

Assuming we have a system S with n layers (S1,…,Sn), its input is I, and output is O, represented as: I =>S1=>S2=>…=>Sn => O. If the output O equals the input I, meaning that there is no information loss during the transformation of input I through this system, it indicates that there is no information loss at any layer Si, and at any layer Si, it is another representation of the original information (i.e., input I). For deep learning, we need to learn features automatically. Assuming there are some inputs I (such as images or text), and a system S (with n layers) is designed, we adjust the parameters in the system so that its output remains I. This allows us to automatically obtain a series of hierarchical features of the input I, namely S1,…,Sn. The idea of deep learning is to achieve a hierarchical representation of the input information through stacked multiple layers. When the output requirements change, a different type of deep learning method will also emerge.

Figure 1 shows the process of recognizing a Samoyed dog (bottom left) using its RGB (Red, Green, Blue) image as input, employing a typical convolutional neural network. The rectangular images correspond to the feature maps detected at each image position. Information is transmitted from bottom to top, where low-level features act as directional edge detectors. After calculating scores for each image class, the category of the input image can be determined.

Decoding Deep Learning: Understanding Its Core Concepts

Figure 1: Recognizing a Samoyed Dog Using Convolutional Neural Networks

There are many factors that affect whether a deep learning system can successfully “learn,” and the key components mainly includethree. As noted by Professor Lenka Zdeborová of Paris-Saclay University in her commentary article titled “Understanding deep learning is also a job for physicists” published in Nature Physics, the three main elements of current deep learning systems are neural network architecture, algorithms, and structured data (as shown in Figure 2). Decoding these three elements means decoding the deep learning system.

02

The Three Elements of Deep Learning

Decoding Deep Learning: Understanding Its Core Concepts

Figure 2: Interaction of Three Key Components

2.1

Neural Network Architecture

As mentioned earlier, the neural networks used for deep learning consist of multiple layers, with the number of layers referred to as “depth.” Each layer has a specific dimension known as width, and is associated with a non-linear function called the activation function. Layers can be of different types, such as fully connected or convolutional. One needs to decide the architecture of the neural network by selecting the depth, width, activation function, and type of layers. The architecture of neural networks is complex and varied, mainly divided into three categories: feedforward neural networks, recurrent networks, and symmetric connection networks. Feedforward neural networks are the most common type of neural networks in practical applications. The first layer is the input, and the last layer is the output. If there are multiple hidden layers, it is referred to as a “deep” neural network. Each layer performs a series of transformations to alter sample similarity, and the activity of neurons in each layer is a non-linear function of the activity in the previous layer. Recurrent neural networks are a type of recursive neural network that takes sequential data as input and recurses in the direction of sequence evolution, with all nodes (recurrent units) connected in a chain. Specifically, the network remembers previous information and applies it to the current output calculation, meaning that the nodes between hidden layers are interconnected, and the input to the hidden layer includes not only the output from the input layer but also the output from the previous hidden layer. Symmetric connection networks are similar to recurrent networks, but the connections between units are symmetric (the weights in both directions are the same). Symmetric weights limit the variability of network models, thus restricting the network’s capability, but also making it easier to analyze than recurrent neural networks.
Basic neural networks mainly include: Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), Deep Belief Networks (DBN), Generative Adversarial Networks (GAN), etc.

2.2

Algorithms

Once the data and network structure are defined, an algorithm is needed to set the weight matrix so that the network’s output matches the correct values of the input data. This is typically achieved by minimizing a function on the training set that quantifies the mismatch between the current network output and the expected output. The corresponding algorithm (commonly referred to as backpropagation) is based on the gradient descent of that function concerning the weights. Mathematically, the direction of the gradient is the direction in which the function increases fastest, so the opposite direction of the gradient is where the function decreases fastest. Therefore, if one wants to calculate the minimum of a function, one can use the idea of gradient descent. Common algorithms include first-order derivative-based stochastic gradient descent, adaptive learning rate algorithms, and second-order derivative-based Newton’s method, conjugate gradient method, etc.

2.3

Structured Data

Structured data simply refers to databases. For supervised learning in deep learning, the training data samples consist of input data and label pairs. When classifying images of dogs and cats, a sample includes an image and a label indicating whether the image is of a dog or a cat (for example, label 1 indicates dog, -1 indicates cat). The goal of the neural network is to learn the function from input to output, but the key is that the input must be of the same type as the training set. Therefore, the input data cannot be arbitrary; it needs to be a specific type of data, such as a photo of a cat or a dog. Data samples are extremely important for deep learning; good samples equate to half the success.

These three elements—architecture, algorithms, and structured data—are essentially interdependent, as the choice of network architecture is meant to represent the structure of the data in a way that the given algorithm can learn.

03

Scientists Should Embrace Deep Learning

Although deep learning has achieved tremendous success in research across various scientific disciplines, there are still some unresolved issues: For instance, when can deep neural networks, often described as black boxes, provide satisfactory answers, and when can they not? What is the best way to incorporate relevant domain knowledge, constraints, and symmetries when applying deep learning systems? How can we adapt existing learning tools to new problems, and how can we scientifically interpret the results obtained? How can we determine and quantify whether the training and testing data come from different sources, etc.

As a hot concept today, deep learning has gained wide recognition in both academia and industry. We have already seen its extensive applications and astonishing results across various fields, with many predictions about the scientific problems we can solve using deep learning in the near future. Some even question whether future learning systems will be able to directly collect suitable data and fully automatically infer natural laws from that data. As scientific researchers, we should embrace machine learning, utilize it as a new tool, actively study and understand why and how it truly works, and integrate it as a part of our research field, fully mobilizing our research spirit to explain it.

References
[1] Zdeborova, L. Understanding deep learning is also a job for physicists. Nature Physics (2020).
[2] LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature 521, 436 (2015).
[3] Deep learning. Wikipedia

[4] https://ml4physicalsciences.github.io/

The public account timeline has been revised, and many readers have reported not seeing updated articles. According to the latest rules, it is recommended to frequently visit the “China Optics” public account, read 3-5 articles to become a “regular reader,” to receive timely updates.

Leave a Comment