Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN

Introduction: Deep learning can be applied in various fields, and the shapes of deep neural networks vary according to different application scenarios.

The common deep learning models mainly include Fully Connected (FC), Convolutional Neural Network (CNN), and Recurrent Neural Network (RNN).

Each of these has its own characteristics and plays an important role in different scenarios.This article will introduce the basic concepts of these three models and the scenarios suitable for each.

Authors: Liu Xianglong, Yang Qinghong, Hu Xiaoguang, Yu Tianhai, Bai Haojie, National Engineering Laboratory for Deep Learning Technology and Applications, Baidu Technical Academy
Source: Big Data DT (ID: hzdashuju)

Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN

01 Fully Connected Network Structure

The Fully Connected (FC) network structure is the most basic layer of neural networks/deep neural networks, where each node in the fully connected layer is connected to all nodes in the previous layer.

Initially, fully connected layers were mainly used for classifying extracted features; however, since all outputs and inputs of the fully connected layer are interconnected, it generally has the most parameters, requiring considerable storage and computational space.

The redundancy of parameters makes conventional neural networks composed purely of FC layers rarely applicable in more complex scenarios. Conventional neural networks are generally used in simple scenarios that rely on all features, such as housing price prediction models and online advertising recommendation models that use relatively standard fully connected neural networks. The specific form of conventional neural networks composed of FC layers is shown in Figure 2-7.

Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN

▲ Figure 2-7 Conventional neural network composed of FC layers

02 Convolutional Neural Network

A Convolutional Neural Network (CNN) is a type of neural network specifically designed to process data with a grid-like topology, such as image data (which can be viewed as a two-dimensional pixel grid). Unlike FC, the neurons in the upper and lower layers of CNN are not all directly connected; instead, they are connected through a “convolution kernel” as an intermediary, significantly reducing the parameters in the hidden layers through the sharing of kernels.

A simple CNN consists of a series of layers, each transforming one quantity into another through a differentiable function. These layers mainly include convolutional layers, pooling layers, and fully connected layers.

Convolutional networks have shown excellent performance in many application fields, particularly excelling in large image processing scenarios. Figure 2-8 illustrates the structure of a CNN, where a neuron is arranged in three dimensions to form a convolutional neural network (width, height, and depth). As shown in one of the layers, each layer of the CNN transforms a 3D input into a 3D output.

Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN

▲ Figure 2-8 Structure of CNN

03 Recurrent Neural Network

A Recurrent Neural Network (RNN) is also one of the commonly used deep learning models (as shown in Figure 2-9). Just as CNNs are specialized for processing grid-like data (such as an image), RNNs are designed for handling sequential data.

Since audio contains time components, audio can be represented as a one-dimensional time series; similarly, words in language appear one by one, so the representation of language is also sequential data. RNNs perform exceptionally well in areas such as machine translation and speech recognition.

Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN

▲ Figure 2-9 Simple RNN structure

About the Authors:Liu Xianglong, Associate Professor, PhD Supervisor, currently working at the National Key Laboratory of Software Development Environment at Beihang University, mainly researching big data retrieval, large-scale visual analysis, and trustworthy deep learning.
Yang Qinghong, PhD, Senior Engineer Series at the Chinese Academy of Sciences, Principal Lecturer in Artificial Intelligence at Beihang University, Visiting Scholar at Southern Connecticut State University, USA, and Data Analysis Expert at Yale University Technology Innovation Lab.
Hu Xiaoguang, Outstanding R&D Architect at Baidu, with over 10 years of experience in natural language processing R&D, participated in machine translation projects that won the National Science and Technology Progress Award (Second Class), currently responsible for the development of the core training framework and model algorithms of PaddlePaddle, dedicated to building the most user-friendly deep learning platform.
Yu Tianhai, Chief Architect of Baidu’s Deep Learning Platform PaddlePaddle. Graduated from Peking University in 2008 and joined Baidu, focusing on technology R&D and platform construction related to machine learning and natural language processing.
Bai Haojie, Special Lecturer at Beihang University and Dalian University of Technology, Baidu Certified Deep Learning Evangelist, Visiting Scholar at Florida International University High-Performance Data Lab, focusing on research in mobile object databases, data visualization, machine learning, and deep learning.
This article is excerpted from “Practical Deep Learning with PaddlePaddle”, published with permission from the publisher.
Further Reading“Practical Deep Learning with PaddlePaddle”
Click the image above to learn more and purchase
For reprints, please contact WeChat: DoctorData
Recommendation:This book is officially produced by Baidu, comprehensive in content, gradually in-depth, and emphasizes practice. It covers the foundational knowledge necessary for learning deep learning technology and the main core technologies of deep learning, including relevant mathematical foundations, Python programming basics, machine learning basics, and forward/backward propagation algorithms, convolutional neural networks, recurrent neural networks, etc., aiming to achieve mastery from “zero foundation” to “full proficiency” with just one book.
Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN
Key Points👇
Direct Access to Valuable Content👇
  • Latest! Digital RMB launched, 50,000 people in Shenzhen share a 10 million red envelope!

  • The “Digital Twin Technology White Paper” is released (with full version download attached)

  • Two example problems explaining Bayes’ theorem

  • What is HBase? How does it work? Finally, someone explained it clearly

More Exciting Content👇
In the public account dialogue box enter the followingkeywords
to see more quality content!
PPT | Reading | Book List | Hardcore | Valuable Content | Clearly Explained | Smart Operations
Big Data | Cloud Computing | Database | Python | Visualization
AI | Artificial Intelligence | Machine Learning | Deep Learning | NLP
5G | Mid-Platform |User Profile | 1024 | Mathematics | Algorithms | Digital Twin
Statistics show that 99% of experts have completed this smart operation
👇

Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN

Illustration of 3 Common Deep Learning Network Structures: FC, CNN, RNN

Leave a Comment