A Comprehensive Overview of Deep Learning
Original article source: AIML.com
Introduction:
Deep Learning is a subset of machine learning that is characterized by the use of deep neural networks, with multiple layers (hence the term "deep" learning) to perform tasks that typically require human intelligence. It is inspired by the structure and function of the human brain, where each layer of neurons processes and transforms the input data to progressively extract higher-level features.
Deep neural networks (DNNs), consist of interconnected layers of artificial neurons called nodes. Each node receives input from the previous layer, applies a mathematical transformation to it, and passes the transformed output to the next layer. The layers closer to the input are responsible for learning low-level features, while the deeper layers learn more abstract and complex representations.
This phenomenon of automatically learning meaningful and informative features (or representations) from raw data is also referred to as representation learning, which stands as one of the key strengths of DNNs.
Key characteristics and working of Deep Neural Network
Deep learning works by using artificial neural networks, which are composed of layers of interconnected nodes (neurons) that process and transform the data through neural network training.
Key characteristics and working of deep learning include the following:
(1) The Perceptron,
(2) Deep architecture,
(3) Neural Networks, and
(4) Training
Deep Learning Models
Deep learning encompasses several key architectures, each designed for specific types of data and tasks. These architectures serve as building blocks for solving a wide range of tasks in artificial intelligence and machine learning. Here are some of the key deep learning architectures:
Applications of Deep Learning
For complete list of applications, go to: https://aiml.com/what-is-deep-learning/
Evolution of Deep Learning: A brief history and Resurgence
A brief history:
Deep Learning might appear as a novel discovery in the field of machine learning, given its recent name and fame. However, , the history of Deep Learning spans several decades, dating back to 1940s as presented below:
领英推荐
1950s --> Alan Turing, a British mathematician, first presented the idea that computers would achieve human-level intelligence
1957 --> Frank Rosenblatt, an American psychologist, introduced the perceptron, a single-layer neural network
1965 --> Alexey Ivakhnenko, a Soviet mathematician, created a small functional neural network
1970s --> Limited progress, referred to as the AI winter
1980s --> Backpropagation, a method for training neural networks, was rediscovered by Dr. Geoffrey Hinton, ?a British-Canadian psychologist and computer scientist
1989 --> Yann LeCun’s invents machine that can read handwritten digits
1990s --> Multi-layer perceptrons, the inception of CNNs, and LSTM
1999 --> GPUs (Graphics Processing Units)?were developed
2000s --> Limited progress in the field of Deep Learning
2012 --> Deep neural network, AlexNet, outperformed other methods for image recognition and led to the resurgence of Neural Network. Several notable neural network models and frameworks followed
2017 --> Introduction of Transformer architecture, a game-changer in the field of Deep Learning models for solving Natural Language Processing tasks
2018 onwards --> Revolution in the AI space took place with the introduction of BERT, GPT-3, Stable Diffusion models, and systems such as ChatGPT, Bard, Perplexity etc.
The resurgence was catalyzed by three key factors:
Video Explanation (Playlist):
This playlist contains the following videos in the recommended order:
--
For more such articles, visit https://aiml.com
Looking for practice quizzes, https://aiml.com/quiz-category/technical/
(PS: Do sign up to take practice quizzes and bookmark your favorite questions)
#deeplearning #machineLearning