A Comprehensive Overview of Deep Learning

A Comprehensive Overview of Deep Learning

Original article source: AIML.com

https://aiml.com/what-is-deep-learning/


Introduction:

Deep Learning is a subset of machine learning that is characterized by the use of deep neural networks, with multiple layers (hence the term "deep" learning) to perform tasks that typically require human intelligence. It is inspired by the structure and function of the human brain, where each layer of neurons processes and transforms the input data to progressively extract higher-level features.

What is Artificial Intelligence, Machine Learning and Deep Learning


Deep neural networks (DNNs), consist of interconnected layers of artificial neurons called nodes. Each node receives input from the previous layer, applies a mathematical transformation to it, and passes the transformed output to the next layer. The layers closer to the input are responsible for learning low-level features, while the deeper layers learn more abstract and complex representations.

How neural networks progressively extract higher-level features from the raw input | Source:


This phenomenon of automatically learning meaningful and informative features (or representations) from raw data is also referred to as representation learning, which stands as one of the key strengths of DNNs.

Key characteristics and working of Deep Neural Network

Deep learning works by using artificial neural networks, which are composed of layers of interconnected nodes (neurons) that process and transform the data through neural network training.

Key characteristics and working of deep learning include the following:

(1) The Perceptron,

(2) Deep architecture,

(3) Neural Networks, and

(4) Training

How a Deep Learning Algorithm works | Source:


  1. The Perceptron (Neuron)A perceptron is the structural building block of a deep learning model. It refers to a simple type of artificial neuron or node in a neural network. It operates by calculating a weighted sum of its inputs, adding a bias term, and then applying an activation function to this sum
  2. Deep architecture The term "deep" refers to the depth of the network, which means it has more than one hidden layer. Deep architectures enable DNNs to learn and represent intricate features from data.
  3. Neural Networks Deep learning is built upon the concept of artificial neural networks (ANNs), which comprises of interconnected nodes, called neurons or units, organized into layers: an input layer, hidden layers, and an output layer.
  4. Training Deep Neural networks are trained using large datasets. The training process comprises of five main steps: (a) Sampling mini-batch of data and weights initialization, (b) Forward propagation and Loss calculation, (c) Backpropagation and Optimization, (d) Repeat the training loop, and (e) Infer


Deep Learning Models

Deep learning encompasses several key architectures, each designed for specific types of data and tasks. These architectures serve as building blocks for solving a wide range of tasks in artificial intelligence and machine learning. Here are some of the key deep learning architectures:

Deep Learning Models | Source:


Applications of Deep Learning

Applications of Deep Learning | Source:

For complete list of applications, go to: https://aiml.com/what-is-deep-learning/


Evolution of Deep Learning: A brief history and Resurgence

A brief history:

Deep Learning might appear as a novel discovery in the field of machine learning, given its recent name and fame. However, , the history of Deep Learning spans several decades, dating back to 1940s as presented below:

Evolution of Deep Learning from 1940-2010 | Source:


1950s --> Alan Turing, a British mathematician, first presented the idea that computers would achieve human-level intelligence

1957 --> Frank Rosenblatt, an American psychologist, introduced the perceptron, a single-layer neural network

1965 --> Alexey Ivakhnenko, a Soviet mathematician, created a small functional neural network

1970s --> Limited progress, referred to as the AI winter

1980s --> Backpropagation, a method for training neural networks, was rediscovered by Dr. Geoffrey Hinton, ?a British-Canadian psychologist and computer scientist

1989 --> Yann LeCun’s invents machine that can read handwritten digits

1990s --> Multi-layer perceptrons, the inception of CNNs, and LSTM

1999 --> GPUs (Graphics Processing Units)?were developed

2000s --> Limited progress in the field of Deep Learning

2012 --> Deep neural network, AlexNet, outperformed other methods for image recognition and led to the resurgence of Neural Network. Several notable neural network models and frameworks followed

2017 --> Introduction of Transformer architecture, a game-changer in the field of Deep Learning models for solving Natural Language Processing tasks

2018 onwards --> Revolution in the AI space took place with the introduction of BERT, GPT-3, Stable Diffusion models, and systems such as ChatGPT, Bard, Perplexity etc.


The resurgence was catalyzed by three key factors:

  • Big data The digital age brought about an unprecedented amount of data. Deep learning models thrive on vast datasets, and having access to such data allowed for more effective training of deep neural networks.

  • Hardware (GPU): Neural networks are commonly trained on massive datasets, and often comprises of millions to billions of parameters. The introduction of Graphics Processing Units (GPUs) has been instrumental in facilitating this complex computation by offering accelerated processing power and parallel computing capabilities. Unlike Central Processing Units (CPUs), which have a limited number of cores capable of handling a few software threads at a time, GPUs consist of hundreds of cores capable of simultaneously managing thousands of threads. The increased availability of high-performance GPUs at affordable prices has played a pivotal role in the popularity and success of deep learning.

  • Software Breakthrough in Deep Learning architectures such as Transformers for Language Modeling, CNNs for Computer vision, made it possible to handle complexities in deep neural network and train them effectively Development of Deep Learning frameworks such as Pytorch, Tensorflow made it easier for developers to work with deep networks


Video Explanation (Playlist):

This playlist contains the following videos in the recommended order:

  • "The 'MIT Introduction to Deep Learning' is an introductory lecture from MIT's Deep Learning course, taught by Alexander Amini and Ava Amini. It explores the meaning of deep learning, its significance, applications, and the fundamentals of neural network training and regularization. This video provides a well-rounded understanding of deep learning and neural networks. https://www.youtube.com/watch?v=QDX-1M5Nj7s&list=PLtBw6njQRU-rwp5__7C0oIVt26ZgjG9NI
  • The video 'But what is a neural network?' by 3Blue1Brown explains the concept of deep learning using an image recognition example. The video helps you understand how a deep learning model learns for real-world applications. https://www.youtube.com/watch?v=aircAruvnKk


--

For more such articles, visit https://aiml.com

Looking for practice quizzes, https://aiml.com/quiz-category/technical/

(PS: Do sign up to take practice quizzes and bookmark your favorite questions)

#deeplearning #machineLearning









要查看或添加评论,请登录

Ritika Dokania的更多文章

社区洞察

其他会员也浏览了