Personal tools

DL Research and Applications

Deep Learning Neural Network_030822A
[Deep Learning Neural Network - Pinterest]
 
 

Deep Learning: A Technique for Implementing Machine Learning

 

 

- Overview

Deep learning (DL) is a type of machine learning (ML) that uses artificial neural networks (ANNs) to teach computers to process data and make predictions based on complex patterns. 

DL models are trained on large amounts of data to learn to associate features in the data with the correct labels. For example, a DL model might learn to associate the shape and color of an object with the correct label, such as "dog" or "cat". 

DL can be used to automate tasks that typically require human intelligence, such as image recognition, natural language processing, and speech recognition. 

DL models use multilayered neural networks, called deep neural networks, with three or more layers. The adjective "deep" refers to the use of multiple layers in the network. 

DL models can be trained using supervised, semi-supervised, or unsupervised learning. Unsupervised learning allows DL models to extract characteristics, features, and relationships from raw, unstructured data. 

Graphics processing units (GPUs) are optimized for training DL models because they can process multiple computations simultaneously.

Organizations are increasingly turning to DL because it allows computers to learn independently and perform tasks with little supervision, bringing extraordinary benefits to science and industry. 

Unlike traditional ML, DL attempts to mimic the way our brains learn and process information by creating ANNs that can extract complex concepts and relationships from data. DL models improve on complex pattern recognition of pictures, text, sounds, and other data to generate more accurate insights and predictions. 

Some common DL architectures include convolutional neural networks, recurrent neural networks, transformers, and generative adversarial networks.

All recent advances in AI in recent years are due to DL. Without DL, we would not have self-driving cars, chatbots or personal assistants like Alexa and Siri. The Google Translate app would continue to be as primitive as 10+ years ago (before Google switched to neural networks for this App), and Netflix or Youtube would have no idea which movies or TV series we like or dislike. Behind all these technologies are neural networks.

Please refer to the following for more information:

 

- Deep Learning and Multi-layered ANNs

Deep Learning (DL) is a subset of Machine Learning (ML), which is a subset of Artificial Intelligence AI). DL is just a type of ML, inspired by the structure of a human brain. The adjective "deep" refers to the use of multiple layers in the network. Methods used can be either supervised, semi-supervised or unsupervised.

DL algorithms attempt to draw similar conclusions as humans would by continually analyzing data with a given logical structure. To achieve this, DL uses a multi-layered structure of algorithms called neural networks.  

DL uses ANNs to learn things. There are three types of ANNs: input layer, hidden layer, and output layer. DL is said to be deep because of its hidden layers. Before DL, in neural networks, it was difficult to solve a complex problem with only two layers. In DL, there are multiple hidden layers that form a larger network. It has the potential to solve complex problems.

Most modern DL models are based on multi-layered neural networks such as convolutional neural networks.

 

- Deep Learning with Big Data on GPUs

The AI boom took off when people realized that they can utilize GPU technology to train DL models much faster than waiting days for a general-purpose CPU to complete one cycle of model training. 

Since 2016, when GPU manufacturers like NVIDIA, Intel, and others created their first AI-optimized GPUs, most AI development has been related to how to train models to be as accurate and predictive as possible.

DL is a type of ML that trains a computer to perform human-like tasks, such as recognizing speech, identifying images or making predictions. Instead of organizing data to run through predefined equations, DL sets up basic parameters about the data and trains the computer to learn on its own by recognizing patterns using many layers of processing.

DL is probably one of the hottest tech topics right now. Large corporations and young startups alike are all gold-rushing this fancy field. If you think big data is important, then you should care about DL. The Economist says that data is the new oil in the 21st Century. If data is the crude oil, databases and data warehouses are the drilling rigs that digs and pumps the data on the Internet, then think of DL as the oil refinery that finally turns crude oil into all the useful and insightful final products. 

There could be a lot of “fossil fuels” hidden underground, and there are a lot of drills and pumps in the market, but without the right refinery tools, you ain’t gonna get anything valuable. That’s why DL is important. It’s part of the data-driven big picture.

The good news is, we are not going to run out of data and our “refinery machine” is getting better and better. Today, just about doing anything online will generate data. In the meantime, as long as the data isn’t garbage-in, then there’s no garbage-out from DL. Also, this “oil refinery” is improving on both software and hardware. 

DL algorithms have improved over the past few decades and developers around the world have contributed to open source frameworks like TensorFlow, Theano, Keras, and Torch, all of which make it easy for people to build DL algorithms as if playing with LEGO pieces. 

And thanks to the demand from gamers around the world, GPUs (graphics processing units) make it possible for us to leverage DL algorithms to build and train models with impressive results in a time-efficient manner.

 
DL Vs. ML_122623A
[Deep Learning Vs. Machine Learning - XenonStack]

- Deep Learning Accelerators 

Deep learning (DL) accelerators are hardware or computer systems that are designed to speed up the performance of DL algorithms. They are also known as AI accelerators, deep learning processors, or neural processing units (NPUs). 

DL accelerators are designed to improve performance, reduce latency, and reduce the cost of deploying ML applications. DL accelerators are used to improve the performance of ML applications, such as computer vision and ANNs. They are often used in data-intensive or sensor-driven tasks, such as robotics and the Internet of Things (IoTs).

Machine learning (ML) is widely used in many modern AI applications. Various hardware platforms are implemented to support such applications. Among them, graphics processing unit (GPU) is the most widely used one due to its fast computation speed and compatibility with various algorithms. 

Field programmable gate arrays (FPGA) show better energy efficiency compared with GPU when computing ML algorithm at the cost of low speed. Deep learning accelerators such as GPUs, FPGAs, and more recently TPUs. More companies have been announcing plans to design their own accelerators, which are widely used in data centers. 

There is also an opportunity to deploy them at the edge, initially for inference and for limited training over time. This also includes accelerators for very low power devices. The development of these technologies will allow ML (or smart devices) to be used in many IoT devices and appliances.

Some examples of DL accelerators include:

  • NVIDIA Deep Learning Accelerator (DLA): A fixed-function hardware accelerator engine that supports various layers, including convolution, deconvolution, and activation.
  • NVIDIA A100 GPU: A GPU developed by NVIDIA to execute the Megatron model, a transformer-based language neural network model for natural language processing. 

 

- Deep Learning and the Human Brain

Neural networks are based on the principles of the human brain. A neural network consists of many neurons and connections between them. A neuron can be represented as a function with several inputs and one output. Each neuron takes parameters from inputs (each input may have a different weight, which determines its importance), performs a specific function on them and gives the result to the output. The output of one neuron can be the input for another. Thus, multi-layer neural networks are formed, which are the subject of deep learning.   

DL is ML. More specifically, DL is considered an evolution of ML. It uses a programmable neural network that enables machines to make accurate decisions without help from humans. 

DL achieves great power and flexibility by learning to represent the world as nested hierarchy of concepts, with each concept defined in relation to simpler concepts, and more abstract representations computed in terms of less abstract ones.

DL software attempts to mimic the activity in layers of neurons in the neocortex, the wrinkly 80 percent of the brain where thinking occurs. The software learns, in a very real sense, to recognize patterns in digital representations of sounds, images, and other data.

The basic idea - that software can simulate the neocortex’s large array of neurons in an artificial “neural network” - is decades old, and it has led to as many disappointments as breakthroughs. 

But because of improvements in mathematical formulas and increasingly powerful computers, computer scientists can now model many more layers of virtual neurons than ever before.

 

- Deep Learning Architectures

While deep learning (DL) is certainly not new, it is experiencing explosive growth because of the intersection of deeply layered neural networks and the use of GPUs to accelerate their execution. 

Big data has also fed this growth. Because DL relies on training neural networks with example data and rewarding them based on their success, the more data, the better to build these DL structures.

The number of architectures and algorithms that are used in DL is wide and varied. Artificial neural network (ANN) is the underlying architecture behind DL. Based on ANN, several variations of the algorithms have been invented.

DL models, including various architectures like recurrent neural networks (RNN), convolutional neural networks (CNN), and deep belief networks (DBN), are structured in a specific manner to enable learning from complex data and making predictions or classifications. 

The architecture of a DL model typically consists of several interconnected layers, each serving a specific purpose in processing and transforming input data to generate useful outputs. 

 

- The Major Deep Learning Algorithms (Architectures)

A deep learning (DL) architecture is a neural network structure with multiple layers that allows machines to learn, understand, and make decisions. DL algorithms (architectures) are inspired by the human brain's neural networks.

DL architectures are made up of layers of artificial neurons that are trained to process data. The layers include an input layer, hidden layers, and an output layer.

DL architectures are used to create complex representations of data and solve problems in a variety of fields, including computer vision, natural language processing, and speech recognition.

DL architectures have led to advancements in how machines interact with and interpret information. They can extract meaningful features from data and continually learn. 

The design of the neural network is based on the structure of the human brain. Just as we use our brains to identify patterns and classify different types of information, neural networks can be taught to perform the same tasks on data.

There are five of the most popular DL algorithms (architectures):  

  • Recurrent neural networks
  • Long short-term memory/gated recurrent unit
  • Convolutional neural networks
  • Deep belief networks
  • Deep stacking networks
  • Transformers
  • Generative adversarial networks

 

[More to come ...]

 

Document Actions