Hi, I'm Ankit currently I enrolled for AWS Training and Certification.
Transcript
Interactive Transcript - Enable basic transcript mode by pressing the escape key
You may navigate through the transcript using tab. To save a note for a section of text press CTRL + S. To expand your selection you may use CTRL + arrow key. You may contract your selection using shift + CTRL + arrow key. For screen readers that are incompatible with using arrow keys for shortcuts, you can replace them with the H J K L keys. Some screen readers may require using CTRL in conjunction with the alt key
Hi, I'm Dan Mbaga with AWS AI.
Welcome to the Introduction to Deep Learning Course.
I have been with as for four years and I'm currently
responsible for business development
management in machine learning.
As part of AWS AI Team,
have worked with our customers
to build their machine learning
products from conception to production on AWS.
In this video, you will learn about
deep-learning or DL and about
the services that AWS offers for
developing deep-learning based applications.
We'll also discuss a case study where
one of our customers is innovating with Deep Learning.
DL is a subset of machine learning,
which itself is a subset of Artificial Intelligence.
Scientists started studying deep-learning in 1950s and
devoted significant resources to
wait for the next 70 years.
The foundation for the current error
was laid in the 1980s and
the 1990s with research from
Yann LeCun on Convolutional Neural Networks,
and along short-term memories or
LSTM by Sepp Hochreiter and Juergen Schmidhuber.
In 1986, the rediscovery of
the backpropagation training algorithm
marked a significant milestone
in the study of Deep Learning.
The backpropagation algorithm helps the model learn from
his mistakes by leveraging the chain rule of derivatives.
But in the decades that followed called
a neural winter research into Deep Learning dropped off.
This was partly due to limitations on data and compute.
The introduction of the Internet, smartphones, smart TVs,
and the availability of inexpensive digital cameras,
meant that more and more data was
available.. Computing power was on the rise.
CPUs were becoming faster,
and GPUs became a general-purpose computing tool.
Both trends made neural networks progress.
In 1998, Yann LeCun publish a paper
on convolutional neural networks
for image recognition tasks.
But it wasn't until 2007
when the research began to accelerate again.
The advent of GPU and reduction in training time,
ushered in the mainstay
of Neural Networks and Deep Learning.
Both data and computing power made the task
that neural networks tackled more and more interesting.
With GPS becoming increasingly popular,
Neural Networks resurface in 2008.
Deep-learning uses Artificial Neural Networks to
process and evaluate its reference data
and come to a conclusion.
Artificial Neural Networks or ANN
are different from traditional
Compute Processing Architectures.
In that, they're designed to
operate more like human brain.
The more flexible and better
at handling unanticipated anomalies,
and novelty is in a data.
We'll talk more about Artificial Neural Networks later.
Deep Learning is a subset of machine learning algorithms.
Deep-learning uses layers of
non-linear Processing Units for
features extraction and transformation.
Each successive layer uses
the output from the previous layer as an input.
The algorithms may be supervised or
unsupervised and applications include pattern analysis,
which is unsupervised, and
classification which could be supervised or unsupervised.
These algorithms are also
based on the unsupervised learning of
multiple levels of features
or representations of the data.
Higher-level features are derived from
low-level features to form a hierarchical representation.
Deep learning algorithms are part of
a broader Machine Learning field
of learning representations of data,
and they learn multiple levels of
representations that correspond to
different levels of abstraction.
Where traditional Machine Learning
focuses on feature engineering,
Deep Learning focuses on
end-to-end learning based on raw features.
Each layer is responsible for
analyzing additional complex features in the data.
A Neural network is a collection of
simple trainable mathematical unit
that collectively learn complex functions.
With enough training data,
a Neural Network can perform a decent job of mapping
input data and features to output decisions.
It consists of multiple layers.
There is an input layer,
some hidden layers, and an output layer.
The basic unit of an Artificial Neural Network is
Artificial Neuron sometimes also called a node.
Like the biological neurons for which they are named,
artificial neurons have several input channels.
A neuron sums these inputs inside of processing stage,
and produces one output that can
fan out to multiple other artificial neurons.
In this simplified example,
the input values are multiplied by
the weight to get their weighted value.
Then if appropriate, the node adds
an offset vector to the sum called the bias,
which adjusts to sum to
generate more accurate predictions,
based on the success or
failure of the product predictions.
Once the inputs have been weighted then sumed,
and the bias is added if appropriate,
the neuron activates if the final value produced by
the preceding steps meets or
exceeds the determined activation threshold.
That's called the activation step.
It's the final step before an output is deliver.
The Feedforward Neural Network is
any neural network that doesn't
form a cycle between neurons.
This means data moves from one input
to output without looping backward.
This another Neural Network that does look backwards,
and that's called Recurrent Neural Network.
The primary value of a recurrent neural network
comes when processing
sequential information such as text,
or speech, or handwriting.
Where the ability to predict the next word or
later is vastly improve if you're factoring in the words,
or letters that come before it.
Recurrent Neural Networks became much more popular
after 2007 when Long Short Term
Memory or LSTM approaches
revolutionized speech recognition programs.
LSTM is now the basis for many of
today's most successful applications
in the speech recognition domain,
text-to-speech domain,
and handwriting recognition domain.
Use cases for Deep Learning span multiple industries.
Let's have a look at each.
You can find text analysis use cases in a finance,
social, CRM, and insurance domains to name a few.
It's used to detect insider training.
Check for regulatory compliance.
Brand affinity.
Sentiment analysis.
Intent Analysis, and more
by essentially analyzing blobs of text.
Deep Learning is also used to solve problems
around time-series and predictive analysis.
It's using datacenters for
Log Analysis and risk fraud detection,
by the supply chain industry for resource planning,
and in the IoT field for
predictive analysis using sensor data.
It's also used in social media,
and e-commerce for building recommendation engines.
It's using sound analysis too.
You find Deep Learning being used in
the security domain for voice recognition,
voice analysis, and in
the CRM domain for sentiment analysis.
You'll also find deep learning in
both the automotive and aviation industries,
where it's used for Engine
and instrument floor detection.
You'll even find deep learning
in the finance industry for
credit card fraud detection among other things.
Finally, it's used for image analysis.
In the security domain,
Deep Learning is used for things like facial recognition.
In social media, it's used for
tagging and identifying people in pictures..
The challenge of course is scale.
In 2012, AlexNet won
the convolutional neural network ImageNet competition.
It consisted of eight layers,
650,000 interconnected neurons,
and almost 60 million parameters.
Today, the complexity of
Neural Networks has increased significantly.
With recent networks such as Resnet 152,
a Deep Residual Neural Network which has a 152 layers,
and millions more connected neurons in parameters.
The AWS Platform offers
three advanced Deep Learning
enabled managed API services.
Amazon Lex, Amazon Polly, and Amazon Rekognition.
Amazon Lex is a service for building
conversational interfaces into any application
using voice and text.
It provides that advanced Deep Learning functionalities
of automatic speech recognition,
for converting speech-to-text, and natural
language understanding to recognize
the intent of the input.
That enables you to build applications with
highly engaging user experiences,
and life-like conversational interactions.
Amazon Polly turns tags into lifelike speech.
Allowing you to create applications that talk,
and build entirely new categories
of speech enabled products.
Amazon Rekognition makes it easy to
add image analysis to your applications,
so that your application can detect objects,
scenes, and faces, and images.
You can also search and compare faces,
recognize celebrities,
and identify inappropriate content.
Deep Learning can often be technically challenging.
Requiring you to understand
the math of the models themselves,
and the experience in skating, training,
and inference across large distributed systems.
As a result, several
Deep Learning frameworks have emerged,
which allow you to define
models and then train them at scale.
You can build custom models
using the Amazon deep-learning AMIs.
Built for Amazon Linux and Ubuntu.
The AWS Deep Learning AMIs come
pre-configured with Apache MXnet TensorFlow,
the Microsoft Cognitive Toolkit Caffe,
Caffe2, theano, torch, Pytorch and Keras.
The Deep Learning AMIs enable you to quickly
deploy and run any of these frameworks at scale.
The Deep Learning AMIs can help you get started quickly.
They're provisioned with many deep learning frameworks
including tutorials that demonstrate proper installation,
configuration, and model accuracy.
The Deep Learning AMIs install dependencies,
track library versions, and validate code compatibility.
With updates to the AMIs every month,
you always have the latest versions of
the engines in data science libraries.
Whether you need Amazon EC2 GPU or CPU Instances.
There's no additional charge for the deep learning AMIs.
You only pay for the AWS resources that you
need to store and run your applications.
There are two ways to get started
with AWS Deep Learning AMIs;
you can deploy a deep-learning
Compute Instance in one click.
The AWS Deep Learning AMIs can
quickly be launched from AWS marketplace.
You have the choice of GPUs for large-scale training,
and CPUs for running predictions or inferences.
Both of them give you a stable, secured,
and high-performance execution environment to run
your applications with pay-as-you-go pricing model.
The other way to get started with
the Deep Learning AMIs is
to launch an AWS Cloud Formation Deep Learning template.
To train over multiple instances,
you use the Deep Learning CloudFormation template
for a simple way to
launch all of your resources
quickly using the deep learning AMIs.
Now, let's talk about a use case.
C-Span is non-for-profits surveys focused on
broadcasting and archiving US government proceedings.
C-span had developed
an automated facial recognition solution
to help human indexers, but it will slow.
They could only index half of
the incoming content by speaker
limiting the ability of users to search archive content.
So what are the solutions and benefit that they had
using Amazon Rekognition as a Deep Learning Service?
They implemented Amazon recognition
to automatically match,
uploaded screenshots of a collection
of 97,000 known faces.
That enables C-span to
more than double the video index
from $3,500-7,500 per year.
It drove down to labor required to index an hour of
video content from 60 minutes to 20 minutes.
They deployed it in less than three weeks,
and index at 97,000
image collection in less than two hours.
I hope you learned a little something,
and we'll continue to explore other courses.
I'm Dan Mbaga with AWS AI, and thanks for watching.
Transcript Help Us Translate Interactive Transcript - Enable basic transcript mode by pressing the escape key You may navigate through
0 Comments