What Is Deep Learning (The Rise of Artificial Intelligence)


 

To support the production of more 

high-quality content 

consider supporting us on 

Patreon or 

YouTube membership. 

Additionally, consider visiting our 

parent company, EarthOne. 

For sustainable living made simple! 

Deep learning is a sub-field of AI 

that has taken the world by storm, 

in large part, since the start of this decade. 

In this sixth video in our artificial intelligence series 

and as for the purpose of this deep learning series, 

we'll explore why it has exploded in popularity, 

how deep learning systems work 

and their future applications. 

So sit back, relax and join me on an exploration 

into the field of deep learning. 

As I hope I've gotten across 

in past videos in this AI series, 

artificial intelligence refers to any model that can mimic, 

develop or demonstrate human thinking, 

perception or actions. 

In the case of AI that we all refer to today 

it is about computational intelligence, 

that being intelligence 

demonstrated by computing technology. 

Machine learning then is the ability 

for computers to learn without being explicitly programmed. 

In other words, 

having the ability to derive structure from data, 

that being the features of a data set, 

to be able to predict future outputs, 

whether that output be a value or a category. 

We discussed both supervised 

and unsupervised learning techniques 

quite in-depth in our machine learning series 

and going a step further 

also discussed the 5 tribes of machine learning, 

highlighting the different thought philosophies 

in giving computers learning capabilities. 

Over the past few years, 

the rise in artificial intelligence in mainstream popularity 

has been attributed to the successes 

of the connectionist tribe of machine learning, 

spawning a new field, deep learning. 

The goal of connectionism 

is to digitally re-engineer the brain 

and all of its connections. 

The title given to this digital reconstruction 

of the connections in a brain is a neural network. 

Deep learning then is the ability 

to learn the underlying features in a data set 

using neural networks. 

By this meaning, the ability to learn features 

directly from data rather than being hand-engineered. 

As quoted from a seminal paper on deep learning 

from the godfathers of the deep learning revolution, 

Geoffrey Hinton, Yoshua Bengio and Yann LeCun: 

Conventional machine-learning techniques 

were limited in their ability 

to process natural data in their raw form. 

For decades, constructing a pattern-recognition 

or machine-learning system required careful engineering 

and considerable domain expertise 

to design a feature extractor tht transformed the raw data, 

such as the pixel values of an image, 

into a suitable internal representation 

or feature vector from which the learning subsystem, 

often a classifier, could detect 

or classify patterns in the input. 

Representation learning is a set of methods 

that allows a machine to be feed with raw data 

and to automatically discover the representations needed 

for detection or classification. 

Deep-learning methods are representation-learning methods 

with multiple levels of representation, 

obtained by composing simple but non-linear modules 

that each transform the representation at one level, 

starting with the raw input, 

into a representation at a higher, 

slightly more abstract level. 

With the composition of enough such transformations, 

very complex functions can be learned. 

For classification tasks, 

higher layers of representation amplify aspects of the input 

that are important for discrimination 

and suppress irrelevant variations. 

Let's now take this definition 

of representation-based learning and apply it to an example, 

say of a neural net detecting a face in an input image. 

First off, this image is actually composed 

of an array of pixel values as the input to our system. 

Following that, the first few layers of representation 

learn low-level features like the presence 

or absence of lines and edges at particular orientations 

and locations in the image. The next layers of 

representation begin discerning mid-level features 

based off the low-level features previously found. 

This involves finding particular arrangements 

of edges and lines and producing features like eyes, 

noses and ears. 

Finally, the last layers of representation 

discover high-level features, 

that being a combination of mid-level features, 

to identify different facial structures. 

The key aspect of deep learning 

is that these layers of features 

are not chosen by human engineers, 

they are learned from data 

using a type of general-purpose learning procedure. 

This as opposed to the hand-engineered features 

of traditional machine learning algorithms, 

which is time consuming, costly, 

and overall not scalable for more complex applications. 

The true birth of deep learning, 

that being when the mainstream got wind of it, 

and where an algorithm publicly demonstrated 

this feature detection capability came back in 2012 

at the ImageNet competition. 

Since 2010, 

the annual ImageNet Large Scale Visual Recognition Challenge 

is a competition where research teams 

evaluate their algorithms on image data sets. 

These image data sets are comprised of millions of images 

in approximately 1000 image categories. 

Along with a variety of changing completion milestones 

the general year to year goal 

is to get as low as a possible error rate 

in the classification of images in the data set. 

In 2010, the inception of the challenge, 

the error rate was 28%. 

Progressing forward, in 2011, 

this error rate for the winning team dropped to 26%. 

However, as you can see there was quite a bit of spread 

in between the teams. 

In 2012, most teams began converging 

around the 25 to 30% range in error rate 

through the use of traditional image recognition, 

in other words classification machine learning algorithms, 

like those we discussed in our machine learning series. 

There was one major outlier in 2012 however, 

the winner, an algorithm dubbed SuperVision, 

with an error rate of 16%, 

nearly 10% better than its closest competitor. 

SuperVision, more academically referred to as AlexNet, 

is a deep convolutional neural network with eight layers, 

terms we will dissect later in this series. 

It was a network designed 

by University of Toronto professor Geoffrey Hinton 

and his graduate students. 

By the next year, 2013, 

nearly every top five team 

had switched to a deep learning approach, 

with the winning team having an error rate of 12% 

and eight layers. 

Progressing forward, 

in 2014 the winner had 22 layers and an error rate of 7% 

and by the next year it was the first time 

that the human record error rate on the dataset of 4.94% 

was breached by a network dubbed ResNet, 

comprised of 152 layers, and having an error rate of 4%. 

As an additional note, 

2016's winner had an error rate of 3% and 2017's of 2.25%. 

Had it not been for AlexNet in 2012, 

the error rate may have lingered around 20% 

for another few years. 

This because while most machine learning algorithms 

that have been in use for decades 

work great for smaller, structured data sets, 

when it comes to more real-word datasets, 

large and unstructured, the performance 

of machine learning algorithms falls drastically. 

Deep learning acts like a turbocharger of sorts 

for machine learning algorithms. 

As stated earlier, instead of manual feature extraction, 

the neural network acquires its own internal representations 

of the data, that being, the relevant features 

that describe it. 

To add to this, these networks can be trained 

in an unsupervised or supervised manner 

for both unsupervised and supervised tasks. 

As quoted from the paper discussing 

this breakthrough in 2012 titled, 

"ImageNet Classification 

"With Deep Convolutional Neural Networks": 

Our results show that a large, 

deep convolutional neural network 

is capable of achieving record-breaking results 

on a highly challenging dataset 

using purely supervised learning. 

To simplify our experiments, 

we did not use any unsupervised pre-training 

even though we expect that it will help, 

especially if we obtain enough computational power 

to significantly increase the size of the network 

without obtaining a corresponding increase 

in the amount of labeled data. 

Our results have improved 

as we've made our network larger and trained it longer. 

So, to reiterate, the AlexNet victory in 2012 

was a neural net trained in a supervised manner 

executing on an unsupervised dataset. 

As I'm sure is clear now, 

these networks and the results they produced 

translated to nearly every task 

machine learning algorithm's tackled 

beyond just image recognition. 

Neural nets in principle have existed for decades 

as we saw in the Birth of AI video, 

so why their sudden rise in effectiveness 

and popularity now? 

Well, as we discussed in the last video 

in this series on machine learning, 

some major factors include the pervasiveness of big data 

and increases in computing storage and power. 

Other major factors include 

increasingly parallelizable applications, 

streamlined software interfaces like TensorFlow, 

the list can go on and on. 

As stated earlier, 

conventional machine-learning techniques 

were limited in their ability 

to process natural data in its raw form, 

whereas deep learning benefits 

from vast amounts of training data to produce its models. 

In other words, deep learning is able to leverage big data. 

Also, deep learning, unlike traditional machine learning, 

excels at processing unstructured data, 

as the network's underlying structure 

derives the representations and features of a data set. 

Additionally, the composition of the networks, 

that being the mathematics these neural nets operate on, 

matrix operations, benefit greatly from parallel operation 

and therefore increase the computing performance. 

While this is a lot of terminology all at once, 

I promise this will all make more sense 

by the end of this deep learning series 

after we've discussed the structure of neural 

networks and how they work. 

Coming back on topic, as I hope you can see now, 

deep learning really is the culmination of advances 

in many fields and the hard work and dedication 

of various groups and individuals. 

When most people in the mainstream say AI now-a-days 

or when businesses use the term to hype a product, 

they are really referring to machine learning 

and when they say machine learning 

they really mean deep learning. 

Just like Matryoshka dolls, 

they are further subsets of one another. 

As I hope you took away from our machine learning series, 

machine learning on a high-level 

is essentially pattern recognition, 

a subset of intelligence. 

Deep learning then is turbo-charged pattern recognition. 

Over the past videos 

we have seen the field of connectionism evolve 

from the single-layer perceptron nets of the '60s, 

to the shallow multi-layer networks of the '80s, 

and many other innovations 

leading to the development of modern deep networks 

with tens to hundreds of layers, 

each building layers of representation 

and abstraction on top of one another. 

Now while it may take a lot of work 

to build a deep learning system 

as we'll see in the next videos in this series, 

If you want to learn more about 

deep learning, and I mean 

really learn about the field, 

from how these artificial learning algorithms 

were inspired from the brain 

to their foundational building blocks, 

the perceptron. Scaling up to 

multi-layer networks, different types 

of networks such as 

convolutional networks and recurrent 

networks and much more, 

then Brilliant.org is the place 

for you to go! 

Now what we love about how 

the topics in these courses are 

presented is that first an 

intuitive explanation is given and then 

you are taken through related problems. 

If you get a problem wrong 

you can see an explanation for where you 

went wrong, and how to rectify that flaw! 

In a world were automation through 

algorithms will increasingly replace more jobs, 

it is up to us as individuals to keep our 

brains sharp and think of creative solutions 

to multi-disciplinary problems! 

To support Futurology and learn more 

about Brilliant, go to: 

Brilliant.org Futurology 

and sign up for free! 

Additionally, the first 200 people that 

go to that link will get 20% off 

their annual premium subscription! 

(bright music) 

At this point the video has concluded, 

we'd like to thank you for taking 

the time to watch it! 

If you enjoyed it, consider 

supporting us on Patreon or 

YouTube membership to keep 

this brand growing! 

And if you have any topic 

suggestions, please leave them in 

the comments below. 

Consider subscribing for more content 

and check out our website and 

parent company, EarthOne, for 

more information! 

This has been Ankur, 

you've been watching Futurology 

and we'll see you again soon! 

(upbeat percussive music) 

SUBSCRIBE TO OUR NEWSLETTER

Seorang Blogger pemula yang sedang belajar