Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Deep Learning - Basic Elements #8

Open
NicolaBernini opened this issue Feb 27, 2019 · 2 comments
Open

Deep Learning - Basic Elements #8

NicolaBernini opened this issue Feb 27, 2019 · 2 comments
Assignees
Labels
body of knowledge Basic Elements of Foundational Body of Knowledge

Comments

@NicolaBernini
Copy link
Owner

Overview

Basic Elements about Deep Learning

@NicolaBernini
Copy link
Owner Author

Deep Learning vs Machine Learning

Similarities

  • In both cases the goal is to train a model hence at high level it deals with the same tools: dataset, loss function, optimization, ...

Differencies

  • Machine Learning Models have typically much less learning capacity of Deep Learning Models

  • The reason for this is their input domain is typically low dimensional (e.g. scalar time series, not images, videos, ...) and that's because

    • or they aim to work on that specific domain
    • or they rely on manually engineered features
      • feature extraction is essentially a projection into a lower dimensional space
  • Deep Learning Models target high dimensional spaces as their input (e.g. images, videos) and they do not need any manually engineered feature

  • In fact they automatically learn the most appropriate representation for their data

    • most appropriate, for them to achieve their goal which is essentially minimizing the loss function values
  • This automatic representation learning is the result of the inductive bias related to the architecture consisting of an hierarchy of layers : this choice forces the network to learn a hierarchy of layer specific representations so that each layer representation depends on the previous layer one and the first layer depends on the high dimensional input representation

  • This requires the Deep Learning Models to have a much higher learning capacity (as they need to learn both the hierarchical representation and how to solve their problem) and it creates a lot of additional complexities with respect to machine learning models like

    • loss function landscape becomes highly non-linear hence optimization becomes a hard problem (this is approached training the network)
    • higher learning capacity increases significantly the chances of overfitting, hence the model becomes very data hungry and specific techniques for data augmentation, regularization, ... are required to achieve good results

@NicolaBernini NicolaBernini self-assigned this Feb 27, 2019
@NicolaBernini NicolaBernini added the body of knowledge Basic Elements of Foundational Body of Knowledge label Feb 27, 2019
@NicolaBernini
Copy link
Owner Author

RNN

  • The Recurrent Neural Network is a kind of NN whose topology changes according to the specific perspective it is observed
    • in a time agnostic perspective it has a recurrent topology
    • in a time specific perspective it has a directed topology (directed edges) as each new state depends on the previous state

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
body of knowledge Basic Elements of Foundational Body of Knowledge
Projects
None yet
Development

No branches or pull requests

1 participant