Python complete tutorial
  • Python Complete Tutorial
  • About this book
  • What you need to prepare
  • 1️⃣Try python for the first time
    • Install python
    • Hello world!
    • Hello world in a nutshell
    • The first simple python project
    • most useful libraries
    • Recommended websites
  • 2️⃣Data structure and basic operations
    • Python data structure
    • Data structure without hash table
    • Data structure with hash table
    • Variability and address
    • basic python programming
    • basic python programming 2
    • basic python programming 3
    • some additions
    • Fibonacci sequence
    • Judging prime numbers
    • txt/csv file operation
  • 🐍Practice program
    • 🚩fancy print
    • 🚩Remove duplicate elements
    • 🚩Palindrome detection
  • 😎leetcode
    • what is leetcode
  • 3️⃣Data mining and machine learning
    • What is data mining
    • iris data set
    • Mean median mode
    • Harmonic mean
    • Histogram
    • Correlation algorithm
    • Gaussian distribution data set
    • projection
    • PCA
    • MDS
    • Bayesian and Frequentist
    • Data normalization
    • binary SVM
    • One Hot Encoding
    • Multi-class SVM
    • Accuracy and error rate
    • Confusion matrix & Accuracy, Precision, Recall
    • F1 score
    • ROC and AUC
  • 4️⃣big data and data visualization
    • line chart
    • Parallel coordinates
    • Histogram chart
  • 5️⃣Mathematical algorithm and digital signal processing
    • Mathematical constants and basic operations
    • Normal distribution
    • Permutation and combination
    • Bernoulli distribution
    • Chaotic system
  • 6️⃣Classes and design patterns
    • Classes and design patterns
  • 7️⃣Operate the database with python
    • MySQL
      • Install MySQL
      • First try MySQL
      • MySQL Architecture
      • database operations
      • database
  • 8️⃣Cryptography
    • beginning of Cryptography
  • 9️⃣deep learning
    • What is Deep Learning
    • basic
  • 💔algorithm
    • Algorithms and Data Structures
Powered by GitBook
On this page
  1. deep learning

What is Deep Learning

Previousbeginning of CryptographyNextbasic

Last updated 3 years ago

Deep learning is a very popular concept recently. Such a concept is always hyped up by the media, but in fact deep learning or big data or blockchain or internet of things is something that already exists but has been newly packaged in the 21st century.

In the models we mentioned before, similar to SVM or random forest or Bayesian or Gaussian maximum likelihood Bayesian, the parameters in a model are generally less than 1000, but deep learning generally starts from 1000000 parameters.

This requires a lot of computation to update these million parameters, and also requires a good algorithm, otherwise it is very likely that it will not converge.

The 21st century solved both of these problems, especially after 2012, when large-scale graphics cards made large-scale computing possible.

The above is the gradient descent algorithm. This algorithm runs on the graphics card and calculates thousands of parameters at the same time until more than 1 million parameters are all updated.

Deep learning requires such a large amount of computation, so what is the benefit?

First of all, deep learning can perform multi-core computing, but algorithms such as SVM are not easy to perform multi-core computing, so many super-large data can be computed using deep learning, but using SVM will get stuck.

Second, deep learning can have better accuracy.

ILSVRC is an image classification challenge. Pursue the lowest error rate and the best accuracy rate. There are a total of 1000 different categories of pictures. The model needs to use the training set to train the labeled pictures and predict the pictures of the validation set.

ILSVRC is a subset of imagenet. The entire ILSVRC dataset is over 100GB in size and requires registration to download (not commercially available). When registering, you can only use the edu mailbox to obtain the data of the original image dataset.

AlexNet in 2012 was ILSVRC's first winning deep learning network, trained on two graphics cards, and deep learning has created one accuracy miracle after another.

Last but not least, deep learning requires a lot of computation during the training phase, but once the model is trained, the validation phase will require much less computation.

In addition to that, between 2012 and 2016 people have been going after bigger models, but after 2016 it was found that bigger models don't necessarily give better results, so people started checking at smaller models (of course loss some of the accuracy), but as technology improves, many small computational models now achieve better results than the large models of 2012-2016.

This allows many small boards or embedded devices to be loaded with deep learning. These deep learning models are more efficient, require less computation and are easy to optimize. More small embedded devices now even have deep learning cores to accelerate deep learning.

In later chapters I will introduce the basic neuron model and the backpropagation algorithm.

Statistics

Start time of this page: February 2, 2022

Completion time of this page: February 12, 2022

9️⃣
Page cover image
Gradient Descent Algorithm
ILSVRC challenge
imagenet
NVIDIA Jetson Nano