Skip to content

Implementing Deep Learning Theoretical Principles with Python

License

Notifications You must be signed in to change notification settings

KXCY-AI/Python-DeepLearning

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Python-DeepLearning

Implementing Deep Learning Theoretical Principles with Python

中文说明

Deep learning, as one of the three pillars of next-generation AI--the other two being big data and cloud computing, is a critical theoretical component for implementing a plethora of machine learning and AI technologies, ranging from Autopilot to ChatGPT.

Simultaneously, the innovation of Python has grown dramatically in the last decade. You can use Python in statistics, website development, and facial recognition. On top of that, Python is also used in groundbreaking areas of IT, like machine learning and data science. And it makes Python one of the most practical tools for deep learning implementation.

In this repository, I only used Python to demonstrate the fundamentals of deep learning, with no assistance from third-party frameworks like TensorFlow or PyTorch. It is your first step toward deep learning. Initially, I wrote these notebooks and inline annotations in Chinese for my studies. And I'll keep this page updated with relevant notes for each topic. Please keep checking back later.

You can find the datasets in the notebooks by this link.

List of Contents

  1. Python Base for Deep Learning
  2. Perceptron Neural Network
  3. Activation Functions
  4. Three-Layer Neural Networks
  5. Softmax Activation Function
  6. Batch Size
  7. Loss Functions
  8. Gradient Descent
  9. Learning Algorithms
  10. Backpropagation
  11. Backpropagation in Affine and Softmax Layers
  12. Backpropagation in Two-Layer Neural Networks
  13. Gradient Descent Stochastic and Momentum
  14. Gradient Descent Optimization
  15. Gradient Descent Optimization with MNIST
  16. Weights and Biases initialization
  17. Weights and Biases initialization with MNIST
  18. Batch Normalization
  19. Underfitting and Overfitting
  20. Weight Decay
  21. Dropout Algorithm
  22. Implementing Convolutional Neural Networks

References:

About

Implementing Deep Learning Theoretical Principles with Python

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 100.0%