{ "nbformat": 4, "nbformat_minor": 0, "metadata": { "colab": { "name": "Index.ipynb", "version": "0.3.2", "provenance": [], "collapsed_sections": [] }, "kernelspec": { "name": "python3", "display_name": "Python 3" } }, "cells": [ { "cell_type": "markdown", "metadata": { "id": "8wN5BpUKYFHA", "colab_type": "text" }, "source": [ "# Deep Learning with Tensorflow 2.0" ] }, { "cell_type": "markdown", "metadata": { "id": "fwY_CgcMYFAr", "colab_type": "text" }, "source": [ "![Img](https://raw.githubusercontent.com/adhiraiyan/DeepLearningWithTF2.0/master/logo/DLBookCover2.png)\n", "\n", "A practical guide to [Deep Learning with Tensorflow 2.0](https://www.adhiraiyan.org/DeepLearningWithTensorflow.html) by Mukesh Mithrakumar. \n", "The content is available [on GitHub](https://github.com/adhiraiyan/DeepLearningWithTF2.0) and you can run it in [Google Colaboratory](https://colab.research.google.com/github/adhiraiyan/DeepLearningWithTF2.0/blob/master/notebooks/Index.ipynb) as well. The code is released under the [MIT license](https://opensource.org/licenses/MIT)\n", "\n", "---" ] }, { "cell_type": "markdown", "metadata": { "id": "w0FfjM53YE4C", "colab_type": "text" }, "source": [ "# Table of Contents\n", "\n", "## [1. Introduction](01.00-Introduction.ipynb)\n", "- 01.01 Who should read this book\n", "- 01.02 Historical Trends in Deep Learning\n", "\n", "## [2. Linear Algebra](02.00-Linear-Algebra.ipynb)\n", "- 02.01 Scalars, Vectors, Matrices and Tensors\n", "- 02.02 Multiplying Matrices and Vectors\n", "- 02.03 Identity and Inverse Matrices\n", "- 02.04 Linear Dependence and Span\n", "- 02.05 Norms\n", "- 02.06 Special Kinds of Matrices and Vectors\n", "- 02.07 Eigendecomposition\n", "- 02.08 Singular Value Decomposition\n", "- 02.09 The Moore-Penrose Pseudoinverse\n", "- 02.10 The Trace Operator\n", "- 02.11 The Determinant\n", "- 02.12 Example: Principal Components Analysis\n", "\n", "\n", "## [3. Probability and Information Theory](03.00-Probability-and-Information-Theory.ipynb)\n", "- 03.01 Why Probability?\n", "- 03.02 Random Variables\n", "- 03.03 Probability Distributions\n", "- 03.04 Marginal Probability\n", "- 03.05 Conditional Probability\n", "- 03.06 The Chain Rule of Conditional Probabilities\n", "- 03.07 Independence and Conditional Independence\n", "- 03.08 Expectation, Variance and Covariance\n", "- 03.09 Common Probability Distributions\n", "- 03.10 Useful Properties of Common Functions\n", "- 03.11 Bayes' Rule\n", "- 03.12 Technical Details of Continuous Variables\n", "- 03.13 Information Theory\n", "- 03.14 Structured Probabilistic Models\n", "\n", "\n", "## [4. Numerical Computation](04.00-Numerical-Computation.ipynb)\n", "- 04.01 Overflow and Underflow\n", "- 04.02 Poor Conditioning\n", "- 04.03 Gradient-Based Optimization\n", "- 04.04 Constrained Optimization\n", "- 04.05 Example: Linear Least Squares\n", "\n", "\n", "## [5. Machine Learning Basics](05.00-Machine-Learning-Basics.ipynb)\n", "- 05.01 Learning Algorithms\n", "- 05.02 Capacity, Overfitting and Underfitting\n", "- 05.03 Hyperparameters and Validation Sets\n", "- 05.04 Estimators, Bias and Variance\n", "- 05.05 Maximum Likelihood Estimation\n", "- 05.06 Bayesian Statistics\n", "- 05.07 Supervised Learning Algorithms\n", "- 05.08 Unsupervised Learning Algorithms\n", "- 05.09 Stochastic Gradient Descent\n", "- 05.10 Building a Machine Learning Algorithm\n", "- 05.11 Challenges Motivating Deep Learning\n", "\n", "\n", "## [6. Deep Feedforward Networks](06.00-Deep-Feedforward-Networks.ipynb)\n", "- 06.01 Example: Learning XOR\n", "- 06.02 Gradient-Based Learning\n", "- 06.03 Hidden Units\n", "- 06.04 Architecture Design\n", "- 06.05 Back-Propagation and Other Differentiation Algorithms\n", "- 06.06 Historical Notes\n", "\n", "\n", "## [7. Regularization for Deep Learning](07.00-Regularization-for-Deep-Learning.ipynb)\n", "- 07.01 Parameter Norm Penalties\n", "- 07.02 Norm Penalties as Constrained Optimization\n", "- 07.03 Regularization and Under-Constrained Problems\n", "- 07.04 Dataset Augmentation\n", "- 07.05 Noise Robustness\n", "- 07.06 Semi-Supervised Learning\n", "- 07.07 Multitask Learning\n", "- 07.08 Early Stopping\n", "- 07.09 Parameter Tying and Parameter Sharing\n", "- 07.10 Sparse Representations\n", "- 07.11 Bagging and Other Ensemble Methods\n", "- 07.12 Dropout\n", "- 07.13 Adversarial Training\n", "- 07.14 Tangent Distance, Tangent Prop and Manifold Tangent Classifier\n", "\n", "\n", "## [8. Optimization for Training Deep Models](08.00-Optimization-for-Training-Deep-Models.ipynb)\n", "- 08.01 How Learning Differs from Pure Optimization\n", "- 08.02 Challenges in Neural Network Optimization\n", "- 08.03 Basic Algorithms\n", "- 08.04 Parameter Initialization Strategies\n", "- 08.05 Algorithms with Adaptive Learning Rates\n", "- 08.06 Approximate Second-Order Methods\n", "- 08.07 Optimization Strategies and Meta-Algorithms\n", "\n", "\n", "## [9. Convolutional Networks](09.00-Convolutional-Networks.ipynb)\n", "- 09.01 The Convolution Operation\n", "- 09.02 Motivation\n", "- 09.03 Pooling\n", "- 09.04 Convolution and Pooling as an Infinitely Strong Prior\n", "- 09.05 Variants of the Basic Convolution Function\n", "- 09.06 Structured Outputs\n", "- 09.07 Data Types\n", "- 09.08 Efficient Convolution Algorithms\n", "- 09.09 Random or Unsupervised Features\n", "- 09.10 The Neuroscientific Basis for Convolutional Networks\n", "- 09.11 Convolutional Networks and the History of Deep Learning\n", "\n", "\n", "## [10. Sequence Modeling: Recurrent and Recursive Nets](10.00-Sequence-Modeling-Recurrent-and-Recursive-Nets.ipynb)\n", "- 10.01 Unfolding Computational Graphs\n", "- 10.02 Recurrent Neural Networks\n", "- 10.03 Bidirectional RNNs\n", "- 10.04 Encoder-Decoder Sequence-to-Sequence Architectures\n", "- 10.05 Deep Recurrent Networks\n", "- 10.06 Recursive Neural Networks\n", "- 10.07 The Challenge of Long-Term Dependencies\n", "- 10.08 Echo State Networks\n", "- 10.09 Leaky Units and Other Strategies for Multiple Time Scales\n", "- 10.10 The Long Short-Term Memory and Other Gated RNNs\n", "- 10.11 Optimization for Long-Term Dependencies\n", "- 10.12 Explicit Memory\n", "\n", "\n", "## [11. Practical Methodology](11.00-Practical-Methodology.ipynb)\n", "- 11.01 Performance Metrics\n", "- 11.02 Default Baseline Models\n", "- 11.03 Determining Whether to Gather More Data\n", "- 11.04 Selecting Hyperparameters\n", "- 11.05 Debugging Strategies\n", "- 11.06 Example: Multi-Digit Number Recognition\n", "\n", "\n", "## [12. Applications](12.00-Applications.ipynb)\n", "- 12.01 Large-Scale Deep Learning\n", "- 12.02 Computer Vision\n", "- 12.03 Speech Recognition\n", "- 12.04 Natural Language Processing\n", "- 12.05 Other Applications\n", "\n", "\n", "## [13. Linear Factor Models](13.00-Linear-Factor-Models.ipynb)\n", "- 13.01 Probabilistic PCA and Factor Analysis\n", "- 13.02 Independent Component Analysis\n", "- 13.03 Slow Feature Analysis\n", "- 13.04 Sparse Coding\n", "- 13.05 Manifold Interpretation of PCA\n", "\n", "\n", "## [14. Autoencoders](14.00-Autoencoders.ipynb)\n", "- 14.01 Undercomplete Autoencoders\n", "- 14.02 Regularized Autoencoders\n", "- 14.03 Representational Power, Layer Size and Depth\n", "- 14.04 Stochastic Encoders and Decoders\n", "- 14.05 Denoising Autoencoders\n", "- 14.06 Learning Manifolds with Autoencoders\n", "- 14.07 Contractive Autoencoders\n", "- 14.08 Predictive Sparse Decomposition\n", "- 14.09 Applications of Autoencoders\n", "\n", "\n", "## [15. Representation Learning](15.00-Representation-Learning.ipynb)\n", "- 15.01 Greedy Layer-Wise Unsupervised Pretraining\n", "- 15.02 Transfer Learning and Domain Adaptation\n", "- 15.03 Semi-Supervised Disentangling of Causal Factors\n", "- 15.04 Distributed Representation\n", "- 15.05 Exponential Gains from Depth\n", "- 15.06 Providing Clues to Discover Underlying Causes\n", "\n", "\n", "## [16. Structured Probabilistic Models for Deep Learning](16.00-Structured-Probabilistic-Models-for-Deep-Learning.ipynb)\n", "- 16.01 The Challenge of Unstructured Modeling\n", "- 16.02 Using Graphs to Describe Model Structure\n", "- 16.03 Sampling from Graphical Models\n", "- 16.04 Advantages of Structured Modeling\n", "- 16.05 Learning about Dependencies\n", "- 16.06 Inference and Approximate Inference\n", "- 16.07 The Deep Learning Approach to Structured Probabilistic Models\n", "\n", "\n", "## [17. Monte Carlo Methods](17.00-Monte-Carlo-Methods.ipynb)\n", "- 17.01 Sampling and Monte Carlo Methods\n", "- 17.02 Importance Sampling\n", "- 17.03 Markov Chain Monte Carlo Methods\n", "- 17.04 Gibbs Sampling\n", "- 17.05 The Challenge of Mixing between Separated Modes\n", "\n", "\n", "## [18. Confronting the Partition Function](18.00-Confronting-the-Partition-Function.ipynb)\n", "- 18.01 The Log-Likelihood Gradient\n", "- 18.02 Stochastic Maximum Likelihood and Contrastive Divergence\n", "- 18.03 Pseudolikelihood\n", "- 18.04 Score Matching and Ratio Matching\n", "- 18.05 Denoising Score Matching\n", "- 18.06 Noise-Contrastive Estimation\n", "- 18.07 Estimating the Partition Function\n", "\n", "\n", "## [19. Approximate Inference](19.00-Approximate-Inference.ipynb)\n", "- 19.01 Inference as Optimization\n", "- 19.02 Expectation Maximization\n", "- 19.03 MAP Inference and Sparse Coding\n", "- 19.04 Variational Inference and Learning\n", "- 19.05 Learned Approximate Inference\n", "\n", "\n", "## [20. Deep Generative Models](20.00-Deep-Generative-Models.ipynb)\n", "- 20.01 Boltzmann Machines\n", "- 20.02 Restricted Boltzmann Machines\n", "- 20.03 Deep Belief Networks\n", "- 20.04 Deep Boltzmann Machines\n", "- 20.05 Boltzmann Machines for Real-Valued Data\n", "- 20.06 Convolutional Boltzmann Machines\n", "- 20.07 Boltzmann Machines for Structured or Sequential Outputs\n", "- 20.08 Other Boltzmann Machines\n", "- 20.09 Back-Propagation through Random Operations\n", "- 20.10 Directed Generative Nets\n", "- 20.11 Drawing Samples from Autoencoders\n", "- 20.12 Generative Stochastic Networks\n", "- 20.13 Other Generation Schemes\n", "- 20.14 Evaluating Generative Models\n", "- 20.15 Conclusion" ] } ] }