<div><p>With the reinvigoration of neural networks in the 2000s, deep learning has become an extremely active area of research, one that’s paving the way for modern machine learning. In this practical book, author Nikhil Buduma provides examples and clear explanations to guide you through major conc
Fundamentals of Deep Learning: Designing Next-Generation Machine Intelligence Algorithms
✍ Scribed by Nithin Buduma, Nikhil Buduma, Joe Papa
- Publisher
- O'Reilly Media
- Year
- 2022
- Tongue
- English
- Leaves
- 390
- Edition
- 2
- Category
- Library
No coin nor oath required. For personal study only.
✦ Synopsis
We're in the midst of an AI research explosion. Deep learning has unlocked superhuman perception to power our push toward creating self-driving vehicles, defeating human experts at a variety of difficult games including Go, and even generating essays with shockingly coherent prose. But deciphering these breakthroughs often takes a PhD in machine learning and mathematics.
The updated second edition of this book describes the intuition behind these innovations without jargon or complexity. Python-proficient programmers, software engineering professionals, and computer science majors will be able to reimplement these breakthroughs on their own and reason about them with a level of sophistication that rivals some of the best developers in the field.
• Learn the mathematics behind machine learning jargon
• Examine the foundations of machine learning and neural networks
• Manage problems that arise as you begin to make networks deeper
• Build neural networks that analyze complex images
• Perform effective dimensionality reduction using autoencoders
• Dive deep into sequence analysis to examine language
• Explore methods in interpreting complex machine learning models
• Gain theoretical and practical knowledge on generative modeling
• Understand the fundamentals of reinforcement learning
✦ Table of Contents
Cover
Copyright
Table of Contents
Preface
Prerequisites and Objectives
How Is This Book Organized?
Conventions Used in This Book
Using Code Examples
O’Reilly Online Learning
How to Contact Us
Acknowledgements
Nithin and Nikhil
Joe
Chapter 1. Fundamentals of Linear Algebra for Deep Learning
Data Structures and Operations
Matrix Operations
Vector Operations
Matrix-Vector Multiplication
The Fundamental Spaces
The Column Space
The Null Space
Eigenvectors and Eigenvalues
Summary
Chapter 2. Fundamentals of Probability
Events and Probability
Conditional Probability
Random Variables
Expectation
Variance
Bayes’ Theorem
Entropy, Cross Entropy, and KL Divergence
Continuous Probability Distributions
Summary
Chapter 3. The Neural Network
Building Intelligent Machines
The Limits of Traditional Computer Programs
The Mechanics of Machine Learning
The Neuron
Expressing Linear Perceptrons as Neurons
Feed-Forward Neural Networks
Linear Neurons and Their Limitations
Sigmoid, Tanh, and ReLU Neurons
Softmax Output Layers
Summary
Chapter 4. Training Feed-Forward Neural Networks
The Fast-Food Problem
Gradient Descent
The Delta Rule and Learning Rates
Gradient Descent with Sigmoidal Neurons
The Backpropagation Algorithm
Stochastic and Minibatch Gradient Descent
Test Sets, Validation Sets, and Overfitting
Preventing Overfitting in Deep Neural Networks
Summary
Chapter 5. Implementing Neural Networks in PyTorch
Introduction to PyTorch
Installing PyTorch
PyTorch Tensors
Tensor Init
Tensor Attributes
Tensor Operations
Gradients in PyTorch
The PyTorch nn Module
PyTorch Datasets and Dataloaders
Building the MNIST Classifier in PyTorch
Summary
Chapter 6. Beyond Gradient Descent
The Challenges with Gradient Descent
Local Minima in the Error Surfaces of Deep Networks
Model Identifiability
How Pesky Are Spurious Local Minima in Deep Networks?
Flat Regions in the Error Surface
When the Gradient Points in the Wrong Direction
Momentum-Based Optimization
A Brief View of Second-Order Methods
Learning Rate Adaptation
AdaGrad—Accumulating Historical Gradients
RMSProp—Exponentially Weighted Moving Average of Gradients
Adam—Combining Momentum and RMSProp
The Philosophy Behind Optimizer Selection
Summary
Chapter 7. Convolutional Neural Networks
Neurons in Human Vision
The Shortcomings of Feature Selection
Vanilla Deep Neural Networks Don’t Scale
Filters and Feature Maps
Full Description of the Convolutional Layer
Max Pooling
Full Architectural Description of Convolution Networks
Closing the Loop on MNIST with Convolutional Networks
Image Preprocessing Pipelines Enable More Robust Models
Accelerating Training with Batch Normalization
Group Normalization for Memory Constrained Learning Tasks
Building a Convolutional Network for CIFAR-10
Visualizing Learning in Convolutional Networks
Residual Learning and Skip Connections for Very Deep Networks
Building a Residual Network with Superhuman Vision
Leveraging Convolutional Filters to Replicate Artistic Styles
Learning Convolutional Filters for Other Problem Domains
Summary
Chapter 8. Embedding and Representation Learning
Learning Lower-Dimensional Representations
Principal Component Analysis
Motivating the Autoencoder Architecture
Implementing an Autoencoder in PyTorch
Denoising to Force Robust Representations
Sparsity in Autoencoders
When Context Is More Informative than the Input Vector
The Word2Vec Framework
Implementing the Skip-Gram Architecture
Summary
Chapter 9. Models for Sequence Analysis
Analyzing Variable-Length Inputs
Tackling seq2seq with Neural N-Grams
Implementing a Part-of-Speech Tagger
Dependency Parsing and SyntaxNet
Beam Search and Global Normalization
A Case for Stateful Deep Learning Models
Recurrent Neural Networks
The Challenges with Vanishing Gradients
Long Short-Term Memory Units
PyTorch Primitives for RNN Models
Implementing a Sentiment Analysis Model
Solving seq2seq Tasks with Recurrent Neural Networks
Augmenting Recurrent Networks with Attention
Dissecting a Neural Translation Network
Self-Attention and Transformers
Summary
Chapter 10. Generative Models
Generative Adversarial Networks
Variational Autoencoders
Implementing a VAE
Score-Based Generative Models
Denoising Autoencoders and Score Matching
Summary
Chapter 11. Methods in Interpretability
Overview
Decision Trees and Tree-Based Algorithms
Linear Regression
Methods for Evaluating Feature Importance
Permutation Feature Importance
Partial Dependence Plots
Extractive Rationalization
LIME
SHAP
Summary
Chapter 12. Memory Augmented Neural Networks
Neural Turing Machines
Attention-Based Memory Access
NTM Memory Addressing Mechanisms
Differentiable Neural Computers
Interference-Free Writing in DNCs
DNC Memory Reuse
Temporal Linking of DNC Writes
Understanding the DNC Read Head
The DNC Controller Network
Visualizing the DNC in Action
Implementing the DNC in PyTorch
Teaching a DNC to Read and Comprehend
Summary
Chapter 13. Deep Reinforcement Learning
Deep Reinforcement Learning Masters Atari Games
What Is Reinforcement Learning?
Markov Decision Processes
Policy
Future Return
Discounted Future Return
Explore Versus Exploit
ϵ-Greedy
Annealed ϵ-Greedy
Policy Versus Value Learning
Pole-Cart with Policy Gradients
OpenAI Gym
Creating an Agent
Building the Model and Optimizer
Sampling Actions
Keeping Track of History
Policy Gradient Main Function
PGAgent Performance on Pole-Cart
Trust-Region Policy Optimization
Proximal Policy Optimization
Q-Learning and Deep Q-Networks
The Bellman Equation
Issues with Value Iteration
Approximating the Q-Function
Deep Q-Network
Training DQN
Learning Stability
Target Q-Network
Experience Replay
From Q-Function to Policy
DQN and the Markov Assumption
DQN’s Solution to the Markov Assumption
Playing Breakout with DQN
Building Our Architecture
Stacking Frames
Setting Up Training Operations
Updating Our Target Q-Network
Implementing Experience Replay
DQN Main Loop
DQNAgent Results on Breakout
Improving and Moving Beyond DQN
Deep Recurrent Q-Networks
Asynchronous Advantage Actor-Critic Agent
UNsupervised REinforcement and Auxiliary Learning
Summary
Index
About the Authors
Colophon
✦ Subjects
Machine Learning; Neural Networks; Deep Learning; Reinforcement Learning; Regression; Python; Convolutional Neural Networks; Recurrent Neural Networks; Autoencoders; Principal Component Analysis; TensorFlow; Gradient Descent; Logistic Regression; Long Short-Term Memory; Overfitting; Turing Machine; word2vec; Bellman Equation; Markov Models; Backpropagation
📜 SIMILAR VOLUMES
The neural network -- Training feed-forward neural networks -- Implementing neural networks in TensorFlow -- Beyond gradient descent -- Convolutional neural networks -- Embedding and representation learning -- Models for sequence analysis -- Memory augmented neural networks -- Deep reinforcement lea
With the reinvigoration of neural networks in the 2000s, deep learning has become an extremely active area of research that is paving the way for modern machine learning. This book uses exposition and examples to help you understand major concepts in this complicated field.<br><br>Large companies su
With the reinvigoration of neural networks in the 2000s, deep learning has become an extremely active area of research that is paving the way for modern machine learning. This book uses exposition and examples to help you understand major concepts in this complicated field.<br><br>Large companies su
<p>Machine learning (ML) and deep learning (DL) algorithms are invaluable resources for Industry 4.0 and allied areas and are considered as the future of computing. A subfield called neural networks, to recognize and understand patterns in data, helps a machine carry out tasks in a manner similar to