Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Save more on your purchases! discount-offer-chevron-icon
Savings automatically calculated. No voucher code required.
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletter Hub
Free Learning
Arrow right icon
timer SALE ENDS IN
0 Days
:
00 Hours
:
00 Minutes
:
00 Seconds
Arrow up icon
GO TO TOP
Deep Learning By Example

You're reading from   Deep Learning By Example A hands-on guide to implementing advanced machine learning algorithms and neural networks

Arrow left icon
Product type Paperback
Published in Feb 2018
Publisher Packt
ISBN-13 9781788399906
Length 450 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Author (1):
Arrow left icon
 Menshawy Menshawy
Author Profile Icon Menshawy
Menshawy
Arrow right icon
View More author details
Toc

Table of Contents (23) Chapters Close

Title Page
Copyright and Credits
Packt Upsell
Contributors
Preface
1. Data Science - A Birds' Eye View FREE CHAPTER 2. Data Modeling in Action - The Titanic Example 3. Feature Engineering and Model Complexity – The Titanic Example Revisited 4. Get Up and Running with TensorFlow 5. TensorFlow in Action - Some Basic Examples 6. Deep Feed-forward Neural Networks - Implementing Digit Classification 7. Introduction to Convolutional Neural Networks 8. Object Detection – CIFAR-10 Example 9. Object Detection – Transfer Learning with CNNs 10. Recurrent-Type Neural Networks - Language Modeling 11. Representation Learning - Implementing Word Embeddings 12. Neural Sentiment Analysis 13. Autoencoders – Feature Extraction and Denoising 14. Generative Adversarial Networks 15. Face Generation and Handling Missing Labels 16. Implementing Fish Recognition 1. Other Books You May Enjoy Index

Index

A

  • activation function
    • about / Biological motivation and connections, Activation functions
    • sigmoid / Sigmoid
    • Tanh / Tanh
    • ReLU / ReLU
  • advertising example, linear regression model
    • about / Advertising – a financial example
    • dependencies / Dependencies
    • data, importing with pandas / Importing data with pandas
    • advertising data / Understanding the advertising data
    • data analysis / Data analysis and visualization
    • visualization / Data analysis and visualization
    • simple regression model / Simple regression model
  • alpha / Leaky ReLU
  • apparent error / Apparent (training set) error
  • architecture design, feed-forward neural network
    • input layer / Hidden units and architecture design
    • hidden layer / Hidden units and architecture design
    • output layer / Hidden units and architecture design
  • autoencoder, for compressing MNIST dataset
    • building / Compressing the MNIST dataset
    • MNIST dataset, obtaining / The MNIST dataset
    • model, building / Building the model
    • model, training / Model training
  • autoencoders
    • about / Introduction to autoencoders
    • examples / Examples of autoencoders
    • architecture / Autoencoder architectures
    • hyperparameters / Autoencoder architectures
    • convolutional autoencoder (CAE) / Convolutional autoencoder
    • denoising autoencoders / Denoising autoencoders
    • application / Applications of autoencoders
    • for image colorization / Image colorization
    • for image enhancement / More applications

B

  • backpropagation, of gradients / Exploding and vanishing gradients - recap
  • backpropagation algorithm, MLP
    • about / Training our MLP – the backpropagation algorithm
    • working / Training our MLP – the backpropagation algorithm
    • forward propagation / Step 1 – forward propagation
    • weight updation / Step 2 – backpropagation and weight updation
  • backpropagation through time (BPTT) / The intuition behind RNNs
  • batch normalization
    • about / Data analysis and preprocessing
    • benefits / Data analysis and preprocessing
  • bias-variance decomposition / Bias-variance decomposition

C

  • Cabin feature / Cabin
  • categorical variables / Feature transformations
  • CelebA dataset / Face generation
  • challenges, machine learning
    • overview / Challenges of learning
    • feature extraction / Feature extraction – feature engineering
    • noise / Noise
    • overfitting / Overfitting
    • machine learning algorithm selection / Selection of a machine learning algorithm
    • prior knowledge / Prior knowledge
    • missing values / Missing values
  • character-level language models
    • about / Character-level language models
    • Shakespeare data, using / Language model using Shakespeare data
  • cheat sheet / Model training
  • checkpoints
    • reference / Saving checkpoints
  • CIFAR-10 object detection
    • pre-trained model, using as feature extractor / CIFAR-10 object detection – revisited
    • solution outline / Solution outline
    • packages, exploring / Loading and exploring CIFAR-10
    • packages, loading / Loading and exploring CIFAR-10
    • inception model transfer values, calculating / Inception model transfer values
    • transfer values, analyzing / Analysis of transfer values
    • model, training / Model building and training
    • model, building / Model building and training
  • computational graphs
    • about / Computational graphs
    • data types / TensorFlow data types, variables, and placeholders
    • variables / TensorFlow data types, variables, and placeholders, Variables
    • placeholders / TensorFlow data types, variables, and placeholders, Placeholders
    • mathematical operations / Mathematical operations
  • conda / Importing data with pandas
  • Continuous Bag-of-Words (CBOW) / Building Word2Vec model
  • convolutional autoencoder (CAE)
    • about / Convolutional autoencoder
    • MNIST dataset, using / Dataset
    • model, building / Building the model
    • model, training / Model training
  • convolutional neural network (CNN)
    • about / What this book covers, Face generation
    • motivation / Motivation
    • applications / Applications of CNNs
    • layers / Different layers of CNNs
    • MNIST digit classification example / CNN basic example – MNIST digit classification
    • URL / Face generation
  • convolution operation
    • about / The convolution operation
    • stride / The convolution operation
    • zero-padding / The convolution operation
  • convolution step, CNN / Convolution step
  • CUDA 8
    • installing / Installing NVIDIA drivers and CUDA 8
    • URL / Installing NVIDIA drivers and CUDA 8
  • curse of dimensionality
    • about / The curse of dimensionality
    • avoiding / Avoiding the curse of dimensionality

D

  • DataFrame / Importing data with pandas
  • data science
    • example / Understanding data science by an example
    • data size / Data size and industry needs
    • industry needs / Data size and industry needs
  • data science algorithms
    • design procedure / Design procedure of data science algorithms
  • data types / TensorFlow data types, variables, and placeholders
  • Deconvolutional Network (DNN)
    • URL / Face generation
    • about / Face generation
  • Deep Convolution Neural Network (DCNN) / Semi-supervised learning with Generative Adversarial Networks (GANs)
  • deep learning framework
    • need for / The TensorFlow environment
  • denoising autoencoders
    • about / Denoising autoencoders
    • architecture / Denoising autoencoders
    • model, building / Building the model
    • model, training / Model training
  • derived features
    • about / Derived features
    • name variable / Name
    • Cabin feature / Cabin
    • Ticket feature / Ticket
  • design procedure, data science algorithms
    • data, pre-processing / Data pre-processing, Data pre-processing
    • data cleaning / Data cleaning
    • feature selection / Feature selection
    • model selection / Model selection
    • learning process / Learning process
    • model evaluation / Evaluating your model
  • digit classification model
    • building / Digit classification – model building and training, Building the model
    • training / Digit classification – model building and training, Model training
    • data analysis / Data analysis
  • downsampling step / The pooling step
  • dropout / Building an LSTM cell

E

  • embedding / Introduction to representation learning, Building the model
  • embedding lookup / Skip-gram Word2Vec implementation
  • errors
    • about / Different types of errors
    • apparent error / Apparent (training set) error
    • generalization error / Generalization/true error
  • exploding, of gradients / Exploding and vanishing gradients - recap

F

  • face generation
    • about / Face generation
    • data, obtaining / Getting the data
    • data, exploring / Exploring the Data
    • model, building / Building the model
    • model inputs, defining / Model inputs
    • discriminator, implementing / Discriminator
    • generator, implementing / Generator
    • model losses, calculating / Model losses
    • model optimizer, implementing / Model optimizer
    • model, training / Training the model
  • feature detector / The convolution operation
  • feature engineering
    • about / Feature engineering
    • types / Types of feature engineering
    • Titanic example / Titanic example revisited, Titanic example revisited – all together
    • missing values / Missing values
    • feature transformations / Feature transformations
    • derived features / Derived features
    • interaction features / Interaction features
  • feature map / The convolution operation
  • feature transformations
    • quantitative features / Feature transformations
    • qualitative variables / Feature transformations
    • dummy features / Dummy features
    • factorizing / Factorizing
    • scaling / Scaling
    • binning / Binning
  • feed-forward neural network (FNN)
    • about / Feed-forward neural network, The intuition behind RNNs
    • example / Feed-forward neural network
    • hidden units / Hidden units and architecture design
    • architecture design / Hidden units and architecture design
  • filter / The convolution operation
  • fish recognition/detection model implementation
    • overview / Implementing the fish recognition/detection model
    • knowledge base/dataset / Knowledge base/dataset
    • data analysis pre-processing / Data analysis pre-processing
    • model, building / Model building
    • model training / Model training and testing
    • testing / Model training and testing
    • fish recognition / Fish recognition – all together
    • code / Code for fish recognition
  • Flickr
    • URL / Semi-supervised learning
    • about / Semi-supervised learning
  • forget gate layer / Why does LSTM work?
  • fully connected layer, CNN
    • about / Fully connected layer
    • logits layer / Logits layer

G

  • GAN network
    • building / Building the GAN network
    • model hyperparameters, defining / Model hyperparameters
    • generator, defining / Defining the generator and discriminator
    • discriminator, defining / Defining the generator and discriminator
    • generator loss, defining / Discriminator and generator losses
    • discriminator loss, defining / Discriminator and generator losses
    • optimizers, defining / Optimizers
    • model, training / Model training
    • model performance, testing / Generator samples from training
    • new images, generating with checkpoints / Sampling from the generator
  • Gate / RNNs – sentiment analysis context
  • Gated Recurrent Unit (GRU) / Building the model
  • generalization error / Generalization/true error
  • Generative Adversarial Networks (GANs)
    • about / An intuitive introduction
    • Generator / An intuitive introduction
    • Discriminator / An intuitive introduction
    • implementation / Simple implementation of GANs
    • MNIST dataset, using / Simple implementation of GANs
    • model inputs, defining / Model inputs
    • variable scope, of TensorFlow / Variable scope
    • leaky ReLU / Leaky ReLU
    • generator, defining / Generator
    • discriminator, building / Discriminator
  • global average pooling (GAP) / Discriminator
  • gradient / Classification and logistic regression
  • gradient clipping / Optimizer
  • gradient descent / The intuition behind RNNs

H

  • Hadamard product / Defining multidimensional arrays using TensorFlow

I

  • input gate layer / Why does LSTM work?
  • input layer, CNN / Input layer

K

  • k-fold cross-validation / Model building
  • Keras
    • used, for sentiment analysis / Keras
  • kernel / The convolution operation

L

  • label smooth / Discriminator and generator losses
  • language model
    • implementation / Implementation of the language model
    • mini-batch generation, for training / Mini-batch generation for training
    • building / Building the model
    • architecture / Model architecture
    • inputs, defining / Inputs
    • LSTM cell, building / Building an LSTM cell
    • RNN output layer, creating / RNN output
    • training loss, calculating / Training loss
    • optimizer / Optimizer
    • network, building / Building the network
    • hyperparameters / Model hyperparameters
    • training / Training the model
    • checkpoints, saving / Saving checkpoints
    • text, generating / Generating text
  • latent space / Model inputs
  • layers, convolutional neural network (CNN)
    • about / Different layers of CNNs
    • input layer / Input layer
    • convolution step / Convolution step
    • non-linearity / Introducing non-linearity
    • pooling step / The pooling step
    • fully connected layer / Fully connected layer
  • lazy evaluation / Getting output from TensorFlow
  • leaky ReLU / Leaky ReLU
  • learning visibility / Learning visibility
  • linear discriminant / Linear models for classification
  • linear models, for classification / Linear models for classification
  • linear regression, with TensorFlow / Linear regression with TensorFlow
  • linear regression model
    • about / Linear models for regression, Linear regression model – building and training
    • key factors / Motivation
    • advertising / Advertising – a financial example
    • building / Linear regression model – building and training
    • training / Linear regression model – building and training
  • logistic regression
    • about / Linear models for classification
    • classification / Classification and logistic regression
    • model, building / Logistic regression model – building and training
    • model, training / Logistic regression model – building and training
  • logistic regression model, in TensorFlow
    • utilizing / Utilizing logistic regression in TensorFlow
    • placeholders, using / Why use placeholders?
    • model weights and bias, setting / Set model weights and bias
    • running / Logistic regression model
    • training / Training
    • cost function, defining / Cost function
  • logits layer / Logits layer
  • long-term dependencies
    • problem / The problem of long-term dependencies
  • Long Short Term Networks (LSTM)
    • about / The problem of long-term dependencies
    • reference / LSTM networks
    • architecture / LSTM networks
    • information, processing / Why does LSTM work?

M

  • machine learning
    • challenges / Getting to learn
    • types / Different learning types
    • error types / Different types of errors
  • macOS X
    • TensorFlow CPU installation / TensorFlow CPU installation for macOS X
  • mathematical operations / Mathematical operations
  • Matplotlib
    • URL / Data analysis and visualization
  • maximum likelihood method
    • about / Building Word2Vec model
    • reference / Building Word2Vec model
  • max pool / The pooling step
  • missing values, feature engineering
    • sample with missing values, removing / Removing any sample with missing values in it
    • inputting / Missing value inputting
    • average value, assigning / Assigning an average value
    • predicting, with regression model / Using a regression or another simple model to predict the values of missing variables
  • MNIST dataset analysis
    • about / MNIST dataset analysis
    • MNIST data / The MNIST data
  • MNIST digit classification example
    • implementing / CNN basic example – MNIST digit classification
    • model, building / Building the model
    • cost function, defining / Cost function
    • performance measures, defining / Performance measures
    • model, training / Model training
  • model coefficients / Simple regression model
  • multi-layer perceptron (MLP)
    • about / Capacity of a single neuron, The need for multilayer networks
    • example / The need for multilayer networks
    • training / Training our MLP – the backpropagation algorithm
    • backpropagation algorithm / Training our MLP – the backpropagation algorithm
  • multilayer networks
    • need for / The need for multilayer networks
    • Input layer / The need for multilayer networks
    • Hidden layer / The need for multilayer networks
    • Output layer / The need for multilayer networks
  • multinomial logistic regression / Building the model

N

  • name variable / Name
  • Natural Language Processing (NLP) / Introduction to representation learning
  • negative sampling
    • reference / Building Word2Vec model
    • about / Building Word2Vec model
  • neural network
    • about / Capacity of a single neuron
    • biological motivation and connections / Biological motivation and connections
  • neuron / Biological motivation and connections
  • node / Biological motivation and connections
  • noise-contrastive estimation (NCE)
    • about / Building Word2Vec model
    • reference / Building Word2Vec model
  • non-linearity, CNN / Introducing non-linearity
  • NVIDIA CUDA Deep Neural Network library (cuDNN) / Installing NVIDIA drivers and CUDA 8
  • NVIDIA drivers
    • installing / Installing NVIDIA drivers and CUDA 8

O

  • object detection / Object detection
  • object detection example
    • about / CIFAR-10 – modeling, building, and training
    • packages, importing / Used packages
    • CIFAR-10 dataset, loading / Loading the CIFAR-10 dataset
    • data analysis / Data analysis and preprocessing
    • preprocessing / Data analysis and preprocessing
    • network, building / Building the network
    • model, training / Model training
    • model, testing / Testing the model
  • one-hot encoding / Skip-gram Word2Vec implementation
  • overfitting / Understanding data science by an example, Bias-variance decomposition

P

  • pandas
    • about / Importing data with pandas
    • reference link / Importing data with pandas
  • part-of-speech (POS) tagging / A practical example of the skip-gram architecture
  • placeholders / TensorFlow data types, variables, and placeholders, Placeholders
  • pooling step, CNN / The pooling step
  • Principal Component Analysis (PCA) / Dimensionality reduction, Analysis of transfer values

R

  • Rectified linear unit (ReLU) / ReLU
  • recurrent neural network (RNN)
    • intuition / The intuition behind RNNs
    • architecture / Recurrent neural networks architectures
    • about / General sentiment analysis architecture
    • for sentiment analysis / RNNs – sentiment analysis context
    • gradients, exploding / Exploding and vanishing gradients - recap
    • gradients, vanishing / Exploding and vanishing gradients - recap
  • recurrent neural network (RNN), examples
    • character-level language models / Examples of RNNs
    • vanishing gradient problem / The vanishing gradient problem
    • long-term dependencies problem / The problem of long-term dependencies
  • reinforcement learning / Reinforcement learning
  • representation learning / Introduction to representation learning
  • rule of thumb
    • breaking / Breaking the rule of thumb

S

  • semi-supervised learning / Semi-supervised learning, Semi-supervised learning with Generative Adversarial Networks (GANs)
  • semi-supervised learning, with GANs
    • implementing / Semi-supervised learning with Generative Adversarial Networks (GANs)
    • building / Intuition
    • data analysis / Data analysis and preprocessing
    • data, preprocessing / Data analysis and preprocessing
    • model, building / Building the model
    • model inputs, defining / Model inputs
    • generator, implementing / Generator
    • discriminator, implementing / Discriminator
    • model losses, defining / Model losses
    • model optimizer, defining / Model optimizer
    • model, training / Model training
  • sentiment analysis
    • architecture / General sentiment analysis architecture
    • with recurrent neural network (RNN) / RNNs – sentiment analysis context
    • model, implementation / Sentiment analysis – model implementation
    • Keras, using / Keras
    • data analysis / Data analysis and preprocessing
    • data, preprocessing / Data analysis and preprocessing
    • model, building / Building the model
    • model, training / Model training and results analysis
    • results, analyzing / Model training and results analysis
  • session / Getting output from TensorFlow
  • sigmoid activation function / Sigmoid
  • sigmoid function / Classification and logistic regression
  • simple linear regression model
    • about / Simple regression model
    • model coefficients / Learning model coefficients
    • model coefficients, interpreting / Interpreting model coefficients
    • using, for prediction / Using the model for prediction
  • skip-gram architecture
    • practical example / A practical example of the skip-gram architecture
  • skip-gram Word2Vec
    • implementation / Skip-gram Word2Vec implementation
    • data analysis / Data analysis and pre-processing
    • pre-processing / Data analysis and pre-processing
    • model, building / Building the model
    • training / Training
  • stacked LSTM / Building the model, Stacked LSTMs
  • Stochastic Gradient Descent (SGD) / Model building and training
  • stride / The convolution operation
  • subcategories, feature engineering
    • feature selection / Feature selection
    • dimensionality reduction / Dimensionality reduction
    • feature construction / Feature construction
  • subsampling / The pooling step, Data analysis and pre-processing
  • supervised learning
    • about / Supervised learning, Semi-supervised learning
    • classification / Supervised learning
    • regression / Supervised learning
  • SVHN dataset
    • URL / Data analysis and preprocessing

T

  • Tanh activation function / Tanh
  • TensorBoard
    • about / TensorBoard – visualizing learning
    • using / TensorBoard – visualizing learning
  • TensorFlow
    • about / Data size and industry needs, The TensorFlow environment
    • installation / TensorFlow installation, Installing TensorFlow
    • GPU installation, for Ubuntu 16.04 / TensorFlow GPU installation for Ubuntu 16.04
    • NVIDIA drivers, installing for GPU mode installation / Installing NVIDIA drivers and CUDA 8
    • CUDA 8, installing for GPU mode installation / Installing NVIDIA drivers and CUDA 8
    • CPU installation, for Ubuntu 16.04 / TensorFlow CPU installation for Ubuntu 16.04
    • CPU installation, for macOS X / TensorFlow CPU installation for macOS X
    • GPU/CPU installation, for Windows / TensorFlow GPU/CPU installation for Windows
    • output, obtaining / Getting output from TensorFlow
    • used, for defining multidimensional arrays / Defining multidimensional arrays using TensorFlow
    • variable scope / Variable scope
  • TensorFlow terminologies
    • about / TensorFlow terminologies – recap
    • tensors / Why tensors?
    • variables / Variables
    • placeholders / Placeholders
    • operations / Operations
  • testing phase / Understanding data science by an example
  • tf.contrib.rnn.MultiRNNCell function
    • URL / Building an LSTM cell
  • tf.nn.dynamic_rnn
    • URL / Building the network
  • Ticket feature / Ticket
  • Titanic example
    • about / Titanic example – model building and training
    • data handling / Data handling and visualization
    • data visualization / Data handling and visualization
    • data analysis, supervised machine learning / Data analysis – supervised machine learning
  • tokenizer / Data analysis and preprocessing
  • tokens / General sentiment analysis architecture
  • transfer learning (TL)
    • about / What this book covers, Transfer learning
    • intuition / The intuition behind TL
    • usage scenario / The intuition behind TL
    • versus traditional machine learning / Differences between traditional machine learning and TL
    • conditions / Differences between traditional machine learning and TL
  • types, machine learning
    • about / Different learning types
    • supervised learning / Supervised learning
    • unsupervised learning / Unsupervised learning
    • semi-supervised learning / Semi-supervised learning
    • reinforcement learning / Reinforcement learning

U

  • Ubuntu 16.04
    • TensorFlow GPU installation / TensorFlow GPU installation for Ubuntu 16.04
    • TensorFlow CPU installation / TensorFlow CPU installation for Ubuntu 16.04
  • underfitting / Bias-variance decomposition
  • unit / Biological motivation and connections
  • unseen data / Understanding data science by an example
  • unsupervised learning
    • about / Unsupervised learning, Semi-supervised learning
    • clustering / Unsupervised learning
    • association / Unsupervised learning

V

  • vanishing, of gradients / Exploding and vanishing gradients - recap
  • vanishing gradient problem / The vanishing gradient problem
  • variables / TensorFlow data types, variables, and placeholders, Variables
  • variable scope, TensorFlow
    • about / Variable scope
    • reference / Variable scope

W

  • weight matrix / Skip-gram Word2Vec implementation
  • Windows
    • TensorFlow GPU/CPU installation / TensorFlow GPU/CPU installation for Windows
  • Word2Vec
    • about / Word2Vec
    • model, building / Building Word2Vec model

Y

  • yield
    • URL / Mini-batch generation for training
    • about / Mini-batch generation for training

Z

  • zero-padding / The convolution operation
lock icon The rest of the chapter is locked
arrow left Previous Section
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at £13.99/month. Cancel anytime
Visually different images