1.5 Step 5 – Defining Our Model CS 1.6 Step 6 – Running Our Model CS 1.7 Step 7 – Automatically Finding an Optimised Architecture Using Bayesian Optimisation
15 Case Study 2: Time Series Forecasting with LSTMs CS 2.1 Simple LSTM CS 2.2 Sequence‐to‐Sequence LSTM
16 Case Study 3: Deep Embeddings for Auto‐Encoder‐Based Featurisation
17 Index
List of Tables
1 Chapter 3Table 3.1 A rule of thumb guide for understanding AUC‐ROC scores.
List of Illustrations
1 Chapter 3Figure 3.1 Examples of ROC curves.Figure 3.2 Optimal strategy without knowing the distribution.Figure 3.3 Optimal strategy when you know 50% of galaxies are elliptical and...Figure 3.4 A graphical look at the bias–variance trade‐off.Figure 3.5 A flow chart for dealing with high bias or high‐variance situatio...Figure 3.6 Graphical representation of the holdout‐validation algorithm.Figure 3.7 The effects of different scales on a simple loss function topolog...
2 Chapter 4Figure 4.1 An overview of a single perceptron learning.Figure 4.2 The logistic function.Figure 4.3 Derivatives of the logistic function.Figure 4.4 How learning rate can affect the training, and therefore performa...Figure 4.5 A schematic of a multilayer perceptron.Figure 4.6 Plot of ReLU activation function.Figure 4.7 Plot of leaky ReLU activation function.Figure 4.8 Plot of ELU activation function.Figure 4.9 Bias allows you to shift the activation function along the X‐axis...Figure 4.10 Training vs. validation error.Figure 4.11 Validation error from training model on the Glass dataset.
3 Chapter 5Figure 5.1 A schematic of a RNN cell. X and Y are inputs and outputs, respec...Figure 5.2 Connections in a feedforward layer in an MLP (a) destroy the sequ...Figure 5.3 An example of how sequential information is stored in a recurrent...Figure 5.4 A schematic of information flow through an LSTM cell. As througho...Figure 5.5 An LSTM cell with the flow through the forget gate highlighted.Figure 5.6 An LSTM cell with the flow through the input gate highlighted.Figure 5.7 An LSTM cell with the flow through the output gate highlighted.Figure 5.8 An LSTM cell with peephole connections highlighted.Figure 5.9 A schematic of information flow through a GRU cell. Here, X refer...
4 Chapter 6Figure 6.1 Illustration of convolutional neural network architecture.Figure 6.2 Illustration of average and max pooling algorithms.Figure 6.3 Illustration of average and max pooling on face image.Figure 6.4 Illustration of average and max pooling on handwritten character ...Figure 6.5 Illustration of the effect of stride on change in data volume.Figure 6.6 Illustration of stride.Figure 6.7 Illustration of the impact of sparse connectivity on CNN unit's r...Figure 6.8 Illustration of graph convolutional network.Figure 6.9 Example graph.Figure 6.10 Example adjacency matrix.
5 Chapter 7Figure 7.1 A schematic of a shallow auto‐encoder.Figure 7.2 Representing a neural network as a stack of RBMs for pretraining....Figure 7.3 Training an auto‐encoder from stacked RBMs. (1) Train a stack of ...Figure 7.4 Comparison of standard auto‐encoder and variational auto‐encoder....Figure 7.5 Illustration of sequence to sequence model.
6 Chapter 8Figure 8.1 Schematic for greedy search.Figure 8.2 Bayes rule.
Guide
8 Index
9 Wiley End User License Agreement
Pages
1 iii
2 iv
3 xi
4 xii
5 1
6 2
7 3
8 5
9 6
10 7
11 8
12 9
13 10
14 11
15 12
16 13
17 14
18 15