week 6 Flashcards

(47 cards)

1
Q

What is a deep neural network (DNN)?

A

A composition of layers fθ = fL ∘ fL−1 ∘ … ∘ f1 that maps input x through multiple transformations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the forward pass through layer l?

A

zl+1 = σl(Wl zl + bl).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a shallow neural network?

A

A neural network with one hidden layer: fθ(x)=ηᵀ σ(Wx+b).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are W and b in a neural network layer?

A

W are weights and b are biases for that layer.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are activations?

A

The outputs after applying the activation function: zl = σ(Wl−1 zl−1 + bl−1).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are pre-activations?

A

The linear part before the activation function: al = Wl−1 zl−1 + bl−1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a neuron?

A

A single hidden unit computing σ(wᵀx + b).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is network depth?

A

Number of hidden layers (sometimes +1).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is network width?

A

Number of neurons in a layer.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is network capacity?

A

Total number of neurons across the network.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does a sigmoid activation do?

A

σ(u) = 1 / (1 + e^(−u)).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the ReLU activation?

A

σ(u) = max(0,u).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is tanh activation?

A

σ(u) = tanh(u).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the step activation?

A

σ(u) = 1_{u > 0}.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the Swish activation?

A

σ(u) = u / (1 + e^(−βu)).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What property do neural networks share with RBF kernels?

A

They are universal approximators: can approximate any continuous function on compact sets.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Why use deep instead of shallow networks?

A

Deep networks approximate functions with fewer parameters than shallow networks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is hierarchical feature learning?

A

Hidden layers learn progressively more abstract features (edges → shapes → object parts → concepts).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Why does the last layer behave like a shallow model?

A

It is linear in its parameters: fθ(x)=ηᵀz where z is the final learned feature vector.

20
Q

How can we inspect what features a DNN learns?

A

By analysing which input patterns strongly activate specific neurons.

21
Q

What is data-driven feature learning?

A

Learning features automatically from data rather than manually designing them.

22
Q

Why are GPUs used for DNN training?

A

GPUs perform thousands of operations in parallel, enabling fast training of large models.

23
Q

What is a spurious feature?

A

A feature correlated with the label in training data but not truly relevant to the task.

24
Q

Example of spurious features?

A

Skin lesion classification using presence of a ruler as a signal for malignancy.

25
What is an adversarial example?
Input with imperceptible noise that causes the DNN to misclassify.
26
Why are adversarial examples dangerous?
They can fool autonomous vehicles, security systems, spam filters, etc.
27
Why are adversarial features surprising?
They reveal the model uses patterns humans would never rely on.
28
How does explicit regularisation work in DNNs?
Add penalties (e.g., L2) or use early stopping to avoid overfitting.
29
What is early stopping?
Stop training when validation loss increases; acts as implicit regularisation.
30
What is data augmentation?
Generate modified training samples (rotations, crops, synonym replacement, back-translation).
31
Why is data augmentation useful?
Makes the model robust and reduces overfitting.
32
What is dropout?
Randomly deactivate (e.g., 50%) neurons during training.
33
Why does dropout help?
It prevents reliance on specific neurons and effectively trains many subnetworks.
34
Are neurons dropped at test time?
No, all neurons are active at test time.
35
How do we train DNNs?
Minimise empirical risk using mini-batch SGD or Adam with backpropagation.
36
Why is backpropagation needed?
To compute gradients efficiently through the chain rule in deep composition functions.
37
What is the chain rule for a 1-layer network?
∂l/∂w = (∂l/∂z)(∂z/∂a)(∂a/∂w) with a=wx+b, z=σ(a).
38
Why is backprop efficient?
It reuses intermediate gradients instead of recomputing long chain-rule derivatives.
39
What problem occurs without good initialisation?
Vanishing or exploding gradients.
40
What is He/Kaiming initialisation?
Set W ∼ N(0, 2 / d_l) or N(0, 4/(d_l + d_{l+1})) to stabilise activations and gradients.
41
Why does He initialisation work?
It keeps variances of activations and gradients consistent across layers.
42
What optimiser is typically used?
Adam or mini-batch SGD.
43
What are typical hyperparameters in DNNs?
Number of layers, width, activation, optimiser settings, learning rate, batch size.
44
What are residual layers?
Layers of the form z_{l+1} = z_l + hθ(z_l).
45
Why do residual layers help?
They smooth the loss surface and enable training of very deep networks (e.g., 1000+ layers).
46
What trick helps residual networks avoid exploding gradients?
Batch Normalisation (instead of just He init).
47
What is the purpose of a skip connection?
Allow gradients to flow easily across layers and improve training stability.