diff --git a/lab3/lab-notes.md b/lab3/lab-notes.md index 3e7fc45aa642e8fd301118df6ab441b82c7e54a3..299a716e0f22b2f766c05bd4894efed2166824b7 100644 --- a/lab3/lab-notes.md +++ b/lab3/lab-notes.md @@ -1,3 +1,22 @@ +Assignment 1 + +- What is the kernel trick? +Since we can rewrite the ð¿^2 regularised linear regression formula to a form where non-linear transformations ð“(x) only appear via inner product, we do not have to design a ð‘‘-dimensional vector ð“(x) and derive its inner product. Instead, we can just choose a kernel +ðœ…(x, x') directly where the kernel is the inner product of two non-linear input transformations according to: +ðœ…(x, x') = ð“(x)^Tð“(x'). +This is known as the kernel trick: +If x enters the model as ð“(x)^Tð“(x') only, we can choose a kernel ðœ…(x, x') instead of chosing ð“(x). p. 194 + +- In the literature, it is common to see a formulation of SVMs that makes use of a hyperparameter. What is the purpose of this hyperparameter? +The purpose is to regularize. p. 211 + + +- In neural networks, what do we mean by mini-batch and epoch? +We call a small subsample of data a mini-batch, which typically can contain ð‘›ð‘ = 10, ð‘›ð‘ = 100, or ð‘›ð‘ = 1 000 +data points. One complete pass through the training data is called an epoch, and consequently consists of ð‘›/ð‘›ð‘ iterations. p. 125 + + + Assignment 4 4.1