diff --git a/lab3/lab-notes.md b/lab3/lab-notes.md
index 3e7fc45aa642e8fd301118df6ab441b82c7e54a3..299a716e0f22b2f766c05bd4894efed2166824b7 100644
--- a/lab3/lab-notes.md
+++ b/lab3/lab-notes.md
@@ -1,3 +1,22 @@
+Assignment 1
+
+- What is the kernel trick?
+Since we can rewrite the šæ^2 regularised linear regression formula to a form where non-linear transformations š“(x) only appear via inner product, we do not have to design a š‘‘-dimensional vector š“(x) and derive its inner product. Instead, we can just choose a kernel
+šœ…(x, x') directly where the kernel is the inner product of two non-linear input transformations according to:
+šœ…(x, x') = š“(x)^Tš“(x').
+This is known as the kernel trick:
+If x enters the model as š“(x)^Tš“(x') only, we can choose a kernel šœ…(x, x') instead of chosing š“(x). p. 194
+
+- In the literature, it is common to see a formulation of SVMs that makes use of a hyperparameter. What is the purpose of this hyperparameter?
+The purpose is to regularize. p. 211
+
+
+- In neural networks, what do we mean by mini-batch and epoch?
+We call a small subsample of data a mini-batch, which typically can contain š‘›š‘ = 10, š‘›š‘ = 100, or š‘›š‘ = 1 000
+data points. One complete pass through the training data is called an epoch, and consequently consists of š‘›/š‘›š‘ iterations. p. 125
+
+
+
 Assignment 4
 
 4.1