[IML] class.

This commit is contained in:
RobinB27
2026-03-22 15:54:22 +01:00
parent 110c1b7207
commit 80739c654a
5 changed files with 95 additions and 2 deletions
Binary file not shown.
+5
View File
@@ -8,9 +8,14 @@
\date{HS 2026} \date{HS 2026}
\begin{document} \begin{document}
\textbf{Introduction to Machine Learning}
\input{parts/00_intro.tex} \input{parts/00_intro.tex}
\section{Regression} \section{Regression}
\input{parts/01_regression.tex} \input{parts/01_regression.tex}
\newpage
\section{Classification}
\input{parts/02_classification.tex}
\end{document} \end{document}
+1 -1
View File
@@ -1 +1 @@
\textbf{placeholder} \textit{placeholder}
+1 -1
View File
@@ -95,7 +95,7 @@ The solution is a stationary point, so:
$$ $$
\nabla_w \bigl\Vert y-Xw \bigr\Vert^2 = 2X^\top(X\hat{w}-y) = 0 \nabla_w \bigl\Vert y-Xw \bigr\Vert^2 = 2X^\top(X\hat{w}-y) = 0
$$ $$
Which yields the known \textbf{Normal Equation} Which yields the \textbf{Normal Equation} from linear algebra.
$$ $$
X^\top X\hat{w} = X^\top y X^\top X\hat{w} = X^\top y
$$ $$
+88
View File
@@ -0,0 +1,88 @@
In regression, we search an $\hat{f}: \R^d \to \R$, i.e. $y,\hat{y} \in \R$.\\
In classification, we want $\hat{y} \in \mathcal{Y} \subset \R$, s.t. $\mathcal{Y}$ is discrete.
\subsection{Binary Classification}
We generally use $\mathcal{Y} = \{+1,-1\}$ and set $\hat{y} = \text{sgn}\bigl(\hat{f}(x)\bigr)$.\\
So, a linear classifier where $\hat{f}(x) = w^\top x$ takes the form:
$$
x \mapsto \begin{cases}
1 & w^\top x > 0 \\
-1 & w^\top x < 0
\end{cases}
$$
\definition \textbf{Decision Boundary} $\quad \Bigl\{ x \in \R^d \ \Big|\ \hat{f}(x) = 0 \Bigr\}$
Like in regression, using features is again possible.
\subsection{Surrogate Loss}
We'd like to reuse the loss minimization from regression.\\
A natural metric for accuracy is simply checking if $\hat{y} = y$.
\definition \textbf{Zero-One Loss}
$$
l_{0-1}(\hat{y},y) := \mathbb{I}_{\hat{y}\neq y} = \begin{cases}
1 & \hat{y} \neq y \\
0 & \hat{y} = y
\end{cases}
$$
We could try minimizing this:
$$
\underset{(x,y) \in \mathcal{D}}{\sum} l_{0-1}\bigl( \hat{y},y \bigr) = \underset{(x,y) \in \mathcal{D}}{\sum} \mathbb{I}_{f_w(x) \cdot y < 0}
$$
Unfortunately, $l_{0-1}$ is non-continuous and non-convex.\\
We introduce \textit{surrogate loss} to still apply GD.
Note how $\mathbb{I}_{\hat{y}\neq y} = \mathbb{I}_{\hat{y}\cdot y < 0}$, so $l_{0-1}$ only depends on $z := \hat{y}\cdot y$.\\
We thus define losses over $z$, that are cont. and convex.
\definition \textbf{Surrogate Loss}
$$
l_\text{exp} = e^{-z} \qquad l_\text{log} = \log(1+e^{-z})
$$
A notable difference is that $l_\text{exp}'$ is unbounded,\\
while $l_\text{log}' = \frac{1}{1+e^z} \in (-\frac{1}{2}, -1)$ for $z < 0$.\\
This is better for outliers, thus $l_\text{log}$ is usually preferred.
\newpage
\subsection{Logistic Regression}
\subtext{We assume $w_0 = 0$}
We try to minimize $l_\text{log} = \log(1+e^{-z})$, so:
$$
L(w) = \frac{1}{n}\sum_{i=1}^{n} l_\text{log}(z_i) = \frac{1}{n}\sum_{i=1}^{n}\log\Bigl( 1 + e^{-\overbrace{y_i \cdot w^\top x_i}^{z_i}} \Bigr)
$$
Assume $\{x_i,y_i\}_{i=1}^n$ is linearly seperable, i.e.
$$
\exists w \in \R^d:\quad \underbrace{y_i \cdot w^\top x_i}_{z_i} > 0 \quad \forall i \leq n
$$
Then there are multiple valid decision boundaries.
the distance $x_0$ to the decision boundary is: $\Vert x_0 \Vert_2 \cdot |\cos(\theta)|$.\\
\subtext{$\theta$ between $w,x_0 \in \R^d$}
$$
\Vert x_0 \Vert_2 \cdot |\cos(\theta)| = \Vert x_0 \Vert_2 \cdot \frac{|w^\top x_0 |}{\Vert w \Vert_2 \cdot \Vert x_0 \Vert_2} = \frac{|w^\top x_0|}{\Vert w \Vert_2}
$$
\subtext{Note if $w$ is a unit-vector, this is just $|w^\top x_0|$}
\definition \textbf{Margin} $\quad \text{margin}(w) := \underset{1\leq i\leq n}{\min} y_i \cdot w^\top x_i$
\subsection{Solutions}
\definition \textbf{Maximum Margin Solution}
$$
w_\text{MM} := \underset{\Vert w \Vert_2=1}{\max} \underset{1\leq i\leq n}{\min} \Bigl( y_i \cdot w^\top x_i \Bigr)
$$
If $\mathcal{D}$ is linearly seperable, this is convex.
\definition \textbf{Support Vector Machine}
$$
w_\text{SVM} := \underset{w \in \R^d}{\min} \Vert w \Vert_2 \quad\text{s.t.}\quad y_i \cdot w^\top x_i \geq 1 \quad \forall i \leq n
$$
Solving these problems is actually equivalent, up to scaling:
\lemma $\quad\displaystyle\frac{w_\text{SVM}}{\Vert w_\text{SVM} \Vert_2} = w_\text{MM}$
\subtext{(This also holds for the case $w_0 \neq 0$)}