I am trying to understand the geometry for linear classification models. A linear model, according to Bishop's books, is defined as: $\mathbf{y} = \mathbf{w}^T \mathbf{x} + w_0$.
For instace we have have the following example with a dataset with two features (feature A and feature B) and two classes (class A and class B):
The decision boundary here is the: $\mathbf{y} = \mathbf{w}^T \mathbf{x} + w_0$. The input $\mathbf{x} \in \mathbb{R}^{2}$. Here the decision boundary is a line that can seperate these two classes, however, I am bit puzzled on what $\mathbf{w}$ is. Shouldnt be also $\mathbf{w} \in \mathbb{R}^{2}$ and if yes what is it exactly if not just $w = -1$?
How is this $\mathbf{w}$ perpendidular to the decision boundary?
