Discussion about this post

User's avatar
James McDermott's avatar

This is a very good post, and the diagram at the top should be early in all ML textbooks.

> These are mathematical procedures that compare a data point with a learned pattern.

NB *pattern* here is a vague term. Eg in k-NN, we compare against some particular data points in the training set, the nearest ones. Eg in Naive Bayes, we instead compare against a pseudo-data point which is calculated as the centroid of the class. But "pattern" can mean other things in other contexts.

> The linear regression model is a dot product.

True but a bit misleading in context, as we are not really using this dot product to calculate the *similarity* between the weights and the query point.

Expand full comment
Ted Lorenzen's avatar

I'm not sure I see the distinction between dot product and distance? Isn't the dot product just the unscaled angular separation of the two vectors? I think if we do scale it, the angle between the two vectors will be a distance measure (symmetric, has a zero, obeys the triangle inequality)?

Expand full comment
2 more comments...

No posts