Reading through materials of machine learning problems, I see people treating things like they are doing with probability.
Particularly, consider linear regression. I cannot figure out what is the sample space, outcomes, events, random variables. In what manner they are using the word "probability measure" in the field of machine learning?
For instance, please take a look at this article.
Moreover, in general, why do we need probability for machine learning? I don't see how we are going to calculate any probability of something in machine learning. While the need of linear algebra is obvious, since we are working on lists matrix of numbers. What properties of probability make it essential for machine learning?
Well, linear regression is used to model the relationship between input X and output/outcome Y.
Outcome: The y-value for each sample.
Sample space: all possible outcomes, which could be, like in the figure on the wikipedia article all real numbers or a subset of it. (Compare to a binary classification problem: Sample space {-1, 1})
Event: Everything you want to know a probability for. E.g. How likely is it to observe an outcome between 1 and 3?
random variables: X values.
I'm not sure if I understand "Why do we need probability for machine learning?". Maybe I haven't thought about it enough yet. For me probability is the point of machine learning.
You need probability to find out how "good" ore likely your model (classifier, regressor, ...) is. You need concepts like a probability distribution to make sense of data, to find out which features are actually useful (e.g. PCA) (not all problems have one input variable x and one output variable y, that would be way to nice!)
I used this document and the wikipedia article for explanations.