Let $(X,\mathcal A)$ and $(Y,\mathcal B)$ be measurable space. A ''Markov kernel'' with source $(X,\mathcal A)$ and target $(Y,\mathcal B)$ is a map $\kappa : \mathcal B \times X \to [0,1]$ with the following properties:
(1) For every (fixed) $B \in \mathcal B$, the map $x \mapsto \kappa(B, x)$ is $\mathcal A$-measurable.
(2) For every (fixed) $x \in X$, the map $B \mapsto \kappa(B, x)$ is a probability measure on $(Y, \mathcal B)$.
Can someone please explain what this definition is saying? I am not getting the point, please if someone explain with example that will be great help. Thanks.
Markov kernels are just a way of expressing conditional distributions. The idea is that for each $x \in {X},$ you want to say that conditional law of some random variable $\bf Y$ given an observation of $\mathbf X = x$ is $\kappa(\cdot,x)$ --- this is sometimes denoted $\kappa(\cdot|x)$ instead. This is precisely the point of condition (2). $\kappa(B,x)$ is meant to represent $P(\mathbf Y \in B | \mathbf X = x)$.
In order to work with objects like $\kappa(B,x)$ as $x$ varies, the technical structure of probability theory requires that these conditional probabilities are not wild, in the sense that at least measure-theoretic issues don't arise. We would need this, for instance, to ensure that sets like $E = \{x: \kappa(B,x) \ge \tau\}$ are actually events. Note that such events are something we would often like to be able to deal with --- for example, if you observe $\mathbf Y \in B$, and you want to decide which $x$ are plausible for $\mathbf X$ given this observation, one reasonable answer is precisely $E$ for some value of $\tau$. The condition (1) says (more or less) that no matter what $B \in \mathcal{B}, \tau$ we choose, the sets $E$ belong to $\mathcal{A},$ which is precisely what is needed for things like $P(\mathbf X \in E)$ to make sense.
To sum up, Markov kernels are a formal way to set up conditional distributions. (2) is precisely the part of the definition that captures this aspect, while (1) is needed for technical reasons to ensure that the conditional distributions we treat are nice to work with.