I'm looking for two intuitive examples of random variables X, Y and Z. One where
$ I(X;Y|Z) > I(X;Y) $
and another set of X,Y and Z where
$ I(X;Y|Z) < I(X;Y)$
According to wikipedia conditioning can both reduce and increase mutual information, but I haven't found any simple, clear and intuitive examples of this yet.
$I(X;Y|Z)$ is interpreted as `` the reduction in uncertainty of $X$ due to the knowledge of $Y$ when $Z$ is given''.
The Data Processing inequality tells you if $X \to Y \to Z$ (that is, $X,Y,Z$ form a Markov Chain), then $I(X;Y) \geq I(X;Z)$. Taking $Z=g(Y)$, you get $I(X;Y) \geq I(X;g(Y))$ - that is, functions of $Y$ cannot increase mutual information. Another corollary is, $I(X;Y|Z) \leq I(X;Y)$ where $X \to Y \to Z$.
In the case where $X,Y,Z$ do not follow a Markov chain, you can have $I(X;Y|Z) > I(X;Y)$. Easy example is $X,Y$ are independent Bernoulli(1/2) random variables and $Z=X+Y$. $I(X;Y) = 0$, but $I(X;Y|Z) = 1/2$.
(This is taken from Cover & Thomas, Elements of Information Theory 2e, Chapter 2. There is a nice problem in Chapter 2 on what goes wrong with extending mutual information to multiple variables, but it isn't the same idea.)