In an informatic theoretic sense, complete randomness maximizes information. For instance, an image of randomly distributed black and white pixels has a very high entropy/information. For a human brain, such an image does not contain any valuable information. However, if we remove some randomness and align some pixels to represent a specific shape, the human brain can recognize this shape and process/gather information. Thus, for the human brain, maximum information in the sense of Shannon entropy is not valueable.
My question is: What is the relationship between Shannon entropy and information, understandable by the human brain? Why can we not make sense of a channel with maximum information? Why do we need less than maximum information to gather valuable information?
Thanks!
For convenience, I'll call your concept of an "image of randomly distributed black and white pixels" a "white noise picture".
While there is as yet no satisfactory mathematical characterization of the human brain, we can be confident that the brain's functions arose from evolutionary pressures. It is also widely accepted in academic research psychology that the human brain is a computational device. From this point of view, then, there is no difference between Shannon information and the information processed by the brain.
From this point of view, there is no evolutionary value in recognizing white noise pictures or distinguishing one such picture from another, so the brain doesn't have this functionality. This addresses your second question and much of your third question.
There is plainly evolutionary value in recognizing what we all would call familiar shapes. Since these require less information to describe that a white-noise picture, we get closer to your third question.