I've been trying to prove that estimates of a classifier's performance using training data is a bad thing. Does "bad" mean it is biased? This is part of a larger proof.
If somebody knows of previous work that proves this or a quick proof, any pointers would be much appreciated!
Thanks in advance, Yakka
This issue goes beyond any particular model, like a classifier, to statistical models in general. When you fit a model to training data, you are optimizing its fit or performance relative to that training set. Now, if you took that model and applied to different data, then unless the new data looks exactly like your training set, the performance will uaually be worse.
You can see this even with just the trainig data by running your classifier on a boostrap sample of your training set. You'll see that your traning set performance was biased high.