Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Chapter 2, Page 23 #352

Open
FZhg opened this issue Oct 21, 2021 · 1 comment
Open

Chapter 2, Page 23 #352

FZhg opened this issue Oct 21, 2021 · 1 comment

Comments

@FZhg
Copy link

FZhg commented Oct 21, 2021

“Life is more complicated for the “full” decision tree. Certainly
if it is given a test example that is identical to one of the training
examples, it will do the right thing (assuming no noise). But for
everything else, it will only get about 50% error. This means that
even if every other test point happens to be identical to one of the
training points, it would only get about 25% error. In practice, this is
probably optimistic, and maybe only one in every 10 examples would
match a training example, yielding a 35% error”

The 35% error rate should be 45% error.
Since we have 10 test example, one of which must be right since it matches the training example.
The rest of the 9 examples are just random guess. Then the expected examples that we have got right are 4.5 + 1 = 5.5.
Thus the expected error rate should be 45%. Unless you take into account the condition that the number of the examples have to be intergers instead of decimals.

@meeraray
Copy link

meeraray commented Jan 7, 2022

I had this exact issue too! Unless there's something I'm missing, I think it's a typo.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants