Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

05-clarity #23

Open
ltetrel opened this issue Oct 15, 2019 · 2 comments
Open

05-clarity #23

ltetrel opened this issue Oct 15, 2019 · 2 comments
Assignees

Comments

@ltetrel
Copy link
Collaborator

ltetrel commented Oct 15, 2019

@manojneuro Thanks for your submission.
This issue will review notebook 05.

2. Circular Inference: How to avoid double dipping

The GridSearchCV method that you will learn about below makes it easy (though not guaranteed) to avoid double dipping. In previous exercises we examined cases where double dipping is clear (e.g., training on all of the data and testing on a subset). However, double dipping can be a lot more subtle and hard to detect, for example in situations where you perform feature selection on the entire dataset before classification (as in last week's notebook).

You introduce the term "double dipping" without explaining it. I think you could add few words to explain it ?

3.2 Regularization Example: L2 vs. L1

I think the figure could gain more clarity if it was 3x3, and by labeling the axis ("-th fold" for example)

@ltetrel ltetrel self-assigned this Oct 15, 2019
@manojneuro
Copy link

@ltetrel: We did cover double dipping in the previous notebook, but we can provide references etc. on the topic here.

For the L2 vs. L1, we can enhance the figure and the description to make it clearer.

@ltetrel
Copy link
Collaborator Author

ltetrel commented Oct 28, 2019

Thanks @manojneuro

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants