Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Initialisation induced score variation #6

Open
WeberJulian opened this issue Jul 24, 2020 · 1 comment
Open

[Question] Initialisation induced score variation #6

WeberJulian opened this issue Jul 24, 2020 · 1 comment

Comments

@WeberJulian
Copy link

Hello,

I'm posting this question here because I don't know anywhere else to ask it but I can take it down if it's not the proper medium to ask it.

Looking at the paper, it seemed to me that the relatively large variation in score due to different initialisation could mean that if we average the scoring given by n different initialisation we could have a much more consistent scoring.

image

However, after implementing a quick version of that idea, the result showed no significative improvement.
Can someone explain to me why I was wrong ?

Thanks in advance

@egracheva
Copy link

This is probably because the selected architecture will be the one with the highest score, and the highest scores do not show high variance anyway. There should be a big difference when you select the worse architecture though. You can check it if you're interested. :)

However an important takeaway from this plot is that relatively good archs will show lower variance. This idea is actually implemented here, with modest but consistent results.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants