Skip to content
This repository has been archived by the owner on May 24, 2018. It is now read-only.

V2-refactor - Higher train errors with eval=val #50

Closed
DataCentric1 opened this issue Mar 14, 2015 · 6 comments
Closed

V2-refactor - Higher train errors with eval=val #50

DataCentric1 opened this issue Mar 14, 2015 · 6 comments

Comments

@DataCentric1
Copy link

I just switched to v2-refactor for the Kaggle bowl. My first goal was to match the CNN model with what I had been using with master and see if I get the same train error scores...just as a way to baseline.

I kept getting much higher train errors in v2-refactor (0.33 vs. 0.19 in master). I was able to root cause this to using "eval = val" in bowl.conf. If I change to "eval = train", then train error matches what I see in master.

Just curious to know why this is the case? I was assuming "eval=val" should only affect the validation error score and not train errors.

@DataCentric1
Copy link
Author

Thinking more about it, it makes sense train error could also be affected if we are modifying the params (LR, bias etc) adaptively based on the val error (which is really great!).

Still, the train error on first iteration is much higher with "eval = val" (0.76) vs. "eval = train" (0.61)?

@antinucleon
Copy link
Contributor

Do you use CUDNN? I find there is some unstable stuff in CUDNN pooling, which makes unpredictable result. Now I disabled CuDNN pooling.

@DataCentric1
Copy link
Author

No, I don't use CUDNN. I'm planning to run the exact same model in both master and v2-refac and see if my submission scores in the competition are any different. Will update with what I find.

@DataCentric1
Copy link
Author

Ok, something's definitely amiss. I ran the exact same bowl.conf and pred.conf files (just minor chnages to be rev compatible) in both master and v2-refactor. My train error / val error in Master was .227 / .257 and in V2-refac was .342 / .296.

When I submitted in Kaggle, leaderboard score for master was 0.90 and V2-refac was 0.98. This was only one single submission with no averaging of multiple outputs. Something definitely seems off with using v2-refac or I'm missing some details?

FYI, this is the older master from ~6 weeks back, I haven't updated at all.

@antinucleon
Copy link
Contributor

Thanks very much! I will check my configuration tomorrow. If possible, could you share me your configuration? just email me: antinucleon àt gmail.com so that I will be more clear of what happened. I used V2 for all competition, and I didn't find out any abnormal.

@antinucleon
Copy link
Contributor

I re-run the experiment again. result is:
new:
[312] train-error:0.237291 train-logloss:0.712367 val-error:0.232272 val-logloss:0.725718
old:
[312] train-error:0.244916 train-logloss:0.745804 val-error:0.241366 val-logloss:0.7564

So I don't think it is CXXNET's problem.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants