Data Science Asked by sneaky_lobster on March 31, 2021
I’m performing emotion classification over FER2013 dataset. I’m trying to measure different models performance, and when I checked ImageDataGenerator with a model I had already used I came up with the following situation:
Model without data augmentation got:
Model with data augmentation got:
As you can see, validation accuracy is the same in both models, but train accuracy is significantly different. In this case:
Thanks for your time.
Based on this, both models generalize equally well. However both have overfit, the second one more significantly. You simply want to avoid that. For example, I'd expect the best validation loss to be closer to training loss if you're using early stopping. You can turn up things like dropout.
If you do so, I think you'll find the augmented model ends up producing a better model (lower validation loss).
Answered by Sean Owen on March 31, 2021
Given two models that have the same out-of-sample performance but different in-sample performance, I would go with the simpler model. In other words, you get non-performance gain by going with the second model, but you have a drawback of greater complexity, perhaps even worse overfitting, and increased computing time.
However, accuracy is a flawed metric!
Compare the two models on their cross-entropy loss that I suspect you’re using to optimize them. Cross-entropy loss is a strictly proper scoring rule, while accuracy is not.
Please read this linked post on Cross Validated, the statistics Stack, and read the linked posts on Frank Harrell’s blog.
I also have a post over there about how to talk to your boss about using a proper scoring rule instead of accuracy.
Answered by Dave on March 31, 2021
Get help from others!
Recent Answers
Recent Questions
© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP