Skip to content

Latest commit

 

History

History
15 lines (9 loc) · 904 Bytes

File metadata and controls

15 lines (9 loc) · 904 Bytes

Trying to replicate some of the plots from deep double descent: where bigger models and more data hurt

Currently the speed is 2 slow for 4K epochs :'( need to investigate ! I tried various things, but looks like batch size increase was the best approach

TODO: Train the models with the configs. Currently lambda has no available instances, it would be nice to have a 8GPU machine* Lambda used to have 8 GPU instances for me, but no more :( ended up switching to vast.ai

Results

The current results are without the label noise, since I don't see the exact same phenomena as OpenAI, I assume it might be a cause. The results are still interesting though.

Resnet width / Test Accuracy

Epoch wise test accuracy

Test accuracy over time