-
Notifications
You must be signed in to change notification settings - Fork 177
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question about training and validation split #280
Comments
Hi, line 117 is just because the evaluation on all images is very slow. For the train test split, you need to check with data parser you used. For reconstruction, since we care more about the reconstruction quality and using more images is generally help so we use all available images for training. |
I used the sdfstudio-data parser, that's the one using train data also as eval. I was thinking that because we want to reproduce well just the train scenario, overfitting might not be that bad. But correct me if I'm wrong |
Yes, It might be. But I think since the SDF has strong regulations with eikonal loss, using 10 % more images is complete fine. In general, if you use more images for training NeRF, you will also get better results. Of course, you will need train/test split if you want to do some bench mark. |
Hi, i'm training some bakedangelo models which gave very nice outputs, but I can't seem to wrap my head around one thing:
The comment on row 117 states: # set to a very large model so we don't eval with all images
So i was wondering that there's not a "complete" train/validation split like happens on Nerfstudio (90-10)
Correct me if i'm wrong: Let's say I have 100 images: in sdfstudio these models use all the 100 images for training set and also the same 100 images as validation set, but the 100 images aren't all used because the model is set to a very large number of iterations (in this case 1 million). I tipically stop the training at 200k~ iterations so that should mean that it's at 1/5 of the maximum allowed steps for evaluating all images. Does that mean that I've used 1/5 of all the evaluation dataset -> 20 images?
Second question: why are we using the same images for training and validation?
Sorry if I've been messy but this part has been bugging me for the past week. Can you explain it to me better?
Thanks in advance
The text was updated successfully, but these errors were encountered: