-
Notifications
You must be signed in to change notification settings - Fork 154
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Amount of data for training #6
Comments
I doubt you'll get a response on the weights file; I've asked before and didn't receive a response; after looking into it, I think the data being trained on might be under NDA or otherwise not available for the public, so it's unlikely the trained model will be shared. Depending on what you want to do, I recommend checking out COWC (Cars Overhead With Context), SpaceNet, or xView as possible datasets. In my experience, a couple hundred samples will start to get some good results, although you'll probably spend a lot of time preprocessing the data to figure out what works well for you. |
I have started training for 700 objects of 1 category only for 60,000 epochs, Lets hope it works! |
We've been busy updating the code and writing papers, and updated examples (and weights) will be uploaded in the near future. In the meantime, Table 2 and 3 of https://arxiv.org/abs/1805.09512 give an idea of what you can expect for performance versus training size. |
That's awesome to hear! |
That's Great! |
Just to make sure I am not doing something wrong, |
Give the https://github.com/avanetten/simrdwn/blob/master/core/prep_data_cowc.py a try, hopefully this will clear up your nans issue. As for how much training data you need, often only a few dozen is enough (see https://arxiv.org/pdf/1805.09512, Table 2). |
How much data is required in general for training?
Also, if possible, can you share the pretrained weights file?
The text was updated successfully, but these errors were encountered: