Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to prepare new dataset? #39

Open
jdyjjj opened this issue Oct 17, 2023 · 7 comments
Open

How to prepare new dataset? #39

jdyjjj opened this issue Oct 17, 2023 · 7 comments

Comments

@jdyjjj
Copy link

jdyjjj commented Oct 17, 2023

Great job, I have benefited a lot. I would like to train on the videos I have collected. What should I do if I want to train with my video dataset? Thank you.

@horsto
Copy link

horsto commented Oct 17, 2023

I'd also like to know! Is there an "easy" way to take the tracked points and re-train/re-fine the model?

@nikitakaraevv
Copy link
Contributor

Hi @jdyjjj, @horsto, I think the easiest way to train on a custom dataset right now is to adapt this class to your use case:

class KubricMovifDataset(CoTrackerDataset):

I'm not sure if this is considered an easy way, though :)

@Anderstask1
Copy link

Hi @nikitakaraevv, I adapted the KubrikMovifDataset to a local dataset. In my case, I have 180 points to track per video. However, when sampling data with augmentations, I sometimes end up with fewer points. I these cases, the getitem_helper function return gotit=False. Is that a wished behaviour?

Thanks for your answers!

@nikitakaraevv
Copy link
Contributor

Hi @Anderstask1, some points don't satisfy the sampling criteria: being visible in the first frame or the middle frame of the sequence. That's why we sometimes end up with fewer points, especially if we don't have additional points to sample from. I think you can modify the sampling criteria to make it less strict:

visibile_pts_inds = torch.cat(

@Anderstask1
Copy link

Hi again @nikitakaraevv, I modified the sampling criteria to make it less strict. However, during training, it seems like several points are sampled mid-sequence in the provided prediction videos (even though the ground truth video don't sample points mid-sequence). Do you know the reason for this?

@G1tQuality
Copy link

Hi @Anderstask1 @jdyjjj @horsto guys, did you successfully train on your own datasets? I'd appreciate it if you guys could share me some advice.

@horsto
Copy link

horsto commented Jul 18, 2024

I never tried, but I'd be very curious to know too!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants