Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about implementation details #3

Open
Co4AI opened this issue Jan 6, 2021 · 3 comments
Open

Questions about implementation details #3

Co4AI opened this issue Jan 6, 2021 · 3 comments
Labels
enhancement New feature or request

Comments

@Co4AI
Copy link

Co4AI commented Jan 6, 2021

As introduced in the paper, TransTrack takers composite features (from different frames) as input for transformer during inference.
But I find TransTrack take features from the same frame during training (only supports static-picture training). Moreover, when train current-frame decoder, it doesn't take pre-feature (combine the same features). Am I right?

@Co4AI
Copy link
Author

Co4AI commented Jan 6, 2021

It seems that self.decoder_track is trained under torch.randn(1).item() > 0.0 condition, self.decoder is trained otherwise. There are two questions:

  1. when we train self.decoder, the two part feature maps are the same which different from testing stage.
  2. when we train self.decoder_track, we match outputs of modified images with unchanged annotation to get indices.

I think it's hard to understand. Do I miss something important?

@PeizeSun
Copy link
Owner

PeizeSun commented Jan 6, 2021

It seems that self.decoder_track is trained under torch.randn(1).item() > 0.0 condition, self.decoder is trained otherwise. There are two questions:

  1. when we train self.decoder, the two part feature maps are the same which different from testing stage.
  2. when we train self.decoder_track, we match outputs of modified images with unchanged annotation to get indices.

I think it's hard to understand. Do I miss something important?

Hi~

  1. We tried to train self.decoder with two feature from different images, the result is similar(even a little worse).
  2. We are verifying now changing annotations accordingly, and to see whether it helps. Once we get the result, we will update the result here

@Co4AI
Copy link
Author

Co4AI commented Jan 7, 2021

It seems that self.decoder_track is trained under torch.randn(1).item() > 0.0 condition, self.decoder is trained otherwise. There are two questions:

  1. when we train self.decoder, the two part feature maps are the same which different from testing stage.
  2. when we train self.decoder_track, we match outputs of modified images with unchanged annotation to get indices.

I think it's hard to understand. Do I miss something important?

Hi~

  1. We tried to train self.decoder with two feature from different images, the result is similar(even a little worse).
  2. We are verifying now changing annotations accordingly, and to see whether it helps. Once we get the result, we will update the result here

Thank you for your reply~

  1. About training self.decoder with different features, I got the same result (a little worse). But I wonder why different training strategy can get improvement.
  2. Looking forward for your updating.

@PeizeSun PeizeSun added the enhancement New feature or request label Mar 30, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants