Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

special token "[semantic]" #44

Open
GinnyXiao opened this issue Jan 20, 2025 · 1 comment
Open

special token "[semantic]" #44

GinnyXiao opened this issue Jan 20, 2025 · 1 comment

Comments

@GinnyXiao
Copy link

For training your newly released checkpoint supporting semantic level segmentation, you added a special token "[semantic]" before the input prompt in the training data. I was wondering what difference this token makes? You have not modified your model architecture, which gives 1-1 mapping between your prediction and your prompt. The training strategy for referring seg datasets and semantics seg datasets remain the same, except for this token. Why would adding a "[semantic]" token improve your model ability? Would training without it yield similar performance?

Thank you very much in advance!

@CoderZhangYx
Copy link
Collaborator

When I apply semantic-segmentation datasets like ade20k, I add '[semantic]' because the category name refers to all the instances within the image in those datasets. This is in contrast to original referring datasets where the prompt only refers to a single instance. I apply the special token to balance the conflict, making the model behave better during joint training. By employing semantic segmentation datasets, the model is able to segment multiple objects, and also stuff categories if prompted with '[semantic]'

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants