Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How Triton inference server always compare the current frame infer result with the previous one #7893

Open
Komoro2023 opened this issue Dec 19, 2024 · 1 comment
Labels
question Further information is requested

Comments

@Komoro2023
Copy link

Komoro2023 commented Dec 19, 2024

Hi there, I want to reduce the false alarms of fire and smoke detection model, and our models are deployed on triton inference server, so I try to find if there is a way that on triton inference server it always compare the current infer result with previous infere result, if the area of bbox of detection is changing, so it is probably a real alarm, and give the response, or it is a false alarm, give a empty repsonse. so please help me find a way or is there a feature can solve my problem?
Many thanks.

I found Triton Response Cache feature probably could do this work, but don't know how to get the previous infer result from response cache, and there is no any example for extracting from response cache

@tanmayv25
Copy link
Contributor

tanmayv25 commented Jan 24, 2025

@Komoro2023 You can add a custom python backend model in front of your fire and smoke detection model building pipeline either using ensemble or BLS. See here for more information.

The custom model can store the last result and compare it with the incoming result. If they are identical, it can return empty response, otherwise it can return the needful.

Response Cache is placed in front of the model. If the input to the model is identical to the previously cached input then response data is pulled from the cache instead of executing inference on model.

@tanmayv25 tanmayv25 added the question Further information is requested label Jan 24, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Development

No branches or pull requests

2 participants