-
Notifications
You must be signed in to change notification settings - Fork 20
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comparing between two audio #23
Comments
Hi @leonardltk! The way almost all of these algorithms (included the ones that are not in the repo) work is that they estimate a similarity/dissimilarity score among a collection of songs. These similarity scores are sorted to compute a number of performance metrics common in Information Retrieval tasks, e.g. mean average precision, mean rank, number of relevant items in top-1. As a result, the absolute values of these similarity scores do not necessarily mean something. What we care more is that when we give a query, whether the algorithm returns a relevant item (in our case, a cover) in the first retrieved results or not. Based on the algorithm that you use, you can check the distance (or similarity score) distributions to set some thresholds. For example, if the distances of covers lie between 0 and 0.4, and the distances of non-covers lie between 0.3 and 0.9, you can then set a threshold considering whether precision or recall is more important to you. Keep in mind that these distance distributions are likely to differ depending on the algorithm that you use. I hope this answers your question. Please let me know if you have any further questions! |
Thanks for your reply! In that case, how do you suggest i go about proceeding with this problem ?
An approach im considering right now is this: For example,
Does it mean i can use this function during test time ?
1 problem i foresee using this is that the scores are unnormalised, as i did not take into account |
Hey @leonardltk ! I am currently facing a similar problem as the one you mentioned. How were you able to solve it ? |
Hi,
according to the README, everything seems to be end-to-end when running benchmark across the whole covers80 dataset.
Is there a way to simply compare two audio using any of the algorithms, and determine whether they are indeed cover versions of each other?
The text was updated successfully, but these errors were encountered: