Replies: 1 comment
-
Do you mean it should ignore duplicated words? Fuzz.token_set_ratio is a scorer which ignores duplicated words. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
It doesn't seem like any of the scorers consider the token frequency (how many times a word/substring appears), making it difficult to match phrases with repeating terms. For my current need, I'm able to simply filter out the reoccurring word, but it would be cool to have a frequency-sensitive scorer.
Beta Was this translation helpful? Give feedback.
All reactions