Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Discrepancies in Scores when compared to basic implementations. #19

Open
kiran-vj opened this issue Jan 17, 2025 · 0 comments
Open

Discrepancies in Scores when compared to basic implementations. #19

kiran-vj opened this issue Jan 17, 2025 · 0 comments

Comments

@kiran-vj
Copy link

I was comparing scores for this package against a basic Java implementation which can be found here https://www.geeksforgeeks.org/jaro-and-jaro-winkler-similarity/

While for majority of the cases I observed identical scores- but I also see cases where there are large differences.

For example:

s1 = "aidahaddad", s2 = "thadwindle";

The basic java version returns a score= 0.5666666666666667

While the python package returns a score= 0.6

All the parameters such as prefix_weight/max_prefix_len etc. seem to be in line with the python package.

What could be the root cause of such discrepancies? Are there any corresponding Java packages where we could observe scores identical to the python implementation here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant