-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add parenthesis to score expression #13
Conversation
@bunny-therapist Do you mind checking against the sample on their website? http://yake.inesctec.pt/demo.html?doc=Sample1 It looks like the parentheses causes a decrease in accuracy. |
@bunny-therapist Do you have a discord we can discuss this on? I'm looking at this. The scores look better with this version but the actual results seem slightly worse. There seems to be an increase of results with adjectives and I don't think that's what we want. I'm thinking this could be a scenario where the fix you've posted is uncovering a deficiency somewhere else but I'm not sure. |
I am running it against their sample, but I am not getting the same results with either this PR or the pre-existing yake-rust code. I am trying to create python binding for yake-rust so we can replace LIAAD/yake in our projects. For this reason, I am running tests comparing the yake-rust results to LIAAD/yake - that is how I am finding these issues. Even with this PR, I am not getting the same as LIAAD/yake nor the homepage. However, I believe that is because there are more issues here. I think this PR fixes one issue, but I still think there are issues related to relatedness and frequency (when I am comparing LIAAD/yake and yake-rust, the discrepancies appear to be coming from those two). |
I am pretty sure this is also a bug: #12 But I think there may be more. |
Do we expect agreement with the yake homepage? Because in that case we should just use their scores for the tests in the future. How exactly do you tell the accuracy? I don't know what to look for. |
No I do not have a discord. I have only used the discords of others to discuss their projects. I am not a very experienced discord user. |
If I do the changes from #12 together with these changes, we get
|
@bunny-therapist I can't. I'm going based on what I'd expect so not objective by any measure. For me though the When I originally released this I remember noticing the issue but thinking the results were close enough. To answer your question, yes I think we should shoot for their scores. |
I think the play here is to branch off of main, maybe create a "v1.0.0" branch to work towards 1:1 scores with that. That way if we start going the wrong direction we can easily just keep whats in main because I think it works well enough if we can't get to 1:1 scoring. Let me know your thoughts. And yes, sorry for calling you out like that before. Coffee is kicking in now. I promise to be pleasant. |
To work toward their scores, that sounds like a good idea. But we are merging this and looking into the other bug, right? |
@bunny-therapist I'll create a v1.0.0-alpha branch and release this as that. |
I can post more about the other bugs I reported and what scores we get tomorrow. I have to stop for today since it is getting late here. |
Thanks for your help. I'll release this today sometime. |
Closes #10