Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FactCC Scores - cant replicate scores #4

Open
Lukecn1 opened this issue Jul 28, 2021 · 3 comments
Open

FactCC Scores - cant replicate scores #4

Lukecn1 opened this issue Jul 28, 2021 · 3 comments

Comments

@Lukecn1
Copy link

Lukecn1 commented Jul 28, 2021

Hi Again :)

I was checking my own implementation of the factCC scoring you described in the paper against your data, and noticed that for 90 cases we derived different scores

I suspect this is due to difference in how we split summaries into their individual sentences prior to classification and scoring.

How did you split summary sentences for factCC scoring?

(I use nltk sent_tokenize function)

@artidoro
Copy link
Owner

Hello,
I think you are right and this is probably due to differences in how we split sentences. I used spacy's sentence segmentation https://spacy.io/usage/linguistic-features#sbd

Let me know if the results are significantly different and we could investigate further otherwise we can close the issue.

@Lukecn1
Copy link
Author

Lukecn1 commented Jul 28, 2021

Hello,
I think you are right and this is probably due to differences in how we split sentences. I used spacy's sentence segmentation https://spacy.io/usage/linguistic-features#sbd

Let me know if the results are significantly different and we could investigate further otherwise we can close the issue.

The differences I have found is only for 90 cases so its not a massive difference already. But thanks for the reply, ill test it using spacy and get back to you :)

@Lukecn1
Copy link
Author

Lukecn1 commented Jul 30, 2021

Hello,
I think you are right and this is probably due to differences in how we split sentences. I used spacy's sentence segmentation https://spacy.io/usage/linguistic-features#sbd

Let me know if the results are significantly different and we could investigate further otherwise we can close the issue.

Using spacy i have 30 more differences in scores than using nltk. Do yo do any other preprocessing of data before scoring?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants