-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Strange output tuples #3
Comments
It seems that the value of min_entdist and min_numdist in the Inference class are wrongly initialized. I solve this issue by manually set
|
Hi, I also have the same issue, and I "know" where it comes from: during the fix for the previous issue, I have added See lines 91, 92 and 94 of the original code. However, to be honest, I am not entirely sure why these Would you have time to remove them from your local copy of this code (see lines 97 and 98) and see if training a model + inference works correctly? I don't have time right now to investigate properly, I am sorry. Let me know if this works and I'll update the code (and retrain models once more!) Thanks for your collaboration & efforts, |
Hi, I remove + 1 at:
Then evaluate with run.py, but the number of extracted tuples decreased largely due to this operation. I think +1 is necessary for this process. |
This error occurs when I inference with the pretrained models, now I'm training my own models, no errors occurs by now, I will let you know if it works to inference with the newly trained models. |
Hi, I remove the
It seems everything is fine on my side! |
Hi, I encounter the same issue of strange output tuples as mentioned above. I follow the instructions to remove '+1' in data.py and retrain the model. As instructed, I trained 10 BiLSTM models and 10 CNN models, and respectively selected top-3 models for the ensemble. The training and inference process worked well. However, I am confused about the output result:
The output result seems very high, so I wonder the correct method to calculate 'RG P% and RG #'? Is 'RG P%' equals to 'prec' or 'nodup_prec' ? And is 'RG #' equals to the number of 'nodup correct' divide 'total correct' (the output result is nearly 100, but about 50 in Wiseman's paper.) Could you help me to figure out this issue and I am really grateful for your help. Shuqi |
Did you evaluate the model on the training set? I trained 10 lstm mdoels and select best 5, then evaluated the GT summary of test set and nodup_prec is 0.9286. |
Hi Yunhao, Really thanks for your quick reply. I evaluated the model on the training and test dataset just now and here is the result:
The result is so high that seems kind of weird. Do you know how to compute the RG metric of RG P% and RG # according to the output? Thanks for your help! Shuqi |
Hi Shuqi, RG P% is computed by #correct/#pred, where #pred is the number of valid relations predicted by the model, and #correct is the number of correct relation among the valid relations. For more details, you can refer to https://github.com/KaijuML/rotowire-rg-metric/blob/5c5018cb1cb3feb584a0cf846b2e9f73b7db989b/inference.py#L9. By the way, I evaluated the model on the training set, the output is:
|
hi,can you tell me how to use the model to generate a summary. |
Hi,
I clone the newest code, download the models and run the extract script. The prec is 0.8846444487571716 but the output tuples prep_predictions.h5-tuples.txt seems to be very strange:
Do you have the same problem?
The text was updated successfully, but these errors were encountered: