Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MCAT and USMLE question-answer datasets for benchmarking? #178

Open
karafecho opened this issue Aug 21, 2023 · 0 comments
Open

MCAT and USMLE question-answer datasets for benchmarking? #178

karafecho opened this issue Aug 21, 2023 · 0 comments
Assignees

Comments

@karafecho
Copy link

This issue is to reconsider the work described in this paper, which focused on using MCAT multiple-choice questions for training and performance evaluation of ROBOKOP and several other Translator "Reasoners" during Phase I of the Translator program.

We used Khan Academy for the work reported in the paper. I had also looked into Kaplan. Then, there's the USMLE Step 1 questions, which proved to be a bit too challenging at the time and probably still is.

This morning, I stumbled on MedQA and MedMCQA, which look interesting, although I have not conducted a deep dive. MedQA is based on USMLE questions, so that may be too challenging. MedMCQA is based on Indian medical school entrance exams, which I am not familiar with, but they should be similar to MCAT questions.

Anyway, the main point in this issue is to consider whether we should consider using MCAT, USMLE, and/or other "ground truth" medical testing datasets for benchmarking.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants