| | --- |
| | dataset_info: |
| | features: |
| | - name: id |
| | dtype: string |
| | - name: input |
| | dtype: string |
| | - name: opa |
| | dtype: string |
| | - name: opb |
| | dtype: string |
| | - name: opc |
| | dtype: string |
| | - name: opd |
| | dtype: string |
| | - name: cop |
| | dtype: int64 |
| | - name: choice_type |
| | dtype: string |
| | - name: exp |
| | dtype: string |
| | - name: subject_name |
| | dtype: string |
| | - name: topic_name |
| | dtype: string |
| | - name: output |
| | dtype: string |
| | - name: options |
| | dtype: string |
| | - name: letter |
| | dtype: string |
| | - name: incorrect_letters |
| | list: string |
| | - name: incorrect_answers |
| | list: string |
| | - name: single_incorrect_answer |
| | dtype: string |
| | - name: system_prompt |
| | dtype: string |
| | - name: messages |
| | list: |
| | - name: content |
| | dtype: string |
| | - name: role |
| | dtype: string |
| | splits: |
| | - name: train |
| | num_bytes: 221816870 |
| | num_examples: 164539 |
| | - name: test |
| | num_bytes: 24647517 |
| | num_examples: 18283 |
| | download_size: 144137775 |
| | dataset_size: 246464387 |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: train |
| | path: data/train-* |
| | - split: test |
| | path: data/test-* |
| | dataset_name: mkurman/medmcqa-hard |
| | license: cc |
| | language: |
| | - en |
| | task_categories: |
| | - multiple-choice |
| | - question-answering |
| | - reinforcement-learning |
| | tags: |
| | - medical |
| | - MCQ |
| | - evaluation |
| | - SFT |
| | - DPO |
| | - RL |
| | pretty_name: MedMCQA-Hard |
| | size_categories: |
| | - 10k<n<1M |
| | --- |
| | |
| | # medmcqa-hard |
| |
|
| | **A harder, de-duplicated remix of MedMCQA** designed to reduce memorization and strengthen medical MCQ generalization. |
| |
|
| | ## Why “hard”? |
| |
|
| | * **Answer list variants:** Each correct option appears in **multiple phrasing/list variants** (e.g., reordered enumerations, equivalent wording), so models can’t rely on surface-form recall and must reason over content. |
| | * **RL-friendly targets:** Every item includes **one canonical correct answer** and both **single** and **set** of incorrect answers → plug-and-play for **DPO**, **RLAIF/GRPO**, and contrastive objectives. |
| | * **Chat formatting:** Adds lightweight **`messages`** (and optional `system_prompt`) not present in the original dataset, making it convenient for instruction-tuned models and SFT. |
| |
|
| | ## Intended uses |
| |
|
| | * Robust **eval** of medical QA beyond memorization. |
| | * **SFT** with chat-style prompts. |
| | * **DPO / other RL** setups using `single_incorrect_answer` or `incorrect_answers`. |
| |
|
| | ## Data schema (fields) |
| |
|
| | * `question`: str |
| | * `options`: list[str] (usually 4) |
| | * `letter`: str (A/B/C/D) |
| | * `cop`: int (0-based index of correct option) |
| | * `incorrect_answers`: list[str] |
| | * `single_incorrect_answer`: str |
| | * `messages`: list[{role: "system"|"user"|"assistant", content: str}] |
| | * `system_prompt`: str (optional) |
| |
|
| | ### Example |
| |
|
| | ```json |
| | { |
| | "question": "Which of the following is true about …?", |
| | "options": ["A …", "B …", "C …", "D …"], |
| | "letter": "C", |
| | "cop": 2, |
| | "incorrect_answers": ["A …", "B …", "D …"], |
| | "single_incorrect_answer": "B …", |
| | "messages": [ |
| | {"role":"system","content":"You are a medical tutor."}, |
| | {"role":"user","content":"Q: Which of the following…?\nA) …\nB) …\nC) …\nD) …"} |
| | ] |
| | } |
| | ``` |
| |
|
| | ## Source & attribution |
| |
|
| | Derived from **MedMCQA** (Pal, Umapathi, Sankarasubbu; CHIL 2022). Please cite the original dataset/paper when using this work. |
| |
|
| | > **Safety note:** Research/education only. Not for clinical use. |
| |
|
| |
|