CALSA+ dataset and other additional materials for the paper: Identification of Multiple Logical Interpretations in Counter-Arguments
dataset folder contains the CALSA+ dataset and the related files
processed_merged_results_readable.csv: Raw annotation results of three annotators at the predicate-levelall.jsonl: processed predicate-level results, the answer is determined by taking as manyYESas possible (i.e., answer =YESif any of the annotators selectYES; otherwise,NO)calsaplus_dataset.jsonl: The CALSA+ dataset where each CA has multiple logical interpretations obtained by aggregating the predicate-level results inall.jsonl- The ptn numbers are consistent with the patterns of logical structure defined in original CALSA paper: https://aclanthology.org/2024.findings-emnlp.661/
original_calsa_testset.json: The original CALSA test set where CALSA+ is created from.
grpo folder contains all the data and scripts for RLVR experiments
- For training models with RLVR, run any of the scripts in
grpo/src/scripts/train/, each one is corresponded to a base model- e.g.,
cd <where-you-clone-this-repo> && source grpo/src/scripts/train/qwen25-7b-instruct_deepspeed_zero2.sh
- e.g.,
- For runing inferences, run any of the scripts in
grpo/src/scripts/inference/, each one is corresponded to a base model- e.g.,
cd <where-you-clone-this-repo> && source grpo/src/scripts/inference/qwen25-7b-instruct_deepspeed_zero2.sh
- e.g.,
- For running baseline experiments to compare with RLVR, run the corresponding script in
grpo/src/scripts/inference/baselines/
sft folder contains all the data and scripts for SFT experiments
- For training models with SFT, run any of the scripts in
sft/src/scripts/train/, each one is corresponded to a base model- e.g.,
cd <where-you-clone-this-repo> && source sft/src/scripts/train/qwen25-7b-instruct_deepspeed_zero2.sh
- e.g.,
- For runing inferences, run any of the scripts in
sft/src/scripts/inference/, each one is corresponded to a base model- e.g.,
cd <where-you-clone-this-repo> && source sft/src/scripts/inference/qwen25-7b-instruct_deepspeed_zero2.sh
- e.g.,
prompt_enginneering folder contains all the data and scripts for prompting OpenAI models
- Usage:
python prompt_engineering/src/gpts.py
For evaluting any of the above experiments, run python evaluate_results.py with the corresponding file path where you saved the results
The content of this project itself is licensed under the MIT license, and the dataset provided in the folder dataset is licensed under the CC BY 4.0.