Until 2022, Google was the primary tool for answering questions. Recently, with the rise of Large Language Models (LLMs), there has been a significant shift in how people search for information online. This shift highlights the importance of question-answering abilities for LLMs, and particularly, improving their results and characterizing their knowledge.
This work aims to determine LLMs' level of knowledge on Multiple-Choice Questions (MCQ) by using various elimination strategies and investigating whether those strategies can enhance their abilities. Our findings suggest that LLMs do not improve their performance on MCQ by using elimination strategies, which might be due to their partial knowledge.
-
Run the following commands:
pip install datasets pip install accelerate -U pip install --upgrade transformers
-
Change the
TOKENandPATHconstants (as described inelimination.py). -
Change the
RUN_XXXandANALYZEconstants to control your running options.
Results can be found in 'Plots', 'Figures' and 'Data' directories.
This project was done as part of an advanced NLP course (67664) in the Hebrew University of Jerusalem.
