Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CP for LLMs #65

Closed
4 tasks done
pat-alt opened this issue Jun 1, 2023 · 4 comments · Fixed by #70
Closed
4 tasks done

CP for LLMs #65

pat-alt opened this issue Jun 1, 2023 · 4 comments · Fixed by #70
Assignees
Labels
enhancement New feature or request medium This is expected to be medium.

Comments

@pat-alt
Copy link
Member

pat-alt commented Jun 1, 2023

Issue set up for Experiment Week

  • Study this paper
  • Get data from hugging face
  • Train small transformer model from scratch
  • Look at fine-tuning of pre-trained model
@pat-alt pat-alt added enhancement New feature or request medium This is expected to be medium. labels Jun 1, 2023
@pat-alt pat-alt self-assigned this Jun 1, 2023
@pat-alt
Copy link
Member Author

pat-alt commented Jun 1, 2023

cc @florisdenhengst

@pat-alt
Copy link
Member Author

pat-alt commented Jun 12, 2023

Notes from paper

  • Context: LLMs for the task of multiple-choice question answering (MCQA)
  • Uncertainty, as estimated through CP, highly correlated with model accuracy
  • Working with inductive/split conformal prediction to avoid retraining
  • Model: LLaMA-13B
  • Data: MMLU benchmark containing MCQA question from 57 domains, 4 possible answers each
  • Find that set size is higher on average (closer to 4 for more difficult domains)
  • Find that set size negatively correlates with top-1 accuracy and argue that this can be used to filter low-quality predictions
  • Also find fairly robust size-stratified coverage, even though it does not appear they used adaptive prediction sets
  • Show that coverage guarantees do not hold if exchangeability assumption is violated (see below)

image

@pat-alt
Copy link
Member Author

pat-alt commented Jun 12, 2023

For work in Julia, we probably want to look at Transformers.jl

@pat-alt
Copy link
Member Author

pat-alt commented Jun 12, 2023

Idea: for small dataset/model, check if we can use conformal training

@pat-alt pat-alt linked a pull request Jun 27, 2023 that will close this issue
3 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request medium This is expected to be medium.
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant