Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Addition Request - Paper on Improving Reasoning of Large Language Models with Alignment #3

Closed
Wangpeiyi9979 opened this issue Sep 6, 2023 · 2 comments

Comments

@Wangpeiyi9979
Copy link

Wangpeiyi9979 commented Sep 6, 2023

Hi, thanks for your excellent survey.

We recently proposed a new ranking strategy to align large language models (LLMs) to enhance their reasoning ability.
We also delve deeply into recent ranking-based alignment methods, such as DPO, RRHF, and PRO, and provide some analyses.

Here are the details of our work:

Title: Making Large Language Models Better Reasoners with Alignment
Link: https://arxiv.org/pdf/2309.02144.pdf

We kindly request that you consider adding our work to this repository and the survey.
Thank you for your time and consideration. 😊

@GaryYufei
Copy link
Owner

Thanks for sharing this great work with us. I have just added your paper into repo (see https://github.com/GaryYufei/AlignLLMHumanSurvey#rank-based-training), which will be incorporated into our next version of this survey!

@Wangpeiyi9979
Copy link
Author

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants