Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Paper Digest: NeurIPS-2023 Highlights (Full List) #333

Open
1 task
irthomasthomas opened this issue Jan 13, 2024 · 0 comments
Open
1 task

Paper Digest: NeurIPS-2023 Highlights (Full List) #333

irthomasthomas opened this issue Jan 13, 2024 · 0 comments
Labels
finetuning Tools for finetuning of LLMs e.g. SFT or RLHF llm Large Language Models llm-experiments experiments with large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately Papers Research papers

Comments

@irthomasthomas
Copy link
Owner

Paper Digest: NeurIPS 2023 Highlights

https://www.paperdigest.org

1, Toolformer: Language Models Can Teach Themselves to Use Tools
Timo Schick; Jane Dwivedi-Yu; Roberto Dessi; Roberta Raileanu; Maria Lomeli; Eric Hambro; Luke Zettlemoyer; Nicola Cancedda; Thomas Scialom;
Related Papers   Related Patents   Related Grants   Related Venues   Related Experts   Related Code   View
Highlight: In this paper, we show that LMs can teach themselves to use external tools via simple APIs and achieve the best of both worlds.

2, Self-Refine: Iterative Refinement with Self-Feedback
Aman Madaan; Niket Tandon; Prakhar Gupta; Skyler Hallinan; Luyu Gao; Sarah Wiegreffe; Uri Alon; Nouha Dziri; Shrimai Prabhumoye; Yiming Yang; Shashank Gupta; Bodhisattwa Prasad Majumder; Katherine Hermann; Sean Welleck; Amir Yazdanbakhsh; Peter Clark;
Related Papers   Related Patents   Related Grants   Related Venues   Related Experts   Related Code   View
Highlight: Motivated by how humans refine their written text, we introduce Self-Refine, an approach for improving initial outputs from LLMs through iterative feedback and refinement.

3, Vicuna Evaluation: Exploring LLM-as-a-Judge and Chatbot Arena
Lianmin Zheng; Wei-Lin Chiang; Ying Sheng; Siyuan Zhuang; Zhanghao Wu; Yonghao Zhuang; Zi Lin; Zhuohan Li; Dacheng Li; Eric Xing; Hao Zhang; Joseph Gonzalez; Ion Stoica;
Related Papers   Related Patents   Related Grants   Related Venues   Related Experts   View
Highlight: To address this, we explore using strong LLMs as judges to evaluate these models on more open-ended questions. We examine the usage and limitations of LLM-as-a-judge, including position, verbosity, and self-enhancement biases, as well as limited reasoning ability, and propose solutions to mitigate some of them.

Suggested labels

{ "key": "LLM-Applications", "value": "Topics related to practical applications of Large Language Models in various fields" }

@irthomasthomas irthomasthomas added llm Large Language Models finetuning Tools for finetuning of LLMs e.g. SFT or RLHF llm-experiments experiments with large language models Papers Research papers New-Label Choose this option if the existing labels are insufficient to describe the content accurately labels Jan 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
finetuning Tools for finetuning of LLMs e.g. SFT or RLHF llm Large Language Models llm-experiments experiments with large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately Papers Research papers
Projects
None yet
Development

No branches or pull requests

1 participant