CodeUltraFeedback: aligning large language models to coding preferences
-
Updated
Mar 17, 2024 - Python
CodeUltraFeedback: aligning large language models to coding preferences
This is the repo for the survey of Bias and Fairness in IR with LLMs.
Code and data for ACL ARR 2024 paper "Benchmarking Cognitive Biases in Large Language Models as Evaluators"
Notebooks for evaluating LLM based applications using the Model (LLM) as a judge pattern.
Add a description, image, and links to the llm-as-a-judge topic page so that developers can more easily learn about it.
To associate your repository with the llm-as-a-judge topic, visit your repo's landing page and select "manage topics."