- LegalEvalHub: Code for this website. See here for instructions on how to contribute tasks, leaderboards, and evaluation runs.
- legal-eval-harness: A harness for evaluating LLMs on legal tasks.
- legal-ml-datasets: A collection of legal datasets for machine learning.
- HELM: A framework for evaluating LLMs on legal tasks, run by Stanford CRFM.