Skip to content
View Darcos923's full-sized avatar
πŸ’­
πŸ€– Al human on working!
πŸ’­
πŸ€– Al human on working!

Block or report Darcos923

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Stars

πŸ§ͺ Evaluator | LLM

5 repositories

Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.

Python 15,583 2,670 Updated Dec 18, 2024

Evaluation tool for LLM QA chains

Python 1,070 93 Updated May 10, 2023

Build, evaluate, understand, and fix LLM-based apps

Jupyter Notebook 484 33 Updated Jan 16, 2024

Supercharge Your LLM Application Evaluations πŸš€

Python 8,337 855 Updated Feb 24, 2025

ACL 2023: Evaluating Open-Domain Question Answering in the Era of Large Language Models

Python 43 1 Updated Jan 12, 2024