From e7bfb63a581c15064c48b29588315299ec0359ed Mon Sep 17 00:00:00 2001 From: JIMMY ZHAO Date: Thu, 8 Aug 2024 10:55:41 -0400 Subject: [PATCH] add evalscope --- README.md | 1 + 1 file changed, 1 insertion(+) diff --git a/README.md b/README.md index 7d96399f..4f8b04a7 100644 --- a/README.md +++ b/README.md @@ -629,6 +629,7 @@ This repository contains a curated list of awesome open source libraries that wi * [DeepEval](https://github.com/confident-ai/deepeval) ![](https://img.shields.io/github/stars/confident-ai/deepeval.svg?style=social) - DeepEval is a simple-to-use, open-source evaluation framework for LLM applications. * [EvalAI](https://github.com/Cloud-CV/EvalAI) ![](https://img.shields.io/github/stars/Cloud-CV/EvalAI.svg?style=social) - EvalAI is an open source platform for evaluating and comparing AI algorithms at scale. * [Evals](https://github.com/openai/evals) ![](https://img.shields.io/github/stars/openai/evals.svg?style=social) - Evals is a framework for evaluating OpenAI models and an open-source registry of benchmarks. +* [evalscope](https://github.com/modelscope/evalscope) ![](https://img.shields.io/github/stars/modelscope/evalscope.svg?style=social) - evalscope is a streamlined and customizable framework for efficient large model evaluation and performance benchmarking. * [EvalPlus](https://github.com/evalplus/evalplus) ![](https://img.shields.io/github/stars/evalplus/evalplus.svg?style=social) - EvalPlus is a rigorous evaluation framework for LLM4Code. * [Evaluate](https://github.com/huggingface/evaluate) ![](https://img.shields.io/github/stars/huggingface/evaluate.svg?style=social) - Evaluate is a library that makes evaluating and comparing models and reporting their performance easier and more standardized. * [Evalverse](https://github.com/UpstageAI/evalverse) ![](https://img.shields.io/github/stars/UpstageAI/evalverse.svg?style=social) - Evalverse is a framework to effortlessly evaluate and report LLMs with no-code requests and comprehensive reports.