#大语言模型#Test your prompts, agents, and RAGs. Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command ...
The LLM Evaluation Framework
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
The official evaluation suite and dynamic data release for MixEval.
#大语言模型#LangFair is a Python library for conducting use-case level LLM bias and fairness assessments
#大语言模型#Python SDK for experimenting, testing, evaluating & monitoring LLM-powered applications - Parea AI (YC S23)
#大语言模型# MIT-licensed Framework for LLMs, RAGs, Chatbots testing. Configurable via YAML and integrable into CI pipelines for automated testing.
#大语言模型#[ACL'24] A Knowledge-grounded Interactive Evaluation Framework for Large Language Models
#大语言模型#Develop reliable AI apps
#大语言模型#Benchmarking Large Language Models for FHIR
FM-Leaderboard-er allows you to create leaderboard to find the best LLM/prompt for your own business use case based on your data, task, prompts
Realign is a testing and simulation framework for AI applications.
Code for "Prediction-Powered Ranking of Large Language Models", NeurIPS 2024.
#大语言模型#Create an evaluation framework for your LLM based app. Incorporate it into your test suite. Lay the monitoring foundation.
#大语言模型#TypeScript SDK for experimenting, testing, evaluating & monitoring LLM-powered applications - Parea AI (YC S23)
#大语言模型#Community Plugin for Genkit to use Promptfoo
#大语言模型#Evaluating LLMs with Multiple Problems at once: A New Paradigm for Probing LLM Capabilities
#大语言模型#Shin Rakuda is a comprehensive framework for evaluating and benchmarking Japanese large language models, offering researchers and developers a flexible toolkit for assessing LLM performance across div...