evaluations topic
adversarial-explainable-ai
💡 Adversarial attacks on explanations and how to defend them
evaluations
This library implements various metrics (including Kaggle Competition, Medicine) for evaluating ML, DL, AI models, and algorithms. 📐📊📈📉📏
42-Evaluations
42 School Projects Evaluation Marking Criteria
Crunch
The fastest java expression compiler/evaluator
log10
Python client library for improving your LLM app accuracy
leaf-playground
A framework to build scenario simulation projects where human and LLM based agents can participant in, with a user-friendly web UI to visualize simulation, support automatically evaluation on agent ac...
langtrace
Langtrace 🔍 is an open-source, Open Telemetry based end-to-end observability tool for LLM applications, providing real-time tracing, evaluations and metrics for popular LLMs, LLM frameworks, vectorD...
evalkit
The TypeScript LLM Evaluation Library