A simple template module for evaluating user/runtime-unknown value expressions in a safe manner, using Python's 'eval'.
-
Updated
Oct 25, 2018
A simple template module for evaluating user/runtime-unknown value expressions in a safe manner, using Python's 'eval'.
LLM evaluation framework
Benchmark for assessing contextual-semantic sentence models in Brazilian legal domain.
A tool to perform functional testing and performance testing of the Dhruva Platform
CHECKLIST-style test cases and the testing of three Hungarian Named Entity Recognition tools.
Flight Delay using Machine Learning
Most popular metrics used to evaluate object detection algorithms.
This repository contains codes for NLP project "In-context Learning of Pre-trained Language Models for Controlled Dialogue Summarization: A Holistic Benchmark and Empirical Analysis"
Integrated Evaluation Framework - Front-End Web Application
Official repository for the paper *Are Models Biased on Text Without Gender-related Language?*, published in ICLR 2024.
Calculate calibration of a model on DataSHIELD servers
drug repositioning method evaluation
N-Compariw: End-to-End Workflow for Neural Networks Comparison
A hybrid search engine based on the BM25 and VSM retrieval models.
Evaluation Framework for Graph Sample Clustering
REgistration PErformance Assessment Tools
A Visual Dashboard for Fundamental Benchmarking of LLMs
Add a description, image, and links to the evaluation-framework topic page so that developers can more easily learn about it.
To associate your repository with the evaluation-framework topic, visit your repo's landing page and select "manage topics."