There are 2 repositories under benchmarking-framework topic.
GERBIL - General Entity annotatoR Benchmark
Benchmark framework of compute-in-memory based accelerators for deep neural network (on-chip training chip focused)
Python Multi-Process Execution Pool: concurrent asynchronous execution pool with custom resource constraints (memory, timeouts, affinity, CPU cores and caching), load balancing and profiling capabilities of the external apps on NUMA architecture
A toolkit for auto-generation of OpenAI Gym environments from RDDL description files.
SustainDC is a set of Python environments for Data Center simulation and control using Heterogeneous Multi Agent Reinforcement Learning. Includes customizable environments for workload scheduling, cooling optimization, and battery management, with integration into Gymnasium.
Benchmark framework of compute-in-memory based accelerators for deep neural network (on-chip training chip focused)
A guidance for the design and evaluation of motion planners for quadrotors in Environments with Varying Complexities
PIMeval simulator and PIMbench suite
Arline Benchmarks platform allows to benchmark various algorithms for quantum circuit mapping/compression against each other on a list of predefined hardware types and target circuit classes
Benchmarking machine learning inferencing on embedded hardware.
The greatest collection of the worst code
Command execution time meter
Benchmark framework of compute-in-memory based accelerators for deep neural network (inference engine focused)
PHP Micro & Router Framework Benchmark
Benchmarking optimization solvers.
A library to measure and profile some metrics of your code (execution time, average execution time, garbage collections)
CLAMH (Cross-LAnguage Microbenchmark Harness) is a language-independent benchmark harness design and the implementation of that design for different languages.
Framework for benchmarking deep learning operators for Apache MXNet
Observatory is an extensible Java framework for benchmarking InfiniBand verbs implementations.
Consistent, repeatable CS2 benchmarks – Automated CapFrameX capture with precise timing for reliable, comparable results every run.
Cbench is a header only micro micro-benchmarking toolkit for c. Designed to remain in the code base, by being eliminated during compile time when not enabled.
Crossbar Parasitics Simulator – A tool for benchmarking parasitic resistance models in RRAM crossbars and evaluating neural networks under realistic hardware constraints.
A client-server code to bench-mark dpdk in different configurations
PARROT (Performance Assessment of Reasoning and Responses On Trivia) is a novel benchmarking framework designed to evaluate Large Language Models (LLMs) on real-world, complex, and ambiguous QA tasks.
A lightweight benchmarking and visualization framework to analyze long-context failures in large language models (LLMs) using synthetic datasets, retrieval-augmented methods, and evaluation metrics.
How To Measure And Improve Code Efficiency with Pytest Benchmark (The Ultimate Guide)