Randl / MultiPL-E

A multi-programming language benchmark for evaluating the performance of large language model of code.

Home Page:https://nuprl.github.io/MultiPL-E/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Multi-Programming Language Evaluation of Large Language Models of Code (MultiPL-E)

MultiPL-E is a system for translating unit test-driven neural code generation benchmarks to new languages. We have used MultiPL-E to translate two popular Python benchmarks (HumanEval and MBPP) to 18 other programming languages.

For more information:

Versions

  • Version 0.3.0 (work in progress)

  • Version 0.2.0: used to evaluate SantaCoder

About

A multi-programming language benchmark for evaluating the performance of large language model of code.

https://nuprl.github.io/MultiPL-E/

License:Other


Languages

Language:Python 86.4%Language:Lua 10.3%Language:Jupyter Notebook 2.1%Language:Shell 0.9%Language:Dockerfile 0.2%Language:C++ 0.0%Language:Makefile 0.0%