Skip to content

Latest commit

 

History

History
26 lines (19 loc) · 1.17 KB

README.md

File metadata and controls

26 lines (19 loc) · 1.17 KB

Multi-Programming Language Evaluation of Large Language Models of Code (MultiPL-E)

MultiPL-E is a system for translating unit test-driven neural code generation benchmarks to new languages. We have used MultiPL-E to translate two popular Python benchmarks (HumanEval and MBPP) to 18 other programming languages.

For more information:

Versions

  • Version 0.3.0 (work in progress)

  • Version 0.2.0: used to evaluate SantaCoder