bigcode-project/bigcode-evaluation-harness
bigcode-evaluation-harness
A framework for the evaluation of autoregressive code generation language models.
Builder

bigcode-project
bigcode-project • individual
Stars
1,029
Using upstream star count
Forks
257
Using upstream fork count
Open Issues
0
Activity Score
0/100
0 commits in 30d
Created
Aug 9, 2022
Project creation date
README Summary
The BigCode Evaluation Harness is a comprehensive framework designed to evaluate autoregressive code generation language models across multiple programming tasks and benchmarks. It provides standardized evaluation metrics and datasets to assess the performance of AI models in generating functional code. The framework supports various code generation tasks and enables reproducible evaluation of large language models specialized in programming.
AI Dev Skills
Unmapped
Tags
Taxonomy
Deployment Context
Skill Areas
Recent Activity
Updated 8 months ago
7 Days
0
30 Days
0
90 Days
0
Quality
research- Quality
- high
- Maturity
- research
Categories
PM Skills
Languages
Timeline
- Project created
- Aug 9, 2022
- Forked
- Mar 22, 2026
- Your last push
- 8 months ago
- Upstream last push
- 8 months ago
- Tracked since
- Jul 22, 2025
Similar Repos
pgvector cosine similarity · $0
Loading…