SWE-MERA: A Dynamic Benchmark for Agenticly Evaluating Large Language Models on Software Engineering Tasks Paper • 2507.11059 • Published Jul 15 • 6
MERA Code: A Unified Framework for Evaluating Code Generation Across Tasks Paper • 2507.12284 • Published Jul 16 • 7
SWE-MERA: A Dynamic Benchmark for Agenticly Evaluating Large Language Models on Software Engineering Tasks Paper • 2507.11059 • Published Jul 15 • 6
MERA Code: A Unified Framework for Evaluating Code Generation Across Tasks Paper • 2507.12284 • Published Jul 16 • 7
SWE-MERA: A Dynamic Benchmark for Agenticly Evaluating Large Language Models on Software Engineering Tasks Paper • 2507.11059 • Published Jul 15 • 6
MERA Code: A Unified Framework for Evaluating Code Generation Across Tasks Paper • 2507.12284 • Published Jul 16 • 7
CoLES: Contrastive Learning for Event Sequences with Self-Supervision Paper • 2002.08232 • Published Feb 19, 2020 • 2
Leveraging Large Language Models in Code Question Answering: Baselines and Issues Paper • 2411.03012 • Published Nov 5, 2024