Zobrazeno 1 - 1
of 1
pro vyhledávání: '"Dristi, Simantika"'
Large Language Models (LLMs) are gaining popularity among software engineers. A crucial aspect of developing effective code generation LLMs is to evaluate these models using a robust benchmark. Evaluation benchmarks with quality issues can provide a
Externí odkaz:
http://arxiv.org/abs/2404.10155