SOTAVerified

Code Completion

Papers

Showing 151200 of 212 papers

TitleStatusHype
FEA-Bench: A Benchmark for Evaluating Repository-Level Code Generation for Feature Implementation0
From Copilot to Pilot: Towards AI Supported Software Development0
Full Line Code Completion: Bringing AI to Desktop0
GenAI for Simulation Model in Model-Based Systems Engineering0
Generation Probabilities Are Not Enough: Uncertainty Highlighting in AI Code Completions0
HiLDe: Intentional Code Generation via Human-in-the-Loop Decoding0
Horizon-Length Prediction: Advancing Fill-in-the-Middle Capabilities for Code Generation with Lookahead Planning0
Identifying and Mitigating the Security Risks of Generative AI0
Identifying and Mitigating Vulnerabilities in LLM-Integrated Applications0
Improving Code Autocompletion with Transfer Learning0
Improving FIM Code Completions via Context & Curriculum Based Learning0
IntelliCode Compose: Code Generation Using Transformer0
Interpretability Illusions in the Generalization of Simplified Models0
Is Next Token Prediction Sufficient for GPT? Exploration on Code Logic Comprehension0
Unveiling Code Pre-Trained Models: Investigating Syntax and Semantics Capacities0
Jailbreak Attacks and Defenses Against Large Language Models: A Survey0
JudgeRank: Leveraging Large Language Models for Reasoning-Intensive Reranking0
Laminar: A New Serverless Stream-based Framework with Semantic Code Search and Code Completion0
Learning to Extend Program Graphs to Work-in-Progress Code0
Learning to Complete Code with Sketches0
HierarchyNet: Learning to Summarize Source Code with Heterogeneous Representations0
LibEvolutionEval: A Benchmark and Study for Version-Specific Code Generation0
Long-Range Modeling of Source Code Files with eWASH: Extended Window Access by Syntax Hierarchy0
M2rc-Eval: Massively Multilingual Repository-level Code Completion Evaluation0
MarsCode Agent: AI-native Automated Bug Fixing0
MathFimer: Enhancing Mathematical Reasoning by Expanding Reasoning Steps through Fill-in-the-Middle Task0
Measuring memorization in RLHF for code completion0
Mechanistic Understanding of Language Models in Syntactic Code Completion0
Model Cascading for Code: A Cascaded Black-Box Multi-Model Framework for Cost-Efficient Code Completion with Self-Testing0
MultiCoder: Multi-Programming-Lingual Pre-Training for Low-Resource Code Completion0
Natural Language Generation and Understanding of Big Code for AI-Assisted Programming: A Review0
Neural Machine Translation for Code Generation0
Neural Models for Source Code Synthesis and Completion0
NoEsis: Differentially Private Knowledge Transfer in Modular LLM Adaptation0
OMPGPT: A Generative Pre-trained Transformer Model for OpenMP0
On Explaining (Large) Language Models For Code Using Global Code-Based Explanations0
Past as a Guide: Leveraging Retrospective Learning for Python Code Completion0
Plan for Speed -- Dilated Scheduling for Masked Diffusion Language Models0
Procedural Memory Is Not All You Need: Bridging Cognitive Gaps in LLM-Based Agents0
Protect Your Secrets: Understanding and Measuring Data Exposure in VSCode Extensions0
R2C2-Coder: Enhancing and Benchmarking Real-world Repository-level Code Completion Abilities of Code Large Language Models0
Repoformer: Selective Retrieval for Repository-Level Code Completion0
REPOFUSE: Repository-Level Code Completion with Fused Dual Context0
RepoFusion: Training Code Models to Understand Your Repository0
RepoMasterEval: Evaluating Code Completion via Real-World Repositories0
Rethinking Software Engineering in the Foundation Model Era: From Task-Driven AI Copilots to Goal-Driven AI Pair Programmers0
Retrieval-augmented code completion for local projects using large language models0
RTLRepoCoder: Repository-Level RTL Code Completion through the Combination of Fine-Tuning and Retrieval Augmentation0
SecureFalcon: Are We There Yet in Automated Software Vulnerability Detection with LLMs?0
Sequence Model Design for Code Completion in the Modern IDE0
Show:102550
← PrevPage 4 of 5Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1deepseek-coder-33b-baseAverage69.01Unverified
2deepseek-coder-6.7b-baseAverage63.4Unverified
3starcoderbaseAverage55.54Unverified
4gpt-4-1106-previewAverage53.28Unverified
5CodeLlama-13b-hfAverage52.78Unverified
6deepseek-coder-1.3b-baseAverage52.63Unverified
7CodeLlama-34b-hfAverage49.66Unverified
8CodeLlama-7b-hfAverage45Unverified
9gpt-3.5-turbo-0301Average40.86Unverified
10incoder-6BAverage33.79Unverified
#ModelMetricClaimedVerifiedStatus
1CodeGPT-adaptedAccuracy (token-level)77.13Unverified
2CodeT5+ 770MEM (line-level)37.9Unverified
3CodeT5+ 220MEM (line-level)35.17Unverified
#ModelMetricClaimedVerifiedStatus
1CodeGPT-adaptedAccuracy (token-level)75.11Unverified
2CodeT5+ 770MEM (line-level)44.86Unverified
3CodeT5+ 220MEM (line-level)43.42Unverified
#ModelMetricClaimedVerifiedStatus
1SantaCoder-MGDCompilation Rate73.03Unverified
2SantaCoderCompilation Rate59.97Unverified
3SantaCoderCompilation Rate59.79Unverified
#ModelMetricClaimedVerifiedStatus
1RamboCompilation Rate76.47Unverified
2RepoCoderCompilation Rate74.02Unverified
#ModelMetricClaimedVerifiedStatus
1RamboCompilation Rate61.7Unverified
2RepoCoderCompilation Rate58.09Unverified