SOTAVerified

Code Completion

Papers

Showing 76100 of 212 papers

TitleStatusHype
Dataflow-Guided Retrieval Augmentation for Repository-Level Code CompletionCode1
Superposed Decoding: Multiple Generations from a Single Autoregressive Inference PassCode1
Model Cascading for Code: A Cascaded Black-Box Multi-Model Framework for Cost-Efficient Code Completion with Self-Testing0
Optimizing Large Language Models for OpenAPI Code CompletionCode2
AutoCoder: Enhancing Code Large Language Model with AIEV-InstructCode4
A Transformer-Based Approach for Smart Invocation of Automatic Code CompletionCode0
Full Line Code Completion: Bringing AI to Desktop0
Does Your Neural Code Completion Model Use My Code? A Membership Inference ApproachCode0
Rethinking Software Engineering in the Foundation Model Era: From Task-Driven AI Copilots to Goal-Driven AI Pair Programmers0
Is Next Token Prediction Sufficient for GPT? Exploration on Code Logic Comprehension0
Stable Code Technical Report0
PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language ModelsCode3
Investigating the Performance of Language Models for Completing Code in Functional Programming Languages: a Haskell Case StudyCode0
Repoformer: Selective Retrieval for Repository-Level Code Completion0
Token Alignment via Character Matching for Subword Completion0
CodeAttack: Revealing Safety Generalization Challenges of Large Language Models via Code CompletionCode2
RepoHyper: Search-Expand-Refine on Semantic Graphs for Repository-Level Code CompletionCode2
Evaluation of LLMs on Syntax-Aware Code Fill-in-the-Middle TasksCode1
IRCoder: Intermediate Representations Make Language Models Robust Multilingual Code GeneratorsCode1
StarCoder 2 and The Stack v2: The Next GenerationCode7
Insights from the Usage of the Ansible Lightspeed Code Completion Service0
CodeChameleon: Personalized Encryption Framework for Jailbreaking Large Language ModelsCode1
Language Models for Code Completion: A Practical EvaluationCode1
REPOFUSE: Repository-Level Code Completion with Fused Dual Context0
Context Composing for Full Line Code Completion0
Show:102550
← PrevPage 4 of 9Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1deepseek-coder-33b-baseAverage69.01Unverified
2deepseek-coder-6.7b-baseAverage63.4Unverified
3starcoderbaseAverage55.54Unverified
4gpt-4-1106-previewAverage53.28Unverified
5CodeLlama-13b-hfAverage52.78Unverified
6deepseek-coder-1.3b-baseAverage52.63Unverified
7CodeLlama-34b-hfAverage49.66Unverified
8CodeLlama-7b-hfAverage45Unverified
9gpt-3.5-turbo-0301Average40.86Unverified
10incoder-6BAverage33.79Unverified
#ModelMetricClaimedVerifiedStatus
1CodeGPT-adaptedAccuracy (token-level)77.13Unverified
2CodeT5+ 770MEM (line-level)37.9Unverified
3CodeT5+ 220MEM (line-level)35.17Unverified
#ModelMetricClaimedVerifiedStatus
1CodeGPT-adaptedAccuracy (token-level)75.11Unverified
2CodeT5+ 770MEM (line-level)44.86Unverified
3CodeT5+ 220MEM (line-level)43.42Unverified
#ModelMetricClaimedVerifiedStatus
1SantaCoder-MGDCompilation Rate73.03Unverified
2SantaCoderCompilation Rate59.97Unverified
3SantaCoderCompilation Rate59.79Unverified
#ModelMetricClaimedVerifiedStatus
1RamboCompilation Rate76.47Unverified
2RepoCoderCompilation Rate74.02Unverified
#ModelMetricClaimedVerifiedStatus
1RamboCompilation Rate61.7Unverified
2RepoCoderCompilation Rate58.09Unverified