SOTAVerified

Code Completion

Papers

Showing 101150 of 212 papers

TitleStatusHype
Protect Your Secrets: Understanding and Measuring Data Exposure in VSCode Extensions0
LibEvolutionEval: A Benchmark and Study for Version-Specific Code Generation0
FastDraft: How to Train Your Draft0
CoCoP: Enhancing Text Classification with LLM through Code Completion Prompt0
JudgeRank: Leveraging Large Language Models for Reasoning-Intensive Reranking0
M2rc-Eval: Massively Multilingual Repository-level Code Completion Evaluation0
KV Prediction for Improved Time to First TokenCode0
CodeCipher: Learning to Obfuscate Source Code Against LLMs0
Horizon-Length Prediction: Advancing Fill-in-the-Middle Capabilities for Code Generation with Lookahead Planning0
Statically Contextualizing Large Language Models with Typed Holes0
MarsCode Agent: AI-native Automated Bug Fixing0
Retrieval-augmented code completion for local projects using large language models0
RepoMasterEval: Evaluating Code Completion via Real-World Repositories0
Black-Box Adversarial Attacks on LLM-Based Code Completion0
TaskEval: Assessing Difficulty of Code Generation Tasks for Large Language Models0
Evaluating Long Range Dependency Handling in Code Generation Models using Multi-Step Key Retrieval0
Curriculum Learning for Small Code Language Models0
TPIA: Towards Target-specific Prompt Injection Attack against Code-oriented Large Language Models0
Jailbreak Attacks and Defenses Against Large Language Models: A Survey0
Measuring memorization in RLHF for code completion0
CodeGemma: Open Code Models Based on Gemma0
Chain of Agents: Large Language Models Collaborating on Long-Context Tasks0
R2C2-Coder: Enhancing and Benchmarking Real-world Repository-level Code Completion Abilities of Code Large Language Models0
Model Cascading for Code: A Cascaded Black-Box Multi-Model Framework for Cost-Efficient Code Completion with Self-Testing0
A Transformer-Based Approach for Smart Invocation of Automatic Code CompletionCode0
Full Line Code Completion: Bringing AI to Desktop0
Does Your Neural Code Completion Model Use My Code? A Membership Inference ApproachCode0
Rethinking Software Engineering in the Foundation Model Era: From Task-Driven AI Copilots to Goal-Driven AI Pair Programmers0
Is Next Token Prediction Sufficient for GPT? Exploration on Code Logic Comprehension0
Stable Code Technical Report0
Investigating the Performance of Language Models for Completing Code in Functional Programming Languages: a Haskell Case StudyCode0
Repoformer: Selective Retrieval for Repository-Level Code Completion0
Token Alignment via Character Matching for Subword Completion0
Insights from the Usage of the Ansible Lightspeed Code Completion Service0
REPOFUSE: Repository-Level Code Completion with Fused Dual Context0
Context Composing for Full Line Code Completion0
Neural Models for Source Code Synthesis and Completion0
Do Large Code Models Understand Programming Concepts? Counterfactual Analysis for Code Predicates0
Enhancing LLM-Based Coding Tools through Native Integration of IDE-Derived Static Context0
OMPGPT: A Generative Pre-trained Transformer Model for OpenMP0
When Neural Code Completion Models Size up the Situation: Attaining Cheaper and Faster Completion through Dynamic Model InferenceCode0
Traces of Memorisation in Large Language Models for CodeCode0
A Review of Repository Level Prompting for LLMs0
Breaking the Silence: the Threats of Using LLMs in Software EngineeringCode0
INSPECT: Intrinsic and Systematic Probing Evaluation for Code TransformersCode0
Interpretability Illusions in the Generalization of Simplified Models0
GenCodeSearchNet: A Benchmark Test Suite for Evaluating Generalization in Programming Language UnderstandingCode0
Past as a Guide: Leveraging Retrospective Learning for Python Code Completion0
Identifying and Mitigating Vulnerabilities in LLM-Integrated Applications0
Attention Alignment and Flexible Positional Embeddings Improve Transformer Length Extrapolation0
Show:102550
← PrevPage 3 of 5Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1deepseek-coder-33b-baseAverage69.01Unverified
2deepseek-coder-6.7b-baseAverage63.4Unverified
3starcoderbaseAverage55.54Unverified
4gpt-4-1106-previewAverage53.28Unverified
5CodeLlama-13b-hfAverage52.78Unverified
6deepseek-coder-1.3b-baseAverage52.63Unverified
7CodeLlama-34b-hfAverage49.66Unverified
8CodeLlama-7b-hfAverage45Unverified
9gpt-3.5-turbo-0301Average40.86Unverified
10incoder-6BAverage33.79Unverified
#ModelMetricClaimedVerifiedStatus
1CodeGPT-adaptedAccuracy (token-level)77.13Unverified
2CodeT5+ 770MEM (line-level)37.9Unverified
3CodeT5+ 220MEM (line-level)35.17Unverified
#ModelMetricClaimedVerifiedStatus
1CodeGPT-adaptedAccuracy (token-level)75.11Unverified
2CodeT5+ 770MEM (line-level)44.86Unverified
3CodeT5+ 220MEM (line-level)43.42Unverified
#ModelMetricClaimedVerifiedStatus
1SantaCoder-MGDCompilation Rate73.03Unverified
2SantaCoderCompilation Rate59.97Unverified
3SantaCoderCompilation Rate59.79Unverified
#ModelMetricClaimedVerifiedStatus
1RamboCompilation Rate76.47Unverified
2RepoCoderCompilation Rate74.02Unverified
#ModelMetricClaimedVerifiedStatus
1RamboCompilation Rate61.7Unverified
2RepoCoderCompilation Rate58.09Unverified