SOTAVerified

Text-to-Code Generation

Text-to-Code Generation is a task where we can generate code based on the natural language description.

Source: Text-to-code Generation with TensorFlow, πŸ€— & MBPP

Papers

Showing 11–20 of 20 papers

TitleStatusHype
SparsePO: Controlling Preference Alignment of LLMs via Sparse Token Masksβ€”0
Reranking Laws for Language Generation: A Communication-Theoretic Perspectiveβ€”0
Can OpenSource beat ChatGPT? -- A Comparative Study of Large Language Models for Text-to-Code Generationβ€”0
Generating Unseen Code Tests In Infinitumβ€”0
Text-to-Code Generation with Modality-relative Pre-trainingβ€”0
Fine-Tuning Large Language Models for Answering Programming Questions with Code Snippetsβ€”0
Code Execution with Pre-trained Language ModelsCode0
C3PO: A Lightweight Copying Mechanism for Translating Pseudocode to CodeCode0
PanGu-Coder: Program Synthesis with Function-Level Language ModelingCode0
Compilable Neural Code Generation with Compiler Feedbackβ€”0
Show:102550
← PrevPage 2 of 2Next β†’

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1CodeT5BLEU41.48β€”Unverified
2CodeGPT-adaptedBLEU32.79β€”Unverified