SOTAVerified

A Case Study of Web App Coding with OpenAI Reasoning Models

2024-09-19Code Available1· sign in to hype

Yi Cui

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper presents a case study of coding tasks by the latest reasoning models of OpenAI, i.e. o1-preview and o1-mini, in comparison with other frontier models. The o1 models deliver SOTA results for WebApp1K, a single-task benchmark. To this end, we introduce WebApp1K-Duo, a harder benchmark doubling number of tasks and test cases. The new benchmark causes the o1 model performances to decline significantly, falling behind Claude 3.5. Moreover, they consistently fail when confronted with atypical yet correct test cases, a trap non-reasoning models occasionally avoid. We hypothesize that the performance variability is due to instruction comprehension. Specifically, the reasoning mechanism boosts performance when all expectations are captured, meanwhile exacerbates errors when key expectations are missed, potentially impacted by input lengths. As such, we argue that the coding success of reasoning models hinges on the top-notch base model and SFT to ensure meticulous adherence to instructions.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
WebApp1k-Duo-Reactclaude-3-5-sonnetpass@10.68Unverified
WebApp1k-Duo-Reacto1-minipass@10.67Unverified
WebApp1k-Duo-Reacto1-previewpass@10.65Unverified
WebApp1k-Duo-Reactgpt-4o-2024-08-06pass@10.53Unverified
WebApp1k-Duo-Reactdeepseek-v2.5pass@10.49Unverified
WebApp1k-Duo-Reactmistral-large-2pass@10.45Unverified
WebApp1K-Reacto1-previewpass@10.95Unverified
WebApp1K-Reacto1-minipass@10.94Unverified
WebApp1K-Reactdeepseek-v2.5pass@10.83Unverified

Reproductions