SOTAVerified

CPU

Papers

Showing 651675 of 2231 papers

TitleStatusHype
A Simple Sparse Matrix Vector Multiplication Approach to Padded ConvolutionCode0
Improving Accuracy and Generalization for Efficient Visual Tracking0
An Integrated Artificial Intelligence Operating System for Advanced Low-Altitude Aviation Applications0
A Runtime-Adaptive Transformer Neural Network Accelerator on FPGAsCode0
KVPR: Efficient LLM Inference with I/O-Aware KV Cache Partial RecomputationCode0
A Data-Driven Approach to Dataflow-Aware Online Scheduling for Graph Neural Network Inference0
Plastic Arbor: a modern simulation framework for synaptic plasticity x2013 from single synapses to networks of morphological neuronsCode0
OPMOS: Ordered Parallel Algorithm for Multi-Objective Shortest-Paths0
SMM-Conv: Scalar Matrix Multiplication with Zero Packing for Accelerated Convolution0
Deep operator network models for predicting post-burn contraction0
MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs0
Generative AI on the Edge: Architecture and Performance Evaluation0
Llama Guard 3-1B-INT4: Compact and Efficient Safeguard for Human-AI Conversations0
Towards Accurate and Efficient Sub-8-Bit Integer Training0
Pie: Pooling CPU Memory for LLM Inference0
Offline Adaptation of Quadruped Locomotion using Diffusion ModelsCode0
Input-Based Ensemble-Learning Method for Dynamic Memory Configuration of Serverless Computing Functions0
TinyML Security: Exploring Vulnerabilities in Resource-Constrained Machine Learning Systems0
Project Tracyn: Generative Artificial Intelligence based Peripherals Trace Synthesizer0
P-MOSS: Learned Scheduling For Indexes Over NUMA Servers Using Low-Level Hardware Statistics0
DeepContext: A Context-aware, Cross-platform, and Cross-framework Tool for Performance Profiling and Analysis of Deep Learning Workloads0
Map++: Towards User-Participatory Visual SLAM Systems with Efficient Map Expansion and Sharing0
AI-Ready Energy Modelling for Next Generation RANCode0
NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM InferenceCode0
DynaSplit: A Hardware-Software Co-Design Framework for Energy-Aware Inference on Edge0
Show:102550
← PrevPage 27 of 90Next →

No leaderboard results yet.