Model Compression
Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.
Papers
Showing 101–110 of 1356 papers
Benchmark Results
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | ADLIK-MO-ResNet50+W4A4 | Top-1 | 77.88 | — | Unverified |
| 2 | ADLIK-MO-ResNet50+W3A4 | Top-1 | 77.34 | — | Unverified |
| 3 | ResNet-18 + 4bit-1dim model compression using DKM | Top-1 | 70.52 | — | Unverified |
| 4 | MobileNet-v1 + 4bit-1dim model compression using DKM | Top-1 | 69.63 | — | Unverified |
| 5 | ResNet-18 + 2bit-1dim model compression using DKM | Top-1 | 68.63 | — | Unverified |
| 6 | MobileNet-v1 + 2bit-1dim model compression using DKM | Top-1 | 67.62 | — | Unverified |
| 7 | ResNet-18 + 4bit-4dim model compression using DKM | Top-1 | 66.1 | — | Unverified |
| 8 | ResNet-18 + 2bit-2dim model compression using DKM | Top-1 | 64.7 | — | Unverified |
| 9 | MobileNet-v1 + 4bit-4dim model compression using DKM | Top-1 | 61.4 | — | Unverified |
| 10 | ResNet-18 + 1bit-1dim model compression using DKM | Top-1 | 59.7 | — | Unverified |
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | MobileBERT + 2bit-1dim model compression using DKM | Accuracy | 82.13 | — | Unverified |
| 2 | MobileBERT + 1bit-1dim model compression using DKM | Accuracy | 63.17 | — | Unverified |