diff --git a/README.md b/README.md
index 0d5078c..3b6adce 100644
--- a/README.md
+++ b/README.md
@@ -73,7 +73,7 @@ An original FP32 source model is quantized either using post-training quantizati
RetinaNet |
GitHub Repo |
Pretrained Model |
- See Example |
+ See Example |
(COCO) mAP FP32: 0.35 INT8: 0.349 Detailed Results |
RetinaNet.md |
1.15 |
@@ -250,7 +250,7 @@ An original FP32 source model is quantized either using post-training quantizati
Pytorch Torchvision |
Pytorch Torchvision |
Quantized Model |
- (ImageNet) Top-1 Accuracy FP32: 69.75% INT8: 69.54%
|
+ (ImageNet) Top-1 Accuracy FP32: 69.75% INT8: 69.54% INT4: 69.1%
|
Classification.md |
@@ -258,7 +258,7 @@ An original FP32 source model is quantized either using post-training quantizati
Pytorch Torchvision |
Pytorch Torchvision |
Quantized Model |
- (ImageNet) Top-1 Accuracy FP32: 76.14% INT8: 75.81%
|
+ (ImageNet) Top-1 Accuracy FP32: 76.14% INT8: 75.81% INT4: 75.63%
|
Classification.md |
@@ -266,30 +266,30 @@ An original FP32 source model is quantized either using post-training quantizati
Pytorch Torchvision |
Pytorch Torchvision |
Quantized Model |
- (ImageNet) Top-1 Accuracy FP32: 78.36% INT8: 78.10%
|
+ (ImageNet) Top-1 Accuracy FP32: 78.36% INT8: 78.10% INT4: 77.70%
|
Classification.md |
EfficientNet-lite0 |
GitHub Repo |
Pretrained Model |
- Quantized Model |
- (ImageNet) Top-1 Accuracy FP32: 75.40% INT8: 75.36% |
+ Quantized Model |
+ (ImageNet) Top-1 Accuracy FP32: 75.40% INT8: 75.36% INT4: 74.46% |
EfficientNet-lite0.md |
DeepLabV3+ |
GitHub Repo |
Pretrained Model |
- Quantized Model |
- (PascalVOC) mIOU FP32: 72.91% INT8: 72.44% |
+ Quantized Model |
+ (PascalVOC) mIOU FP32: 72.91% INT8: 72.44% INT4: 72.18% |
DeepLabV3.md |
MobileNetV2-SSD-Lite |
GitHub Repo |
Pretrained Model |
- Quantized Model |
+ Quantized Model |
(PascalVOC) mAP FP32: 68.7% INT8: 68.6% |
MobileNetV2-SSD-lite.md |
@@ -306,7 +306,7 @@ An original FP32 source model is quantized either using post-training quantizati
Based on Ref. |
FP32 Model |
Quantized Model |
- (COCO) mAP FP32: 0.765 INT8: 0.763 mAR FP32: 0.793 INT8: 0.792 |
+ (COCO) mAP FP32: 0.765 INT8: 0.763 INT4: 0.762 mAR FP32: 0.793 INT8: 0.792 INT4: 0.791 |
Hrnet-posenet.md |
SRGAN |
@@ -320,8 +320,8 @@ An original FP32 source model is quantized either using post-training quantizati
DeepSpeech2 |
GitHub Repo |
Pretrained Model |
- See Example |
- (Librispeech Test Clean) WER FP32 9.92% INT8: 10.22% |
+ See Example |
+ (Librispeech Test Clean) WER FP32: 9.92% INT8: 10.22% |
DeepSpeech2.md |
@@ -353,7 +353,7 @@ An original FP32 source model is quantized either using post-training quantizati
GitHub Repo |
Original model weight not available |
See Example |
- (Cityscapes) mIOU FP32 81.04% INT8: 80.78% |
+ (Cityscapes) mIOU FP32: 81.04% INT8: 80.65% INT4: 80.07% |
HRNet-w48.md |
@@ -361,7 +361,7 @@ An original FP32 source model is quantized either using post-training quantizati
GitHub Repo |
Pretrained Model |
See Example |
- (Cityscapes) mIOU FP32 77.81% INT8: 77.17% |
+ (Cityscapes) mIOU FP32: 77.81% INT8: 77.17% |
InverseForm.md |
@@ -369,9 +369,17 @@ An original FP32 source model is quantized either using post-training quantizati
GitHub Repo |
Pretrained Model |
See Example |
- (Cityscapes) mIOU FP32 86.31% INT8: 86.21% |
+ (Cityscapes) mIOU FP32: 86.31% INT8: 86.21% |
InverseForm.md |
+
+ FFNets |
+ Github Repo |
+ Prepared Models (5 in total) |
+ See Example |
+ (Cityscapes) mIOU segmentation_ffnet78S_dBBB_mobile FP32: 81.3% INT8: 80.7% segmentation_ffnet54S_dBBB_mobile FP32: 80.8% INT8: 80.1% segmentation_ffnet40S_dBBB_mobile FP32: 79.2% INT8: 78.9% segmentation_ffnet78S_BCC_mobile_pre_down FP32: 80.6% INT8: 80.4% segmentation_ffnet122NS_CCC_mobile_pre_down FP32: 79.3% INT8: 79.0% |
+ FFNet.md |
+
*[1]* Original FP32 model source
@@ -479,7 +487,7 @@ All results below used a *Scaling factor (LR-to-HR upscaling) of 2x* and the *Se
### Install AIMET
Before you can run the example script for a specific model, you need to install the AI Model Efficiency ToolKit (AIMET) software. Please see this [Getting Started](https://github.com/quic/aimet#getting-started) page for an overview. Then install AIMET and its dependencies using these [Installation instructions](https://github.com/quic/aimet/blob/develop/packaging/install.md).
-> **NOTE:** To obtain the exact version of AIMET software that was used to test this model zoo, please install release [1.13.0](https://github.com/quic/aimet/releases/tag/1.13.0) when following the above instructions *except where specified otherwise within the individual model documentation markdown file*.
+> **NOTE:** To obtain the exact version of AIMET software that was used to test this model zoo, please install release [1.22.2](https://github.com/quic/aimet/releases/tag/1.22.2) when following the above instructions *except where specified otherwise within the individual model documentation markdown file*.
### Running the scripts
Download the necessary datasets and code required to run the example for the model of interest. The examples run quantized evaluation and if necessary apply AIMET techniques to improve quantized model performance. They generate the final accuracy results noted in the table above. Refer to the Docs for [TensorFlow](zoo_tensorflow/Docs) or [PyTorch](zoo_torch/Docs) folder to access the documentation and procedures for a specific model.
diff --git a/zoo_tensorflow/Docs/SRGAN.md b/zoo_tensorflow/Docs/SRGAN.md
index a94a4d6..7227213 100644
--- a/zoo_tensorflow/Docs/SRGAN.md
+++ b/zoo_tensorflow/Docs/SRGAN.md
@@ -26,7 +26,7 @@ pip install tensorflow-gpu==2.4.0
## Model Weights
- The original SRGAN model is available at:
- - [krasserm](https://github.com/krasserm/super-resolution")
+ - [krasserm](https://github.com/krasserm/super-resolution)
## Usage
```bash
diff --git a/zoo_torch/Docs/Classification.md b/zoo_torch/Docs/Classification.md
index d3a9016..c49cb92 100644
--- a/zoo_torch/Docs/Classification.md
+++ b/zoo_torch/Docs/Classification.md
@@ -1,13 +1,9 @@
# PyTorch Classification models
This document describes evaluation of optimized checkpoints for Resnet18, Resnet50 and Regnet_x_3_2gf.
-## AIMET installation and setup
-Please [install and setup AIMET](https://github.com/quic/aimet/blob/release-aimet-1.21/packaging/install.md) (*Torch GPU* variant) before proceeding further.
-
-**NOTE**
-- All AIMET releases are available here: https://github.com/quic/aimet/releases
-- This model has been tested using AIMET version *1.21.0* (i.e. set `release_tag="1.21.0"` in the above instructions).
-- This model is compatible with the PyTorch GPU variant of AIMET (i.e. set `AIMET_VARIANT="torch_gpu"` in the above instructions).
+## Setup AI Model Efficiency Toolkit (AIMET)
+Please [install and setup AIMET](https://github.com/quic/aimet/blob/release-aimet-1.22/packaging/install.md) before proceeding further.
+This model was tested with the `torch_gpu` variant of AIMET 1.22.2.
## Additional Setup Dependencies
```
@@ -15,22 +11,41 @@ sudo -H pip install torchvision==0.11.2 --no-deps
sudo -H chmod 777 -R /dist-packages/*
```
-## Obtaining model checkpoint, ImageNet validation dataset and calibration dataset
-- [Pytorch Torchvision hub](https://pytorch.org/vision/0.11/models.html#classification) instances of Resnet18, Resnet50 and Regnet_x_3_2gf are used as refernce FP32 models. These instances are optimized using AIMET to obtain quantized optimized checkpoints.
-- Optimized Resnet18, Resnet50 and Regnet_x_3_2gf checkpoint can be downloaded from the [Releases](/../../releases) page.
-- ImageNet can be downloaded from here:
- - http://www.image-net.org/
-- Use standard validation set of ImageNet dataset (50k images set) for evaluting performance of FP32 and quantized models.
+## Obtain the Original Model for Comparison
+- [Pytorch Torchvision hub](https://pytorch.org/vision/0.11/models.html#classification) instances of Resnet18, Resnet50 and Regnet_x_3_2gf are used as reference FP32 models. These instances are optimized using AIMET to obtain quantized optimized checkpoints.
+
+## Experiment setup
+```python
+export PYTHONPATH=$PYTHONPATH:/aimet-model-zoo
+```
-For the quantization task, we require the model path, evaluation dataset path and calibration dataset path - which is a subset of validation dataset to be used for computing the encodings and AdaRound optimizaiton.
+## Dataset
+This evaluation was designed for the 2012 ImageNet Large Scale Visual Recognition Challenge (ILSVRC2012), which can be obtained from: http://www.image-net.org/
+The dataset directory is expected to have 3 subdirectories: train, valid, and test (only the valid test is used, hence if the other subdirectories are missing that is ok).
+Each of the {train, valid, test} directories is then expected to have 1000 subdirectories, each containing the images from the 1000 classes present in the ILSVRC2012 dataset, such as in the example below:
+
+```
+ train/
+ ├── n01440764
+ │ ├── n01440764_10026.JPEG
+ │ ├── n01440764_10027.JPEG
+ │ ├── ......
+ ├── ......
+ val/
+ ├── n01440764
+ │ ├── ILSVRC2012_val_00000293.JPEG
+ │ ├── ILSVRC2012_val_00002138.JPEG
+ │ ├── ......
+ ├── ......
+```
## Usage
-- To run evaluation with QuantSim in AIMET, use the following
+To run evaluation with QuantSim in AIMET, use the following
```bash
cd classification
python classification_quanteval.py\
--fp32-model \
- --default-param-bw \
+ --default-param-bw \
--default-output-bw