You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: models_v2/pytorch/stable_diffusion/inference/cpu/README.md
+8-5
Original file line number
Diff line number
Diff line change
@@ -43,10 +43,6 @@ export DATASET_DIR=<directory where the dataset will be saved>
43
43
bash download_dataset.sh
44
44
```
45
45
46
-
### **NOTE**:Int8 model
47
-
48
-
Please get a quant_model.pt before run INT8-BF16 model or INT8-FP32 model. Please refer the [link](https://github.com/intel/intel-extension-for-transformers/blob/v1.5/examples/huggingface/pytorch/text-to-image/quantization/qat/README.md).
@@ -61,7 +57,6 @@ Please get a quant_model.pt before run INT8-BF16 model or INT8-FP32 model. Pleas
61
57
```
62
58
5. Install the latest CPU versions of [torch, torchvision and intel_extension_for_pytorch](https://intel.github.io/intel-extension-for-pytorch/index.html#installation)
63
59
64
-
65
60
6. Setup required environment paramaters
66
61
67
62
| **Parameter** | **export command** |
@@ -79,6 +74,14 @@ Please get a quant_model.pt before run INT8-BF16 model or INT8-FP32 model. Pleas
79
74
| **LOCAL_BATCH_SIZE** (optional for DISTRIBUTED) | `export LOCAL_BATCH_SIZE=64` |
80
75
7. Run `run_model.sh`
81
76
77
+
* NOTE:
78
+
Please get quantized model before running `INT8-BF16` or `INT8-FP32`.
79
+
For `ipex-jit` mode, please refer the [link](https://github.com/intel/intel-extension-for-transformers/blob/v1.5/examples/huggingface/pytorch/text-to-image/quantization/qat/README.md).
80
+
For `compile-inductor` mode, please do calibration first:
0 commit comments