Skip to content

Commit e2fbae0

Browse files
authored
add known issue of EmbeddingBag INT8 accuracy loss (#1017)
* add known issue restate the embedding issue * add release notes for 1.12.100
1 parent f092200 commit e2fbae0

File tree

2 files changed

+6
-0
lines changed

2 files changed

+6
-0
lines changed

docs/tutorials/performance_tuning/known_issues.md

+2
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,8 @@
11
Known Issues
22
============
33

4+
- Supporting of EmbeddingBag with INT8 when bag size > 1 is working in progress.
5+
46
- Compiling with gcc 11 might result in `illegal instruction` error.
57

68
- `RuntimeError: Overflow when unpacking long` when a tensor's min max value exceeds int range while performing int8 calibration. Please customize QConfig to use min-max calibration method.

docs/tutorials/releases.md

+4
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,10 @@
11
Releases
22
=============
33

4+
## 1.12.100
5+
6+
This is a patch release to fix the AVX2 issue that blocks running on non-AVX512 platforms.
7+
48
## 1.12.0
59

610
We are excited to bring you the release of Intel® Extension for PyTorch\* 1.12.0-cpu, by tightly following PyTorch [1.12](https://github.com/pytorch/pytorch/releases/tag/v1.12.0) release. In this release, we matured the automatic int8 quantization and made it a stable feature. We stabilized runtime extension and brought about a MultiStreamModule feature to further boost throughput in offline inference scenario. We also brought about various enhancements in operation and graph which are positive for performance of broad set of workloads.

0 commit comments

Comments
 (0)