You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jan 15, 2024. It is now read-only.
sxjscience
changed the title
[TVM Integration] Support converting models with FP16 data type
[TVM Integration] Support TVM conversion with FP16 data type
Jan 29, 2021
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Description
We fixed TVM integration in GluonNLP recently for fp32 dtype. However, we still do not support fp16 dtype. We should
gluon-nlp/tests/test_models.py
Lines 65 to 77 in e8d4c8a
gluon-nlp/scripts/benchmarks/benchmark_utils.py
Lines 610 to 623 in e8d4c8a
The text was updated successfully, but these errors were encountered: