Please update the NPU driver to the latest version to fully utilize the library features
PIP package: https://pypi.org/project/intel-npu-acceleration-library/1.3.0/
What's Changed
- Fix export error with trust_remote_code by @alessandropalla in #43
- Create warnings if driver is old by @alessandropalla in #46
- Fix int4 quantization for llama and gemma by @alessandropalla in #47
- Add C++ example by @alessandropalla in #54
- adding new operations by @SarahByrneIntel in #50
- Fix for NaNs in LLM inference by @alessandropalla in #58
- Change function generate_with_static_shape by @xduzhangjiayu in #60
- Native convolution and dw convolution by @alessandropalla in #61
- Sarah/feature/constant operation support by @SarahByrneIntel in #62
- Add memory operation and tensor class by @alessandropalla in #63
- Adding support for L2 normalisation operation by @SarahByrneIntel in #65
- Better torch integration by @alessandropalla in #66
- Add torch.nn.functional.conv2d by @alessandropalla in #70
- fix BatchNorm layer by @alessandropalla in #71
- Sarah/feature/operations by @SarahByrneIntel in #68
- Add torch NPU device by @alessandropalla in #72
- Automatic handling of output layers by @alessandropalla in #73
- Sarah/feature/reduce ops by @SarahByrneIntel in #74
- Hotfix for module by @alessandropalla in #76
- Fix SDPA in case attn_mask == None by @alessandropalla in #78
New Contributors
- @SarahByrneIntel made their first contribution in #50
- @xduzhangjiayu made their first contribution in #60
Full Changelog: v1.2.0...v1.3.0