Skip to content

History / inference_with_transformers_zh

Revisions

  • update inference_with_transformers

    @iMountTai iMountTai committed Dec 25, 2023
  • Add usage instruction of the parameter `--flash_attn`

    @GoGoJoestar GoGoJoestar committed Oct 25, 2023
  • add speculative sampling

    @airaria airaria committed Oct 8, 2023
  • Updated inference_with_transformers_zh (markdown)

    @airaria airaria committed Aug 8, 2023
  • Updated inference_with_transformers_zh (markdown)

    @airaria airaria committed Aug 3, 2023
  • Merge branch 'master' of https://github.com/ymcui/Chinese-LLaMA-Alpaca-2.wiki

    @ymcui ymcui committed Aug 2, 2023
  • prioritize full mode usage, fix style

    @ymcui ymcui committed Aug 2, 2023
  • reserve load_in_8bit

    @iMountTai iMountTai committed Aug 2, 2023
  • 修改load_in_8bit为load_in_kbit及部分说明信息

    @iMountTai iMountTai committed Aug 2, 2023
  • Updated inference_with_transformers_zh (markdown)

    @GoGoJoestar GoGoJoestar committed Aug 1, 2023
  • Updated inference_with_transformers_zh (markdown)

    @airaria airaria committed Aug 1, 2023
  • Updated inference_with_transformers_zh (markdown)

    @airaria airaria committed Aug 1, 2023
  • Updated inference_with_transformers_zh (markdown)

    @ymcui ymcui committed Jul 31, 2023
  • init

    @ymcui ymcui committed Jul 31, 2023