Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

support for T4 #2620

Open
4 tasks
krishnanpooja opened this issue Dec 24, 2024 · 1 comment
Open
4 tasks

support for T4 #2620

krishnanpooja opened this issue Dec 24, 2024 · 1 comment
Labels
triaged Issue has been triaged by maintainers

Comments

@krishnanpooja
Copy link

System Info

T4 GPU
TensorRT-LLM: 0.15.0.dev2024111200

Who can help?

@ncomly-nvidia , @juney-nvidia , @byshiue

Information

  • The official example scripts
  • My own modified scripts

Tasks

  • An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
  • My own task or dataset (give details below)

Reproduction

Build models for tensorrt-llm engine on T4 GPU.

Expected behavior

Should work on T4 GPU

actual behavior

I am seeing warning message
[12/24/2024-10:21:22] [TRT-LLM] [W] Failed to infer cluster info for Tesla T4, treat it as a L40 node with 15 GB memory. This setting makes no effect if you do not use auto parallel.

How do I get it to recognize T4?

additional notes

The model works on A100. Both single and mGPU setup

@krishnanpooja krishnanpooja added the bug Something isn't working label Dec 24, 2024
@nv-guomingz
Copy link
Collaborator

Hi @krishnanpooja , T4 is no longer a support hardware since 0.14 release.

@nv-guomingz nv-guomingz added triaged Issue has been triaged by maintainers and removed bug Something isn't working labels Dec 24, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
triaged Issue has been triaged by maintainers
Projects
None yet
Development

No branches or pull requests

2 participants