-
Notifications
You must be signed in to change notification settings - Fork 309
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to do inference without connecting to HuggingFace? #35
Labels
FAQ
Frequently asked question
Comments
@ForestsKing typically, using a HF model prefix should not have a significant overhead. However, if you're facing issues with your connection, you might try downloading the model first and loading from a local path. Here's how to do it:
import torch
from chronos import ChronosPipeline
pipeline = ChronosPipeline.from_pretrained(
"./checkpoints/chronos-t5-small",
device_map="cuda",
torch_dtype=torch.bfloat16,
) |
Thank! |
lostella
changed the title
Close the connection to Hugging Face before inference
How to do inference without connecting to Hugging Face
Mar 31, 2024
Leaving open as FAQ |
lostella
changed the title
How to do inference without connecting to Hugging Face
How to do inference without connecting to Hugging Face?
Mar 31, 2024
lostella
changed the title
How to do inference without connecting to Hugging Face?
How to do inference without connecting to HuggingFace?
May 21, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
The connection between my server and Hugging Face is not very smooth. I have downloaded the model weights. I would like to know if it is possible to close the connection to Hugging Face before calling Chronos, it often takes a lot of time and may fail. Thanks!
The text was updated successfully, but these errors were encountered: