-
Notifications
You must be signed in to change notification settings - Fork 158
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Deployment to SageMaker and/or HuggingFace Inference Endpoints Fails With Error #94
Comments
unfortunatley we do not have sagemaker experts in our team. Could you check with AWS team for more details? Or share a scripts that can reproduce the error locally? |
Absolutely! I don't believe its a sagemaker issue, it seems like a lack of support for the custom config Here is a simple script that will immediately produce the issue when trying to load the model via the hugging face transformers library:
|
i think the problem is that we haven't tested with vqa-pipeline yet. Could you check with our offical inference impl? |
Even simpler example.
|
I copied what I needed from run_vila.py and it worked.
it should fix the llava_llama issue. |
When attempting to manually deploy the model to sagemaker via a deployment script or automatically deploying the model via the huggingface inference endpoints UI, I receive the same error:
"ValueError: The checkpoint you are trying to load has model type
llava_llama
but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date."The text was updated successfully, but these errors were encountered: