-
-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Truncation not explicitly mention #813
Comments
I also try to run a query face the same problem, but the system only shows "Setting |
I got the same message and the query takes forever... |
I find the problem is, this author build the program in serial, instead of parallel, while you compile run_localGPT, you can also monitor you CPU usage(by top, or htop instructions).
In my aspect, I only utilize 1~2 cpu cores to run the program, that’s the reason why it run so slow.
… On Jul 25, 2024, at 4:20 PM, KansaiTraining ***@***.***> wrote:
I got the same message and the query takes forever...
Any explanation of the error and if it has influence on the query results?
—
Reply to this email directly, view it on GitHub <#813 (comment)>, or unsubscribe <https://github.com/notifications/unsubscribe-auth/BJC5EIKMTLAJJ6K2KGHHCG3ZOCYMJAVCNFSM6AAAAABKDZOUTCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDENBZG42DGMJXGM>.
You are receiving this because you commented.
|
Same issue here... I also see my SSD reading a lot because of python 3.10, even after getting : Truncation was not explicitly activated but Has anyone found a solution? |
I get this error when i Try to run a query
Truncation was not explicitly activated but
max_length
is provided a specific value, please usetruncation=True
to explicitly truncate examples to max length. Defaulting to 'longest_first' truncation strategy. If you encode pairs of sequences (GLUE-style) with the tokenizer you can select this strategy more precisely by providing a specific strategy totruncation
.Setting
pad_token_id
toeos_token_id
:128001 for open-end generation.C:\Users\Tarun Sridhar.conda\envs\mummy\lib\site-packages\transformers\models\llama\modeling_llama.py:648: UserWarning: 1Torch was not compiled with flash attention. (Triggered internally at ..\aten\src\ATen\native\transformers\cuda\sdp_utils.cpp:455.)
attn_output = torch.nn.functional.scaled_dot_product_attention(
What can be possible fixes?
The text was updated successfully, but these errors were encountered: