Anyone was able to run it succesfully? #168
Replies: 8 comments 27 replies
-
ran it on 8x a100 80g with the code in this repo(no modification, i just added a loop to get input from terminal), using 524GB of vram during single batch inference with nearly no context(10~100 tokens input), speed is only 7 tokens per second ish |
Beta Was this translation helpful? Give feedback.
-
Has anyone tried on Mac M3 with 128GB shared memory? |
Beta Was this translation helpful? Give feedback.
-
I created a pytorch implementation: https://github.com/DominiqueGarmier/grok-pytorch, you can't really use it at this point but maybe it would be a good starting point. |
Beta Was this translation helpful? Give feedback.
-
I finally was able to run it (though PyTorch version of it https://huggingface.co/hpcai-tech/grok-1 ) on Akash Network. |
Beta Was this translation helpful? Give feedback.
-
Is it consuming all 8 GPU when you ask a question? We got it working but
it seems super inefficient.
George
…On Fri, Mar 22, 2024 at 1:25 PM Andrey Arapov ***@***.***> wrote:
I finally was able to run it (though PyTorch version of it
https://huggingface.co/hpcai-tech/grok-1 ) on Akash Network.
Details akash-network/awesome-akash#507 (comment)
<akash-network/awesome-akash#507 (comment)>
image.png (view on web)
<https://github.com/xai-org/grok-1/assets/107317698/70909402-a77b-41cd-bcc6-21b9a571344a>
—
Reply to this email directly, view it on GitHub
<#168 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AR5KMMHZLTEJR5RPZGLIXXLYZRSQZAVCNFSM6AAAAABE4O2YYGVHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM4DQOBRGAZTE>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
I have made the model successfully working on 8 A100 80GB, but the response from model was not satisfactory. I am going to try to do second time and more properly set query. Instructions: #246 |
Beta Was this translation helpful? Give feedback.
-
https://www.fedml.ai/models/941?owner=xai-org Try here. I just use it for the completion task. The current Grok-1 does not support chat template. It is much better now. |
Beta Was this translation helpful? Give feedback.
-
Hello can me someone explain what is that all thanks
Odoslané z aplikácie Outlook pre Android<https://aka.ms/AAb9ysg>
…________________________________
From: iamdeepakgit ***@***.***>
Sent: Tuesday, June 25, 2024 6:06:09 AM
To: xai-org/grok-1 ***@***.***>
Cc: Subscribed ***@***.***>
Subject: Re: [xai-org/grok-1] Anyone was able to run it succesfully? (Discussion #168)
Yes it's running good and prefect 😉
—
Reply to this email directly, view it on GitHub<#168 (reply in thread)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/BITFOVUQYN7AOITQJKFDKWLZJDUDDAVCNFSM6AAAAABE4O2YYGVHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM4TQNRWG4ZTE>.
You are receiving this because you are subscribed to this thread.Message ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
I want to know what is look like when running? What you can do with it? Checkpoints are not finetuned or anything.. so what to do next with this code and trained data?
Beta Was this translation helpful? Give feedback.
All reactions