You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-8 (128g)? I tried 8,1 and 4,1 but they don't seem to work.
The text was updated successfully, but these errors were encountered:
joytianya
changed the title
Can the for 30B LLama model, server be supported by configuring mesh_dims on tpu v3-8 (128g)? I tried 8,1 and 4,1 but they don't seem to work.
for 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-8 (128g)? I tried 8,1 and 4,1 but they don't seem to work.
Apr 19, 2023
joytianya
changed the title
for 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-8 (128g)? I tried 8,1 and 4,1 but they don't seem to work.
for 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-256 (128g)? I tried 1, 64,4 and 1,32,8 but they don't seem to work.
Jun 16, 2023
joytianya
changed the title
for 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-256 (128g)? I tried 1, 64,4 and 1,32,8 but they don't seem to work.
For 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-8 (128g)? I tried 8,1 and 4,1 but they don't seem to work.
Jun 16, 2023
Any luck training the 30B on a single TPU v3-8 so far? Does it even fit? The 7B needs 84GB of VRAM, so I would expect the 30B to need at least 4 times that.
For 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-8 (128g)? I tried 8,1 and 4,1 but they don't seem to work.
The text was updated successfully, but these errors were encountered: