-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Parallel sharding #21
Commits on Apr 9, 2024
-
Configuration menu - View commit details
-
Copy full SHA for d75ba94 - Browse repository at this point
Copy the full SHA d75ba94View commit details -
feat: import transformer's gemma modeling code
It will be used to adapt it for sharding. Only imports have been adapted, and only code relevant for GemmaForCausalLM has been added.
Configuration menu - View commit details
-
Copy full SHA for 0ee7430 - Browse repository at this point
Copy the full SHA 0ee7430View commit details -
Configuration menu - View commit details
-
Copy full SHA for ca88068 - Browse repository at this point
Copy the full SHA ca88068View commit details -
Configuration menu - View commit details
-
Copy full SHA for a3de4d7 - Browse repository at this point
Copy the full SHA a3de4d7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 80170a9 - Browse repository at this point
Copy the full SHA 80170a9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9a9bcf8 - Browse repository at this point
Copy the full SHA 9a9bcf8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5bf6c70 - Browse repository at this point
Copy the full SHA 5bf6c70View commit details -
fix(TpuGemma): avoid using device_map when loading model
It seems that device_map parameter triggers a chain of calls that will try to use accelerate to load the model using less memory. The problem is that it skips the load state pre-hooks, making the weights loading impossible.
Configuration menu - View commit details
-
Copy full SHA for 9dfb7b6 - Browse repository at this point
Copy the full SHA 9dfb7b6View commit details -
It will now be running in parallel. More changes to come.
Configuration menu - View commit details
-
Copy full SHA for ec3b752 - Browse repository at this point
Copy the full SHA ec3b752View commit details -
Configuration menu - View commit details
-
Copy full SHA for a7d7c0b - Browse repository at this point
Copy the full SHA a7d7c0bView commit details -
Configuration menu - View commit details
-
Copy full SHA for b6fe32e - Browse repository at this point
Copy the full SHA b6fe32eView commit details -
Configuration menu - View commit details
-
Copy full SHA for e13d9ec - Browse repository at this point
Copy the full SHA e13d9ecView commit details -
Configuration menu - View commit details
-
Copy full SHA for 6cdede2 - Browse repository at this point
Copy the full SHA 6cdede2View commit details -
feat: model il loaded using pytorch_dtype from config
This will lead to loading the model in bfloat16 when specified in the config.
Configuration menu - View commit details
-
Copy full SHA for cd99226 - Browse repository at this point
Copy the full SHA cd99226View commit details -
Configuration menu - View commit details
-
Copy full SHA for 550e1fb - Browse repository at this point
Copy the full SHA 550e1fbView commit details -
Configuration menu - View commit details
-
Copy full SHA for 2215595 - Browse repository at this point
Copy the full SHA 2215595View commit details
Commits on Apr 10, 2024
-
Configuration menu - View commit details
-
Copy full SHA for fe888a9 - Browse repository at this point
Copy the full SHA fe888a9View commit details -
fix: get_generation_mode is now a method of generation_config
API change when transformers was updated.
Configuration menu - View commit details
-
Copy full SHA for dbf11f7 - Browse repository at this point
Copy the full SHA dbf11f7View commit details -
Configuration menu - View commit details
-
Copy full SHA for a96903b - Browse repository at this point
Copy the full SHA a96903bView commit details -
fix(generator): fix sample generation again
I wrongly chose the model's generation config instead of the one to the token selector.
Configuration menu - View commit details
-
Copy full SHA for 6e6b44e - Browse repository at this point
Copy the full SHA 6e6b44eView commit details -
fix: better handle torch_dtype
bfloat16 will be set by default in gemma models, other models will still load in float32 by default.
Configuration menu - View commit details
-
Copy full SHA for 92e9e31 - Browse repository at this point
Copy the full SHA 92e9e31View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7901d91 - Browse repository at this point
Copy the full SHA 7901d91View commit details