Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How are multiple datasets loaded? #1330

Open
fxnie opened this issue Dec 17, 2024 · 4 comments
Open

How are multiple datasets loaded? #1330

fxnie opened this issue Dec 17, 2024 · 4 comments
Labels
feature request New feature or request

Comments

@fxnie
Copy link

fxnie commented Dec 17, 2024

local_setup.yml:
"data_path": "/workspace/mnt/cm-nfx/gpt-neox/data/mamba/mamba-2.8b/algorithmic_corpus/algorithmic_corpus_text_document,/workspace/mnt/cm-nfx/gpt-neox/data/mamba/mamba-2.8b/opencode_sft/opencode_sft_text_document,/workspace/mnt/cm-nfx/gpt-neox/data/mamba/mamba-2.8b/synthetic_code_snippet/synthetic_code_snippet_text_document,/workspace/mnt/cm-nfx/gpt-neox/data/mamba/mamba-2.8b/synthetic_qa/synthetic_qa_text_document",

Another question is how to reduce the ratio of the test set and validation set to 0 when setting the training set ratio?

@fxnie fxnie added the feature request New feature or request label Dec 17, 2024
@iPRET
Copy link

iPRET commented Dec 18, 2024

Not sure if this helps, cause I myself am a user, not a maintainer, but:
I think a nice alternative to doing the data splits is to just define the training, validation and test datasets explicitly. Like e.g.:

{
  "train_data_paths": ["/scratch/project_465001281/IP/tokenized/gpt_neox/en3_text_document", "/scratch/project_465001281/IP/tokenized/gpt_neox/lv3_text_document"],
  "valid_data_paths": ["/scratch/project_465001281/IP/tokenized/gpt_neox/en3_flores_text_document"],
  "test_data_paths": ["/scratch/project_465001281/IP/tokenized/gpt_neox/en3_flores_text_document"]
}

Also in the example above it shows how one can load two datasets in parallel ("train_data_paths").

@fxnie
Copy link
Author

fxnie commented Dec 19, 2024

How should I set up using only the training set and not the test and validation sets?

@iPRET
Copy link

iPRET commented Dec 19, 2024

A workaround could be to just generate some small dummy datasets, like e.g. from a .jsonl file with the line {"text": "eeeeeeeeee"}, and set eval_interval to a large number, if you really don't care about testing the model.

@fxnie
Copy link
Author

fxnie commented Dec 25, 2024

Is there a way to load only the data from these two files as described by you? "/cramch/project-465001281/IP/tokenized/gpt_neox/en3_text.dideo" and "/cramch/project-465001281/IP/tokenized/gpt_neox/lv3_text.dideo"

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants