Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sync #903

Merged
merged 26 commits into from
Jul 24, 2024
Merged

Sync #903

merged 26 commits into from
Jul 24, 2024

Conversation

sarthakpati
Copy link
Collaborator

Fixes #ISSUE_NUMBER

Proposed Changes

Checklist

  • CONTRIBUTING guide has been followed.
  • PR is based on the current GaNDLF master .
  • Non-breaking change (does not break existing functionality): provide as many details as possible for any breaking change.
  • Function/class source code documentation added/updated (ensure typing is used to provide type hints, including and not limited to using Optional if a variable has a pre-defined value).
  • Code has been blacked for style consistency and linting.
  • If applicable, version information has been updated in GANDLF/version.py.
  • If adding a git submodule, add to list of exceptions for black styling in pyproject.toml file.
  • Usage documentation has been updated, if appropriate.
  • Tests added or modified to cover the changes; if coverage is reduced, please give explanation.
  • If customized dependency installation is required (i.e., a separate pip install step is needed for PR to be functional), please ensure it is reflected in all the files that control the CI, namely: python-test.yml, and all docker files [1,2,3].

VukW and others added 26 commits May 9, 2024 01:03
1. return output of whole batch, not just one item
2. make ground truth & predictions array to take into account `q_samples_per_volume` (the whole dataset size during 1 epoch is equal to len(data) * q_samples_per_volume; so if dataset df contains 100 records and q_samples_per_volume = 10 (by default) and batch size is 4, there would be 250 batches by 4 elements
3. make ground truth take into account that train_dataloader is shuffled. So now ground truth is sorted in the same order as predictions and as train_dataloader.
To ensure values in csv are always written in the same order as header
was turned off as workaround at #870
fixes test_train_inference_classification_histology_large_2d (35)
(for one of classes per-label metrics are not counted thus metric shape may differ)
Corrected forward operations order
- the same metric error was occuring in the loops in forward_pass.py -
  now it is fixed
- entire epoch completes successfully
Implemented by Szymon Mazurek [email protected]
@sarthakpati sarthakpati requested a review from a team as a code owner July 24, 2024 14:55
Copy link
Contributor

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@sarthakpati sarthakpati reopened this Jul 24, 2024
@github-actions github-actions bot locked and limited conversation to collaborators Jul 24, 2024
@sarthakpati sarthakpati merged commit ba385e0 into older_apis Jul 24, 2024
37 of 46 checks passed
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants