-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flows for moving back forth between cfs and hpss #59
Comments
I'm taking the time to think about this before jumping into the code since there are a few big pieces. HPSS
Considering the first point, maybe we should avoid moving individual scans to HPSS, and instead wait until after a beamtime/experiment is completed, and then bundle and send it using Either way, I can add a new class EDIT: It looks like it is possible to transfer directly from HPSS to a Globus endpoint. Upon further reading, it is not recommended to go this route for a few reasons. Instead we can use the SFAPI EDIT: Thinking about how to refactor Pruning SciCat
Then, for each beamline, we can define specific SciCat implementations with the types of metadata and derived datasets we want to track. |
…quires thorough testing. Includes a new transfer controller CFSToHPSSTransferController() with logic for handling single files vs directories using HPSS best practices. Moves create_sfapi_client() to the same level as transfer_controller.py, such that it can be easily accessed by multiple components. Includes new documentation in MkDocs for HPSS. Added an HPSS endpoint to config.yml. Updates orchestration/_tests/test_sfapi_flow.py to reflect the new location of create_sfapi_client().
Now that we have a working implementation of the sfapi in this repo, we can start looking at other things we can do with it at NERSC.
We are using the CFS for all strorage. But much of our data should be moved to tape. That system is called HPSS. Software at NERSC called
hsi
is the way of moving data between the two file systems.While NERSC has a globus DTN for HPSS, it doesn't work for our use case because we depend on the
alsdev
collab account as a type of service account for our jobs, and there is no concept of collab accounts in HPSS. All data put there would be granted permissions for our personal accounts.So, now that we have
sfapi
from prefect, we can start using HPSS.Some use cases:
New file created:
Prune job:
Recovery job
Assigning David to start thinking about.
The text was updated successfully, but these errors were encountered: