-
Notifications
You must be signed in to change notification settings - Fork 341
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
enumerate chunks for a given pattern #44
Comments
found a flaw in my theory: metadata chunks by truncating all chunks, I effectively broken the ability of duplicacy to operate instead, I am achieving my goal with unionfs-fuse & rclone mount (gcsfuse was not operating in a manner that allowed me to be successful) thanks for the great software, hopefully my evaluation proves successful and I can buy a few hundred licenses |
Actually, the truncation portion of my scheme may not be workable, but priming the pump may still be valuable for a discrete set of data (given by an include filter). That would make enumerating relevant chunks a useful function. |
I agree enumerating relevant chunks is useful, but I think you can just parse the output of the cat command to find relevant chunks. The output is the snapshot file in json format, so parsing it won't be too hard. For each file the 'content' field specifies the chunks referened by this file. For example:
means the content of this file starts at chunk 1333 offset 5630902 while ends at chunk 1336 offset 972724, so chunks 1333-1336 are needed for this file. |
I can work on a simple script to parse the snapshot file output from the cat command and print all relevant chunks. |
Hello,
I'm evaluating duplicacy and am narrowing on a particular mode of operation:
In this mode of operation, I can quickly download chunks via parallelism. When I have the chunks (and snapshot files) downloaded and available locally, I can recover quickly. This does not work, however, if I do not have sufficient local disk space to bring the entire collection of chunks down.
If duplicacy could enumerate chunks necessary for a specific restore pattern (say, all my cat pictures only), I could prime the restore by first obtaining only those chunks necessary for the restore prior to running the restore job. At present, I am only able to enumerate chunks for a given snapshot. It is feasible to try to cut my dataset into smaller logical repos, each independently backed up/snapshotted, such that a full restore could be performed repo by repo, but that is laborious.
So the feature request is to add chunk enumeration for a given restore pattern.
The text was updated successfully, but these errors were encountered: