-
Notifications
You must be signed in to change notification settings - Fork 1
free memory for data load #30
Comments
Please provide description of issue. |
@huard description provided |
I also would like to use Ocgis calculation with chunks. Meanwhile I just use workaround, with ocgis as well, but not in sophisticated way... Define the size of the dataset (with simple function |
It is definitely time to address this. Let's be sure and work out a plan on the next call! Related to bird-house/bird-house.github.io#17. |
Listed as a "hot topic". |
the standalone is updated: @bekozi still a bug in ocgis, when calc=None its not possible to compute in chunks once the freememory and dataload comparison is running basically, we can integrate it properly in eggshell and add more complex cases (polygon subset etc ... ) |
Thanks for the reminder. I'll look into it this morning (NCPP/ocgis#402). |
@nilshempelmann I pushed changes to this |
@bekozi runing into this error: .... AttributeError: datatype conda ocgis-next was able to handle the datatype. |
Unless this is a blocker, let's hold off on debugging. An improved version of compute is almost ready, and it would be best to test against that. |
@bekozi is the improved version to calculate chunks ready? Could you provide a code snipppet? |
@nilshempelmann Could you provide more context for the |
standalone script moved to eggshell and is adopted |
@bekozi And there is a BIG difference in perfomance if data are performed directly or in chunks: operation performed with execute in 0.636506 sec. /tiles progress: [########################################] So swithing to |
There are cases, where the dataload for the memory is bigger the the available memory.
For that the data can't be processed in one block. Ocgis provides calculation in chunks (chunk size needs to be determined as well) to be switched on in case of potential memory overload.
Two parts of the problem:
existing function to be optimized: utils.FreeMemory()
There is a get_base_request_size and "large_array.compute" in ocgis on operations which should be revisited. There can be metadata on the calculations themselves used for defining the decompositions:
check dataload is tricky, when the process includes:
test standalone script is here:
https://github.com/bird-house/flyingpigeon/blob/next/scripts/ocgis_freememory.py
The text was updated successfully, but these errors were encountered: