-
Notifications
You must be signed in to change notification settings - Fork 510
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Could not use cudf or cuml when rapids-runtime = DASK #1039
Comments
Thank you for this report! @nvliyuan do you want to take a look at this? |
Actually, nvliyuan has been contributing to the spark runtime. Not certain who to tap about the dask runtime. I'll check the commit history shortly and get back to you. |
Hey folks! I work on RAPIDS and Dask, happy to help. We are currently in the process of documenting and testing deploying RAPIDS on cloud platforms but I expect we will not get to Dataproc until after the holidays. But we will definitely dig into this as part of that work. Pinging @mroeschke who may have some quick thoughts about the Pandas error. I expect pandas needs upgrading/downgrading. |
I suspect your environment has pandas>=1.5 installed, and cudf was not compatible with that version of pandas until 22.10. Therefore if you downgrade pandas<1.5 or upgrade cudf>22.10 the error |
Thank you Jacob and Matt! @blis-teng - please let us know if this solves this issue for you so we can mark the issue resolved or otherwise offer an appropriate solution. |
@blis-teng - are you able to share the If you've got a support contract with GCP, I'd appreciate if you could open a support case and provide me the case #. By doing this, we can track our work and share case details privately rather than on the permanent record for the initialization-actions repository. Please do not open development cases as P2 or P1, as those are reserved for production outage situations, and development is by definition not a production environment. C.J. in Cloud Support, Seattle |
I have tried, but it will not work.
|
I used the cmd line from the given documentation in https://github.com/GoogleCloudDataproc/initialization-actions/blob/master/rapids/README.md
|
okay, I'll try to reproduce it now. |
With these arguments, it is installing pandas-1.2.5 and libcudf-22.04.00-cuda11. I think I found a bug in the rapids.sh script. I'll see if patching it improves the situation. |
In order to use 22.10 with pandas>=1.5, I need to upgrade these python packages: "cuspatial=${CUSPATIAL_VERSION}" "rope>=0.9.4" "gdal>3.5.0" And gdal>3.5.0 is not available in bullseye. Backports only go up to 3.2, so I'm going to try ubuntu20. |
|
so no, it looks like pandas >= 1.5 is not stable. I'll try doing the lower numbers. |
Now it looks like the python interpreter we install with dataproc is too new for the rapids release. I'll try 22.06 and 22.08 to see if either of those versions work. |
Okay, I was able to get this working on 2.0-debian10 with dask-rapids 22.06 I had to specify this mamba command: mamba install -n 'dask-rapids' -y --no-channel-priority -c 'conda-forge' -c 'nvidia' -c 'rapidsai' I'm testing the change with dask-rapids 22.08 ; if that works as well, I will submit a PR. |
@blis-teng - please try replacing the rapids.sh you link to from your project's initialization-actions checkout with this one. https://github.com/cjac/initialization-actions/raw/dask-rapids-202212/rapids/rapids.sh I am working with the product team to review this change. I should be able to close up PR #1041 pretty quick here. |
You may have mentioned that you have not yet read the README.md[1] from the initialization-actions repository. Can you please review and confirm for me that you understand where you would like to copy rapids.sh[2] from my pre-release branch for testing? [1] https://github.com/GoogleCloudDataproc/initialization-actions/blob/master/README.md#how-initialization-actions-are-used |
* Update to work with dask-rapids 10.06 Fix for issue #1039 * Incremental changes * 22.08 tested on 2.0-debian10 * added -m argument to mamba install ; previous test included dataproc:conda.env.config.uri which pre-defined the environment * tested with rapids version 22.10 * rapids works with rocky
@blis-teng can you re-try using the latest rapids/rapids.sh from github? |
hi, @cjac sorry for the late reply, I will re-try the new rapids.sh and get back to you next week, thanks! |
Thank you. Standing by for confirmation! I am presently not able to reproduce your problem. If there is still a change to be made, I'd like to know that information early in the week, please. |
Please remember to read the README I referenced. You are violating the guidance by using --initialization-actions gs://goog-dataproc-initialization-actions-${REGION}/gpu/install_gpu_driver.sh,gs://goog-dataproc-initialization-actions-${REGION}/rapids/rapids.sh \ |
Hi @cjac , could you please update to work with latest |
Not last I checked. What versions are you pinning to? |
I was wondering if you can upgrade |
I'm about to go on vacation, and I'm trying to put projects down. Can you open a new issue or better yet a GCP support case so I don't lose track of the work item, please? This issue is about the action not working. I think it's working now, but not patched up to latest release. A separate issue would be appropriate. |
I am trying to setup a dataproc cluster with GPU attached, to use cuml and cudf, I followed the instruction https://github.com/GoogleCloudDataproc/initialization-actions/blob/master/rapids/README.md
And able to setup the cluster, with nvidia driver successfully installed. But when I try
It throws out the error
I follow the instruction here: https://docs.rapids.ai/notices/rsn0020/
But after the downgraded version, another error show up when import cudf which is
The dask rapids installation version in rapids.sh is 22.04
The text was updated successfully, but these errors were encountered: