-
Notifications
You must be signed in to change notification settings - Fork 51
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Main 2406 release #397
Main 2406 release #397
Conversation
Signed-off-by: Tim Liu <[email protected]>
[auto-merge] branch-24.04 to branch-24.06 [skip ci] [bot]
[auto-merge] branch-24.04 to branch-24.06 [skip ci] [bot]
[auto-merge] branch-24.04 to branch-24.06 [skip ci] [bot]
Signed-off-by: Bobby Wang <[email protected]>
After installing latest Jupyter notebooke, the json format check is strick, due to passing a string value "true" for the "use_gpu":"true", "use_gpu": True is required. ``` params = { "tree_method": "gpu_hist", "grow_policy": "depthwise", "num_workers": 1, "use_gpu": "true", } ``` TypeError: Invalid param value given for param "use_gpu". Boolean Param requires value of type bool. Found <class 'str'>. More over, "use_gpu": True is deprecated in dmlc/XGBoost, change to user "device": "cuda" instead Signed-off-by: Tim Liu <[email protected]>
Signed-off-by: Tim Liu <[email protected]>
* Update Databricks Tools Notebooks for 24.04.0 Signed-off-by: Partho Sarthi <[email protected]> * Add limitation for single event logs * Revert "Add limitation for single event logs" * Update Databricks Tools Notebooks for 24.06.0 Signed-off-by: Partho Sarthi <[email protected]> --------- Signed-off-by: Partho Sarthi <[email protected]>
Signed-off-by: liyuan <[email protected]>
# Conflicts: # examples/ML+DL-Examples/Spark-cuML/pca/README.md # examples/ML+DL-Examples/Spark-cuML/pca/pom.xml # examples/ML+DL-Examples/Spark-cuML/pca/spark-submit.sh # examples/UDF-Examples/RAPIDS-accelerated-UDFs/pom.xml # examples/UDF-Examples/Spark-cuSpatial/gpu-run.sh # examples/UDF-Examples/Spark-cuSpatial/notebooks/cuspatial_sample_standalone.ipynb # examples/UDF-Examples/Spark-cuSpatial/pom.xml # examples/XGBoost-Examples/mortgage/notebooks/python/cv-mortgage-gpu.ipynb # examples/XGBoost-Examples/mortgage/notebooks/scala/mortgage-ETL.ipynb
This is missing #388. |
Yes, I mean that commit. It's weird in that I can see the commit in the commits to be merged, yet when I diff this PR against branch-24.06 there are some missing changes that I thought were coming from #388. For example: $ git fetch https://github.com/nvliyuan/spark-rapids-examples.git main-2406-release
$ git diff branch-24.06 FETCH_HEAD
diff --git a/docs/get-started/xgboost-examples/csp/databricks/databricks.md b/docs/get-started/xgboost-examples/csp/databricks/databricks.md
index 1377a15..2d03013 100644
--- a/docs/get-started/xgboost-examples/csp/databricks/databricks.md
+++ b/docs/get-started/xgboost-examples/csp/databricks/databricks.md
@@ -21,8 +21,7 @@ Navigate to your home directory in the UI and select **Create** > **File** from
create an `init.sh` scripts with contents:
```bash
#!/bin/bash
- sudo wget -O /databricks/jars/rapids-4-spark_2.12-24.06.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/24.06.0/rapids-4-spark_2.12-24.06.0.jar
- ```
+ sudo wget -O /databricks/jars/rapids-4-spark_2.12-24.04.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/24.04.0/rapids-4-spark_2.12-24.04.0.jar
1. Select the Databricks Runtime Version from one of the supported runtimes specified in the
Prerequisites section.
2. Choose the number of workers that matches the number of GPUs you want to use.
[...] We can see from the diff above that we're rolling back changes in branch-24.06 as part of this merge of that branch to main. Not quite sure why these changes were lost despite #388 being part of the commit list in this PR (bad merge conflict resolution maybe?), but it needs to be fixed. You can do a diff like I show above to verify all of the changes between this PR and branch-24.06 make sense. |
Signed-off-by: liyuan <[email protected]>
Signed-off-by: liyuan <[email protected]>
Signed-off-by: liyuan <[email protected]>
Signed-off-by: liyuan <[email protected]>
Signed-off-by: liyuan <[email protected]>
Hi @jlowe, thanks for pointing it put, I just checked the diff details, basically there are three part:
part2: There are something wrong in branch-2406, since PCA demo only compatible with spark-rapids v23.04, cuspecial demo only compatible with spark-rapids v23.02, we miss this during the previous release, for example:
part3: it is ok to exist some snapshot version diff since we will remove all -snapshot in the main branch, for example:
For part1, I already push some commits to fix that |
Yes, snapshot diffs are expected. The other two need to be fixed. Nit: Given there were a number of merge mistakes on this PR, it would be cleaner to wait for #398 to be fixed and then re-merge branch-24.06 to main in a fresh PR. We're not squashing commits for this, and all these "fix" commits will appear in the mainline release. The mainline history ideally should have a minimal number of divergent commits from the dev branches. Not must-fix for me, just would be nice. |
close this pr and draft new pr #401 |
update main branch for v2406 release.
Please create merge commit not squash.