Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

rapids to 23.06 #307

Merged
merged 7 commits into from
Jul 10, 2023
Merged

Conversation

eordentlich
Copy link
Collaborator

local (fast) tests pass, need to verify cloud.
Note that dataproc image had to be bumped to get python to > 3.8 which is dropped in rapids 23.06.
rapids-4-spark plugin 23.06 update pending push to maven.

@lijinf2
Copy link
Collaborator

lijinf2 commented Jun 24, 2023

Looks good overall.

Does dataproc work with RAPIDS-23.06? Only the RAPIDS version gets upgraded, and python version remains the same (python 3.8?).

docker/Dockerfile Outdated Show resolved Hide resolved
@wbo4958
Copy link
Collaborator

wbo4958 commented Jun 25, 2023

I'm wondering how you make this PR. seems there're lots of places need to be updated.

Signed-off-by: Erik Ordentlich <[email protected]>
Signed-off-by: Erik Ordentlich <[email protected]>
@eordentlich
Copy link
Collaborator Author

build

@eordentlich
Copy link
Collaborator Author

Verified that after patch-release updates of rapids pip packages, benchmarks run on databricks, dataproc and emr.
Dataproc required patching pyarrow in the conda environment, update to 2.1, and a minor edit to the spark-rapids install script which was added to our dataproc benchmark script directory.

@eordentlich
Copy link
Collaborator Author

build

1 similar comment
@pxLi
Copy link
Collaborator

pxLi commented Jul 10, 2023

build

@pxLi
Copy link
Collaborator

pxLi commented Jul 10, 2023

re-triggered, seems some internal networking issue

The TLS connection was non-properly terminated

@eordentlich
Copy link
Collaborator Author

build

@pxLi
Copy link
Collaborator

pxLi commented Jul 10, 2023

failed

[2023-07-10T04:43:03.414Z] �[0m�[91mTraceback (most recent call last):
[2023-07-10T04:43:03.414Z]   File "/root/miniconda3/bin/mamba", line 7, in <module>
[2023-07-10T04:43:03.414Z]     from mamba.mamba import main
[2023-07-10T04:43:03.414Z]   File "/root/miniconda3/lib/python3.8/site-packages/mamba/mamba.py", line 53, in <module>
[2023-07-10T04:43:03.414Z]     import mamba.mamba_api as api
[2023-07-10T04:43:03.414Z] ImportError: libarchive.so.13: cannot open shared object file: No such file or directory
[2023-07-10T04:43:03.414Z] �[0mThe command '/bin/sh -c conda install -c conda-forge mamba &&     mamba install -y -c rapidsai -c nvidia -c conda-forge cuml=$CUML_VER python=3.9 cuda-toolkit=11.5     && mamba clean --all -f -y' returned a non-zero code: 1

we saw this often in nightly docker images build, seems some bad timing about remote mamba cache, could try later

@pxLi
Copy link
Collaborator

pxLi commented Jul 10, 2023

build

2 similar comments
@eordentlich
Copy link
Collaborator Author

build

@pxLi
Copy link
Collaborator

pxLi commented Jul 10, 2023

build

Signed-off-by: Erik Ordentlich <[email protected]>
@eordentlich
Copy link
Collaborator Author

build

@lijinf2
Copy link
Collaborator

lijinf2 commented Jul 10, 2023

It seems have to copy third-party code in spark_rapids.sh. Will be wonderful if there is a way to avoid copy (can be addressed in the future). Overall looks good.

@eordentlich
Copy link
Collaborator Author

build

Copy link
Collaborator

@lijinf2 lijinf2 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for testing all of databricks, gcp dataproc and aws emr. Looks nice!

@eordentlich eordentlich merged commit 2ce27b5 into NVIDIA:branch-23.06 Jul 10, 2023
1 check passed
@eordentlich eordentlich deleted the eo_23.6_rapids branch July 10, 2023 21:37
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants