Skip to content

Commit

Permalink
[ci][train] Remove unnecessary xgboost_ray/lightgbm_ray reinstall…
Browse files Browse the repository at this point in the history
…s for release tests (#43176)

Pin `xgboost_ray` and `lightgbm_ray` versions to the current master commit in order to fix release tests that were still using the latest release.

---------

Signed-off-by: Justin Yu <[email protected]>
  • Loading branch information
justinvyu committed Feb 16, 2024
1 parent 2bd0fd4 commit 71d37ff
Show file tree
Hide file tree
Showing 11 changed files with 16 additions and 45 deletions.
2 changes: 2 additions & 0 deletions release/lightgbm_tests/workloads/tune_16x4.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@
as the Ray Tune results table. No trials should error. All trials should
run in parallel.
"""

from collections import Counter
import json
import os
Expand Down Expand Up @@ -58,6 +59,7 @@ def train_wrapper(config, ray_params):
config=search_space,
num_samples=16,
resources_per_trial=ray_params.get_tune_resources(),
storage_path="/mnt/cluster_storage",
)
taken = time.time() - start

Expand Down
2 changes: 2 additions & 0 deletions release/lightgbm_tests/workloads/tune_4x16.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@
as the Ray Tune results table. No trials should error. All trials should
run in parallel.
"""

from collections import Counter
import json
import os
Expand Down Expand Up @@ -58,6 +59,7 @@ def train_wrapper(config, ray_params):
config=search_space,
num_samples=4,
resources_per_trial=ray_params.get_tune_resources(),
storage_path="/mnt/cluster_storage",
)
taken = time.time() - start

Expand Down
2 changes: 2 additions & 0 deletions release/lightgbm_tests/workloads/tune_small.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@
as the Ray Tune results table. No trials should error. All trials should
run in parallel.
"""

from collections import Counter
import json
import os
Expand Down Expand Up @@ -58,6 +59,7 @@ def train_wrapper(config, ray_params):
config=search_space,
num_samples=4,
resources_per_trial=ray_params.get_tune_resources(),
storage_path="/mnt/cluster_storage",
)
taken = time.time() - start

Expand Down
2 changes: 1 addition & 1 deletion release/ray_release/byod/byod_xgboost_cuj_test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -4,4 +4,4 @@

set -exo pipefail

pip3 install -U --force-reinstall --no-deps xgboost_ray
pip install -U "git+https://github.com/ray-project/xgboost_ray@5a840af05d487171883dadbfdd37b138b607bed8#egg=xgboost_ray"
12 changes: 0 additions & 12 deletions release/ray_release/byod/byod_xgboost_master_test.sh

This file was deleted.

2 changes: 1 addition & 1 deletion release/ray_release/byod/byod_xgboost_test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@

set -exo pipefail

pip3 install -U --force-reinstall --no-deps xgboost_ray lightgbm_ray
pip3 install -U --force-reinstall --no-deps "git+https://github.com/ray-project/xgboost_ray@5a840af05d487171883dadbfdd37b138b607bed8#egg=xgboost_ray" "git+https://github.com/ray-project/lightgbm_ray@4c4d3413f86db769bddb6d08e2480a04bc75d712#egg=lightgbm_ray"
sudo mkdir -p /data || true
sudo chown ray:1000 /data || true
rm -rf /data/classification.parquet || true
Expand Down
2 changes: 1 addition & 1 deletion release/ray_release/byod/byod_xgboost_tune_test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@

set -exo pipefail

pip3 install -U --force-reinstall --no-deps xgboost_ray
pip3 install -U --force-reinstall --no-deps "git+https://github.com/ray-project/xgboost_ray@5a840af05d487171883dadbfdd37b138b607bed8#egg=xgboost_ray" "git+https://github.com/ray-project/lightgbm_ray@4c4d3413f86db769bddb6d08e2480a04bc75d712#egg=lightgbm_ray"
sudo mkdir -p /data || true
sudo chown ray:1000 /data || true
rm -rf /data/train.parquet || true
Expand Down
31 changes: 1 addition & 30 deletions release/release_tests.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -1647,35 +1647,6 @@

alert: default

- name: ml_user_xgboost_gpu_connect_latest
group: ML user tests
working_dir: ml_user_tests

frequency: nightly-3x
team: ml

cluster:
byod:
type: gpu
post_build_script: byod_xgboost_test.sh
cluster_compute: xgboost/tpl_gpu_small_scaling_aws.yaml

run:
timeout: 1200
script: python xgboost/train_gpu_connect.py
wait_for_nodes:
num_nodes: 5

variations:
- __suffix__: aws
- __suffix__: gce
env: gce
frequency: manual
cluster:
cluster_compute: xgboost/tpl_gpu_small_scaling_gce.yaml

alert: default

- name: ml_user_xgboost_gpu_connect_master
group: ML user tests
working_dir: ml_user_tests
Expand All @@ -1686,7 +1657,7 @@
cluster:
byod:
type: gpu
post_build_script: byod_xgboost_master_test.sh
post_build_script: byod_xgboost_test.sh
cluster_compute: xgboost/tpl_gpu_small_scaling_aws.yaml

run:
Expand Down
2 changes: 2 additions & 0 deletions release/xgboost_tests/workloads/tune_32x4.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@
as the Ray Tune results table. No trials should error. All trials should
run in parallel.
"""

from collections import Counter
import json
import os
Expand Down Expand Up @@ -58,6 +59,7 @@ def train_wrapper(config, ray_params):
config=search_space,
num_samples=32,
resources_per_trial=ray_params.get_tune_resources(),
storage_path="/mnt/cluster_storage",
)
taken = time.time() - start

Expand Down
2 changes: 2 additions & 0 deletions release/xgboost_tests/workloads/tune_4x32.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@
as the Ray Tune results table. No trials should error. All trials should
run in parallel.
"""

from collections import Counter
import json
import os
Expand Down Expand Up @@ -58,6 +59,7 @@ def train_wrapper(config, ray_params):
config=search_space,
num_samples=4,
resources_per_trial=ray_params.get_tune_resources(),
storage_path="/mnt/cluster_storage",
)
taken = time.time() - start

Expand Down
2 changes: 2 additions & 0 deletions release/xgboost_tests/workloads/tune_small.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@
as the Ray Tune results table. No trials should error. All trials should
run in parallel.
"""

from collections import Counter
import json
import os
Expand Down Expand Up @@ -58,6 +59,7 @@ def train_wrapper(config, ray_params):
config=search_space,
num_samples=4,
resources_per_trial=ray_params.get_tune_resources(),
storage_path="/mnt/cluster_storage",
)
taken = time.time() - start

Expand Down

0 comments on commit 71d37ff

Please sign in to comment.