Skip to content

[pull] master from ray-project:master#4065

Merged
pull[bot] merged 4 commits intomiqdigital:masterfrom
ray-project:master
Apr 18, 2026
Merged

[pull] master from ray-project:master#4065
pull[bot] merged 4 commits intomiqdigital:masterfrom
ray-project:master

Conversation

@pull
Copy link
Copy Markdown

@pull pull bot commented Apr 18, 2026

See Commits and Changes for more details.


Created by pull[bot] (v2.0.0-alpha.4)

Can you help keep this open source service alive? 💖 Please sponsor : )

kyuds and others added 4 commits April 17, 2026 19:21
## Description
Necessary clean up before removing ExecutionPlan copy and deepcopy.

## Related issues
#60358

## Additional information
N/A

---------

Signed-off-by: Daniel Shin <kyuseung1016@gmail.com>
…62578)

## Description

Adds a Cursor Bugbot rule to enforce the separation between unit tests
(`python/ray/data/tests/unit/`) and integration tests
(`python/ray/data/tests/`).

The rule flags misplaced tests during PR review:
- Tests in `unit/` that call `ray.*` APIs at runtime, use
cluster-starting fixtures, or use `time.sleep()` → flagged as
integration tests that should be moved out.
- Tests in top-level files that make no runtime `ray.*` calls →
suggested to move into `unit/`.

This is a lighter-weight alternative to the pytest fixture/hook approach
explored in #61505.

## Related issues

Related to #61505

## Additional information

Doc-only change — only modifies `src/ray/.cursor/BUGBOT.md`.

---------

Signed-off-by: Hyunoh-Yeo <hyunoh.yeo@gmail.com>
…gies (#62526)

## Description
Follow up to #62306 per the
discussion
[here](#62306 (comment)).

This change introduces a `chips_per_vm` override parameter to resolve
ambiguity in TPU slice shapes (like v6e 2x4) where a topology can be
provisioned as either a single 8-chip VM or multiple 4-chip VMs. This is
an edge case for [v6e TPU
topologies](https://docs.cloud.google.com/tpu/docs/v6e#configurations),
but it would be useful to add an optional parameter to support it so
that users can request their desired topology without mismatches in the
number of bundles created by the `SlicePlacementGroup`.

What we changed
- Adjusted `get_chips_per_host` defaults to align with both v6e and v5e
machine types. If # of chips is <= 8 with a sub or single-host topology,
we should always just return the chip count. For chips > 8, these are
multi-host configurations which are always 4 chips per host/VM.
- Added a `chips_per_vm` parameter to `SlicePlacementGroup`. If
provided, `chips_per_vm` is used to calculate the number of bundles to
provision for the PG. Otherwise, we default to the expected number of
hosts based on the topology.

## Related issues


## Additional information
> Optional: Add implementation details, API changes, usage examples,
screenshots, etc.

---------

Signed-off-by: ryanaoleary <ryanaoleary@google.com>
Signed-off-by: Ryan O'Leary <113500783+ryanaoleary@users.noreply.github.com>
After #62492 we no longer set CUDA_VISIBLE_DEVIES ="" when num_gpus=0 or
not set. Torch if it detects that CUDA_VISIBLE_DEVIES ="" throws a
runtime error, however now that CUDA_VISIBLE_DEVIES is not set at all it
falls back to the nvidia driver to get the device ids. Following up on
#62653 and instead checking for the default cuda:0 gpu id in these
tests.

---------

Signed-off-by: Joshua Lee <joshlee@anyscale.com>
@pull pull bot locked and limited conversation to collaborators Apr 18, 2026
@pull pull bot added the ⤵️ pull label Apr 18, 2026
@pull pull bot merged commit 64385a5 into miqdigital:master Apr 18, 2026
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants