Skip to content

Commit

Permalink
Use np.prod instead of np.product (#7301)
Browse files Browse the repository at this point in the history
  • Loading branch information
JackCaoG authored Jun 17, 2024
1 parent f50e951 commit 5fa23c1
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
4 changes: 2 additions & 2 deletions test/spmd/test_sharding_strategies.py
Original file line number Diff line number Diff line change
Expand Up @@ -67,9 +67,9 @@

num_devices = xr.global_runtime_device_count()

assert np.product(dcn_parallelism) * np.product(
assert np.prod(dcn_parallelism) * np.prod(
ici_parallelism) == num_devices, f"Number of devices {num_devices} \
does not match the product of the parallelism {np.product(dcn_parallelism) * np.product(ici_parallelism)}"
does not match the product of the parallelism {np.prod(dcn_parallelism) * np.prod(ici_parallelism)}"

# Use HybridMesh to optimize multislice topology
mesh = xs.HybridMesh(
Expand Down
2 changes: 1 addition & 1 deletion torch_xla/distributed/spmd/xla_sharding.py
Original file line number Diff line number Diff line change
Expand Up @@ -268,7 +268,7 @@ def _create_device_mesh_for_nd_torus(
indices = itertools.combinations(
range(len(assignable_physical_mesh)), num_axes)
for c_axes, c_indices in zip(axes, indices):
if np.product(c_axes) == logical_axis_size:
if np.prod(c_axes) == logical_axis_size:
assignment[logical_axis_index] = c_indices
# Zero the assigned physical axes.
assignable_physical_mesh = [
Expand Down

0 comments on commit 5fa23c1

Please sign in to comment.