Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inconsistency in ttnn.reshape for single vs multi-device tensors #14767

Open
avoraTT opened this issue Nov 5, 2024 · 5 comments
Open

Inconsistency in ttnn.reshape for single vs multi-device tensors #14767

avoraTT opened this issue Nov 5, 2024 · 5 comments

Comments

@avoraTT
Copy link
Contributor

avoraTT commented Nov 5, 2024

Description

It seems that ttnn.reshape is inconsistent when being performed on single device vs multi-device. Specifically, the inconsistency can be seen when doing a [W, X, Y*Z] -> [W, X, Y, Z] reshape on a tensor that is on a single device, vs. a tensor that is replicated across devices.

For single device tensor, the outputs match torch, but for a multi-device tensor, they do not.

Repro steps

  • Machine: any T3K
  • Branch: avora/reshape_bug
  • Command: pytest models/demos/t3000/llama2_70b/tests/test_rope_reshape.py
@avoraTT
Copy link
Contributor Author

avoraTT commented Nov 5, 2024

cc @cglagovichTT

@cglagovichTT
Copy link
Contributor

FYI The models team has been having a few correctness issues with multidevice reshape/transpose. They are tough to repro, but now Ammar has found that unit tests may not be catching them because they exist on multidevice.

@ntarafdar
Copy link
Contributor

Thanks @avoraTT and @cglagovichTT will look at this early next week

@ntarafdar ntarafdar assigned jvegaTT and unassigned sjameelTT Nov 15, 2024
@uaydonat
Copy link
Contributor

any updates here?

@ntarafdar
Copy link
Contributor

The team is looking at fixing single device reshape this week still, that has gone overtime afterwards we can look at this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

6 participants