Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ban tensor constants in tracing. #1844

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

Ban tensor constants in tracing. #1844

wants to merge 1 commit into from

Conversation

zhxchen17
Copy link
Contributor

Summary:
There're two patterns resulting in the similar constant lifting effect during tracing:

def forward(self, x):
    return x + torch.tensor(...)

and

class Model(torch.nn.Module):
     def __init__(self):
         super().__init__()
         self.a = torch.tensor(...)
     def forward(self, x):
         return x + self.a

In theory and practice, it will be too implicit to promote them as buffers in tracing, due to two reasons:

  1. They don't have a formal FQN.
  2. They likely result in baked-in values without being noticed by users.
  3. Their values can be easily mutated without being noticed by users.

Therefore we want to make a policy decision to ban any form of tensor constants to show up in the traced graph. To proceed in the tracing, users either need to register these constants explicitly as buffers, or use constructors to properly initialize them. In some cases, users can get away without even wrapping scalars inside a torch.tensor() call.

Differential Revision: D53436364

Summary:
There're two patterns resulting in the similar constant lifting effect during tracing:
```
def forward(self, x):
    return x + torch.tensor(...)
```
and
```
class Model(torch.nn.Module):
     def __init__(self):
         super().__init__()
         self.a = torch.tensor(...)
     def forward(self, x):
         return x + self.a
```
In theory and practice, it will be too implicit to promote them as buffers in tracing, due to two reasons:

1. They don't have a formal FQN.
2. They likely result in baked-in values without being noticed by users.
3. Their values can be easily mutated without being noticed by users.

Therefore we want to make a policy decision to ban any form of tensor constants to show up in the traced graph. To proceed in the tracing, users either need to register these constants explicitly as buffers, or use constructors to properly initialize them. In some cases, users can get away without even wrapping scalars inside a torch.tensor() call.

Differential Revision: D53436364
Copy link

pytorch-bot bot commented Feb 5, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/1844

Note: Links to docs will display an error until the docs builds have been completed.

✅ You can merge normally! (8 Unrelated Failures)

As of commit 13c2f92 with merge base 2827c24 (image):

FLAKY - The following job failed but was likely due to flakiness present on trunk:

BROKEN TRUNK - The following jobs failed but was present on the merge base:

👉 Rebase onto the `viable/strict` branch to avoid these failures

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Feb 5, 2024
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D53436364

facebook-github-bot pushed a commit to pytorch/pytorch that referenced this pull request Feb 6, 2024
Summary:
X-link: pytorch/executorch#1844


There're two patterns resulting in the similar constant lifting effect during tracing:
```
def forward(self, x):
    return x + torch.tensor(...)
```
and
```
class Model(torch.nn.Module):
     def __init__(self):
         super().__init__()
         self.a = torch.tensor(...)
     def forward(self, x):
         return x + self.a
```
In theory and practice, it will be too implicit to promote them as buffers in tracing, due to two reasons:

1. They don't have a formal FQN.
2. They likely result in baked-in values without being noticed by users.
3. Their values can be easily mutated without being noticed by users.

Therefore we want to make a policy decision to ban any form of tensor constants to show up in the traced graph. To proceed in the tracing, users either need to register these constants explicitly as buffers, or use constructors to properly initialize them. In some cases, users can get away without even wrapping scalars inside a torch.tensor() call.

Test Plan: buck test mode/opt caffe2/test:test_export -- -r test_export_tensor_constants

Differential Revision: D53436364
facebook-github-bot pushed a commit to pytorch/pytorch that referenced this pull request Feb 6, 2024
Summary:
X-link: pytorch/executorch#1844


There're two patterns resulting in the similar constant lifting effect during tracing:
```
def forward(self, x):
    return x + torch.tensor(...)
```
and
```
class Model(torch.nn.Module):
     def __init__(self):
         super().__init__()
         self.a = torch.tensor(...)
     def forward(self, x):
         return x + self.a
```
In theory and practice, it will be too implicit to promote them as buffers in tracing, due to two reasons:

1. They don't have a formal FQN.
2. They likely result in baked-in values without being noticed by users.
3. Their values can be easily mutated without being noticed by users.

Therefore we want to make a policy decision to ban any form of tensor constants to show up in the traced graph. To proceed in the tracing, users either need to register these constants explicitly as buffers, or use constructors to properly initialize them. In some cases, users can get away without even wrapping scalars inside a torch.tensor() call.

Test Plan: buck test mode/opt caffe2/test:test_export -- -r test_export_tensor_constants

Differential Revision: D53436364
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants