torch.backends.fp32_precision setter propagate to cudnn.conv/rnn#179750
torch.backends.fp32_precision setter propagate to cudnn.conv/rnn#179750khushali9 wants to merge 8 commits intopytorch:mainfrom
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/179750
Note: Links to docs will display an error until the docs builds have been completed. ❗ 1 Active SEVsThere are 1 currently active SEVs. If your PR is affected, please view them below: ❌ 3 Awaiting Approval, 1 Cancelled Job, 1 Unrelated FailureAs of commit f11490c with merge base 781bb5c ( AWAITING APPROVAL - The following workflows need approval before CI can run:
CANCELLED JOB - The following job was cancelled. Please retry:
BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
|
@pytorchbot label "topic: not user facing" |
|
To add the ciflow label This helps ensure we don't trigger CI on this PR until it is actually authorized to do so. Please ping one of the reviewers if you do not have access to approve and run workflows. |
|
@pytorchmergebot label ciflow/h100 ciflow/b200 |
|
@eqy can you rerun the CI , I added another push to fix lint and it got canceled. |
|
@pytorchmergebot ciflow/h100 ciflow/b200 |
|
❌ 🤖 pytorchbot command failed: Try |
|
@pytorchmergebot label ciflow/h100 ciflow/b200 |
|
@eqy can you please review it, failing test is timeout and not due to my change. Thanks. |
|
@eqy PR is ready to be reviews, removed some more complexity around test and code. Thanks for pointing it out. |
|
@eqy PR is ready to be reviewed again. |
eqy
left a comment
There was a problem hiding this comment.
can you also check in the new test that the default for conv/rnn is tf32 before the decorator?
|
@eqy added test update to check for tf32. |
|
@eqy can we merge this. Thanks |
|
Can someone provide a tl;dr of what's going on with precision settings - I see constant issues about it, warnings warn about nonexistent things, settings don't work, but I don't have a full picture, and I don't understand which part of the full picture (if any) this PR is fixing |
@ngimel I can give you tldr, basically when we set torch.backends.fp32_precision =ieee, it should set for conv/rnn too , but they were not being set, only cunn, matmul were set automatically. |
d684e7e to
c4af9d8
Compare
|
@ngimel this PR is hopefully orthogonal to user complaints about warning spam with fp32 precision settings I'm trying to migrate some of the remaining uses over soon (tm) #180230 |
|
@pytorchmergebot rebase |
|
@pytorchbot started a rebase job onto refs/remotes/origin/viable/strict. Check the current status here |
|
Successfully rebased |
c4af9d8 to
f11490c
Compare
fixes 179445
issue#1
torch.backends.fp32_precision setter doesn't propagate to cudnn.conv/rnn, I have tried to invoke default to handle this.
issue#2 not resolved as that seems like legit to me. we need to error out if someone tries to use legacy and new API at the same time.
suggested workaround to @ydshieh
Let me know what you think.