mirror of
https://github.com/zebrajr/pytorch.git
synced 2026-01-15 12:15:51 +00:00
Back out "Make PT2 compile backprop through custom op without autograd key a hard error (#166367)" (#168142)
Summary: Original commit changeset: 7148dc4803f5 Original Phabricator Diff: D86736500 Differential Revision: D87407335 Pull Request resolved: https://github.com/pytorch/pytorch/pull/168142 Approved by: https://github.com/wdvr
This commit is contained in:
committed by
PyTorch MergeBot
parent
a5436a5e8e
commit
5a607febc0
@@ -55,7 +55,6 @@ MUTABLE_OPS_THAT_CANNOT_GET_AN_OUT_VARIANT = [
|
||||
|
||||
# All of these operators don't have any tensor like returns
|
||||
FUNCTIONAL_OPS_THAT_CANNOT_GET_AN_OUT_VARIANT = [
|
||||
"_async_error",
|
||||
"_assert_async", # no return
|
||||
"_assert_async.msg", # no return
|
||||
"_assert_tensor_metadata", # no return
|
||||
|
||||
Reference in New Issue
Block a user