jbilcke-hf HF Staff commited on
Commit
c81c2d8
·
1 Parent(s): 23edbfb

trying a quick fix

Browse files
docs/.DS_Store ADDED
Binary file (6.15 kB). View file
 
networks/op/fused_act.py CHANGED
@@ -19,7 +19,6 @@ else:
19
  ],
20
  )
21
 
22
-
23
  class FusedLeakyReLUFunctionBackward(Function):
24
  @staticmethod
25
  def forward(ctx, grad_output, out, bias, negative_slope, scale):
@@ -128,3 +127,5 @@ def fused_leaky_relu(input, bias=None, negative_slope=0.2, scale=2 ** 0.5):
128
  return FusedLeakyReLUFunction.apply(
129
  input.contiguous(), bias, negative_slope, scale
130
  )
 
 
 
19
  ],
20
  )
21
 
 
22
  class FusedLeakyReLUFunctionBackward(Function):
23
  @staticmethod
24
  def forward(ctx, grad_output, out, bias, negative_slope, scale):
 
127
  return FusedLeakyReLUFunction.apply(
128
  input.contiguous(), bias, negative_slope, scale
129
  )
130
+
131
+ torch.compiler.allow_in_graph(fused.fused_bias_act)
networks/op/upfirdn2d.py CHANGED
@@ -210,3 +210,5 @@ def upfirdn2d_native(
210
  out_w = (in_w * up_x + pad_x0 + pad_x1 - kernel_w + down_x) // down_x
211
 
212
  return out.view(-1, channel, out_h, out_w)
 
 
 
210
  out_w = (in_w * up_x + pad_x0 + pad_x1 - kernel_w + down_x) // down_x
211
 
212
  return out.view(-1, channel, out_h, out_w)
213
+
214
+ torch.compiler.allow_in_graph(upfirdn2d_op.upfirdn2d)