1 | #pragma once |
2 | |
3 | // @generated by torchgen/gen.py from Function.h |
4 | |
5 | #include <ATen/Context.h> |
6 | #include <ATen/DeviceGuard.h> |
7 | #include <ATen/TensorUtils.h> |
8 | #include <ATen/TracerMode.h> |
9 | #include <ATen/core/Generator.h> |
10 | #include <ATen/core/Reduction.h> |
11 | #include <ATen/core/Tensor.h> |
12 | #include <c10/core/Scalar.h> |
13 | #include <c10/core/Storage.h> |
14 | #include <c10/core/TensorOptions.h> |
15 | #include <c10/util/Deprecated.h> |
16 | #include <c10/util/Optional.h> |
17 | |
18 | |
19 | |
20 | #include <ATen/ops/sigmoid_backward_ops.h> |
21 | |
22 | namespace at { |
23 | |
24 | |
25 | // aten::sigmoid_backward.grad_input(Tensor grad_output, Tensor output, *, Tensor(a!) grad_input) -> Tensor(a!) |
26 | inline at::Tensor & sigmoid_backward_out(at::Tensor & grad_input, const at::Tensor & grad_output, const at::Tensor & output) { |
27 | return at::_ops::sigmoid_backward_grad_input::call(grad_output, output, grad_input); |
28 | } |
29 | // aten::sigmoid_backward.grad_input(Tensor grad_output, Tensor output, *, Tensor(a!) grad_input) -> Tensor(a!) |
30 | inline at::Tensor & sigmoid_backward_outf(const at::Tensor & grad_output, const at::Tensor & output, at::Tensor & grad_input) { |
31 | return at::_ops::sigmoid_backward_grad_input::call(grad_output, output, grad_input); |
32 | } |
33 | |
34 | // aten::sigmoid_backward(Tensor grad_output, Tensor output) -> Tensor |
35 | inline at::Tensor sigmoid_backward(const at::Tensor & grad_output, const at::Tensor & output) { |
36 | return at::_ops::sigmoid_backward::call(grad_output, output); |
37 | } |
38 | |
39 | } |
40 | |