1 | #pragma once |
2 | |
3 | // @generated by torchgen/gen.py from Function.h |
4 | |
5 | #include <ATen/Context.h> |
6 | #include <ATen/DeviceGuard.h> |
7 | #include <ATen/TensorUtils.h> |
8 | #include <ATen/TracerMode.h> |
9 | #include <ATen/core/Generator.h> |
10 | #include <ATen/core/Reduction.h> |
11 | #include <ATen/core/Tensor.h> |
12 | #include <c10/core/Scalar.h> |
13 | #include <c10/core/Storage.h> |
14 | #include <c10/core/TensorOptions.h> |
15 | #include <c10/util/Deprecated.h> |
16 | #include <c10/util/Optional.h> |
17 | |
18 | |
19 | |
20 | #include <ATen/ops/float_power_ops.h> |
21 | |
22 | namespace at { |
23 | |
24 | |
25 | // aten::float_power.Tensor_Tensor_out(Tensor self, Tensor exponent, *, Tensor(a!) out) -> Tensor(a!) |
26 | inline at::Tensor & float_power_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & exponent) { |
27 | return at::_ops::float_power_Tensor_Tensor_out::call(self, exponent, out); |
28 | } |
29 | // aten::float_power.Tensor_Tensor_out(Tensor self, Tensor exponent, *, Tensor(a!) out) -> Tensor(a!) |
30 | inline at::Tensor & float_power_outf(const at::Tensor & self, const at::Tensor & exponent, at::Tensor & out) { |
31 | return at::_ops::float_power_Tensor_Tensor_out::call(self, exponent, out); |
32 | } |
33 | |
34 | // aten::float_power.Tensor_Tensor(Tensor self, Tensor exponent) -> Tensor |
35 | inline at::Tensor float_power(const at::Tensor & self, const at::Tensor & exponent) { |
36 | return at::_ops::float_power_Tensor_Tensor::call(self, exponent); |
37 | } |
38 | |
39 | // aten::float_power.Scalar_out(Scalar self, Tensor exponent, *, Tensor(a!) out) -> Tensor(a!) |
40 | inline at::Tensor & float_power_out(at::Tensor & out, const at::Scalar & self, const at::Tensor & exponent) { |
41 | return at::_ops::float_power_Scalar_out::call(self, exponent, out); |
42 | } |
43 | // aten::float_power.Scalar_out(Scalar self, Tensor exponent, *, Tensor(a!) out) -> Tensor(a!) |
44 | inline at::Tensor & float_power_outf(const at::Scalar & self, const at::Tensor & exponent, at::Tensor & out) { |
45 | return at::_ops::float_power_Scalar_out::call(self, exponent, out); |
46 | } |
47 | |
48 | // aten::float_power.Scalar(Scalar self, Tensor exponent) -> Tensor |
49 | inline at::Tensor float_power(const at::Scalar & self, const at::Tensor & exponent) { |
50 | return at::_ops::float_power_Scalar::call(self, exponent); |
51 | } |
52 | |
53 | // aten::float_power.Tensor_Scalar_out(Tensor self, Scalar exponent, *, Tensor(a!) out) -> Tensor(a!) |
54 | inline at::Tensor & float_power_out(at::Tensor & out, const at::Tensor & self, const at::Scalar & exponent) { |
55 | return at::_ops::float_power_Tensor_Scalar_out::call(self, exponent, out); |
56 | } |
57 | // aten::float_power.Tensor_Scalar_out(Tensor self, Scalar exponent, *, Tensor(a!) out) -> Tensor(a!) |
58 | inline at::Tensor & float_power_outf(const at::Tensor & self, const at::Scalar & exponent, at::Tensor & out) { |
59 | return at::_ops::float_power_Tensor_Scalar_out::call(self, exponent, out); |
60 | } |
61 | |
62 | // aten::float_power.Tensor_Scalar(Tensor self, Scalar exponent) -> Tensor |
63 | inline at::Tensor float_power(const at::Tensor & self, const at::Scalar & exponent) { |
64 | return at::_ops::float_power_Tensor_Scalar::call(self, exponent); |
65 | } |
66 | |
67 | } |
68 | |