1 | #pragma once |
2 | |
3 | // @generated by torchgen/gen.py from Function.h |
4 | |
5 | #include <ATen/Context.h> |
6 | #include <ATen/DeviceGuard.h> |
7 | #include <ATen/TensorUtils.h> |
8 | #include <ATen/TracerMode.h> |
9 | #include <ATen/core/Generator.h> |
10 | #include <ATen/core/Reduction.h> |
11 | #include <ATen/core/Tensor.h> |
12 | #include <c10/core/Scalar.h> |
13 | #include <c10/core/Storage.h> |
14 | #include <c10/core/TensorOptions.h> |
15 | #include <c10/util/Deprecated.h> |
16 | #include <c10/util/Optional.h> |
17 | |
18 | |
19 | |
20 | #include <ATen/ops/_make_per_tensor_quantized_tensor_ops.h> |
21 | |
22 | namespace at { |
23 | |
24 | |
25 | // aten::_make_per_tensor_quantized_tensor(Tensor self, float scale, int zero_point) -> Tensor |
26 | inline at::Tensor _make_per_tensor_quantized_tensor(const at::Tensor & self, double scale, int64_t zero_point) { |
27 | return at::_ops::_make_per_tensor_quantized_tensor::call(self, scale, zero_point); |
28 | } |
29 | |
30 | // aten::_make_per_tensor_quantized_tensor.out(Tensor self, float scale, int zero_point, *, Tensor(a!) out) -> Tensor(a!) |
31 | inline at::Tensor & _make_per_tensor_quantized_tensor_out(at::Tensor & out, const at::Tensor & self, double scale, int64_t zero_point) { |
32 | return at::_ops::_make_per_tensor_quantized_tensor_out::call(self, scale, zero_point, out); |
33 | } |
34 | // aten::_make_per_tensor_quantized_tensor.out(Tensor self, float scale, int zero_point, *, Tensor(a!) out) -> Tensor(a!) |
35 | inline at::Tensor & _make_per_tensor_quantized_tensor_outf(const at::Tensor & self, double scale, int64_t zero_point, at::Tensor & out) { |
36 | return at::_ops::_make_per_tensor_quantized_tensor_out::call(self, scale, zero_point, out); |
37 | } |
38 | |
39 | } |
40 | |