#pragma once // @generated by torchgen/gen_executorch.py from NativeFunctions.h #ifdef TORCH_ASSERT_NO_OPERATORS #error This change adds a dependency on native_functions.yaml, \ meaning the file will need to be re-compiled every time an operator \ is changed or added. Consider if your change would be better placed in \ another file, or if a more specific header might achieve the same goal. \ See NOTE: [Tensor vs. TensorBase] #endif #if defined(AT_PER_OPERATOR_HEADERS) && defined(TORCH_ASSERT_ONLY_METHOD_OPERATORS) #error This change adds a dependency on all pytorch operators, meaning the \ file will need to be re-compiled every time an operator is changed or added. \ Consider including a specific operator from \ and see NOTE [TORCH_ASSERT_ONLY_METHOD_OPERATORS]. #endif #include #include #include #include #include #include #include #include #include #include namespace at { namespace native { TORCH_API at::Tensor & mkldnn_add_out(const at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha, at::Tensor & out); TORCH_API at::Tensor & baddbmm_out_xpu(const at::Tensor & self, const at::Tensor & batch1, const at::Tensor & batch2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out); TORCH_API at::Tensor & bmm_out_xpu(const at::Tensor & self, const at::Tensor & mat2, at::Tensor & out); TORCH_API at::Tensor & cat_out_quantized_cpu(const at::ITensorListRef & tensors, int64_t dim, at::Tensor & out); TORCH_API at::Tensor & clamp_out_mps(const at::Tensor & self, const ::std::optional & min, const ::std::optional & max, at::Tensor & out); TORCH_API at::Tensor & cumsum_out_mps(const at::Tensor & self, int64_t dim, ::std::optional dtype, at::Tensor & out); TORCH_API at::Tensor & div_out_mps(const at::Tensor & self, const at::Tensor & other, at::Tensor & out); TORCH_API at::Tensor & exp_out(const at::Tensor & self, at::Tensor & out); TORCH_API at::Tensor & floor_divide_out_mps(const at::Tensor & self, const at::Tensor & other, at::Tensor & out); TORCH_API at::Tensor & index_out(const at::Tensor & self, const c10::List<::std::optional> & indices, at::Tensor & out); TORCH_API at::Tensor & mean_out_quantized_cpu(const at::Tensor & self, at::OptionalIntArrayRef dim, bool keepdim, ::std::optional dtype, at::Tensor & out); TORCH_API at::Tensor & mm_out_xpu(const at::Tensor & self, const at::Tensor & mat2, at::Tensor & out); TORCH_API at::Tensor & mkldnn_mul_out(const at::Tensor & self, const at::Tensor & other, at::Tensor & out); TORCH_API ::std::tuple batch_norm_mps_out(const at::Tensor & input, const ::std::optional & weight, const ::std::optional & bias, const ::std::optional & running_mean, const ::std::optional & running_var, bool training, double momentum, double eps, at::Tensor & out, at::Tensor & save_mean, at::Tensor & save_invstd); TORCH_API at::Tensor & round_out(const at::Tensor & self, at::Tensor & out); TORCH_API at::Tensor & gelu_out_mps(const at::Tensor & self, c10::string_view approximate, at::Tensor & out); TORCH_API at::Tensor & sigmoid_out_mps(const at::Tensor & self, at::Tensor & out); TORCH_API at::Tensor & logit_out_mps(const at::Tensor & self, ::std::optional eps, at::Tensor & out); TORCH_API at::Tensor & softmax_mps_out(const at::Tensor & self, int64_t dim, bool half_to_float, at::Tensor & out); TORCH_API at::Tensor & stack_out(at::TensorList tensors, int64_t dim, at::Tensor & out); TORCH_API at::Tensor & sum_out_mps(const at::Tensor & self, at::OptionalIntArrayRef dim, bool keepdim, ::std::optional dtype, at::Tensor & out); TORCH_API at::Tensor & tanh_out(const at::Tensor & self, at::Tensor & out); TORCH_API at::Tensor & sub_out_mps(const at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha, at::Tensor & out); TORCH_API at::Tensor & addmm_out_xpu(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out); TORCH_API at::Tensor & bitwise_and_out_mps(const at::Tensor & self, const at::Tensor & other, at::Tensor & out); TORCH_API at::Tensor & ne_out_quantized_cpu(const at::Tensor & self, const at::Scalar & other, at::Tensor & out); TORCH_API at::Tensor & eq_out_quantized_cpu(const at::Tensor & self, const at::Scalar & other, at::Tensor & out); TORCH_API at::Tensor & eq_out_quantized_cpu(const at::Tensor & self, const at::Tensor & other, at::Tensor & out); TORCH_API at::Tensor & gt_out_quantized_cpu(const at::Tensor & self, const at::Scalar & other, at::Tensor & out); TORCH_API at::Tensor & index_select_out_cuda(const at::Tensor & self, int64_t dim, const at::Tensor & index, at::Tensor & out); TORCH_API at::Tensor & nonzero_out_mps(const at::Tensor & self, at::Tensor & out); TORCH_API at::Tensor & remainder_out(const at::Tensor & self, const at::Scalar & other, at::Tensor & out); TORCH_API at::Tensor & max_quantized_unary_out(const at::Tensor & self, at::Tensor & out); TORCH_API at::Tensor & minimum_out_mps(const at::Tensor & self, const at::Tensor & other, at::Tensor & out); TORCH_API ::std::tuple sort_out(const at::Tensor & self, int64_t dim, bool descending, at::Tensor & values, at::Tensor & indices); TORCH_API ::std::tuple topk_out_mps(const at::Tensor & self, int64_t k, int64_t dim, bool largest, bool sorted, at::Tensor & values, at::Tensor & indices); TORCH_API at::Tensor & leaky_relu_out_quantized_cpu(const at::Tensor & self, const at::Scalar & negative_slope, at::Tensor & out); TORCH_API at::Tensor & softplus_out_mps(const at::Tensor & self, const at::Scalar & beta, const at::Scalar & threshold, at::Tensor & out); TORCH_API at::Tensor & mkldnn_avg_pool2d_out(const at::Tensor & self, at::IntArrayRef kernel_size, at::IntArrayRef stride, at::IntArrayRef padding, bool ceil_mode, bool count_include_pad, ::std::optional divisor_override, at::Tensor & out); TORCH_API ::std::tuple max_pool2d_with_indices_out_mps(const at::Tensor & self, at::IntArrayRef kernel_size, at::IntArrayRef stride, at::IntArrayRef padding, at::IntArrayRef dilation, bool ceil_mode, at::Tensor & out, at::Tensor & indices); TORCH_API at::Tensor & upsample_nearest2d_out_mps(const at::Tensor & self, at::IntArrayRef output_size, ::std::optional scales_h, ::std::optional scales_w, at::Tensor & out); TORCH_API ::std::tuple linalg_inv_ex_out_mps(const at::Tensor & A, bool check_errors, at::Tensor & inverse, at::Tensor & info); TORCH_API void unbind_copy_int_out(const at::Tensor & self, int64_t dim, at::TensorList out); TORCH_API void split_copy_Tensor_out(const at::Tensor & self, int64_t split_size, int64_t dim, at::TensorList out); TORCH_API void split_with_sizes_copy_out_cuda(const at::Tensor & self, at::IntArrayRef split_sizes, int64_t dim, at::TensorList out); TORCH_API at::Tensor & constant_pad_nd_out_symint(const at::Tensor & self, c10::SymIntArrayRef pad, const at::Scalar & value, at::Tensor & out); TORCH_API at::Tensor & convolution_out_symint(const at::Tensor & input, const at::Tensor & weight, const ::std::optional & bias, c10::SymIntArrayRef stride, c10::SymIntArrayRef padding, c10::SymIntArrayRef dilation, bool transposed, c10::SymIntArrayRef output_padding, c10::SymInt groups, at::Tensor & out); TORCH_API at::Tensor & embedding_out_symint(const at::Tensor & weight, const at::Tensor & indices, c10::SymInt padding_idx, bool scale_grad_by_freq, bool sparse, at::Tensor & out); TORCH_API at::Tensor & grid_sampler_2d_out(const at::Tensor & input, const at::Tensor & grid, int64_t interpolation_mode, int64_t padding_mode, bool align_corners, at::Tensor & out); TORCH_API at::Tensor & index_put_out(const at::Tensor & self, const c10::List<::std::optional> & indices, const at::Tensor & values, bool accumulate, at::Tensor & out); TORCH_API ::std::tuple native_layer_norm_out_symint(const at::Tensor & input, c10::SymIntArrayRef normalized_shape, const ::std::optional & weight, const ::std::optional & bias, double eps, at::Tensor & out0, at::Tensor & out1, at::Tensor & out2); TORCH_API at::Tensor & pixel_shuffle_out(const at::Tensor & self, int64_t upscale_factor, at::Tensor & out); TORCH_API at::Tensor & repeat_out_symint(const at::Tensor & self, c10::SymIntArrayRef repeats, at::Tensor & out); TORCH_API at::Tensor & relu_out(const at::Tensor & self, at::Tensor & out); TORCH_API void unsafe_split_Tensor_out_symint(const at::Tensor & self, c10::SymInt split_size, int64_t dim, at::TensorList out); TORCH_API ::std::tuple _unique2_out(const at::Tensor & self, bool sorted, bool return_inverse, bool return_counts, at::Tensor & out0, at::Tensor & out1, at::Tensor & out2); TORCH_API at::Tensor & zeros_like_out(const at::Tensor & self, ::std::optional memory_format, at::Tensor & out); TORCH_API at::Tensor & clone_out(const at::Tensor & self, ::std::optional memory_format, at::Tensor & out); TORCH_API at::Tensor & rsub_Scalar_out(const at::Tensor & self, const at::Scalar & other, const at::Scalar & alpha, at::Tensor & out); TORCH_API ::std::tuple _fake_quantize_per_tensor_affine_cachemask_tensor_qparams_out(const at::Tensor & self, const at::Tensor & scale, const at::Tensor & zero_point, const at::Tensor & fake_quant_enabled, int64_t quant_min, int64_t quant_max, at::Tensor & out0, at::Tensor & out1); TORCH_API at::Tensor & _to_copy_out(const at::Tensor & self, bool non_blocking, ::std::optional memory_format, at::Tensor & out); TORCH_API at::Tensor & masked_fill_Scalar_out(const at::Tensor & self, const at::Tensor & mask, const at::Scalar & value, at::Tensor & out); TORCH_API at::Tensor & upsample_bilinear2d_vec_out_symint(const at::Tensor & input, at::OptionalSymIntArrayRef output_size, bool align_corners, ::std::optional> scale_factors, at::Tensor & out); TORCH_API at::Tensor & upsample_nearest2d_vec_out_symint(const at::Tensor & input, at::OptionalSymIntArrayRef output_size, ::std::optional> scale_factors, at::Tensor & out); TORCH_API at::Tensor & expand_copy_out_symint(const at::Tensor & self, c10::SymIntArrayRef size, bool implicit, at::Tensor & out); TORCH_API at::Tensor & permute_copy_out(const at::Tensor & self, at::IntArrayRef dims, at::Tensor & out); TORCH_API at::Tensor & _reshape_alias_copy_out_symint(const at::Tensor & self, c10::SymIntArrayRef size, c10::SymIntArrayRef stride, at::Tensor & out); TORCH_API at::Tensor & select_copy_int_out_symint(const at::Tensor & self, int64_t dim, c10::SymInt index, at::Tensor & out); TORCH_API at::Tensor & detach_copy_out(const at::Tensor & self, at::Tensor & out); TORCH_API at::Tensor & slice_copy_Tensor_out_symint(const at::Tensor & self, int64_t dim, ::std::optional start, ::std::optional end, c10::SymInt step, at::Tensor & out); TORCH_API at::Tensor & transpose_copy_int_out(const at::Tensor & self, int64_t dim0, int64_t dim1, at::Tensor & out); TORCH_API at::Tensor & unsqueeze_copy_out(const at::Tensor & self, int64_t dim, at::Tensor & out); TORCH_API at::Tensor & view_copy_out_symint(const at::Tensor & self, c10::SymIntArrayRef size, at::Tensor & out); } // namespace native } // namespace at namespace custom { namespace native { TORCH_API at::Tensor & add_3_out(const at::Tensor & a, const at::Tensor & b, const at::Tensor & c, at::Tensor & out); } // namespace native } // namespace custom