// required for old g++ to compile PRId64 macros, see
// https://github.com/pytorch/pytorch/issues/3571
// for context
#ifndef __STDC_FORMAT_MACROS
#define __STDC_FORMAT_MACROS
#endif

// an external backend might generate file within its code tree
// and check all the source files within the tree with clang-format.
// so, disable it since the backend might have a different config.
// clang-format off

// NOTE: This condition is true for all PyTorch internal libraries, it
//       just excludes external projects such as torch_xla which
//       re-use some of the PyTorch codegen machinery.
#if defined(CAFFE2_BUILD_MAIN_LIB)        || \
    defined(TORCH_CUDA_BUILD_MAIN_LIB)    || \
    defined(TORCH_HIP_BUILD_MAIN_LIB)     || \
    defined(TORCH_XPU_BUILD_MAIN_LIB)     || \
    defined(TORCH_CUDA_CU_BUILD_MAIN_LIB) || \
    defined(TORCH_CUDA_CPP_BUILD_MAIN_LIB)
#define TORCH_ASSERT_ONLY_METHOD_OPERATORS
#endif

// @generated by torchgen/gen.py from RegisterDispatchKey.cpp

#include <c10/core/TensorImpl.h>
#include <c10/core/Allocator.h>
#include <ATen/DeviceGuard.h>
#include <ATen/NamedTensorUtils.h>
#include <ATen/Utils.h>
#include <ATen/WrapDimUtils.h>
#include <ATen/Dispatch.h>
#include <c10/util/ExclusivelyOwned.h>
#include <c10/util/Half.h>
#include <c10/core/UndefinedTensorImpl.h>
#include <optional>
#include <ATen/Tensor.h>
#include <ATen/native/Resize.h>

#include <cstddef>
#include <functional>
#include <memory>
#include <utility>

#include <ATen/Config.h>
#include <ATen/core/op_registration/adaption.h>
#include <torch/library.h>
#include <c10/cuda/CUDAGuard.h>
#include <ATen/cuda/ATenCUDAGeneral.h>
#include <ATen/cuda/CUDADevice.h>
#include <ATen/cuda/CUDAContext.h>

#include <ATen/ops/as_strided_native.h>
#include <ATen/ops/empty.h>
#include <ATen/ops/empty_strided.h>
#include <ATen/ops/_copy_from_and_resize.h>
#include <ATen/ops/_copy_from.h>
#include <c10/macros/Macros.h>
#include <ATen/ops/_conj_physical_native.h>
#include <ATen/ops/_nnz_native.h>
#include <ATen/ops/_sparse_csr_prod_native.h>
#include <ATen/ops/_sparse_csr_sum_native.h>
#include <ATen/ops/_spsolve_native.h>
#include <ATen/ops/_to_dense_native.h>
#include <ATen/ops/_to_sparse_bsc_native.h>
#include <ATen/ops/_to_sparse_bsr_native.h>
#include <ATen/ops/_to_sparse_csc_native.h>
#include <ATen/ops/_to_sparse_csr_native.h>
#include <ATen/ops/_to_sparse_native.h>
#include <ATen/ops/abs_native.h>
#include <ATen/ops/add_native.h>
#include <ATen/ops/addmm_native.h>
#include <ATen/ops/addmv_native.h>
#include <ATen/ops/angle_native.h>
#include <ATen/ops/asin_native.h>
#include <ATen/ops/asinh_native.h>
#include <ATen/ops/atan_native.h>
#include <ATen/ops/atanh_native.h>
#include <ATen/ops/baddbmm_native.h>
#include <ATen/ops/bmm_native.h>
#include <ATen/ops/ccol_indices_native.h>
#include <ATen/ops/ceil_native.h>
#include <ATen/ops/clone_native.h>
#include <ATen/ops/col_indices_native.h>
#include <ATen/ops/conj_physical_native.h>
#include <ATen/ops/copy_native.h>
#include <ATen/ops/crow_indices_native.h>
#include <ATen/ops/deg2rad_native.h>
#include <ATen/ops/dense_dim_native.h>
#include <ATen/ops/empty_like_native.h>
#include <ATen/ops/empty_native.h>
#include <ATen/ops/erf_native.h>
#include <ATen/ops/erfinv_native.h>
#include <ATen/ops/expm1_native.h>
#include <ATen/ops/fill_native.h>
#include <ATen/ops/floor_native.h>
#include <ATen/ops/frac_native.h>
#include <ATen/ops/isinf_native.h>
#include <ATen/ops/isnan_native.h>
#include <ATen/ops/isneginf_native.h>
#include <ATen/ops/isposinf_native.h>
#include <ATen/ops/log1p_native.h>
#include <ATen/ops/mm_native.h>
#include <ATen/ops/mul_native.h>
#include <ATen/ops/neg_native.h>
#include <ATen/ops/normal_native.h>
#include <ATen/ops/rad2deg_native.h>
#include <ATen/ops/relu_native.h>
#include <ATen/ops/resize_as_sparse_native.h>
#include <ATen/ops/resize_native.h>
#include <ATen/ops/round_native.h>
#include <ATen/ops/row_indices_native.h>
#include <ATen/ops/select_copy_native.h>
#include <ATen/ops/select_native.h>
#include <ATen/ops/sgn_native.h>
#include <ATen/ops/sign_native.h>
#include <ATen/ops/signbit_native.h>
#include <ATen/ops/sin_native.h>
#include <ATen/ops/sinh_native.h>
#include <ATen/ops/sparse_dim_native.h>
#include <ATen/ops/sparse_mask_native.h>
#include <ATen/ops/sparse_sampled_addmm_native.h>
#include <ATen/ops/sqrt_native.h>
#include <ATen/ops/sum_native.h>
#include <ATen/ops/tan_native.h>
#include <ATen/ops/tanh_native.h>
#include <ATen/ops/threshold_backward_native.h>
#include <ATen/ops/triangular_solve_native.h>
#include <ATen/ops/trunc_native.h>
#include <ATen/ops/values_native.h>
#include <ATen/ops/zero_native.h>

namespace at {
namespace {
C10_DIAGNOSTIC_PUSH_AND_IGNORED_IF_DEFINED("-Wunused-function")

void resize_out(const Tensor &out, IntArrayRef sizes, IntArrayRef strides, const TensorOptions &options) {
  TORCH_CHECK(options.dtype() == out.dtype(),
      "Expected out tensor to have dtype ", options.dtype(), ", but got ", out.dtype(), " instead");
  TORCH_CHECK(options.device() == out.device(),
      "Expected out tensor to have device ", options.device(), ", but got ", out.device(), " instead");
  const bool resized = at::native::resize_output(out, sizes);
  // Only restride if a resize occurred; otherwise we ignore the (advisory)
  // strides from the meta function and directly use the output tensor's
  // preexisting strides
  if (resized) {
    if (!strides.empty()) {
      TORCH_INTERNAL_ASSERT(!options.memory_format_opt().has_value());
      // TODO: avoid the redispatch here
      out.as_strided_(sizes, strides);
    } else if (options.memory_format_opt().has_value()) {
      out.unsafeGetTensorImpl()->empty_tensor_restride(*options.memory_format_opt());
    }
  }
}

void check_inplace(const Tensor &self, IntArrayRef sizes, const TensorOptions &options) {
  // These checks are needed on those operators that:
  //   1) don't use 'TensorIterator' (e.g. 'addmm' and 'baddbmm')
  //   2) have particular typing rules (e.g. 'cumsum' and 'cumprod')
  // For other operators (e.g. 'add'), 'TensorIterator' already checks
  // these things separately.
  TORCH_CHECK(options.dtype() == self.dtype(),
      "Bad in-place call: ",
      "input tensor dtype ", self.dtype(), " and output tensor dtype ", options.dtype(), " should match");
  TORCH_CHECK(options.device() == self.device(),
      "Bad in-place call: ",
      "input tensor device ", self.device(), " and output tensor device ", options.device(), " should match");
  TORCH_CHECK(sizes == self.sizes(),
      "Bad in-place call: ",
      "input tensor size ", self.sizes(), " and output tensor size ", sizes, " should match");
}
C10_DIAGNOSTIC_POP()
} // namespace
} // namespace at

// See template file RegisterDispatchDefinitions.ini
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__abs(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::abs_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_abs_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::abs_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__abs_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::abs_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("abs",
TORCH_FN(wrapper_SparseCsrCUDA__abs));
m.impl("abs.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_abs_out));
m.impl("abs_",
TORCH_FN(wrapper_SparseCsrCUDA__abs_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor abs(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__abs(self);
}
at::Tensor & abs_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_abs_out(self, out);
}
at::Tensor & abs_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_abs_out(self, out);
}
at::Tensor & abs_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__abs_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__angle(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::angle_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_angle_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::angle_sparse_csr_out(self, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("angle",
TORCH_FN(wrapper_SparseCsrCUDA__angle));
m.impl("angle.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_angle_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor angle(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__angle(self);
}
at::Tensor & angle_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_angle_out(self, out);
}
at::Tensor & angle_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_angle_out(self, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sgn(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__sgn", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sgn_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_sgn_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_sgn_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_sgn_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sgn_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__sgn_(at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__sgn_", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sgn_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sgn",
TORCH_FN(wrapper_SparseCsrCUDA__sgn));
m.impl("sgn.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_sgn_out));
m.impl("sgn_",
TORCH_FN(wrapper_SparseCsrCUDA__sgn_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sgn(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__sgn(self);
}
at::Tensor & sgn_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_sgn_out(self, out);
}
at::Tensor & sgn_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_sgn_out(self, out);
}
at::Tensor & sgn_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__sgn_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___conj_physical(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA___conj_physical", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::conj_physical_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_conj_physical",
TORCH_FN(wrapper_SparseCsrCUDA___conj_physical));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _conj_physical(const at::Tensor & self) {
return wrapper_SparseCsrCUDA___conj_physical(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_conj_physical_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_conj_physical_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_conj_physical_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::conj_physical_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__conj_physical_(at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__conj_physical_", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::conj_physical_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("conj_physical.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_conj_physical_out));
m.impl("conj_physical_",
TORCH_FN(wrapper_SparseCsrCUDA__conj_physical_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & conj_physical_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_conj_physical_out(self, out);
}
at::Tensor & conj_physical_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_conj_physical_out(self, out);
}
at::Tensor & conj_physical_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__conj_physical_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_Tensor_add(const at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::add_sparse_csr(self, other, alpha);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_add_out(const at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::add_out_sparse_compressed_cuda(self, other, alpha, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_Tensor_add_(at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::add_sparse_csr_(self, other, alpha);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("add.Tensor",
TORCH_FN(wrapper_SparseCsrCUDA_Tensor_add));
m.impl("add.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_add_out));
m.impl("add_.Tensor",
TORCH_FN(wrapper_SparseCsrCUDA_Tensor_add_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor add(const at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA_Tensor_add(self, other, alpha);
}
at::Tensor & add_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA_out_add_out(self, other, alpha, out);
}
at::Tensor & add_outf(const at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_add_out(self, other, alpha, out);
}
at::Tensor & add_(at::Tensor & self, const at::Tensor & other, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA_Tensor_add_(self, other, alpha);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_addmv_out(const at::Tensor & self, const at::Tensor & mat, const at::Tensor & vec, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_addmv_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_addmv_out", "self");
  c10::impl::check_and_update_common_device(common_device, mat, "wrapper_SparseCsrCUDA_out_addmv_out", "mat");
  c10::impl::check_and_update_common_device(common_device, vec, "wrapper_SparseCsrCUDA_out_addmv_out", "vec");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::addmv_out_sparse_compressed_cuda(self, mat, vec, beta, alpha, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("addmv.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_addmv_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & addmv_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & mat, const at::Tensor & vec, const at::Scalar & beta, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA_out_addmv_out(self, mat, vec, beta, alpha, out);
}
at::Tensor & addmv_outf(const at::Tensor & self, const at::Tensor & mat, const at::Tensor & vec, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_addmv_out(self, mat, vec, beta, alpha, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__asinh(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__asinh", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::asinh_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_asinh_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_asinh_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_asinh_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::asinh_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__asinh_(at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__asinh_", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::asinh_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("asinh",
TORCH_FN(wrapper_SparseCsrCUDA__asinh));
m.impl("asinh.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_asinh_out));
m.impl("asinh_",
TORCH_FN(wrapper_SparseCsrCUDA__asinh_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor asinh(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__asinh(self);
}
at::Tensor & asinh_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_asinh_out(self, out);
}
at::Tensor & asinh_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_asinh_out(self, out);
}
at::Tensor & asinh_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__asinh_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__atanh(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__atanh", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::atanh_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_atanh_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_atanh_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_atanh_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::atanh_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__atanh_(at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__atanh_", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::atanh_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("atanh",
TORCH_FN(wrapper_SparseCsrCUDA__atanh));
m.impl("atanh.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_atanh_out));
m.impl("atanh_",
TORCH_FN(wrapper_SparseCsrCUDA__atanh_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor atanh(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__atanh(self);
}
at::Tensor & atanh_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_atanh_out(self, out);
}
at::Tensor & atanh_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_atanh_out(self, out);
}
at::Tensor & atanh_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__atanh_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__asin(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::asin_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_asin_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::asin_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__asin_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::asin_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("asin",
TORCH_FN(wrapper_SparseCsrCUDA__asin));
m.impl("asin.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_asin_out));
m.impl("asin_",
TORCH_FN(wrapper_SparseCsrCUDA__asin_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor asin(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__asin(self);
}
at::Tensor & asin_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_asin_out(self, out);
}
at::Tensor & asin_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_asin_out(self, out);
}
at::Tensor & asin_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__asin_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__atan(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::atan_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_atan_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::atan_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__atan_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::atan_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("atan",
TORCH_FN(wrapper_SparseCsrCUDA__atan));
m.impl("atan.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_atan_out));
m.impl("atan_",
TORCH_FN(wrapper_SparseCsrCUDA__atan_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor atan(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__atan(self);
}
at::Tensor & atan_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_atan_out(self, out);
}
at::Tensor & atan_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_atan_out(self, out);
}
at::Tensor & atan_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__atan_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_baddbmm_out(const at::Tensor & self, const at::Tensor & batch1, const at::Tensor & batch2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_baddbmm_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_baddbmm_out", "self");
  c10::impl::check_and_update_common_device(common_device, batch1, "wrapper_SparseCsrCUDA_out_baddbmm_out", "batch1");
  c10::impl::check_and_update_common_device(common_device, batch2, "wrapper_SparseCsrCUDA_out_baddbmm_out", "batch2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::baddbmm_out_sparse_csr_cuda(self, batch1, batch2, beta, alpha, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("baddbmm.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_baddbmm_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & baddbmm_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & batch1, const at::Tensor & batch2, const at::Scalar & beta, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA_out_baddbmm_out(self, batch1, batch2, beta, alpha, out);
}
at::Tensor & baddbmm_outf(const at::Tensor & self, const at::Tensor & batch1, const at::Tensor & batch2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_baddbmm_out(self, batch1, batch2, beta, alpha, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_bmm_out(const at::Tensor & self, const at::Tensor & mat2, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_bmm_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_bmm_out", "self");
  c10::impl::check_and_update_common_device(common_device, mat2, "wrapper_SparseCsrCUDA_out_bmm_out", "mat2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::bmm_out_sparse_csr_cuda(self, mat2, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("bmm.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_bmm_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & bmm_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & mat2) {
return wrapper_SparseCsrCUDA_out_bmm_out(self, mat2, out);
}
at::Tensor & bmm_outf(const at::Tensor & self, const at::Tensor & mat2, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_bmm_out(self, mat2, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__ceil(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::ceil_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_ceil_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::ceil_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__ceil_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::ceil_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("ceil",
TORCH_FN(wrapper_SparseCsrCUDA__ceil));
m.impl("ceil.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_ceil_out));
m.impl("ceil_",
TORCH_FN(wrapper_SparseCsrCUDA__ceil_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor ceil(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__ceil(self);
}
at::Tensor & ceil_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_ceil_out(self, out);
}
at::Tensor & ceil_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_ceil_out(self, out);
}
at::Tensor & ceil_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__ceil_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA__copy_(at::Tensor & self, const at::Tensor & src, bool non_blocking) {
    // No device check
  // DeviceGuard omitted
  return at::native::copy_sparse_compressed_(self, src, non_blocking);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("copy_",
TORCH_FN(wrapper_SparseCsrCUDA__copy_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & copy_(at::Tensor & self, const at::Tensor & src, bool non_blocking) {
return wrapper_SparseCsrCUDA__copy_(self, src, non_blocking);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_memory_format_empty(c10::SymIntArrayRef size, ::std::optional<at::ScalarType> dtype, ::std::optional<at::Layout> layout, ::std::optional<at::Device> device, ::std::optional<bool> pin_memory, ::std::optional<at::MemoryFormat> memory_format) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  globalContext().lazyInitDevice(c10::DeviceType::CUDA);
  const DeviceGuard device_guard(device_or_default(device));
  return at::native::empty_sparse_compressed(C10_AS_INTARRAYREF_SLOW(size), dtype, layout, device, pin_memory, memory_format);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("empty.memory_format",
TORCH_FN(wrapper_SparseCsrCUDA_memory_format_empty));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor empty(at::IntArrayRef size, at::TensorOptions options, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA_memory_format_empty(c10::fromIntArrayRefSlow(size), c10::optTypeMetaToScalarType(options.dtype_opt()), options.layout_opt(), options.device_opt(), options.pinned_memory_opt(), c10::impl::check_tensor_options_and_extract_memory_format(options, memory_format));
}
at::Tensor empty(at::IntArrayRef size, ::std::optional<at::ScalarType> dtype, ::std::optional<at::Layout> layout, ::std::optional<at::Device> device, ::std::optional<bool> pin_memory, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA_memory_format_empty(c10::fromIntArrayRefSlow(size), dtype, layout, device, pin_memory, memory_format);
}
at::Tensor empty_symint(c10::SymIntArrayRef size, at::TensorOptions options, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA_memory_format_empty(size, c10::optTypeMetaToScalarType(options.dtype_opt()), options.layout_opt(), options.device_opt(), options.pinned_memory_opt(), c10::impl::check_tensor_options_and_extract_memory_format(options, memory_format));
}
at::Tensor empty_symint(c10::SymIntArrayRef size, ::std::optional<at::ScalarType> dtype, ::std::optional<at::Layout> layout, ::std::optional<at::Device> device, ::std::optional<bool> pin_memory, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA_memory_format_empty(size, dtype, layout, device, pin_memory, memory_format);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
const at::Tensor & wrapper_SparseCsrCUDA__resize_(const at::Tensor & self, c10::SymIntArrayRef size, ::std::optional<at::MemoryFormat> memory_format) {
    // No device check
  // DeviceGuard omitted
  return at::native::resize_sparse_csr_(self, C10_AS_INTARRAYREF_SLOW(size), memory_format);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("resize_",
TORCH_FN(wrapper_SparseCsrCUDA__resize_));
}
} // anonymous namespace
namespace sparsecsrcuda {
const at::Tensor & resize_(const at::Tensor & self, at::IntArrayRef size, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA__resize_(self, c10::fromIntArrayRefSlow(size), memory_format);
}
const at::Tensor & resize__symint(const at::Tensor & self, c10::SymIntArrayRef size, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA__resize_(self, size, memory_format);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__empty_like(const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<at::Layout> layout, ::std::optional<at::Device> device, ::std::optional<bool> pin_memory, ::std::optional<at::MemoryFormat> memory_format) {
    // No device check
  // DeviceGuard omitted
  return at::native::empty_like_sparse_csr(self, dtype, layout, device, pin_memory, memory_format);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("empty_like",
TORCH_FN(wrapper_SparseCsrCUDA__empty_like));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor empty_like(const at::Tensor & self, at::TensorOptions options, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA__empty_like(self, c10::optTypeMetaToScalarType(options.dtype_opt()), options.layout_opt(), options.device_opt(), options.pinned_memory_opt(), c10::impl::check_tensor_options_and_extract_memory_format(options, memory_format));
}
at::Tensor empty_like(const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<at::Layout> layout, ::std::optional<at::Device> device, ::std::optional<bool> pin_memory, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA__empty_like(self, dtype, layout, device, pin_memory, memory_format);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__erf(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::erf_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_erf_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::erf_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__erf_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::erf_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("erf",
TORCH_FN(wrapper_SparseCsrCUDA__erf));
m.impl("erf.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_erf_out));
m.impl("erf_",
TORCH_FN(wrapper_SparseCsrCUDA__erf_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor erf(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__erf(self);
}
at::Tensor & erf_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_erf_out(self, out);
}
at::Tensor & erf_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_erf_out(self, out);
}
at::Tensor & erf_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__erf_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__expm1(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::expm1_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_expm1_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::expm1_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__expm1_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::expm1_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("expm1",
TORCH_FN(wrapper_SparseCsrCUDA__expm1));
m.impl("expm1.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_expm1_out));
m.impl("expm1_",
TORCH_FN(wrapper_SparseCsrCUDA__expm1_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor expm1(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__expm1(self);
}
at::Tensor & expm1_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_expm1_out(self, out);
}
at::Tensor & expm1_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_expm1_out(self, out);
}
at::Tensor & expm1_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__expm1_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA_Scalar_fill_(at::Tensor & self, const at::Scalar & value) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::fill_sparse_csr_(self, value);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("fill_.Scalar",
TORCH_FN(wrapper_SparseCsrCUDA_Scalar_fill_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & fill_(at::Tensor & self, const at::Scalar & value) {
return wrapper_SparseCsrCUDA_Scalar_fill_(self, value);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__floor(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::floor_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_floor_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::floor_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__floor_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::floor_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("floor",
TORCH_FN(wrapper_SparseCsrCUDA__floor));
m.impl("floor.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_floor_out));
m.impl("floor_",
TORCH_FN(wrapper_SparseCsrCUDA__floor_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor floor(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__floor(self);
}
at::Tensor & floor_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_floor_out(self, out);
}
at::Tensor & floor_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_floor_out(self, out);
}
at::Tensor & floor_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__floor_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__frac(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::frac_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_frac_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::frac_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__frac_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::frac_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("frac",
TORCH_FN(wrapper_SparseCsrCUDA__frac));
m.impl("frac.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_frac_out));
m.impl("frac_",
TORCH_FN(wrapper_SparseCsrCUDA__frac_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor frac(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__frac(self);
}
at::Tensor & frac_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_frac_out(self, out);
}
at::Tensor & frac_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_frac_out(self, out);
}
at::Tensor & frac_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__frac_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__isnan(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::isnan_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("isnan",
TORCH_FN(wrapper_SparseCsrCUDA__isnan));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor isnan(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__isnan(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__log1p(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::log1p_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_log1p_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::log1p_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__log1p_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::log1p_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("log1p",
TORCH_FN(wrapper_SparseCsrCUDA__log1p));
m.impl("log1p.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_log1p_out));
m.impl("log1p_",
TORCH_FN(wrapper_SparseCsrCUDA__log1p_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor log1p(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__log1p(self);
}
at::Tensor & log1p_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_log1p_out(self, out);
}
at::Tensor & log1p_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_log1p_out(self, out);
}
at::Tensor & log1p_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__log1p_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__mm(const at::Tensor & self, const at::Tensor & mat2) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__mm", "self");
  c10::impl::check_and_update_common_device(common_device, mat2, "wrapper_SparseCsrCUDA__mm", "mat2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::_sparse_csr_mm(self, mat2);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_mm_out(const at::Tensor & self, const at::Tensor & mat2, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_mm_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_mm_out", "self");
  c10::impl::check_and_update_common_device(common_device, mat2, "wrapper_SparseCsrCUDA_out_mm_out", "mat2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::_sparse_csr_mm_out(self, mat2, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("mm",
TORCH_FN(wrapper_SparseCsrCUDA__mm));
m.impl("mm.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_mm_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor mm(const at::Tensor & self, const at::Tensor & mat2) {
return wrapper_SparseCsrCUDA__mm(self, mat2);
}
at::Tensor & mm_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & mat2) {
return wrapper_SparseCsrCUDA_out_mm_out(self, mat2, out);
}
at::Tensor & mm_outf(const at::Tensor & self, const at::Tensor & mat2, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_mm_out(self, mat2, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_Tensor_mul(const at::Tensor & self, const at::Tensor & other) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::mul_sparse_csr(self, other);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_mul_out(const at::Tensor & self, const at::Tensor & other, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::mul_out_sparse_csr(self, other, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_Tensor_mul_(at::Tensor & self, const at::Tensor & other) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::mul_sparse_csr_(self, other);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("mul.Tensor",
TORCH_FN(wrapper_SparseCsrCUDA_Tensor_mul));
m.impl("mul.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_mul_out));
m.impl("mul_.Tensor",
TORCH_FN(wrapper_SparseCsrCUDA_Tensor_mul_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor mul(const at::Tensor & self, const at::Tensor & other) {
return wrapper_SparseCsrCUDA_Tensor_mul(self, other);
}
at::Tensor & mul_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & other) {
return wrapper_SparseCsrCUDA_out_mul_out(self, other, out);
}
at::Tensor & mul_outf(const at::Tensor & self, const at::Tensor & other, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_mul_out(self, other, out);
}
at::Tensor & mul_(at::Tensor & self, const at::Tensor & other) {
return wrapper_SparseCsrCUDA_Tensor_mul_(self, other);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_Scalar_mul(const at::Tensor & self, const at::Scalar & other) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::mul_scalar_sparse_csr(self, other);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_Scalar_mul_(at::Tensor & self, const at::Scalar & other) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::mul__scalar_sparse_csr(self, other);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("mul.Scalar",
TORCH_FN(wrapper_SparseCsrCUDA_Scalar_mul));
m.impl("mul_.Scalar",
TORCH_FN(wrapper_SparseCsrCUDA_Scalar_mul_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor mul(const at::Tensor & self, const at::Scalar & other) {
return wrapper_SparseCsrCUDA_Scalar_mul(self, other);
}
at::Tensor & mul_(at::Tensor & self, const at::Scalar & other) {
return wrapper_SparseCsrCUDA_Scalar_mul_(self, other);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__rad2deg(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__rad2deg", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::rad2deg_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_rad2deg_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_rad2deg_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_rad2deg_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::rad2deg_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__rad2deg_(at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__rad2deg_", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::rad2deg_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("rad2deg",
TORCH_FN(wrapper_SparseCsrCUDA__rad2deg));
m.impl("rad2deg.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_rad2deg_out));
m.impl("rad2deg_",
TORCH_FN(wrapper_SparseCsrCUDA__rad2deg_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor rad2deg(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__rad2deg(self);
}
at::Tensor & rad2deg_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_rad2deg_out(self, out);
}
at::Tensor & rad2deg_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_rad2deg_out(self, out);
}
at::Tensor & rad2deg_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__rad2deg_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__deg2rad(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__deg2rad", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::deg2rad_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_deg2rad_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_deg2rad_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_deg2rad_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::deg2rad_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__deg2rad_(at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__deg2rad_", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::deg2rad_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("deg2rad",
TORCH_FN(wrapper_SparseCsrCUDA__deg2rad));
m.impl("deg2rad.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_deg2rad_out));
m.impl("deg2rad_",
TORCH_FN(wrapper_SparseCsrCUDA__deg2rad_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor deg2rad(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__deg2rad(self);
}
at::Tensor & deg2rad_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_deg2rad_out(self, out);
}
at::Tensor & deg2rad_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_deg2rad_out(self, out);
}
at::Tensor & deg2rad_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__deg2rad_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__neg(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::neg_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_neg_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::neg_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__neg_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::neg_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("neg",
TORCH_FN(wrapper_SparseCsrCUDA__neg));
m.impl("neg.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_neg_out));
m.impl("neg_",
TORCH_FN(wrapper_SparseCsrCUDA__neg_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor neg(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__neg(self);
}
at::Tensor & neg_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_neg_out(self, out);
}
at::Tensor & neg_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_neg_out(self, out);
}
at::Tensor & neg_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__neg_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__round(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::round_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_round_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::round_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__round_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::round_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("round",
TORCH_FN(wrapper_SparseCsrCUDA__round));
m.impl("round.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_round_out));
m.impl("round_",
TORCH_FN(wrapper_SparseCsrCUDA__round_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor round(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__round(self);
}
at::Tensor & round_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_round_out(self, out);
}
at::Tensor & round_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_round_out(self, out);
}
at::Tensor & round_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__round_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__relu(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::relu_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__relu_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::relu_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("relu",
TORCH_FN(wrapper_SparseCsrCUDA__relu));
m.impl("relu_",
TORCH_FN(wrapper_SparseCsrCUDA__relu_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor relu(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__relu(self);
}
at::Tensor & relu_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__relu_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_int_select(const at::Tensor & self, int64_t dim, c10::SymInt index) {
    // No device check
  // DeviceGuard omitted
  return at::native::select_sparse_csr(self, dim, index.guard_int(__FILE__, __LINE__));
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("select.int",
TORCH_FN(wrapper_SparseCsrCUDA_int_select));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor select(const at::Tensor & self, int64_t dim, int64_t index) {
return wrapper_SparseCsrCUDA_int_select(self, dim, index);
}
at::Tensor select_symint(const at::Tensor & self, int64_t dim, c10::SymInt index) {
return wrapper_SparseCsrCUDA_int_select(self, dim, index);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sin(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sin_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_sin_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sin_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__sin_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sin_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sin",
TORCH_FN(wrapper_SparseCsrCUDA__sin));
m.impl("sin.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_sin_out));
m.impl("sin_",
TORCH_FN(wrapper_SparseCsrCUDA__sin_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sin(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__sin(self);
}
at::Tensor & sin_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_sin_out(self, out);
}
at::Tensor & sin_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_sin_out(self, out);
}
at::Tensor & sin_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__sin_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sinh(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sinh_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_sinh_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sinh_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__sinh_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sinh_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sinh",
TORCH_FN(wrapper_SparseCsrCUDA__sinh));
m.impl("sinh.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_sinh_out));
m.impl("sinh_",
TORCH_FN(wrapper_SparseCsrCUDA__sinh_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sinh(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__sinh(self);
}
at::Tensor & sinh_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_sinh_out(self, out);
}
at::Tensor & sinh_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_sinh_out(self, out);
}
at::Tensor & sinh_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__sinh_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sum(const at::Tensor & self, ::std::optional<at::ScalarType> dtype) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sum_csr(self, dtype);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sum",
TORCH_FN(wrapper_SparseCsrCUDA__sum));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sum(const at::Tensor & self, ::std::optional<at::ScalarType> dtype) {
return wrapper_SparseCsrCUDA__sum(self, dtype);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_dim_IntList_sum(const at::Tensor & self, at::OptionalIntArrayRef dim, bool keepdim, ::std::optional<at::ScalarType> dtype) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sum_sparse_compressed(self, dim, keepdim, dtype);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sum.dim_IntList",
TORCH_FN(wrapper_SparseCsrCUDA_dim_IntList_sum));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sum(const at::Tensor & self, at::OptionalIntArrayRef dim, bool keepdim, ::std::optional<at::ScalarType> dtype) {
return wrapper_SparseCsrCUDA_dim_IntList_sum(self, dim, keepdim, dtype);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sqrt(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sqrt_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_sqrt_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sqrt_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__sqrt_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sqrt_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sqrt",
TORCH_FN(wrapper_SparseCsrCUDA__sqrt));
m.impl("sqrt.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_sqrt_out));
m.impl("sqrt_",
TORCH_FN(wrapper_SparseCsrCUDA__sqrt_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sqrt(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__sqrt(self);
}
at::Tensor & sqrt_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_sqrt_out(self, out);
}
at::Tensor & sqrt_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_sqrt_out(self, out);
}
at::Tensor & sqrt_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__sqrt_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__tan(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::tan_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_tan_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::tan_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__tan_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::tan_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("tan",
TORCH_FN(wrapper_SparseCsrCUDA__tan));
m.impl("tan.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_tan_out));
m.impl("tan_",
TORCH_FN(wrapper_SparseCsrCUDA__tan_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor tan(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__tan(self);
}
at::Tensor & tan_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_tan_out(self, out);
}
at::Tensor & tan_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_tan_out(self, out);
}
at::Tensor & tan_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__tan_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__tanh(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::tanh_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_tanh_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::tanh_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__tanh_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::tanh_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("tanh",
TORCH_FN(wrapper_SparseCsrCUDA__tanh));
m.impl("tanh.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_tanh_out));
m.impl("tanh_",
TORCH_FN(wrapper_SparseCsrCUDA__tanh_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor tanh(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__tanh(self);
}
at::Tensor & tanh_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_tanh_out(self, out);
}
at::Tensor & tanh_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_tanh_out(self, out);
}
at::Tensor & tanh_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__tanh_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__threshold_backward(const at::Tensor & grad_output, const at::Tensor & self, const at::Scalar & threshold) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, grad_output, "wrapper_SparseCsrCUDA__threshold_backward", "grad_output");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__threshold_backward", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::threshold_backward_sparse_compressed(grad_output, self, threshold);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_grad_input_threshold_backward_out(const at::Tensor & grad_output, const at::Tensor & self, const at::Scalar & threshold, at::Tensor & grad_input) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, grad_input, "wrapper_SparseCsrCUDA_grad_input_threshold_backward_out", "grad_input");
  c10::impl::check_and_update_common_device(common_device, grad_output, "wrapper_SparseCsrCUDA_grad_input_threshold_backward_out", "grad_output");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_grad_input_threshold_backward_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::threshold_backward_sparse_compressed_out(grad_output, self, threshold, grad_input);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("threshold_backward",
TORCH_FN(wrapper_SparseCsrCUDA__threshold_backward));
m.impl("threshold_backward.grad_input",
TORCH_FN(wrapper_SparseCsrCUDA_grad_input_threshold_backward_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor threshold_backward(const at::Tensor & grad_output, const at::Tensor & self, const at::Scalar & threshold) {
return wrapper_SparseCsrCUDA__threshold_backward(grad_output, self, threshold);
}
at::Tensor & threshold_backward_out(at::Tensor & grad_input, const at::Tensor & grad_output, const at::Tensor & self, const at::Scalar & threshold) {
return wrapper_SparseCsrCUDA_grad_input_threshold_backward_out(grad_output, self, threshold, grad_input);
}
at::Tensor & threshold_backward_outf(const at::Tensor & grad_output, const at::Tensor & self, const at::Scalar & threshold, at::Tensor & grad_input) {
return wrapper_SparseCsrCUDA_grad_input_threshold_backward_out(grad_output, self, threshold, grad_input);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__trunc(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::trunc_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_trunc_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::trunc_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__trunc_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::trunc_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("trunc",
TORCH_FN(wrapper_SparseCsrCUDA__trunc));
m.impl("trunc.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_trunc_out));
m.impl("trunc_",
TORCH_FN(wrapper_SparseCsrCUDA__trunc_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor trunc(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__trunc(self);
}
at::Tensor & trunc_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_trunc_out(self, out);
}
at::Tensor & trunc_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_trunc_out(self, out);
}
at::Tensor & trunc_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__trunc_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_sum(const at::Tensor & self, at::IntArrayRef dim, bool keepdim, ::std::optional<at::ScalarType> dtype) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_sum", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::_sparse_csr_sum_cuda(self, dim, keepdim, dtype);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_sparse_csr_sum.dim_dtype",
TORCH_FN(wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_sum));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _sparse_csr_sum(const at::Tensor & self, at::IntArrayRef dim, bool keepdim, ::std::optional<at::ScalarType> dtype) {
return wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_sum(self, dim, keepdim, dtype);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_prod(const at::Tensor & self, at::IntArrayRef dim, bool keepdim, ::std::optional<at::ScalarType> dtype) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_prod", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::_sparse_csr_prod_cuda(self, dim, keepdim, dtype);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_sparse_csr_prod.dim_dtype",
TORCH_FN(wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_prod));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _sparse_csr_prod(const at::Tensor & self, at::IntArrayRef dim, bool keepdim, ::std::optional<at::ScalarType> dtype) {
return wrapper_SparseCsrCUDA_dim_dtype__sparse_csr_prod(self, dim, keepdim, dtype);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__clone(const at::Tensor & self, ::std::optional<at::MemoryFormat> memory_format) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__clone", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::clone_sparse_compressed(self, memory_format);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("clone",
TORCH_FN(wrapper_SparseCsrCUDA__clone));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor clone(const at::Tensor & self, ::std::optional<at::MemoryFormat> memory_format) {
return wrapper_SparseCsrCUDA__clone(self, memory_format);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
const at::Tensor & wrapper_SparseCsrCUDA__resize_as_sparse_(const at::Tensor & self, const at::Tensor & the_template) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__resize_as_sparse_", "self");
  c10::impl::check_and_update_common_device(common_device, the_template, "wrapper_SparseCsrCUDA__resize_as_sparse_", "the_template");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::resize_as_sparse_compressed_(self, the_template);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("resize_as_sparse_",
TORCH_FN(wrapper_SparseCsrCUDA__resize_as_sparse_));
}
} // anonymous namespace
namespace sparsecsrcuda {
const at::Tensor & resize_as_sparse_(const at::Tensor & self, const at::Tensor & the_template) {
return wrapper_SparseCsrCUDA__resize_as_sparse_(self, the_template);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA__zero_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::zero_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("zero_",
TORCH_FN(wrapper_SparseCsrCUDA__zero_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & zero_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__zero_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sparse_sampled_addmm(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__sparse_sampled_addmm", "self");
  c10::impl::check_and_update_common_device(common_device, mat1, "wrapper_SparseCsrCUDA__sparse_sampled_addmm", "mat1");
  c10::impl::check_and_update_common_device(common_device, mat2, "wrapper_SparseCsrCUDA__sparse_sampled_addmm", "mat2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_sampled_addmm_sparse_csr_cuda(self, mat1, mat2, beta, alpha);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out", "self");
  c10::impl::check_and_update_common_device(common_device, mat1, "wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out", "mat1");
  c10::impl::check_and_update_common_device(common_device, mat2, "wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out", "mat2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_sampled_addmm_out_sparse_csr_cuda(self, mat1, mat2, beta, alpha, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sparse_sampled_addmm",
TORCH_FN(wrapper_SparseCsrCUDA__sparse_sampled_addmm));
m.impl("sparse_sampled_addmm.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sparse_sampled_addmm(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA__sparse_sampled_addmm(self, mat1, mat2, beta, alpha);
}
at::Tensor & sparse_sampled_addmm_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out(self, mat1, mat2, beta, alpha, out);
}
at::Tensor & sparse_sampled_addmm_outf(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_sparse_sampled_addmm_out(self, mat1, mat2, beta, alpha, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__addmm(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__addmm", "self");
  c10::impl::check_and_update_common_device(common_device, mat1, "wrapper_SparseCsrCUDA__addmm", "mat1");
  c10::impl::check_and_update_common_device(common_device, mat2, "wrapper_SparseCsrCUDA__addmm", "mat2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::addmm_sparse_compressed_dense(self, mat1, mat2, beta, alpha);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_addmm_out(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_addmm_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_addmm_out", "self");
  c10::impl::check_and_update_common_device(common_device, mat1, "wrapper_SparseCsrCUDA_out_addmm_out", "mat1");
  c10::impl::check_and_update_common_device(common_device, mat2, "wrapper_SparseCsrCUDA_out_addmm_out", "mat2");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::addmm_out_sparse_compressed_cuda(self, mat1, mat2, beta, alpha, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("addmm",
TORCH_FN(wrapper_SparseCsrCUDA__addmm));
m.impl("addmm.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_addmm_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor addmm(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA__addmm(self, mat1, mat2, beta, alpha);
}
at::Tensor & addmm_out(at::Tensor & out, const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha) {
return wrapper_SparseCsrCUDA_out_addmm_out(self, mat1, mat2, beta, alpha, out);
}
at::Tensor & addmm_outf(const at::Tensor & self, const at::Tensor & mat1, const at::Tensor & mat2, const at::Scalar & beta, const at::Scalar & alpha, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_addmm_out(self, mat1, mat2, beta, alpha, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sparse_mask(const at::Tensor & self, const at::Tensor & mask) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__sparse_mask", "self");
  c10::impl::check_and_update_common_device(common_device, mask, "wrapper_SparseCsrCUDA__sparse_mask", "mask");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_mask_sparse_compressed(self, mask);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sparse_mask",
TORCH_FN(wrapper_SparseCsrCUDA__sparse_mask));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sparse_mask(const at::Tensor & self, const at::Tensor & mask) {
return wrapper_SparseCsrCUDA__sparse_mask(self, mask);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___to_dense(const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<bool> masked_grad) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA___to_dense", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_compressed_to_dense(self, dtype, masked_grad);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_to_dense",
TORCH_FN(wrapper_SparseCsrCUDA___to_dense));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _to_dense(const at::Tensor & self, ::std::optional<at::ScalarType> dtype, ::std::optional<bool> masked_grad) {
return wrapper_SparseCsrCUDA___to_dense(self, dtype, masked_grad);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
int64_t wrapper_SparseCsrCUDA__sparse_dim(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::sparse_dim_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sparse_dim",
TORCH_FN(wrapper_SparseCsrCUDA__sparse_dim));
}
} // anonymous namespace
namespace sparsecsrcuda {
int64_t sparse_dim(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__sparse_dim(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
int64_t wrapper_SparseCsrCUDA__dense_dim(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::dense_dim_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("dense_dim",
TORCH_FN(wrapper_SparseCsrCUDA__dense_dim));
}
} // anonymous namespace
namespace sparsecsrcuda {
int64_t dense_dim(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__dense_dim(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
int64_t wrapper_SparseCsrCUDA___nnz(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::_nnz_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_nnz",
TORCH_FN(wrapper_SparseCsrCUDA___nnz));
}
} // anonymous namespace
namespace sparsecsrcuda {
int64_t _nnz(const at::Tensor & self) {
return wrapper_SparseCsrCUDA___nnz(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__values(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::values_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("values",
TORCH_FN(wrapper_SparseCsrCUDA__values));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor values(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__values(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__crow_indices(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::crow_indices_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("crow_indices",
TORCH_FN(wrapper_SparseCsrCUDA__crow_indices));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor crow_indices(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__crow_indices(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__col_indices(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::col_indices_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("col_indices",
TORCH_FN(wrapper_SparseCsrCUDA__col_indices));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor col_indices(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__col_indices(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__ccol_indices(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::ccol_indices_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("ccol_indices",
TORCH_FN(wrapper_SparseCsrCUDA__ccol_indices));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor ccol_indices(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__ccol_indices(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__row_indices(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::row_indices_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("row_indices",
TORCH_FN(wrapper_SparseCsrCUDA__row_indices));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor row_indices(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__row_indices(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_sparse_dim__to_sparse(const at::Tensor & self, int64_t sparse_dim) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_sparse_dim__to_sparse", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_compressed_to_sparse(self, sparse_dim);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_to_sparse.sparse_dim",
TORCH_FN(wrapper_SparseCsrCUDA_sparse_dim__to_sparse));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _to_sparse(const at::Tensor & self, int64_t sparse_dim) {
return wrapper_SparseCsrCUDA_sparse_dim__to_sparse(self, sparse_dim);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___to_sparse(const at::Tensor & self, ::std::optional<at::Layout> layout, at::OptionalIntArrayRef blocksize, ::std::optional<int64_t> dense_dim) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA___to_sparse", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_compressed_to_sparse(self, layout, blocksize, dense_dim);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_to_sparse",
TORCH_FN(wrapper_SparseCsrCUDA___to_sparse));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _to_sparse(const at::Tensor & self, ::std::optional<at::Layout> layout, at::OptionalIntArrayRef blocksize, ::std::optional<int64_t> dense_dim) {
return wrapper_SparseCsrCUDA___to_sparse(self, layout, blocksize, dense_dim);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___to_sparse_csr(const at::Tensor & self, ::std::optional<int64_t> dense_dim) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA___to_sparse_csr", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_compressed_to_sparse_csr(self, dense_dim);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_to_sparse_csr",
TORCH_FN(wrapper_SparseCsrCUDA___to_sparse_csr));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _to_sparse_csr(const at::Tensor & self, ::std::optional<int64_t> dense_dim) {
return wrapper_SparseCsrCUDA___to_sparse_csr(self, dense_dim);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___to_sparse_csc(const at::Tensor & self, ::std::optional<int64_t> dense_dim) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA___to_sparse_csc", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_compressed_to_sparse_csc(self, dense_dim);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_to_sparse_csc",
TORCH_FN(wrapper_SparseCsrCUDA___to_sparse_csc));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _to_sparse_csc(const at::Tensor & self, ::std::optional<int64_t> dense_dim) {
return wrapper_SparseCsrCUDA___to_sparse_csc(self, dense_dim);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___to_sparse_bsr(const at::Tensor & self, at::IntArrayRef blocksize, ::std::optional<int64_t> dense_dim) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA___to_sparse_bsr", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_compressed_to_sparse_bsr(self, blocksize, dense_dim);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_to_sparse_bsr",
TORCH_FN(wrapper_SparseCsrCUDA___to_sparse_bsr));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _to_sparse_bsr(const at::Tensor & self, at::IntArrayRef blocksize, ::std::optional<int64_t> dense_dim) {
return wrapper_SparseCsrCUDA___to_sparse_bsr(self, blocksize, dense_dim);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___to_sparse_bsc(const at::Tensor & self, at::IntArrayRef blocksize, ::std::optional<int64_t> dense_dim) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA___to_sparse_bsc", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sparse_compressed_to_sparse_bsc(self, blocksize, dense_dim);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_to_sparse_bsc",
TORCH_FN(wrapper_SparseCsrCUDA___to_sparse_bsc));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _to_sparse_bsc(const at::Tensor & self, at::IntArrayRef blocksize, ::std::optional<int64_t> dense_dim) {
return wrapper_SparseCsrCUDA___to_sparse_bsc(self, blocksize, dense_dim);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
::std::tuple<at::Tensor &,at::Tensor &> wrapper_SparseCsrCUDA_X_triangular_solve_out(const at::Tensor & self, const at::Tensor & A, bool upper, bool transpose, bool unitriangular, at::Tensor & X, at::Tensor & M) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, X, "wrapper_SparseCsrCUDA_X_triangular_solve_out", "X");
  c10::impl::check_and_update_common_device(common_device, M, "wrapper_SparseCsrCUDA_X_triangular_solve_out", "M");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_X_triangular_solve_out", "self");
  c10::impl::check_and_update_common_device(common_device, A, "wrapper_SparseCsrCUDA_X_triangular_solve_out", "A");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::triangular_solve_out_sparse_csr_cuda(self, A, upper, transpose, unitriangular, X, M);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("triangular_solve.X",
TORCH_FN(wrapper_SparseCsrCUDA_X_triangular_solve_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
::std::tuple<at::Tensor &,at::Tensor &> triangular_solve_out(at::Tensor & X, at::Tensor & M, const at::Tensor & self, const at::Tensor & A, bool upper, bool transpose, bool unitriangular) {
return wrapper_SparseCsrCUDA_X_triangular_solve_out(self, A, upper, transpose, unitriangular, X, M);
}
::std::tuple<at::Tensor &,at::Tensor &> triangular_solve_outf(const at::Tensor & self, const at::Tensor & A, bool upper, bool transpose, bool unitriangular, at::Tensor & X, at::Tensor & M) {
return wrapper_SparseCsrCUDA_X_triangular_solve_out(self, A, upper, transpose, unitriangular, X, M);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__erfinv(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::erfinv_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_erfinv_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::erfinv_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__erfinv_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::erfinv_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("erfinv",
TORCH_FN(wrapper_SparseCsrCUDA__erfinv));
m.impl("erfinv.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_erfinv_out));
m.impl("erfinv_",
TORCH_FN(wrapper_SparseCsrCUDA__erfinv_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor erfinv(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__erfinv(self);
}
at::Tensor & erfinv_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_erfinv_out(self, out);
}
at::Tensor & erfinv_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_erfinv_out(self, out);
}
at::Tensor & erfinv_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__erfinv_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__sign(const at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sign_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_sign_out(const at::Tensor & self, at::Tensor & out) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sign_sparse_csr_out(self, out);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA__sign_(at::Tensor & self) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::sign_sparse_csr_(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("sign",
TORCH_FN(wrapper_SparseCsrCUDA__sign));
m.impl("sign.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_sign_out));
m.impl("sign_",
TORCH_FN(wrapper_SparseCsrCUDA__sign_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor sign(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__sign(self);
}
at::Tensor & sign_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_sign_out(self, out);
}
at::Tensor & sign_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_sign_out(self, out);
}
at::Tensor & sign_(at::Tensor & self) {
return wrapper_SparseCsrCUDA__sign_(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__signbit(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__signbit", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::signbit_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_signbit_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_signbit_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_signbit_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::signbit_sparse_csr_out(self, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("signbit",
TORCH_FN(wrapper_SparseCsrCUDA__signbit));
m.impl("signbit.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_signbit_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor signbit(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__signbit(self);
}
at::Tensor & signbit_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_signbit_out(self, out);
}
at::Tensor & signbit_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_signbit_out(self, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor & wrapper_SparseCsrCUDA__normal_(at::Tensor & self, double mean, double std, ::std::optional<at::Generator> generator) {
    // No device check
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::normal_sparse_csr_(self, mean, std, generator);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("normal_",
TORCH_FN(wrapper_SparseCsrCUDA__normal_));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor & normal_(at::Tensor & self, double mean, double std, ::std::optional<at::Generator> generator) {
return wrapper_SparseCsrCUDA__normal_(self, mean, std, generator);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__isinf(const at::Tensor & self) {
    // No device check
  // DeviceGuard omitted
  return at::native::isinf_sparse_csr(self);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("isinf",
TORCH_FN(wrapper_SparseCsrCUDA__isinf));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor isinf(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__isinf(self);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__isposinf(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__isposinf", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::isposinf_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_isposinf_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_isposinf_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_isposinf_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::isposinf_sparse_csr_out(self, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("isposinf",
TORCH_FN(wrapper_SparseCsrCUDA__isposinf));
m.impl("isposinf.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_isposinf_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor isposinf(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__isposinf(self);
}
at::Tensor & isposinf_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_isposinf_out(self, out);
}
at::Tensor & isposinf_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_isposinf_out(self, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA__isneginf(const at::Tensor & self) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA__isneginf", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::isneginf_sparse_csr(self);
}
} // anonymous namespace
namespace {
at::Tensor & wrapper_SparseCsrCUDA_out_isneginf_out(const at::Tensor & self, at::Tensor & out) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, out, "wrapper_SparseCsrCUDA_out_isneginf_out", "out");
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_out_isneginf_out", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::isneginf_sparse_csr_out(self, out);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("isneginf",
TORCH_FN(wrapper_SparseCsrCUDA__isneginf));
m.impl("isneginf.out",
TORCH_FN(wrapper_SparseCsrCUDA_out_isneginf_out));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor isneginf(const at::Tensor & self) {
return wrapper_SparseCsrCUDA__isneginf(self);
}
at::Tensor & isneginf_out(at::Tensor & out, const at::Tensor & self) {
return wrapper_SparseCsrCUDA_out_isneginf_out(self, out);
}
at::Tensor & isneginf_outf(const at::Tensor & self, at::Tensor & out) {
return wrapper_SparseCsrCUDA_out_isneginf_out(self, out);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA___spsolve(const at::Tensor & A, const at::Tensor & B, bool left) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, A, "wrapper_SparseCsrCUDA___spsolve", "A");
  c10::impl::check_and_update_common_device(common_device, B, "wrapper_SparseCsrCUDA___spsolve", "B");
  const OptionalDeviceGuard device_guard(device_of(A));
  return at::native::_sparse_csr_linear_solve(A, B, left);
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("_spsolve",
TORCH_FN(wrapper_SparseCsrCUDA___spsolve));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor _spsolve(const at::Tensor & A, const at::Tensor & B, bool left) {
return wrapper_SparseCsrCUDA___spsolve(A, B, left);
}
} // namespace sparsecsrcuda
} // namespace at
namespace at {
// NB: TORCH_LIBRARY_IMPL must be in an anonymous namespace to avoid
// ambiguity with conflicting identifiers that may have been defined in
// at namespace already.
namespace {
namespace {
at::Tensor wrapper_SparseCsrCUDA_int_select_copy(const at::Tensor & self, int64_t dim, c10::SymInt index) {
  std::optional<Device> common_device = std::nullopt;
(void)common_device; // Suppress unused variable warning
  c10::impl::check_and_update_common_device(common_device, self, "wrapper_SparseCsrCUDA_int_select_copy", "self");
  const OptionalDeviceGuard device_guard(device_of(self));
  return at::native::select_copy_sparse_csr(self, dim, index.guard_int(__FILE__, __LINE__));
}
} // anonymous namespace
TORCH_LIBRARY_IMPL(aten, SparseCsrCUDA, m) {
    m.impl("select_copy.int",
TORCH_FN(wrapper_SparseCsrCUDA_int_select_copy));
}
} // anonymous namespace
namespace sparsecsrcuda {
at::Tensor select_copy(const at::Tensor & self, int64_t dim, int64_t index) {
return wrapper_SparseCsrCUDA_int_select_copy(self, dim, index);
}
at::Tensor select_copy_symint(const at::Tensor & self, int64_t dim, c10::SymInt index) {
return wrapper_SparseCsrCUDA_int_select_copy(self, dim, index);
}
} // namespace sparsecsrcuda
} // namespace at
