/*************************************************************************************************** * Copyright (c) 2023 - 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * 3. Neither the name of the copyright holder nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * **************************************************************************************************/ /*! \file \brief Utilities for initializing workspaces */ #pragma once #if !defined(__CUDACC_RTC__) #include "cuda.h" #include "cuda_runtime.h" #include "cutlass/trace.h" #endif #include "cutlass.h" #include "cutlass/cuda_host_adapter.hpp" namespace cutlass { ///////////////////////////////////////////////////////////////////////////////////////////////// static constexpr int MinWorkspaceAlignment = 16; #if !defined(__CUDACC_RTC__) static Status zero_workspace( void* workspace, size_t workspace_size, cudaStream_t stream = nullptr, [[maybe_unused]] CudaHostAdapter *cuda_adapter = nullptr) { if (workspace_size > 0) { if (workspace == nullptr) { CUTLASS_TRACE_HOST(" error: device workspace must not be null"); return Status::kErrorWorkspaceNull; } CUTLASS_TRACE_HOST(" clearing workspace"); #if defined(CUTLASS_ENABLE_CUDA_HOST_ADAPTER) && CUTLASS_ENABLE_CUDA_HOST_ADAPTER // // Use the cuda host adapter // CUTLASS_ASSERT(cuda_adapter); if (cuda_adapter) { if (Status::kSuccess != cuda_adapter->memsetDevice(workspace, static_cast(0), workspace_size, stream)) { return Status::kErrorInternal; } } else { return Status::kErrorInternal; } #else cudaError_t result = cudaMemsetAsync(workspace, 0, workspace_size, stream); if (cudaSuccess != result) { result = cudaGetLastError(); // to clear the error bit CUTLASS_TRACE_HOST(" cudaMemsetAsync() returned error " << cudaGetErrorString(result)); return Status::kErrorInternal; } #endif } return Status::kSuccess; } #endif #if !defined(__CUDACC_RTC__) template Status fill_workspace(void* workspace, T fill_value, size_t fill_count, cudaStream_t stream = nullptr, CudaHostAdapter *cuda_adapter = nullptr) { static_assert(sizeof(T) == 4 || sizeof(T) == 2 || sizeof(T) == 1, "Unsupported fill type"); if (fill_count > 0) { if (workspace == nullptr) { CUTLASS_TRACE_HOST(" error: device workspace must not be null"); return Status::kErrorWorkspaceNull; } CUTLASS_TRACE_HOST(" filling workspace"); #if defined(CUTLASS_ENABLE_CUDA_HOST_ADAPTER) && CUTLASS_ENABLE_CUDA_HOST_ADAPTER // // Use the cuda host adapter // CUTLASS_ASSERT(cuda_adapter); if (cuda_adapter) { if (Status::kSuccess != cuda_adapter->memsetDevice(workspace, fill_value, fill_count, stream)) { return Status::kErrorInternal; } } else { return Status::kErrorInternal; } #else CUdeviceptr d_workspace = reinterpret_cast(workspace); CUresult result = CUDA_SUCCESS; if (sizeof(T) == 4) { result = cuMemsetD32Async(d_workspace, reinterpret_cast(fill_value), fill_count, stream); } else if (sizeof(T) == 2) { result = cuMemsetD16Async(d_workspace, reinterpret_cast(fill_value), fill_count, stream); } else if (sizeof(T) == 1) { result = cuMemsetD8Async(d_workspace, reinterpret_cast(fill_value), fill_count, stream); } if (CUDA_SUCCESS != result) { const char** error_string_ptr = nullptr; (void) cuGetErrorString(result, error_string_ptr); if (error_string_ptr != nullptr) { CUTLASS_TRACE_HOST(" cuMemsetD" << sizeof(T) * 8 << "Async() returned error " << *error_string_ptr); } else { CUTLASS_TRACE_HOST(" cuMemsetD" << sizeof(T) * 8 << "Async() returned unrecognized error"); } return Status::kErrorInternal; } #endif } return Status::kSuccess; } #endif ///////////////////////////////////////////////////////////////////////////////////////////////// } // namespace cutlass