/*! * Copyright (c) 2017 by Contributors * \file Use external cudnn utils function */ #ifndef TVM_CONTRIB_CUDNN_CUDNN_UTILS_H_ #define TVM_CONTRIB_CUDNN_CUDNN_UTILS_H_ #include <dmlc/logging.h> #include <cudnn.h> #include <tvm/runtime/device_api.h> #include "../../runtime/cuda/cuda_common.h" namespace tvm { namespace contrib { #define CUDNN_CALL(func) \ { \ cudnnStatus_t e = (func); \ CHECK_EQ(e, CUDNN_STATUS_SUCCESS) << "cuDNN: " << cudnnGetErrorString(e); \ } /*! breif Convert DLTensor type to CuDNN type */ struct CuDNNDataType { static cudnnDataType_t DLTypeToCuDNNType(const DLDataType &dtype); template<int v> static const void* GetConst(cudnnDataType_t type); }; // struct CuDNNDataType inline void GetStride(int nbdim, const int *dims, int *strides) { int mul = 1; for (int i = nbdim - 1; i >=0; --i) { mul *= dims[i]; strides[i] = mul; } } struct ConvEntry { cudnnConvolutionDescriptor_t conv_desc; cudnnConvolutionMode_t mode; cudnnFilterDescriptor_t filter_desc; cudnnDataType_t data_type; cudnnTensorFormat_t tensor_format; cudnnTensorDescriptor_t input_desc; cudnnTensorDescriptor_t output_desc; cudnnConvolutionFwdAlgo_t fwd_algo; // cudnnMathType_t math_type; TVMContext ctx; runtime::DeviceAPI *cuda_api; void *workspace{nullptr}; size_t workspace_size{0}; int group_count {0}; ConvEntry(); ~ConvEntry(); void UpdateWorkspace(const size_t wsize); void CleanWorkspace(); }; // ConvThreadEntry struct CuDNNThreadEntry { CuDNNThreadEntry(); ~CuDNNThreadEntry(); cudnnHandle_t handle{nullptr}; ConvEntry conv_entry; runtime::DeviceAPI *cuda_api{nullptr}; static CuDNNThreadEntry* ThreadLocal(); }; // CuDNNThreadEntry } // namespace contrib } // namespace tvm #endif // TVM_CONTRIB_CUDNN_CUDNN_UTILS_H_