1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103
|
/*
* Copyright 2008-2013 NVIDIA Corporation
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#pragma once
#include <thrust/system/cuda/detail/guarded_cuda_runtime_api.h>
#include <thrust/detail/config.h>
#include <thrust/detail/raw_pointer_cast.h>
#include <thrust/detail/raw_reference_cast.h>
#include <thrust/detail/seq.h>
#include <thrust/system/cuda/config.h>
#ifdef THRUST_CACHING_DEVICE_MALLOC
#include <cub/util_allocator.cuh>
#endif
#include <thrust/system/cuda/detail/util.h>
#include <thrust/system/detail/bad_alloc.h>
#include <thrust/detail/malloc_and_free.h>
THRUST_NAMESPACE_BEGIN
namespace cuda_cub {
#ifdef THRUST_CACHING_DEVICE_MALLOC
#define __CUB_CACHING_MALLOC
#ifndef __CUDA_ARCH__
inline cub::CachingDeviceAllocator &get_allocator()
{
static cub::CachingDeviceAllocator g_allocator(true);
return g_allocator;
}
#endif
#endif
// note that malloc returns a raw pointer to avoid
// depending on the heavyweight thrust/system/cuda/memory.h header
template<typename DerivedPolicy>
__host__ __device__
void *malloc(execution_policy<DerivedPolicy> &, std::size_t n)
{
void *result = 0;
if (THRUST_IS_HOST_CODE) {
#if THRUST_INCLUDE_HOST_CODE
#ifdef __CUB_CACHING_MALLOC
cub::CachingDeviceAllocator &alloc = get_allocator();
cudaError_t status = alloc.DeviceAllocate(&result, n);
#else
cudaError_t status = cudaMalloc(&result, n);
#endif
if(status != cudaSuccess)
{
cudaGetLastError(); // Clear global CUDA error state.
throw thrust::system::detail::bad_alloc(thrust::cuda_category().message(status).c_str());
}
#endif
} else {
#if THRUST_INCLUDE_DEVICE_CODE
result = thrust::raw_pointer_cast(thrust::malloc(thrust::seq, n));
#endif
}
return result;
} // end malloc()
template<typename DerivedPolicy, typename Pointer>
__host__ __device__
void free(execution_policy<DerivedPolicy> &, Pointer ptr)
{
if (THRUST_IS_HOST_CODE) {
#if THRUST_INCLUDE_HOST_CODE
#ifdef __CUB_CACHING_MALLOC
cub::CachingDeviceAllocator &alloc = get_allocator();
cudaError_t status = alloc.DeviceFree(thrust::raw_pointer_cast(ptr));
#else
cudaError_t status = cudaFree(thrust::raw_pointer_cast(ptr));
#endif
cuda_cub::throw_on_error(status, "device free failed");
#endif
} else {
#if THRUST_INCLUDE_DEVICE_CODE
thrust::free(thrust::seq, ptr);
#endif
}
} // end free()
} // namespace cuda_cub
THRUST_NAMESPACE_END
|