[boinc] 01/01: removed nvidia_direct_invocation[_2].patch
Guo Yixuan
yixuan-guest at moszumanska.debian.org
Tue May 27 14:51:59 UTC 2014
This is an automated email from the git hooks/post-receive script.
yixuan-guest pushed a commit to branch master
in repository boinc.
commit e600098f7a5867ee5707880aa8253bad044219c4
Author: Guo Yixuan <culu.gyx at gmail.com>
Date: Tue May 27 10:51:13 2014 -0400
removed nvidia_direct_invocation[_2].patch
the file client/coproc_detect.cpp was removed
---
debian/patches/nvidia_direct_invocation.patch | 117 ----------------
debian/patches/nvidia_direct_invocation_2.patch | 176 ------------------------
debian/patches/series | 2 -
3 files changed, 295 deletions(-)
diff --git a/debian/patches/nvidia_direct_invocation.patch b/debian/patches/nvidia_direct_invocation.patch
deleted file mode 100644
index 6c8bc61..0000000
--- a/debian/patches/nvidia_direct_invocation.patch
+++ /dev/null
@@ -1,117 +0,0 @@
-Index: boinc/client/coproc_detect.cpp
-===================================================================
---- boinc.orig/client/coproc_detect.cpp 2012-05-03 14:01:32.344886723 +0200
-+++ boinc/client/coproc_detect.cpp 2012-05-03 14:01:48.736662216 +0200
-@@ -1024,21 +1024,33 @@
- return;
- }
-
-+#ifndef __cuda_cuda_h__
- retval = (*__cuInit)(0);
-+#else
-+ retval = cuInit(0);
-+#endif
- if (retval) {
- sprintf(buf, "NVIDIA drivers present but no GPUs found");
- warnings.push_back(buf);
- return;
- }
-
-+#ifndef __cuda_cuda_h__
- retval = (*__cuDriverGetVersion)(&cuda_version);
-+#else
-+ retval = cuDriverGetVersion(&cuda_version);
-+#endif
- if (retval) {
- sprintf(buf, "cuDriverGetVersion() returned %d", retval);
- warnings.push_back(buf);
- return;
- }
-
-+#ifndef __cuda_cuda_h__
- retval = (*__cuDeviceGetCount)(&cuda_ndevs);
-+#else
-+ retval = cuDeviceGetCount(&cuda_ndevs);
-+#endif
- if (retval) {
- sprintf(buf, "cuDeviceGetCount() returned %d", retval);
- warnings.push_back(buf);
-@@ -1061,14 +1073,23 @@
- nvidia_gpus.push_back(cc);
- }
- #endif
-+
-+#ifndef __cuda_cuda_h__
- retval = (*__cuDeviceGet)(&device, j);
-+#else
-+ retval = cuDeviceGet(&device, j);
-+#endif
- if (retval) {
- sprintf(buf, "cuDeviceGet(%d) returned %d", j, retval);
- warnings.push_back(buf);
- return;
- }
- cc.prop.deviceHandle = device;
-- (*__cuDeviceGetName)(cc.prop.name, 256, device);
-+#ifndef __cuda_cuda_h__
-+ retval = (*__cuDeviceGetName)(cc.prop.name, 256, device);
-+#else
-+ retval = cuDeviceGetName(cc.prop.name, 256, device);
-+#endif
- if (retval) {
- sprintf(buf, "cuDeviceGetName(%d) returned %d", j, retval);
- warnings.push_back(buf);
-@@ -1195,7 +1216,11 @@
- void* ctx;
-
- available_ram = prop.totalGlobalMem;
-+#ifndef __cuda_cuda_h__
- retval = (*__cuDeviceGet)(&device, device_num);
-+#else
-+ retval = cuDeviceGet(&device, device_num);
-+#endif
- if (retval) {
- if (log_flags.coproc_debug) {
- msg_printf(0, MSG_INFO,
-@@ -1206,7 +1231,11 @@
- }
-
- CUcontext ctx; // no immediate better idea
-+#ifndef __cuda_cuda_h__
- retval = (*__cuCtxCreate)(&ctx, 0, device);
-+#else
-+ retval = cuCtxCreate(&ctx, 0, device);
-+#endif
- if (retval) {
- if (log_flags.coproc_debug) {
- msg_printf(0, MSG_INFO,
-@@ -1217,17 +1246,29 @@
- }
-
- size_t memfree, memtotal;
-+#ifndef __cuda_cuda_h__
- retval = (*__cuMemGetInfo)(&memfree, &memtotal);
-+#else
-+ retval = cuMemGetInfo(&memfree, &memtotal);
-+#endif
- if (retval) {
- if (log_flags.coproc_debug) {
- msg_printf(0, MSG_INFO,
- "[coproc] cuMemGetInfo(%d) returned %d", device_num, retval
- );
- }
-+#ifndef __cuda_cuda_h__
- (*__cuCtxDestroy)(ctx);
-+#else
-+ cuCtxDestroy(ctx);
-+#endif
- return;
- }
-+#ifndef __cuda_cuda_h__
- (*__cuCtxDestroy)(ctx);
-+#else
-+ cuCtxDestroy(ctx);
-+#endif
- available_ram = (double) memfree;
- }
-
diff --git a/debian/patches/nvidia_direct_invocation_2.patch b/debian/patches/nvidia_direct_invocation_2.patch
deleted file mode 100644
index e72a08b..0000000
--- a/debian/patches/nvidia_direct_invocation_2.patch
+++ /dev/null
@@ -1,176 +0,0 @@
-Index: boinc/client/coproc_detect.cpp
-===================================================================
---- boinc.orig/client/coproc_detect.cpp 2012-05-03 14:01:48.736662216 +0200
-+++ boinc/client/coproc_detect.cpp 2012-05-03 14:03:36.391187731 +0200
-@@ -872,6 +872,7 @@
- CUDA_MF __cuMemFree = NULL;
- CUDA_MGI __cuMemGetInfo = NULL;
- #else
-+#ifndef __cuda_cuda_h__
- void* cudalib;
- CUresult (*__cuInit)(unsigned int);
- CUresult (*__cuDeviceGetCount)(int*);
-@@ -887,6 +888,7 @@
- CUresult (*__cuMemFree)(CUdeviceptr);
- CUresult (*__cuMemGetInfo)(size_t*, size_t*);
- #endif
-+#endif
-
-
- // NVIDIA interfaces are documented here:
-@@ -931,6 +933,7 @@
- #endif
- #else
-
-+#ifndef __cuda_cuda_h__
- #ifdef __APPLE__
- cudalib = dlopen("/usr/local/cuda/lib/libcuda.dylib", RTLD_NOW);
- #else
-@@ -942,7 +945,6 @@
- }
-
-
--#ifndef __cuda_cuda_h__
- __cuDeviceGetCount = (CUresult(*)(int*)) dlsym(cudalib, "cuDeviceGetCount");
- __cuDriverGetVersion = (CUresult(*)(int*)) dlsym( cudalib, "cuDriverGetVersion" );
- __cuInit = (CUresult(*)(unsigned int)) dlsym( cudalib, "cuInit" );
-@@ -957,6 +959,7 @@
- __cuMemFree = (CUresult(*)(CUdeviceptr)) dlsym( cudalib, "cuMemFree" );
- __cuMemGetInfo = (CUresult(*)(size_t*, size_t*)) dlsym( cudalib, "cuMemGetInfo" );
- #else
-+#if 0
- __cuDeviceGetCount = &cuDeviceGetCount;
- __cuDriverGetVersion = &cuDriverGetVersion;
- __cuInit = &cuInit;
-@@ -970,11 +973,12 @@
- __cuMemAlloc = &cuMemAlloc;
- __cuMemFree = &cuMemFree;
- __cuMemGetInfo = &cuMemGetInfo;
--#endif
--
-+#endif // #if 0
-+#endif // #ifndef __cuda_cuda_h__
-
--#endif // Win32
-+#endif //#ifdef Win32
-
-+#ifndef __cuda_cuda_h__
- if (!__cuDriverGetVersion) {
- warnings.push_back("cuDriverGetVersion() missing from NVIDIA library");
- return;
-@@ -1023,6 +1027,7 @@
- warnings.push_back("cuMemGetInfo() missing from NVIDIA library");
- return;
- }
-+#endif
-
- #ifndef __cuda_cuda_h__
- retval = (*__cuInit)(0);
-@@ -1095,25 +1100,61 @@
- warnings.push_back(buf);
- return;
- }
-- (*__cuDeviceComputeCapability)(&cc.prop.major, &cc.prop.minor, device);
-- (*__cuDeviceTotalMem)(&global_mem, device);
-+
-+#ifndef __cuda_cuda_h__
-+ retval = (*__cuDeviceComputeCapability)(&cc.prop.major, &cc.prop.minor, device);
-+#else
-+ retval = cuDeviceComputeCapability(&cc.prop.major, &cc.prop.minor, device);
-+#endif
-+ if (retval) {
-+ sprintf(buf, "cuDeviceGetName(%d) returned %d", j, retval);
-+ warnings.push_back(buf);
-+ return;
-+ }
-+
-+#ifndef __cuda_cuda_h__
-+ retval = (*__cuDeviceTotalMem)(&global_mem, device);
-+#else
-+ retval = cuDeviceTotalMem(&global_mem, device);
-+#endif
-+ if (retval) {
-+ sprintf(buf, "cuDeviceGetName(%d) returned %d", j, retval);
-+ warnings.push_back(buf);
-+ return;
-+ }
- cc.prop.totalGlobalMem = (double) global_mem;
-- (*__cuDeviceGetAttribute)(&cc.prop.sharedMemPerBlock, CU_DEVICE_ATTRIBUTE_SHARED_MEMORY_PER_BLOCK, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.regsPerBlock, CU_DEVICE_ATTRIBUTE_REGISTERS_PER_BLOCK, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.warpSize, CU_DEVICE_ATTRIBUTE_WARP_SIZE, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.memPitch, CU_DEVICE_ATTRIBUTE_MAX_PITCH, device);
-- retval = (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsPerBlock, CU_DEVICE_ATTRIBUTE_MAX_THREADS_PER_BLOCK, device);
-- retval = (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsDim[0], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_X, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsDim[1], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Y, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsDim[2], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Z, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.maxGridSize[0], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.maxGridSize[1], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Y, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.maxGridSize[2], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Z, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.clockRate, CU_DEVICE_ATTRIBUTE_CLOCK_RATE, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.totalConstMem, CU_DEVICE_ATTRIBUTE_TOTAL_CONSTANT_MEMORY, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.textureAlignment, CU_DEVICE_ATTRIBUTE_TEXTURE_ALIGNMENT, device);
-- (*__cuDeviceGetAttribute)(&cc.prop.deviceOverlap, CU_DEVICE_ATTRIBUTE_GPU_OVERLAP, device);
-- retval = (*__cuDeviceGetAttribute)(&cc.prop.multiProcessorCount, CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT, device);
-+#define ErrorPartialBlock(Name) \
-+ if (retval) { \
-+ sprintf(buf, "cuDeviceGetAttribute(%s,%d) returned %d", Name, j, retval); \
-+ warnings.push_back(buf); \
-+ return; \
-+ }
-+
-+#ifndef __cuda_cuda_h__
-+#define Block(Desc,Attrib,Name) retval=(*__cuDeviceGetAttribute)(Desc, Attrib, device); ErrorPartialBlock(Name);
-+#else
-+#define Block(Desc,Attrib,Name) retval=cuDeviceGetAttribute(Desc, Attrib, device); ErrorPartialBlock(Name);
-+#endif
-+
-+ Block(&cc.prop.sharedMemPerBlock,CU_DEVICE_ATTRIBUTE_SHARED_MEMORY_PER_BLOCK,"CU_DEVICE_ATTRIBUTE_SHARED_MEMORY_PER_BLOCK")
-+ Block(&cc.prop.regsPerBlock, CU_DEVICE_ATTRIBUTE_REGISTERS_PER_BLOCK, "CU_DEVICE_ATTRIBUTE_REGISTERS_PER_BLOCK");
-+ Block(&cc.prop.warpSize, CU_DEVICE_ATTRIBUTE_WARP_SIZE, "CU_DEVICE_ATTRIBUTE_WARP_SIZE");
-+ Block(&cc.prop.memPitch, CU_DEVICE_ATTRIBUTE_MAX_PITCH, "CU_DEVICE_ATTRIBUTE_MAX_PITCH");
-+ Block(&cc.prop.maxThreadsPerBlock, CU_DEVICE_ATTRIBUTE_MAX_THREADS_PER_BLOCK, "CU_DEVICE_ATTRIBUTE_MAX_THREADS_PER_BLOCK");
-+ Block(&cc.prop.maxThreadsDim[0], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_X, "CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_X");
-+ Block(&cc.prop.maxThreadsDim[1], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Y, "CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Y");
-+ Block(&cc.prop.maxThreadsDim[2], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Z, "CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Z");
-+ Block(&cc.prop.maxGridSize[0], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X, "CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X");
-+ Block(&cc.prop.maxGridSize[1], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Y, "CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X");
-+ Block(&cc.prop.maxGridSize[2], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Z, "CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Z");
-+ Block(&cc.prop.clockRate, CU_DEVICE_ATTRIBUTE_CLOCK_RATE, "CU_DEVICE_ATTRIBUTE_CLOCK_RATE");
-+ Block(&cc.prop.totalConstMem, CU_DEVICE_ATTRIBUTE_TOTAL_CONSTANT_MEMORY, "CU_DEVICE_ATTRIBUTE_TOTAL_CONSTANT_MEMORY");
-+ Block(&cc.prop.textureAlignment, CU_DEVICE_ATTRIBUTE_TEXTURE_ALIGNMENT, "CU_DEVICE_ATTRIBUTE_TEXTURE_ALIGNMENT");
-+ Block(&cc.prop.deviceOverlap, CU_DEVICE_ATTRIBUTE_GPU_OVERLAP, "CU_DEVICE_ATTRIBUTE_GPU_OVERLAP");
-+ Block(&cc.prop.multiProcessorCount, CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT, "CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT");
-+#undef Block
-+#undef ErrorPartialBlock
-+
- //retval = (*__cuDeviceGetProperties)(&cc.prop, device);
- if (cc.prop.major <= 0) continue; // major == 0 means emulation
- if (cc.prop.major > 100) continue; // e.g. 9999 is an error
-@@ -1142,7 +1183,7 @@
- // identify the most capable non-ignored instance
- //
- bool first = true;
-- for (i=0; i<nvidia_gpus.size(); i++) {
-+ for (unsigned int i=0; i<nvidia_gpus.size(); i++) {
- if (in_vector(nvidia_gpus[i].device_num, ignore_devs)) continue;
- if (first) {
- *this = nvidia_gpus[i];
-@@ -1156,7 +1197,7 @@
- // and set the "count" and "device_nums" fields
- //
- count = 0;
-- for (i=0; i<nvidia_gpus.size(); i++) {
-+ for (unsigned int i=0; i<nvidia_gpus.size(); i++) {
- if (in_vector(nvidia_gpus[i].device_num, ignore_devs)) {
- nvidia_gpus[i].is_used = COPROC_IGNORED;
- } else if (use_all || !nvidia_compare(nvidia_gpus[i], *this, true)) {
-@@ -1211,9 +1252,7 @@
- //
- void COPROC_NVIDIA::get_available_ram() {
- int retval;
-- size_t memfree, memtotal;
- int device;
-- void* ctx;
-
- available_ram = prop.totalGlobalMem;
- #ifndef __cuda_cuda_h__
diff --git a/debian/patches/series b/debian/patches/series
index 2524236..fcc94cf 100644
--- a/debian/patches/series
+++ b/debian/patches/series
@@ -26,8 +26,6 @@ debian_debian_AdjustBoincTopdirPython.patch
#or this one
#debian/patches/nvidia_detect_experiments.patch
some_extra_clang_warnings.patch
-#tmpA# nvidia_direct_invocation.patch
-#tmpA# nvidia_direct_invocation_2.patch
cppcheck_realloc.patch
MainDocumentWarnings.patch
more_maxpathlen.patch
--
Alioth's /usr/local/bin/git-commit-notice on /srv/git.debian.org/git/pkg-boinc/boinc.git
More information about the pkg-boinc-commits
mailing list