[boinc] 01/01: removed nvidia_direct_invocation[_2].patch

Guo Yixuan yixuan-guest at moszumanska.debian.org
Tue May 27 14:51:59 UTC 2014


This is an automated email from the git hooks/post-receive script.

yixuan-guest pushed a commit to branch master
in repository boinc.

commit e600098f7a5867ee5707880aa8253bad044219c4
Author: Guo Yixuan <culu.gyx at gmail.com>
Date:   Tue May 27 10:51:13 2014 -0400

    removed nvidia_direct_invocation[_2].patch
    
    the file client/coproc_detect.cpp was removed
---
 debian/patches/nvidia_direct_invocation.patch   | 117 ----------------
 debian/patches/nvidia_direct_invocation_2.patch | 176 ------------------------
 debian/patches/series                           |   2 -
 3 files changed, 295 deletions(-)

diff --git a/debian/patches/nvidia_direct_invocation.patch b/debian/patches/nvidia_direct_invocation.patch
deleted file mode 100644
index 6c8bc61..0000000
--- a/debian/patches/nvidia_direct_invocation.patch
+++ /dev/null
@@ -1,117 +0,0 @@
-Index: boinc/client/coproc_detect.cpp
-===================================================================
---- boinc.orig/client/coproc_detect.cpp	2012-05-03 14:01:32.344886723 +0200
-+++ boinc/client/coproc_detect.cpp	2012-05-03 14:01:48.736662216 +0200
-@@ -1024,21 +1024,33 @@
-         return;
-     }
- 
-+#ifndef __cuda_cuda_h__
-     retval = (*__cuInit)(0);
-+#else
-+    retval = cuInit(0);
-+#endif
-     if (retval) {
-         sprintf(buf, "NVIDIA drivers present but no GPUs found");
-         warnings.push_back(buf);
-         return;
-     }
- 
-+#ifndef __cuda_cuda_h__
-     retval = (*__cuDriverGetVersion)(&cuda_version);
-+#else
-+    retval = cuDriverGetVersion(&cuda_version);
-+#endif
-     if (retval) {
-         sprintf(buf, "cuDriverGetVersion() returned %d", retval);
-         warnings.push_back(buf);
-         return;
-     }
- 
-+#ifndef __cuda_cuda_h__
-     retval = (*__cuDeviceGetCount)(&cuda_ndevs);
-+#else
-+    retval = cuDeviceGetCount(&cuda_ndevs);
-+#endif
-     if (retval) {
-         sprintf(buf, "cuDeviceGetCount() returned %d", retval);
-         warnings.push_back(buf);
-@@ -1061,14 +1073,23 @@
-  nvidia_gpus.push_back(cc);
- }
- #endif
-+
-+#ifndef __cuda_cuda_h__
-         retval = (*__cuDeviceGet)(&device, j);
-+#else
-+        retval = cuDeviceGet(&device, j);
-+#endif
-         if (retval) {
-             sprintf(buf, "cuDeviceGet(%d) returned %d", j, retval);
-             warnings.push_back(buf);
-             return;
-         }
-         cc.prop.deviceHandle = device;
--        (*__cuDeviceGetName)(cc.prop.name, 256, device);
-+#ifndef __cuda_cuda_h__
-+        retval = (*__cuDeviceGetName)(cc.prop.name, 256, device);
-+#else
-+        retval = cuDeviceGetName(cc.prop.name, 256, device);
-+#endif
-         if (retval) {
-             sprintf(buf, "cuDeviceGetName(%d) returned %d", j, retval);
-             warnings.push_back(buf);
-@@ -1195,7 +1216,11 @@
-     void* ctx;
-     
-     available_ram = prop.totalGlobalMem;
-+#ifndef __cuda_cuda_h__
-     retval = (*__cuDeviceGet)(&device, device_num);
-+#else
-+    retval = cuDeviceGet(&device, device_num);
-+#endif
-     if (retval) {
-         if (log_flags.coproc_debug) {
-             msg_printf(0, MSG_INFO,
-@@ -1206,7 +1231,11 @@
-     }
- 
-     CUcontext ctx; // no immediate better idea
-+#ifndef __cuda_cuda_h__
-     retval = (*__cuCtxCreate)(&ctx, 0, device);
-+#else
-+    retval = cuCtxCreate(&ctx, 0, device);
-+#endif
-     if (retval) {
-         if (log_flags.coproc_debug) {
-             msg_printf(0, MSG_INFO,
-@@ -1217,17 +1246,29 @@
-     }
- 
-     size_t memfree, memtotal;
-+#ifndef __cuda_cuda_h__
-     retval = (*__cuMemGetInfo)(&memfree, &memtotal);
-+#else
-+    retval = cuMemGetInfo(&memfree, &memtotal);
-+#endif
-     if (retval) {
-         if (log_flags.coproc_debug) {
-             msg_printf(0, MSG_INFO,
-                 "[coproc] cuMemGetInfo(%d) returned %d", device_num, retval
-             );
-         }
-+#ifndef __cuda_cuda_h__
-         (*__cuCtxDestroy)(ctx);
-+#else
-+        cuCtxDestroy(ctx);
-+#endif
-         return;
-     }
-+#ifndef __cuda_cuda_h__
-     (*__cuCtxDestroy)(ctx);
-+#else
-+        cuCtxDestroy(ctx);
-+#endif
-     available_ram = (double) memfree;
- }
- 
diff --git a/debian/patches/nvidia_direct_invocation_2.patch b/debian/patches/nvidia_direct_invocation_2.patch
deleted file mode 100644
index e72a08b..0000000
--- a/debian/patches/nvidia_direct_invocation_2.patch
+++ /dev/null
@@ -1,176 +0,0 @@
-Index: boinc/client/coproc_detect.cpp
-===================================================================
---- boinc.orig/client/coproc_detect.cpp	2012-05-03 14:01:48.736662216 +0200
-+++ boinc/client/coproc_detect.cpp	2012-05-03 14:03:36.391187731 +0200
-@@ -872,6 +872,7 @@
- CUDA_MF __cuMemFree = NULL;
- CUDA_MGI __cuMemGetInfo = NULL;
- #else
-+#ifndef __cuda_cuda_h__
- void* cudalib;
- CUresult (*__cuInit)(unsigned int);
- CUresult (*__cuDeviceGetCount)(int*);
-@@ -887,6 +888,7 @@
- CUresult (*__cuMemFree)(CUdeviceptr);
- CUresult (*__cuMemGetInfo)(size_t*, size_t*);
- #endif
-+#endif
- 
- 
- // NVIDIA interfaces are documented here:
-@@ -931,6 +933,7 @@
- #endif
- #else
- 
-+#ifndef __cuda_cuda_h__
- #ifdef __APPLE__
-     cudalib = dlopen("/usr/local/cuda/lib/libcuda.dylib", RTLD_NOW);
- #else
-@@ -942,7 +945,6 @@
-     }
- 
- 
--#ifndef __cuda_cuda_h__
-     __cuDeviceGetCount = (CUresult(*)(int*)) dlsym(cudalib, "cuDeviceGetCount");
-     __cuDriverGetVersion = (CUresult(*)(int*)) dlsym( cudalib, "cuDriverGetVersion" );
-     __cuInit = (CUresult(*)(unsigned int)) dlsym( cudalib, "cuInit" );
-@@ -957,6 +959,7 @@
-     __cuMemFree = (CUresult(*)(CUdeviceptr)) dlsym( cudalib, "cuMemFree" );
-     __cuMemGetInfo = (CUresult(*)(size_t*, size_t*)) dlsym( cudalib, "cuMemGetInfo" );
- #else
-+#if 0
-     __cuDeviceGetCount = &cuDeviceGetCount;
-     __cuDriverGetVersion = &cuDriverGetVersion;
-     __cuInit = &cuInit;
-@@ -970,11 +973,12 @@
-     __cuMemAlloc = &cuMemAlloc;
-     __cuMemFree = &cuMemFree;
-     __cuMemGetInfo = &cuMemGetInfo;
--#endif
--
-+#endif  // #if 0
-+#endif  // #ifndef __cuda_cuda_h__
- 
--#endif	 // Win32
-+#endif	 //#ifdef Win32
- 
-+#ifndef __cuda_cuda_h__
-     if (!__cuDriverGetVersion) {
-         warnings.push_back("cuDriverGetVersion() missing from NVIDIA library");
-         return;
-@@ -1023,6 +1027,7 @@
-         warnings.push_back("cuMemGetInfo() missing from NVIDIA library");
-         return;
-     }
-+#endif
- 
- #ifndef __cuda_cuda_h__
-     retval = (*__cuInit)(0);
-@@ -1095,25 +1100,61 @@
-             warnings.push_back(buf);
-             return;
-         }
--        (*__cuDeviceComputeCapability)(&cc.prop.major, &cc.prop.minor, device);
--        (*__cuDeviceTotalMem)(&global_mem, device);
-+
-+#ifndef __cuda_cuda_h__
-+        retval = (*__cuDeviceComputeCapability)(&cc.prop.major, &cc.prop.minor, device);
-+#else
-+        retval = cuDeviceComputeCapability(&cc.prop.major, &cc.prop.minor, device);
-+#endif
-+        if (retval) {
-+            sprintf(buf, "cuDeviceGetName(%d) returned %d", j, retval);
-+            warnings.push_back(buf);
-+            return;
-+        }
-+
-+#ifndef __cuda_cuda_h__
-+        retval = (*__cuDeviceTotalMem)(&global_mem, device);
-+#else
-+        retval = cuDeviceTotalMem(&global_mem, device);
-+#endif
-+        if (retval) {
-+            sprintf(buf, "cuDeviceGetName(%d) returned %d", j, retval);
-+            warnings.push_back(buf);
-+            return;
-+        }
-         cc.prop.totalGlobalMem = (double) global_mem;
--        (*__cuDeviceGetAttribute)(&cc.prop.sharedMemPerBlock, CU_DEVICE_ATTRIBUTE_SHARED_MEMORY_PER_BLOCK, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.regsPerBlock, CU_DEVICE_ATTRIBUTE_REGISTERS_PER_BLOCK, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.warpSize, CU_DEVICE_ATTRIBUTE_WARP_SIZE, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.memPitch, CU_DEVICE_ATTRIBUTE_MAX_PITCH, device);
--        retval = (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsPerBlock, CU_DEVICE_ATTRIBUTE_MAX_THREADS_PER_BLOCK, device);
--        retval = (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsDim[0], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_X, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsDim[1], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Y, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.maxThreadsDim[2], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Z, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.maxGridSize[0], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.maxGridSize[1], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Y, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.maxGridSize[2], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Z, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.clockRate, CU_DEVICE_ATTRIBUTE_CLOCK_RATE, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.totalConstMem, CU_DEVICE_ATTRIBUTE_TOTAL_CONSTANT_MEMORY, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.textureAlignment, CU_DEVICE_ATTRIBUTE_TEXTURE_ALIGNMENT, device);
--        (*__cuDeviceGetAttribute)(&cc.prop.deviceOverlap, CU_DEVICE_ATTRIBUTE_GPU_OVERLAP, device);
--        retval = (*__cuDeviceGetAttribute)(&cc.prop.multiProcessorCount, CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT, device);
-+#define ErrorPartialBlock(Name) \
-+        if (retval) { \
-+            sprintf(buf, "cuDeviceGetAttribute(%s,%d) returned %d", Name, j, retval); \
-+            warnings.push_back(buf); \
-+            return; \
-+	}
-+
-+#ifndef __cuda_cuda_h__
-+#define Block(Desc,Attrib,Name) retval=(*__cuDeviceGetAttribute)(Desc, Attrib, device); ErrorPartialBlock(Name);
-+#else
-+#define Block(Desc,Attrib,Name) retval=cuDeviceGetAttribute(Desc, Attrib, device); ErrorPartialBlock(Name);
-+#endif
-+
-+        Block(&cc.prop.sharedMemPerBlock,CU_DEVICE_ATTRIBUTE_SHARED_MEMORY_PER_BLOCK,"CU_DEVICE_ATTRIBUTE_SHARED_MEMORY_PER_BLOCK")
-+        Block(&cc.prop.regsPerBlock, CU_DEVICE_ATTRIBUTE_REGISTERS_PER_BLOCK, "CU_DEVICE_ATTRIBUTE_REGISTERS_PER_BLOCK");
-+        Block(&cc.prop.warpSize, CU_DEVICE_ATTRIBUTE_WARP_SIZE, "CU_DEVICE_ATTRIBUTE_WARP_SIZE");
-+        Block(&cc.prop.memPitch, CU_DEVICE_ATTRIBUTE_MAX_PITCH, "CU_DEVICE_ATTRIBUTE_MAX_PITCH");
-+        Block(&cc.prop.maxThreadsPerBlock, CU_DEVICE_ATTRIBUTE_MAX_THREADS_PER_BLOCK, "CU_DEVICE_ATTRIBUTE_MAX_THREADS_PER_BLOCK");
-+        Block(&cc.prop.maxThreadsDim[0], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_X, "CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_X");
-+        Block(&cc.prop.maxThreadsDim[1], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Y, "CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Y");
-+        Block(&cc.prop.maxThreadsDim[2], CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Z, "CU_DEVICE_ATTRIBUTE_MAX_BLOCK_DIM_Z");
-+        Block(&cc.prop.maxGridSize[0], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X, "CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X");
-+        Block(&cc.prop.maxGridSize[1], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Y, "CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_X");
-+        Block(&cc.prop.maxGridSize[2], CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Z, "CU_DEVICE_ATTRIBUTE_MAX_GRID_DIM_Z");
-+        Block(&cc.prop.clockRate, CU_DEVICE_ATTRIBUTE_CLOCK_RATE, "CU_DEVICE_ATTRIBUTE_CLOCK_RATE");
-+        Block(&cc.prop.totalConstMem, CU_DEVICE_ATTRIBUTE_TOTAL_CONSTANT_MEMORY, "CU_DEVICE_ATTRIBUTE_TOTAL_CONSTANT_MEMORY");
-+        Block(&cc.prop.textureAlignment, CU_DEVICE_ATTRIBUTE_TEXTURE_ALIGNMENT, "CU_DEVICE_ATTRIBUTE_TEXTURE_ALIGNMENT");
-+        Block(&cc.prop.deviceOverlap, CU_DEVICE_ATTRIBUTE_GPU_OVERLAP, "CU_DEVICE_ATTRIBUTE_GPU_OVERLAP");
-+        Block(&cc.prop.multiProcessorCount, CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT, "CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT");
-+#undef Block
-+#undef ErrorPartialBlock
-+
-         //retval = (*__cuDeviceGetProperties)(&cc.prop, device);
-         if (cc.prop.major <= 0) continue;  // major == 0 means emulation
-         if (cc.prop.major > 100) continue;  // e.g. 9999 is an error
-@@ -1142,7 +1183,7 @@
-     // identify the most capable non-ignored instance
-     //
-     bool first = true;
--    for (i=0; i<nvidia_gpus.size(); i++) {
-+    for (unsigned int i=0; i<nvidia_gpus.size(); i++) {
-         if (in_vector(nvidia_gpus[i].device_num, ignore_devs)) continue;
-         if (first) {
-             *this = nvidia_gpus[i];
-@@ -1156,7 +1197,7 @@
-     // and set the "count" and "device_nums" fields
-     //
-     count = 0;
--    for (i=0; i<nvidia_gpus.size(); i++) {
-+    for (unsigned int i=0; i<nvidia_gpus.size(); i++) {
-         if (in_vector(nvidia_gpus[i].device_num, ignore_devs)) {
-             nvidia_gpus[i].is_used = COPROC_IGNORED;
-         } else if (use_all || !nvidia_compare(nvidia_gpus[i], *this, true)) {
-@@ -1211,9 +1252,7 @@
- //
- void COPROC_NVIDIA::get_available_ram() {
-     int retval;
--    size_t memfree, memtotal;
-     int device;
--    void* ctx;
-     
-     available_ram = prop.totalGlobalMem;
- #ifndef __cuda_cuda_h__
diff --git a/debian/patches/series b/debian/patches/series
index 2524236..fcc94cf 100644
--- a/debian/patches/series
+++ b/debian/patches/series
@@ -26,8 +26,6 @@ debian_debian_AdjustBoincTopdirPython.patch
 #or this one
 #debian/patches/nvidia_detect_experiments.patch
 some_extra_clang_warnings.patch
-#tmpA# nvidia_direct_invocation.patch
-#tmpA# nvidia_direct_invocation_2.patch
 cppcheck_realloc.patch
 MainDocumentWarnings.patch
 more_maxpathlen.patch

-- 
Alioth's /usr/local/bin/git-commit-notice on /srv/git.debian.org/git/pkg-boinc/boinc.git



More information about the pkg-boinc-commits mailing list