2016-12-04 21:44:34 -0600 | answered a question | SVM trainAuto v/s train You don't need to first use train. TrainAuto searches the SVM parameter space using cross validation. It doesn't search the parameter space (if you have any) of your features. Try stepping into trainAuto with a debugger to see what it gets you. I think the main problem is that trainAuto can still give you a local optimum. See the libsvm website for advice on a good approach. |
2016-12-04 21:40:09 -0600 | commented question | How to read images from a camera buffer See if you can build with gstreamer or add codecs for your OS. |
2016-12-04 21:35:30 -0600 | answered a question | build without highgui or gstreamer Check highgui.hpp: if there is nothing you need, then see if CMake allows you to not build highgui, if it doesn't, then just let CMake prepare the build files, then you can exclude the highgui module when compiling OpenCV but that will cause a lot of problems for you to solve, make sure to not build any example apps. It would be best to let CMake exclude highgui |
2016-12-04 21:29:19 -0600 | commented question | opencv with gpu crash in debug mode why not move to OpenCV3 ? Which CUDA version do you have? Only 1 CUDA? Driver up to date? |
2016-12-04 20:50:34 -0600 | commented question | Why is the CUDA version slower than the OpenCL version? I mistakenly merged the original channels, not the output channels. By reusing the input channels for output, this saves a little bit of initialization. I also see that my data only exists in one channel, this saves a little more. The cuda version now times in at the low 4. seconds. Still slower than OpenCL unfortunately. |
2016-12-04 19:28:12 -0600 | received badge | ● Editor (source) |
2016-12-04 19:21:58 -0600 | commented question | Why is the CUDA version slower than the OpenCL version? Do you mean, to get to know which function call is the slowest? this one stands out (~64ms). (I step through it repeatedly using the profiler of VS2015). It did occur to me that the CUDAFilters dll is 300MB, but the lookup time should be marginal. What else could it be? I don't think there is much reason for this normalize call to be exceptionally slow on CUDA8, I mean, it's an NVidia GPU and I try all three memory options. Same goes for the other functions compared to OpenCL1.2. I add a detailed timing image. |
2016-12-03 07:11:36 -0600 | asked a question | Why is the CUDA version slower than the OpenCL version? Hi, I have written a CUDA (8 on my machine) version of a program and compared it to an OpenCL(1.2) / T-API version. The former clocks in quite a bit slower even when using Unified Memory (UM). Could someone advise please? The normalize() function is multi-channel in the T-API, but underneath probably isn't. I had expected Shared Virtual Memory (UM in CUDA) to be faster, which I can't do with my PC because it is limited to OpenCL1.2... I read somewhere it can depend on the size or complexity of the filters, whether pixels are reread etc. but that would be the same for the CL version, wouldn't it? CUDA (5-6 sec.) : OpenCL (3-4 sec) Note. The profile image I made is no longer correct. The code here on the forum changed from my initial question. This code is the most optimized version without changing OpenCV3's source code. It must be the CPU GPU data transfers that take up nearly all the time spent. My test data were 2200x1600 images, |
2016-02-27 05:52:58 -0600 | asked a question | 3.1 samples CUDA, compiling with CUDA Hello, I have just started using OpenCV 3.1 (not the contrib branch) and haven't really kept up with OpenCv for a year or more. Now I have built and installed with CUDA: Unfortunately, many error related to using the GPU occur: Are there dependencies for the samples in /bin? (i.e. caltech images) opencv_test_cudaarithm : Similarly I have had to sudo But this still resulted in a large number of errors: :
:
cascadeclassifier_nvidia_api.cpp:(.text._ZNK9NCVVectorI21HaarClassifierNode128E9copySolidERS1_P11CUstream_stm[_ZNK9NCVVectorI21HaarClassifierNode128E9copySolidERS1_P11CUstream_stm]+0x258): undefined reference to |