- 01 Nov, 2019 1 commit
-
-
Alexander Alekhin authored
-
- 31 Oct, 2019 1 commit
-
-
Alexander Alekhin authored
-
- 30 Oct, 2019 2 commits
-
-
Anna Khakimova authored
-
atalaman authored
* Refactor render tests * Fix comment to review * Move ocv render tests stuff to specific file * Add OCV prefix for render tests * Fix comments to review
-
- 29 Oct, 2019 4 commits
-
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
Chip Kerchner authored
* Use circular lut hustory buffer in computeGradient of HOG * Initialize prefetch data outside main loop. Avoid code duplication.
-
- 28 Oct, 2019 4 commits
-
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
- 27 Oct, 2019 2 commits
-
-
Alexander Alekhin authored
-
André Lippok authored
-
- 26 Oct, 2019 2 commits
-
-
Alexander Alekhin authored
-
Alexander Alekhin authored
- OPENCV_TRACE_ITT_ENABLE
-
- 25 Oct, 2019 2 commits
-
-
Dmitry Budnikov authored
-
Alexander Alekhin authored
-
- 24 Oct, 2019 5 commits
-
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
Vitaly Tuzov authored
-
Alexander Alekhin authored
-
Alexander Alekhin authored
- move TLS & instrumentation code out of core/utility.hpp - (*) TLSData lost .gather() method (to dispose thread data on thread termination) - use TLSDataAccumulator for reliable collecting of thread data - prefer using of .detachData() + .cleanupDetachedData() instead of .gather() method (*) API is broken: replace TLSData => TLSDataAccumulator if gather required (objects disposal on threads termination is not available in accumulator mode)
-
- 23 Oct, 2019 6 commits
-
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
ChipKerchner authored
-
Alexander Alekhin authored
-
float13 authored
I think it would help to change all 3 of the the input file arguments to be "positional" for consistency with the other tutorials. This also simplifies the command line input to run this tutorial by reducing typing, and helpfully prints the "usage" info if any of the 3 required inputs are missing. I'm new to OpenCV and working through the tutorials. I kept getting runtime errors with this one until I realized that the arguments weren't positional, and I was missing the "--input1", "--input2, "--input3" flags preceding the filenames. All of the previous tutorials had required filenames as positional arguments and didn't require this. The original code would require each input to be specified like this: ./compareHist_Demo --input1 filename1 --input2 filename2 --input3 filename3 But with this change, the above command is simplified to: ./compareHist_Demo filename1 filename2 filename3 This avoids a confusing runtime error to make things simpler for newcomers like me :)
-
Anna Khakimova authored
-
- 22 Oct, 2019 7 commits
-
-
Alexander Alekhin authored
-
Ruslan Garnov authored
-
Chip Kerchner authored
Fixing bug with comparison of v_int64x2 or v_uint64x2 * Casting v_uint64x2 to v_float64x2 and comparing does NOT work in all cases. Rewrite using epi64 instructions - faster too. * Fix bad merge. * Fix equal comparsion for non-SSE4.1. Add test cases for v_int64x2 comparisons. * Try to fix merge conflict. * Only test v_int64x2 comparisons if CV_SIMD_64F * Fix compiler warning.
-
Alexander Alekhin authored
- gcc 4.8.4 (ARMv7)
-
Alexander Alekhin authored
-
Alexander Alekhin authored
-
nick black authored
The usage function states that the default for match_conf is 0.65 if the default SURF feature finder is used, and 0.3 for orbs. Indeed, if --feature orbs is used, match_conf is set to 0.3f. This is a NOP, because the real default is also set to 0.3f. Change it to 0.65f when SURF is in play.
-
- 21 Oct, 2019 3 commits
-
-
anton-potapov authored
* G-API: Doxygen documentatation for Async API * G-API: Doxygen documentatation for Async API - renamed local variable (reading parameter async) async -> asyncNumReq in object_detection DNN sample to avoid Doxygen erroneous linking the sample to cv::gapi::wip::async documentation
-
Dmitry Kurtaev authored
-
Yashas Samaga B L authored
CUDA backend for the DNN module * stub cuda4dnn design * minor fixes for tests and doxygen * add csl public api directory to module headers * add low-level CSL components * add high-level CSL components * integrate csl::Tensor into backbone code * switch to CPU iff unsupported; otherwise, fail on error * add fully connected layer * add softmax layer * add activation layers * support arbitary rank TensorDescriptor * pass input wrappers to `initCUDA()` * add 1d/2d/3d-convolution * add pooling layer * reorganize and refactor code * fixes for gcc, clang and doxygen; remove cxx14/17 code * add blank_layer * add LRN layer * add rounding modes for pooling layer * split tensor.hpp into tensor.hpp and tensor_ops.hpp * add concat layer * add scale layer * add batch normalization layer * split math.cu into activations.cu and math.hpp * add eltwise layer * add flatten layer * add tensor transform api * add asymmetric padding support for convolution layer * add reshape layer * fix rebase issues * add permute layer * add padding support for concat layer * refactor and reorganize code * add normalize layer * optimize bias addition in scale layer * add prior box layer * fix and optimize normalize layer * add asymmetric padding support for pooling layer * add event API * improve pooling performance for some padding scenarios * avoid over-allocation of compute resources to kernels * improve prior box performance * enable layer fusion * add const layer * add resize layer * add slice layer * add padding layer * add deconvolution layer * fix channelwise ReLU initialization * add vector traits * add vectorized versions of relu, clipped_relu, power * add vectorized concat kernels * improve concat_with_offsets performance * vectorize scale and bias kernels * add support for multi-billion element tensors * vectorize prior box kernels * fix address alignment check * improve bias addition performance of conv/deconv/fc layers * restructure code for supporting multiple targets * add DNN_TARGET_CUDA_FP64 * add DNN_TARGET_FP16 * improve vectorization * add region layer * improve tensor API, add dynamic ranks 1. use ManagedPtr instead of a Tensor in backend wrapper 2. add new methods to tensor classes - size_range: computes the combined size of for a given axis range - tensor span/view can be constructed from a raw pointer and shape 3. the tensor classes can change their rank at runtime (previously rank was fixed at compile-time) 4. remove device code from tensor classes (as they are unused) 5. enforce strict conditions on tensor class APIs to improve debugging ability * fix parametric relu activation * add squeeze/unsqueeze tensor API * add reorg layer * optimize permute and enable 2d permute * enable 1d and 2d slice * add split layer * add shuffle channel layer * allow tensors of different ranks in reshape primitive * patch SliceOp to allow Crop Layer * allow extra shape inputs in reshape layer * use `std::move_backward` instead of `std::move` for insert in resizable_static_array * improve workspace management * add spatial LRN * add nms (cpu) to region layer * add max pooling with argmax ( and a fix to limits.hpp) * add max unpooling layer * rename DNN_TARGET_CUDA_FP32 to DNN_TARGET_CUDA * update supportBackend to be more rigorous * remove stray include from preventing non-cuda build * include op_cuda.hpp outside condition #if * refactoring, fixes and many optimizations * drop DNN_TARGET_CUDA_FP64 * fix gcc errors * increase max. tensor rank limit to six * add Interp layer * drop custom layers; use BackendNode * vectorize activation kernels * fixes for gcc * remove wrong assertion * fix broken assertion in unpooling primitive * fix build errors in non-CUDA build * completely remove workspace from public API * fix permute layer * enable accuracy and perf. tests for DNN_TARGET_CUDA * add asynchronous forward * vectorize eltwise ops * vectorize fill kernel * fixes for gcc * remove CSL headers from public API * remove csl header source group from cmake * update min. cudnn version in cmake * add numerically stable FP32 log1pexp * refactor code * add FP16 specialization to cudnn based tensor addition * vectorize scale1 and bias1 + minor refactoring * fix doxygen build * fix invalid alignment assertion * clear backend wrappers before allocateLayers * ignore memory lock failures * do not allocate internal blobs * integrate NVTX * add numerically stable half precision log1pexp * fix indentation, following coding style, improve docs * remove accidental modification of IE code * Revert "add asynchronous forward" This reverts commit 1154b9da9da07e9b52f8a81bdcea48cf31c56f70. * [cmake] throw error for unsupported CC versions * fix rebase issues * add more docs, refactor code, fix bugs * minor refactoring and fixes * resolve warnings/errors from clang * remove haveCUDA() checks from supportBackend() * remove NVTX integration * changes based on review comments * avoid exception when no CUDA device is present * add color code for CUDA in Net::dump
-
- 20 Oct, 2019 1 commit
-
-
Alexander Alekhin authored
-