multi.cpp 2.28 KB
Newer Older
luz.paz's avatar
luz.paz committed
1
/* This sample demonstrates the way you can perform independent tasks
2 3 4 5 6 7 8 9
   on the different GPUs */

// Disable some warnings which are caused with CUDA headers
#if defined(_MSC_VER)
#pragma warning(disable: 4201 4408 4100)
#endif

#include <iostream>
Ishant Mrinal Haloi's avatar
Ishant Mrinal Haloi committed
10
#include "opencv2/cvconfig.h"
11
#include "opencv2/core.hpp"
12
#include "opencv2/cudaarithm.hpp"
13

14
#ifdef HAVE_TBB
15 16 17 18
#  include "tbb/tbb.h"
#  include "tbb/task.h"
#  undef min
#  undef max
19 20
#endif

21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39
#if !defined(HAVE_CUDA) || !defined(HAVE_TBB)

int main()
{
#if !defined(HAVE_CUDA)
    std::cout << "CUDA support is required (CMake key 'WITH_CUDA' must be true).\n";
#endif

#if !defined(HAVE_TBB)
    std::cout << "TBB support is required (CMake key 'WITH_TBB' must be true).\n";
#endif

    return 0;
}

#else

using namespace std;
using namespace cv;
40
using namespace cv::cuda;
41 42 43 44 45 46 47 48 49 50 51 52 53

struct Worker { void operator()(int device_id) const; };

int main()
{
    int num_devices = getCudaEnabledDeviceCount();
    if (num_devices < 2)
    {
        std::cout << "Two or more GPUs are required\n";
        return -1;
    }
    for (int i = 0; i < num_devices; ++i)
    {
54
        cv::cuda::printShortCudaDeviceInfo(i);
55 56 57 58

        DeviceInfo dev_info(i);
        if (!dev_info.isCompatible())
        {
59
            std::cout << "CUDA module isn't built for GPU #" << i << " ("
60 61
                 << dev_info.name() << ", CC " << dev_info.majorVersion()
                 << dev_info.minorVersion() << "\n";
62 63 64 65 66 67
            return -1;
        }
    }

    // Execute calculation in two threads using two GPUs
    int devices[] = {0, 1};
68
    tbb::parallel_do(devices, devices + 2, Worker());
69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84

    return 0;
}


void Worker::operator()(int device_id) const
{
    setDevice(device_id);

    Mat src(1000, 1000, CV_32F);
    Mat dst;

    RNG rng(0);
    rng.fill(src, RNG::UNIFORM, 0, 1);

    // CPU works
85
    cv::transpose(src, dst);
86 87 88 89

    // GPU works
    GpuMat d_src(src);
    GpuMat d_dst;
90
    cuda::transpose(d_src, d_dst);
91 92

    // Check results
93
    bool passed = cv::norm(dst - Mat(d_dst), NORM_INF) < 1e-3;
94 95 96 97 98 99 100 101 102 103
    std::cout << "GPU #" << device_id << " (" << DeviceInfo().name() << "): "
        << (passed ? "passed" : "FAILED") << endl;

    // Deallocate data here, otherwise deallocation will be performed
    // after context is extracted from the stack
    d_src.release();
    d_dst.release();
}

#endif