// This file is part of OpenCV project. // It is subject to the license terms in the LICENSE file found in the top-level directory // of this distribution and at http://opencv.org/license.html. // // Copyright (C) 2019-2021 Intel Corporation #include "../test_precomp.hpp" #include "../common/gapi_streaming_tests_common.hpp" #include // sleep_for (Delay) #include #include #include #include #include #include #include #include #include #include #include namespace opencv_test { namespace { enum class KernelPackage: int { OCV, OCV_FLUID, OCL, OCL_FLUID, }; std::ostream& operator<< (std::ostream &os, const KernelPackage &e) { switch (e) { #define _C(X) case KernelPackage::X: os << #X; break _C(OCV); _C(OCV_FLUID); _C(OCL); _C(OCL_FLUID); #undef _C default: GAPI_Assert(false); } return os; } struct GAPI_Streaming: public ::testing::TestWithParam>> { GAPI_Streaming() { KernelPackage pkg_kind; std::tie(pkg_kind, cap) = GetParam(); pkg = getKernelPackage(pkg_kind); } const cv::optional& getQueueCapacity() { return cap; } cv::GKernelPackage getKernelPackage(KernelPackage pkg_kind) { using namespace cv::gapi; switch (pkg_kind) { case KernelPackage::OCV: return cv::gapi::combine(core::cpu::kernels(), imgproc::cpu::kernels()); break; case KernelPackage::OCV_FLUID: return cv::gapi::combine(core::cpu::kernels(), imgproc::cpu::kernels(), core::fluid::kernels()); break; // FIXME: OpenCL backend seem to work fine with Streaming // however the results are not very bit exact with CPU // It may be a problem but may be just implementation innacuracy. // Need to customize the comparison function in tests where OpenCL // is involved. case KernelPackage::OCL: return cv::gapi::combine(core::ocl::kernels(), imgproc::ocl::kernels()); break; case KernelPackage::OCL_FLUID: return cv::gapi::combine(core::ocl::kernels(), imgproc::ocl::kernels(), core::fluid::kernels()); break; } throw std::logic_error("Unknown package"); } cv::GCompileArgs getCompileArgs() { using namespace cv::gapi; auto args = cv::compile_args(use_only{pkg}); if (cap) { args += cv::compile_args(cv::gapi::streaming::queue_capacity{cap.value()}); } return args; } cv::GKernelPackage pkg; cv::optional cap; }; G_API_OP(Delay, , "org.opencv.test.delay") { static cv::GMatDesc outMeta(const cv::GMatDesc &in, int) { return in; } }; GAPI_OCV_KERNEL(OCVDelay, Delay) { static void run(const cv::Mat &in, int ms, cv::Mat &out) { std::this_thread::sleep_for(std::chrono::milliseconds{ms}); in.copyTo(out); } }; class TestMediaBGR final: public cv::MediaFrame::IAdapter { cv::Mat m_mat; using Cb = cv::MediaFrame::View::Callback; Cb m_cb; public: explicit TestMediaBGR(cv::Mat m, Cb cb = [](){}) : m_mat(m), m_cb(cb) { } cv::GFrameDesc meta() const override { return cv::GFrameDesc{cv::MediaFormat::BGR, cv::Size(m_mat.cols, m_mat.rows)}; } cv::MediaFrame::View access(cv::MediaFrame::Access) override { cv::MediaFrame::View::Ptrs pp = { m_mat.ptr(), nullptr, nullptr, nullptr }; cv::MediaFrame::View::Strides ss = { m_mat.step, 0u, 0u, 0u }; return cv::MediaFrame::View(std::move(pp), std::move(ss), Cb{m_cb}); } }; class TestMediaNV12 final: public cv::MediaFrame::IAdapter { cv::Mat m_y; cv::Mat m_uv; public: TestMediaNV12(cv::Mat y, cv::Mat uv) : m_y(y), m_uv(uv) { } cv::GFrameDesc meta() const override { return cv::GFrameDesc{cv::MediaFormat::NV12, m_y.size()}; } cv::MediaFrame::View access(cv::MediaFrame::Access) override { cv::MediaFrame::View::Ptrs pp = { m_y.ptr(), m_uv.ptr(), nullptr, nullptr }; cv::MediaFrame::View::Strides ss = { m_y.step, m_uv.step, 0u, 0u }; return cv::MediaFrame::View(std::move(pp), std::move(ss)); } }; class TestMediaGRAY final : public cv::MediaFrame::IAdapter { cv::Mat m_mat; using Cb = cv::MediaFrame::View::Callback; Cb m_cb; public: explicit TestMediaGRAY(cv::Mat m, Cb cb = []() {}) : m_mat(m), m_cb(cb) { } cv::GFrameDesc meta() const override { return cv::GFrameDesc{ cv::MediaFormat::GRAY, cv::Size(m_mat.cols, m_mat.rows) }; } cv::MediaFrame::View access(cv::MediaFrame::Access) override { cv::MediaFrame::View::Ptrs pp = { m_mat.ptr(), nullptr, nullptr, nullptr }; cv::MediaFrame::View::Strides ss = { m_mat.step, 0u, 0u, 0u }; return cv::MediaFrame::View(std::move(pp), std::move(ss), Cb{ m_cb }); } }; class BGRSource : public cv::gapi::wip::GCaptureSource { public: explicit BGRSource(const std::string& pipeline) : cv::gapi::wip::GCaptureSource(pipeline) { } bool pull(cv::gapi::wip::Data& data) { if (cv::gapi::wip::GCaptureSource::pull(data)) { data = cv::MediaFrame::Create(cv::util::get(data)); return true; } return false; } GMetaArg descr_of() const override { return cv::GMetaArg{cv::GFrameDesc{cv::MediaFormat::BGR, cv::util::get( cv::gapi::wip::GCaptureSource::descr_of()).size}}; } }; void cvtBGR2NV12(const cv::Mat& bgr, cv::Mat& y, cv::Mat& uv) { cv::Size frame_sz = bgr.size(); cv::Size half_sz = frame_sz / 2; cv::Mat yuv; cv::cvtColor(bgr, yuv, cv::COLOR_BGR2YUV_I420); // Copy Y plane yuv.rowRange(0, frame_sz.height).copyTo(y); // Merge sampled U and V planes std::vector dims = {half_sz.height, half_sz.width}; auto start = frame_sz.height; auto range_h = half_sz.height/2; std::vector uv_planes = { yuv.rowRange(start, start + range_h) .reshape(0, dims), yuv.rowRange(start + range_h, start + range_h*2).reshape(0, dims) }; cv::merge(uv_planes, uv); } class NV12Source : public cv::gapi::wip::GCaptureSource { public: explicit NV12Source(const std::string& pipeline) : cv::gapi::wip::GCaptureSource(pipeline) { } bool pull(cv::gapi::wip::Data& data) { if (cv::gapi::wip::GCaptureSource::pull(data)) { cv::Mat bgr = cv::util::get(data); cv::Mat y, uv; cvtBGR2NV12(bgr, y, uv); data = cv::MediaFrame::Create(y, uv); return true; } return false; } GMetaArg descr_of() const override { return cv::GMetaArg{cv::GFrameDesc{cv::MediaFormat::NV12, cv::util::get( cv::gapi::wip::GCaptureSource::descr_of()).size}}; } }; class GRAYSource : public cv::gapi::wip::GCaptureSource { public: explicit GRAYSource(const std::string& pipeline) : cv::gapi::wip::GCaptureSource(pipeline) { } bool pull(cv::gapi::wip::Data& data) { if (cv::gapi::wip::GCaptureSource::pull(data)) { cv::Mat bgr = cv::util::get(data); cv::Mat gray; cvtColor(bgr, gray, cv::COLOR_BGR2GRAY); data = cv::MediaFrame::Create(gray); return true; } return false; } GMetaArg descr_of() const override { return cv::GMetaArg{ cv::GFrameDesc{cv::MediaFormat::GRAY, cv::util::get( cv::gapi::wip::GCaptureSource::descr_of()).size} }; } }; void checkPullOverload(const cv::Mat& ref, const bool has_output, cv::util::variant& args) { EXPECT_TRUE(has_output); using runArgs = cv::util::variant; cv::Mat out_mat; switch (args.index()) { case runArgs::index_of(): { auto outputs = util::get(args); EXPECT_EQ(1u, outputs.size()); out_mat = cv::util::get(outputs[0]); break; } case runArgs::index_of(): { auto outputs = util::get(args); EXPECT_EQ(1u, outputs.size()); auto opt_mat = cv::util::get>(outputs[0]); ASSERT_TRUE(opt_mat.has_value()); out_mat = *opt_mat; break; } default: GAPI_Assert(false && "Incorrect type of Args"); } EXPECT_EQ(0., cv::norm(ref, out_mat, cv::NORM_INF)); } } // anonymous namespace TEST_P(GAPI_Streaming, SmokeTest_ConstInput_GMat) { // This graph models the following use-case: // Canny here is used as some "feature detector" // // Island/device layout may be different given the contents // of the passed kernel package. // // The expectation is that we get as much islands in the // graph as backends the GKernelPackage contains. // // [Capture] --> Crop --> Resize --> Canny --> [out] const auto crop_rc = cv::Rect(13, 75, 377, 269); const auto resample_sz = cv::Size(224, 224); const auto thr_lo = 64.; const auto thr_hi = 192.; cv::GMat in; auto roi = cv::gapi::crop(in, crop_rc); auto res = cv::gapi::resize(roi, resample_sz); auto out = cv::gapi::Canny(res, thr_lo, thr_hi); cv::GComputation c(in, out); // Input data cv::Mat in_mat = cv::imread(findDataFile("cv/edgefilter/kodim23.png")); cv::Mat out_mat_gapi; // OpenCV reference image cv::Mat out_mat_ocv; { cv::Mat tmp; cv::resize(in_mat(crop_rc), tmp, resample_sz); cv::Canny(tmp, out_mat_ocv, thr_lo, thr_hi); } // Compilation & testing auto ccomp = c.compileStreaming(cv::descr_of(in_mat), getCompileArgs()); EXPECT_TRUE(ccomp); EXPECT_FALSE(ccomp.running()); ccomp.setSource(cv::gin(in_mat)); ccomp.start(); EXPECT_TRUE(ccomp.running()); // Fetch the result 15 times for (int i = 0; i < 15; i++) { // With constant inputs, the stream is endless so // the blocking pull() should never return `false`. EXPECT_TRUE(ccomp.pull(cv::gout(out_mat_gapi))); // Fluid's and OpenCV's Resizes aren't bit exact. // So 1% is here because it is max difference between them. EXPECT_TRUE(AbsSimilarPoints(0, 1).to_compare_f()(out_mat_gapi, out_mat_ocv)); } EXPECT_TRUE(ccomp.running()); ccomp.stop(); EXPECT_FALSE(ccomp.running()); } TEST_P(GAPI_Streaming, SmokeTest_VideoInput_GMat) { const auto crop_rc = cv::Rect(13, 75, 377, 269); const auto resample_sz = cv::Size(224, 224); const auto thr_lo = 64.; const auto thr_hi = 192.; cv::GMat in; auto roi = cv::gapi::crop(in, crop_rc); auto res = cv::gapi::resize(roi, resample_sz); auto out = cv::gapi::Canny(res, thr_lo, thr_hi); cv::GComputation c(cv::GIn(in), cv::GOut(cv::gapi::copy(in), out)); // OpenCV reference image code auto opencv_ref = [&](const cv::Mat &in_mat, cv::Mat &out_mat) { cv::Mat tmp; cv::resize(in_mat(crop_rc), tmp, resample_sz); cv::Canny(tmp, out_mat, thr_lo, thr_hi); }; // Compilation & testing auto ccomp = c.compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, getCompileArgs()); EXPECT_TRUE(ccomp); EXPECT_FALSE(ccomp.running()); auto path = findDataFile("cv/video/768x576.avi"); try { ccomp.setSource(gapi::wip::make_src(path)); } catch(...) { throw SkipTestException("Video file can not be opened"); } ccomp.start(); EXPECT_TRUE(ccomp.running()); // Process the full video cv::Mat in_mat_gapi, out_mat_gapi; std::size_t frames = 0u; while (ccomp.pull(cv::gout(in_mat_gapi, out_mat_gapi))) { frames++; cv::Mat out_mat_ocv; opencv_ref(in_mat_gapi, out_mat_ocv); // Fluid's and OpenCV's Resizes aren't bit exact. // So 1% is here because it is max difference between them. EXPECT_TRUE(AbsSimilarPoints(0, 1).to_compare_f()(out_mat_gapi, out_mat_ocv)); } EXPECT_LT(0u, frames); EXPECT_FALSE(ccomp.running()); // Stop can be called at any time (even if the pipeline is not running) ccomp.stop(); EXPECT_FALSE(ccomp.running()); } TEST_P(GAPI_Streaming, Regression_CompileTimeScalar) { // There was a bug with compile-time GScalars. Compile-time // GScalars generate their own DATA nodes at GModel/GIslandModel // level, resulting in an extra link at the GIslandModel level, so // GStreamingExecutor automatically assigned an input queue to // such edges. Since there were no in-graph producer for that // data, no data were pushed to such queue what lead to a // deadlock. cv::GMat in; cv::GMat tmp = cv::gapi::copy(in); for (int i = 0; i < 3; i++) { tmp = tmp & cv::gapi::blur(in, cv::Size(3,3)); } cv::GComputation c(cv::GIn(in), cv::GOut(tmp, tmp + 1)); auto ccomp = c.compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,512}}, getCompileArgs()); cv::Mat in_mat = cv::imread(findDataFile("cv/edgefilter/kodim23.png")); cv::Mat out_mat1, out_mat2; // Fetch the result 15 times ccomp.setSource(cv::gin(in_mat)); ccomp.start(); for (int i = 0; i < 15; i++) { EXPECT_TRUE(ccomp.pull(cv::gout(out_mat1, out_mat2))); } ccomp.stop(); } TEST_P(GAPI_Streaming, SmokeTest_StartRestart) { cv::GMat in; auto res = cv::gapi::resize(in, cv::Size{300,200}); auto out = cv::gapi::Canny(res, 95, 220); cv::GComputation c(cv::GIn(in), cv::GOut(cv::gapi::copy(in), out)); auto ccomp = c.compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, getCompileArgs()); EXPECT_TRUE(ccomp); EXPECT_FALSE(ccomp.running()); // Run 1 auto path = findDataFile("cv/video/768x576.avi"); std::size_t num_frames1 = 0u; try { ccomp.setSource(gapi::wip::make_src(path)); } catch(...) { throw SkipTestException("Video file can not be opened"); } ccomp.start(); EXPECT_TRUE(ccomp.running()); cv::Mat out1, out2; while (ccomp.pull(cv::gout(out1, out2))) num_frames1++; EXPECT_FALSE(ccomp.running()); // Run 2 std::size_t num_frames2 = 0u; try { ccomp.setSource(gapi::wip::make_src(path)); } catch(...) { throw SkipTestException("Video file can not be opened"); } ccomp.start(); EXPECT_TRUE(ccomp.running()); while (ccomp.pull(cv::gout(out1, out2))) num_frames2++; EXPECT_FALSE(ccomp.running()); EXPECT_LT(0u, num_frames1); EXPECT_LT(0u, num_frames2); EXPECT_EQ(num_frames1, num_frames2); } TEST_P(GAPI_Streaming, SmokeTest_VideoConstSource_NoHang) { // A video source is a finite one, while const source is not. // Check that pipeline completes when a video source completes. auto refc = cv::GComputation([](){ cv::GMat in; return cv::GComputation(in, cv::gapi::copy(in)); }).compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, getCompileArgs()); auto path = findDataFile("cv/video/768x576.avi"); try { refc.setSource(gapi::wip::make_src(path)); } catch(...) { throw SkipTestException("Video file can not be opened"); } refc.start(); std::size_t ref_frames = 0u; cv::Mat tmp; while (refc.pull(cv::gout(tmp))) ref_frames++; EXPECT_EQ(100u, ref_frames); cv::GMat in; cv::GMat in2; cv::GMat roi = cv::gapi::crop(in2, cv::Rect{1,1,256,256}); cv::GMat blr = cv::gapi::blur(roi, cv::Size(3,3)); cv::GMat out = blr - in; auto testc = cv::GComputation(cv::GIn(in, in2), cv::GOut(out)) .compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{256,256}}, cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, getCompileArgs()); cv::Mat in_const = cv::Mat::eye(cv::Size(256,256), CV_8UC3); testc.setSource(cv::gin(in_const, gapi::wip::make_src(path))); testc.start(); std::size_t test_frames = 0u; while (testc.pull(cv::gout(tmp))) test_frames++; EXPECT_EQ(ref_frames, test_frames); } TEST_P(GAPI_Streaming, SmokeTest_AutoMeta) { cv::GMat in; cv::GMat in2; cv::GMat roi = cv::gapi::crop(in2, cv::Rect{1,1,256,256}); cv::GMat blr = cv::gapi::blur(roi, cv::Size(3,3)); cv::GMat out = blr - in; auto testc = cv::GComputation(cv::GIn(in, in2), cv::GOut(out)) .compileStreaming(getCompileArgs()); cv::Mat in_const = cv::Mat::eye(cv::Size(256,256), CV_8UC3); cv::Mat tmp; // Test with one video source auto path = findDataFile("cv/video/768x576.avi"); try { testc.setSource(cv::gin(in_const, gapi::wip::make_src(path))); } catch(...) { throw SkipTestException("Video file can not be opened"); } testc.start(); std::size_t test_frames = 0u; while (testc.pull(cv::gout(tmp))) test_frames++; EXPECT_EQ(100u, test_frames); // Now test with another one path = findDataFile("cv/video/1920x1080.avi"); try { testc.setSource(cv::gin(in_const, gapi::wip::make_src(path))); } catch(...) { throw SkipTestException("Video file can not be opened"); } testc.start(); test_frames = 0u; while (testc.pull(cv::gout(tmp))) test_frames++; EXPECT_EQ(165u, test_frames); } TEST_P(GAPI_Streaming, SmokeTest_AutoMeta_2xConstMat) { cv::GMat in; cv::GMat in2; cv::GMat roi = cv::gapi::crop(in2, cv::Rect{1,1,256,256}); cv::GMat blr = cv::gapi::blur(roi, cv::Size(3,3)); cv::GMat out = blr - in; auto testc = cv::GComputation(cv::GIn(in, in2), cv::GOut(out)) .compileStreaming(getCompileArgs()); cv::Mat in_const = cv::Mat::eye(cv::Size(256,256), CV_8UC3); cv::Mat tmp; // Test with first image auto in_src = cv::imread(findDataFile("cv/edgefilter/statue.png")); testc.setSource(cv::gin(in_const, in_src)); testc.start(); ASSERT_TRUE(testc.pull(cv::gout(tmp))); testc.stop(); // Now test with second image in_src = cv::imread(findDataFile("cv/edgefilter/kodim23.png")); testc.setSource(cv::gin(in_const, in_src)); testc.start(); ASSERT_TRUE(testc.pull(cv::gout(tmp))); testc.stop(); } TEST_P(GAPI_Streaming, SmokeTest_AutoMeta_VideoScalar) { cv::GMat in_m; cv::GScalar in_s; cv::GMat out_m = in_m * in_s; auto testc = cv::GComputation(cv::GIn(in_m, in_s), cv::GOut(out_m)) .compileStreaming(getCompileArgs()); cv::Mat tmp; // Test with one video source and scalar auto path = findDataFile("cv/video/768x576.avi"); try { testc.setSource(cv::gin(gapi::wip::make_src(path), cv::Scalar{1.25})); } catch(...) { throw SkipTestException("Video file can not be opened"); } testc.start(); std::size_t test_frames = 0u; while (testc.pull(cv::gout(tmp))) test_frames++; EXPECT_EQ(100u, test_frames); // Now test with another one video source and scalar path = findDataFile("cv/video/1920x1080.avi"); try { testc.setSource(cv::gin(gapi::wip::make_src(path), cv::Scalar{0.75})); } catch(...) { throw SkipTestException("Video file can not be opened"); } testc.start(); test_frames = 0u; while (testc.pull(cv::gout(tmp))) test_frames++; EXPECT_EQ(165u, test_frames); } INSTANTIATE_TEST_CASE_P(TestStreaming, GAPI_Streaming, Combine(Values( KernelPackage::OCV //, KernelPackage::OCL // FIXME: Fails bit-exactness check, maybe relax it? , KernelPackage::OCV_FLUID //, KernelPackage::OCL // FIXME: Fails bit-exactness check, maybe relax it? ), Values(cv::optional{}, 1u, 4u)) ); namespace TypesTest { G_API_OP(SumV, (cv::GMat)>, "test.gapi.sumv") { static cv::GArrayDesc outMeta(const cv::GMatDesc &) { return cv::empty_array_desc(); } }; G_API_OP(AddV, )>, "test.gapi.addv") { static cv::GMatDesc outMeta(const cv::GMatDesc &in, const cv::GArrayDesc &) { return in; } }; GAPI_OCV_KERNEL(OCVSumV, SumV) { static void run(const cv::Mat &in, std::vector &out) { CV_Assert(in.depth() == CV_8U); const auto length = in.cols * in.channels(); out.resize(length); const uchar *ptr = in.ptr(0); for (int c = 0; c < length; c++) { out[c] = ptr[c]; } for (int r = 1; r < in.rows; r++) { ptr = in.ptr(r); for (int c = 0; c < length; c++) { out[c] += ptr[c]; } } } }; GAPI_OCV_KERNEL(OCVAddV, AddV) { static void run(const cv::Mat &in, const std::vector &inv, cv::Mat &out) { CV_Assert(in.depth() == CV_8U); const auto length = in.cols * in.channels(); CV_Assert(length == static_cast(inv.size())); for (int r = 0; r < in.rows; r++) { const uchar *in_ptr = in.ptr(r); uchar *out_ptr = out.ptr(r); for (int c = 0; c < length; c++) { out_ptr[c] = cv::saturate_cast(in_ptr[c] + inv[c]); } } } }; GAPI_FLUID_KERNEL(FluidAddV, AddV, false) { static const int Window = 1; static void run(const cv::gapi::fluid::View &in, const std::vector &inv, cv::gapi::fluid::Buffer &out) { const uchar *in_ptr = in.InLineB(0); uchar *out_ptr = out.OutLineB(0); const auto length = in.meta().size.width * in.meta().chan; CV_Assert(length == static_cast(inv.size())); for (int c = 0; c < length; c++) { out_ptr[c] = cv::saturate_cast(in_ptr[c] + inv[c]); } } }; } // namespace TypesTest TEST_P(GAPI_Streaming, SmokeTest_AutoMeta_VideoArray) { cv::GMat in_m; cv::GArray in_v; cv::GMat out_m = TypesTest::AddV::on(in_m, in_v) - in_m; // Run pipeline auto args = cv::compile_args(cv::gapi::kernels()); auto capacity = getQueueCapacity(); if (capacity) { args += cv::compile_args( cv::gapi::streaming::queue_capacity{capacity.value()}); } auto testc = cv::GComputation(cv::GIn(in_m, in_v), cv::GOut(out_m)) .compileStreaming(std::move(args)); cv::Mat tmp; // Test with one video source and vector auto path = findDataFile("cv/video/768x576.avi"); std::vector first_in_vec(768*3, 1); try { testc.setSource(cv::gin(gapi::wip::make_src(path), first_in_vec)); } catch(...) { throw SkipTestException("Video file can not be opened"); } testc.start(); std::size_t test_frames = 0u; while (testc.pull(cv::gout(tmp))) test_frames++; EXPECT_EQ(100u, test_frames); // Now test with another one path = findDataFile("cv/video/1920x1080.avi"); std::vector second_in_vec(1920*3, 1); try { testc.setSource(cv::gin(gapi::wip::make_src(path), second_in_vec)); } catch(...) { throw SkipTestException("Video file can not be opened"); } testc.start(); test_frames = 0u; while (testc.pull(cv::gout(tmp))) test_frames++; EXPECT_EQ(165u, test_frames); } TEST(GAPI_Streaming_Types, InputScalar) { // This test verifies if Streaming works with Scalar data @ input. cv::GMat in_m; cv::GScalar in_s; cv::GMat out_m = in_m * in_s; cv::GComputation c(cv::GIn(in_m, in_s), cv::GOut(out_m)); // Input data cv::Mat in_mat = cv::Mat::eye(256, 256, CV_8UC1); cv::Scalar in_scl = 32; // Run pipeline auto sc = c.compileStreaming(cv::descr_of(in_mat), cv::descr_of(in_scl)); sc.setSource(cv::gin(in_mat, in_scl)); sc.start(); for (int i = 0; i < 10; i++) { cv::Mat out; EXPECT_TRUE(sc.pull(cv::gout(out))); EXPECT_EQ(0., cv::norm(out, in_mat.mul(in_scl), cv::NORM_INF)); } } TEST(GAPI_Streaming_Types, InputVector) { // This test verifies if Streaming works with Vector data @ input. cv::GMat in_m; cv::GArray in_v; cv::GMat out_m = TypesTest::AddV::on(in_m, in_v) - in_m; cv::GComputation c(cv::GIn(in_m, in_v), cv::GOut(out_m)); // Input data cv::Mat in_mat = cv::Mat::eye(256, 256, CV_8UC1); std::vector in_vec; TypesTest::OCVSumV::run(in_mat, in_vec); EXPECT_EQ(std::vector(256,1), in_vec); // self-sanity-check auto opencv_ref = [&](const cv::Mat &in, const std::vector &inv, cv::Mat &out) { cv::Mat tmp = in_mat.clone(); // allocate the same amount of memory as graph does TypesTest::OCVAddV::run(in, inv, tmp); out = tmp - in; }; // Run pipeline auto sc = c.compileStreaming(cv::descr_of(in_mat), cv::descr_of(in_vec), cv::compile_args(cv::gapi::kernels())); sc.setSource(cv::gin(in_mat, in_vec)); sc.start(); for (int i = 0; i < 10; i++) { cv::Mat out_mat; EXPECT_TRUE(sc.pull(cv::gout(out_mat))); cv::Mat ref_mat; opencv_ref(in_mat, in_vec, ref_mat); EXPECT_EQ(0., cv::norm(ref_mat, out_mat, cv::NORM_INF)); } } TEST(GAPI_Streaming_Types, XChangeScalar) { // This test verifies if Streaming works when pipeline steps // (islands) exchange Scalar data. cv::GMat in; cv::GScalar m = cv::gapi::mean(in); cv::GMat tmp = cv::gapi::convertTo(in, CV_32F) - m; cv::GMat out = cv::gapi::blur(tmp, cv::Size(3,3)); cv::GComputation c(cv::GIn(in), cv::GOut(cv::gapi::copy(in), cv::gapi::convertTo(out, CV_8U))); auto ocv_ref = [](const cv::Mat &in_mat, cv::Mat &out_mat) { cv::Scalar ocv_m = cv::mean(in_mat); cv::Mat ocv_tmp; in_mat.convertTo(ocv_tmp, CV_32F); ocv_tmp -= ocv_m; cv::blur(ocv_tmp, ocv_tmp, cv::Size(3,3)); ocv_tmp.convertTo(out_mat, CV_8U); }; // Here we want mean & convertTo run on OCV // and subC & blur3x3 on Fluid. // FIXME: With the current API it looks quite awful: auto ocv_kernels = cv::gapi::core::cpu::kernels(); // convertTo ocv_kernels.remove(); auto fluid_kernels = cv::gapi::combine(cv::gapi::core::fluid::kernels(), // subC cv::gapi::imgproc::fluid::kernels()); // box3x3 fluid_kernels.remove(); fluid_kernels.remove(); // FIXME: Now // - fluid kernels take over ocv kernels (including Copy, SubC, & Box3x3) // - selected kernels (which were removed from the fluid package) remain in OCV // (ConvertTo + some others) // FIXME: This is completely awful. User should easily pick up specific kernels // to an empty kernel package to craft his own but not do it via exclusion. // Need to expose kernel declarations to public headers to enable kernels<..>() // on user side. auto kernels = cv::gapi::combine(ocv_kernels, fluid_kernels); // Compile streaming pipeline auto sc = c.compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, cv::compile_args(cv::gapi::use_only{kernels})); auto path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(gapi::wip::make_src(path)); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); cv::Mat in_frame; cv::Mat out_mat_gapi; cv::Mat out_mat_ref; std::size_t num_frames = 0u; while (sc.pull(cv::gout(in_frame, out_mat_gapi))) { num_frames++; ocv_ref(in_frame, out_mat_ref); EXPECT_EQ(0., cv::norm(out_mat_gapi, out_mat_ref, cv::NORM_INF)); } EXPECT_LT(0u, num_frames); } TEST(GAPI_Streaming_Types, XChangeVector) { // This test verifies if Streaming works when pipeline steps // (islands) exchange Vector data. cv::GMat in1, in2; cv::GMat in = cv::gapi::crop(in1, cv::Rect{0,0,576,576}); cv::GScalar m = cv::gapi::mean(in); cv::GArray s = TypesTest::SumV::on(in2); // (in2 = eye, so s = [1,0,0,1,..]) cv::GMat out = TypesTest::AddV::on(in - m, s); cv::GComputation c(cv::GIn(in1, in2), cv::GOut(cv::gapi::copy(in), out)); auto ocv_ref = [](const cv::Mat &in_mat1, const cv::Mat &in_mat2, cv::Mat &out_mat) { cv::Mat in_roi = in_mat1(cv::Rect{0,0,576,576}); cv::Scalar ocv_m = cv::mean(in_roi); std::vector ocv_v; TypesTest::OCVSumV::run(in_mat2, ocv_v); out_mat.create(cv::Size(576,576), CV_8UC3); cv::Mat in_tmp = in_roi - ocv_m; TypesTest::OCVAddV::run(in_tmp, ocv_v, out_mat); }; // Let crop/mean/sumV be calculated via OCV, // and AddV/subC be calculated via Fluid auto ocv_kernels = cv::gapi::core::cpu::kernels(); ocv_kernels.remove(); ocv_kernels.include(); auto fluid_kernels = cv::gapi::core::fluid::kernels(); fluid_kernels.include(); // Here OCV takes precedense over Fluid, with SubC & SumV remaining // in Fluid. auto kernels = cv::gapi::combine(fluid_kernels, ocv_kernels); // Compile streaming pipeline cv::Mat in_eye = cv::Mat::eye(cv::Size(576, 576), CV_8UC3); auto sc = c.compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, cv::GMatDesc{CV_8U,3,cv::Size{576,576}}, cv::compile_args(cv::gapi::use_only{kernels})); auto path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(cv::gin(gapi::wip::make_src(path), in_eye)); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); cv::Mat in_frame; cv::Mat out_mat_gapi; cv::Mat out_mat_ref; std::size_t num_frames = 0u; while (sc.pull(cv::gout(in_frame, out_mat_gapi))) { num_frames++; ocv_ref(in_frame, in_eye, out_mat_ref); EXPECT_EQ(0., cv::norm(out_mat_gapi, out_mat_ref, cv::NORM_INF)); } EXPECT_LT(0u, num_frames); } TEST(GAPI_Streaming_Types, OutputScalar) { // This test verifies if Streaming works when pipeline // produces scalar data only cv::GMat in; cv::GScalar out = cv::gapi::mean(in); auto sc = cv::GComputation(cv::GIn(in), cv::GOut(out)) .compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,576}}); std::string video_path; video_path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(gapi::wip::make_src(video_path)); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); cv::VideoCapture cap; cap.open(video_path); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::Mat tmp; cv::Scalar out_scl; std::size_t num_frames = 0u; while (sc.pull(cv::gout(out_scl))) { num_frames++; cap >> tmp; cv::Scalar out_ref = cv::mean(tmp); EXPECT_EQ(out_ref, out_scl); } EXPECT_LT(0u, num_frames); } TEST(GAPI_Streaming_Types, OutputVector) { // This test verifies if Streaming works when pipeline // produces vector data only auto pkg = cv::gapi::kernels(); cv::GMat in1, in2; cv::GMat roi = cv::gapi::crop(in2, cv::Rect(3,3,256,256)); cv::GArray out = TypesTest::SumV::on(cv::gapi::mul(roi, in1)); auto sc = cv::GComputation(cv::GIn(in1, in2), cv::GOut(out)) .compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{256,256}}, cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, cv::compile_args(pkg)); auto ocv_ref = [](const cv::Mat &ocv_in1, const cv::Mat &ocv_in2, std::vector &ocv_out) { auto ocv_roi = ocv_in2(cv::Rect{3,3,256,256}); TypesTest::OCVSumV::run(ocv_roi.mul(ocv_in1), ocv_out); }; cv::Mat in_eye = cv::Mat::eye(cv::Size(256, 256), CV_8UC3); std::string video_path; video_path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(cv::gin(in_eye, gapi::wip::make_src(video_path))); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); cv::VideoCapture cap; cap.open(video_path); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::Mat tmp; std::vector ref_vec; std::vector out_vec; std::size_t num_frames = 0u; while (sc.pull(cv::gout(out_vec))) { num_frames++; cap >> tmp; ref_vec.clear(); ocv_ref(in_eye, tmp, ref_vec); EXPECT_EQ(ref_vec, out_vec); } EXPECT_LT(0u, num_frames); } G_API_OP(DimsChans, , cv::GOpaque>(cv::GMat)>, "test.streaming.dims_chans") { static std::tuple outMeta(const cv::GMatDesc &) { return std::make_tuple(cv::empty_array_desc(), cv::empty_gopaque_desc()); } }; GAPI_OCV_KERNEL(OCVDimsChans, DimsChans) { static void run(const cv::Mat &in, std::vector &ov, int &oi) { ov = {in.cols, in.rows}; oi = in.channels(); } }; struct GAPI_Streaming_TemplateTypes: ::testing::Test { // There was a problem in GStreamingExecutor // when outputs were formally not used by the graph // but still should be in place as operation need // to produce them, and host data type constructors // were missing for GArray and GOpaque in this case. // This test tests exactly this. GAPI_Streaming_TemplateTypes() { // Prepare everything for the test: // Graph itself blur = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat blur_d = cv::gapi::streaming::desync(blur); std::tie(vec, opq) = DimsChans::on(blur_d); // Kernel package pkg = cv::gapi::kernels(); // Input mat in_mat = cv::Mat::eye(cv::Size(320,240), CV_8UC3); } cv::GMat in; cv::GMat blur; cv::GArray vec; cv::GOpaque opq; cv::GKernelPackage pkg; cv::Mat in_mat; }; TEST_F(GAPI_Streaming_TemplateTypes, UnusedVectorIsOK) { // Declare graph without listing vec as output auto sc = cv::GComputation(cv::GIn(in), cv::GOut(blur, opq)) .compileStreaming(cv::compile_args(pkg)); sc.setSource(cv::gin(in_mat)); sc.start(); cv::optional out_mat; cv::optional out_int; int counter = 0; while (sc.pull(cv::gout(out_mat, out_int))) { if (counter++ == 10) { // Stop the test after 10 iterations sc.stop(); break; } GAPI_Assert(out_mat || out_int); if (out_int) { EXPECT_EQ(3, out_int.value()); } } } TEST_F(GAPI_Streaming_TemplateTypes, UnusedOpaqueIsOK) { // Declare graph without listing opq as output auto sc = cv::GComputation(cv::GIn(in), cv::GOut(blur, vec)) .compileStreaming(cv::compile_args(pkg)); sc.setSource(cv::gin(in_mat)); sc.start(); cv::optional out_mat; cv::optional > out_vec; int counter = 0; while (sc.pull(cv::gout(out_mat, out_vec))) { if (counter++ == 10) { // Stop the test after 10 iterations sc.stop(); break; } GAPI_Assert(out_mat || out_vec); if (out_vec) { EXPECT_EQ(320, out_vec.value()[0]); EXPECT_EQ(240, out_vec.value()[1]); } } } struct GAPI_Streaming_Unit: public ::testing::Test { cv::Mat m; cv::GComputation cc; cv::GStreamingCompiled sc; cv::GCompiled ref; GAPI_Streaming_Unit() : m(cv::Mat::ones(224,224,CV_8UC3)) , cc([]{ cv::GMat a, b; cv::GMat c = a + b*2; return cv::GComputation(cv::GIn(a, b), cv::GOut(c)); }) { const auto a_desc = cv::descr_of(m); const auto b_desc = cv::descr_of(m); sc = cc.compileStreaming(a_desc, b_desc); ref = cc.compile(a_desc, b_desc); } }; // FIXME: (GAPI_Streaming_Types, InputOpaque) test is missing here! // FIXME: (GAPI_Streaming_Types, XChangeOpaque) test is missing here! // FIXME: (GAPI_Streaming_Types, OutputOpaque) test is missing here! TEST(GAPI_Streaming, TestTwoVideosDifferentLength) { auto desc = cv::GMatDesc{CV_8U,3,{768,576}}; auto path1 = findDataFile("cv/video/768x576.avi"); auto path2 = findDataFile("highgui/video/big_buck_bunny.avi"); cv::GMat in1, in2; auto out = in1 + cv::gapi::resize(in2, desc.size); cv::GComputation cc(cv::GIn(in1, in2), cv::GOut(out)); auto sc = cc.compileStreaming(); try { sc.setSource(cv::gin(gapi::wip::make_src(path1), gapi::wip::make_src(path2))); } catch(...) { throw SkipTestException("Video file can not be found"); } sc.start(); cv::Mat out_mat; std::size_t frames = 0u; while(sc.pull(cv::gout(out_mat))) { frames++; } // big_buck_bunny.avi has 125 frames, 768x576.avi - 100 frames, // expect framework to stop after 100 frames EXPECT_EQ(100u, frames); } TEST_F(GAPI_Streaming_Unit, TestStartWithoutnSetSource) { EXPECT_ANY_THROW(sc.start()); } TEST_F(GAPI_Streaming_Unit, TestStopWithoutStart1) { // It is ok! EXPECT_NO_THROW(sc.stop()); } TEST_F(GAPI_Streaming_Unit, TestStopWithoutStart2) { // It should be ok as well sc.setSource(cv::gin(m, m)); EXPECT_NO_THROW(sc.stop()); } TEST_F(GAPI_Streaming_Unit, StopStartStop) { cv::Mat out; EXPECT_NO_THROW(sc.stop()); EXPECT_NO_THROW(sc.setSource(cv::gin(m, m))); EXPECT_NO_THROW(sc.start()); std::size_t i = 0u; while (i++ < 10u) {EXPECT_TRUE(sc.pull(cv::gout(out)));}; EXPECT_NO_THROW(sc.stop()); } TEST_F(GAPI_Streaming_Unit, ImplicitStop) { EXPECT_NO_THROW(sc.setSource(cv::gin(m, m))); EXPECT_NO_THROW(sc.start()); // No explicit stop here - pipeline stops successfully at the test exit } TEST_F(GAPI_Streaming_Unit, StartStopStart_NoSetSource) { EXPECT_NO_THROW(sc.setSource(cv::gin(m, m))); EXPECT_NO_THROW(sc.start()); EXPECT_NO_THROW(sc.stop()); EXPECT_ANY_THROW(sc.start()); // Should fail since setSource was not called } TEST_F(GAPI_Streaming_Unit, StartStopStress_Const) { // Runs 100 times with no deadlock - assumed stable (robust) enough for (int i = 0; i < 100; i++) { sc.stop(); sc.setSource(cv::gin(m, m)); sc.start(); cv::Mat out; for (int j = 0; j < 5; j++) EXPECT_TRUE(sc.pull(cv::gout(out))); } } TEST_F(GAPI_Streaming_Unit, StartStopStress_Video) { // Runs 100 times with no deadlock - assumed stable (robust) enough sc = cc.compileStreaming(cv::GMatDesc{CV_8U,3,cv::Size{768,576}}, cv::GMatDesc{CV_8U,3,cv::Size{768,576}}); m = cv::Mat::eye(cv::Size{768,576}, CV_8UC3); auto path = findDataFile("cv/video/768x576.avi"); for (int i = 0; i < 100; i++) { sc.stop(); try { sc.setSource(cv::gin(cv::gapi::wip::make_src(path), m)); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); cv::Mat out; for (int j = 0; j < 5; j++) EXPECT_TRUE(sc.pull(cv::gout(out))); } } TEST_F(GAPI_Streaming_Unit, PullNoStart) { sc.setSource(cv::gin(m, m)); cv::Mat out; EXPECT_ANY_THROW(sc.pull(cv::gout(out))); } TEST_F(GAPI_Streaming_Unit, SetSource_Multi_BeforeStart) { cv::Mat eye = cv::Mat::eye (224, 224, CV_8UC3); cv::Mat zrs = cv::Mat::zeros(224, 224, CV_8UC3); // Call setSource two times, data specified last time // should be actually processed. sc.setSource(cv::gin(zrs, zrs)); sc.setSource(cv::gin(eye, eye)); // Run the pipeline, acquire result once sc.start(); cv::Mat out, out_ref; EXPECT_TRUE(sc.pull(cv::gout(out))); sc.stop(); // Pipeline should process `eye` mat, not `zrs` ref(cv::gin(eye, eye), cv::gout(out_ref)); EXPECT_EQ(0., cv::norm(out, out_ref, cv::NORM_INF)); } TEST_F(GAPI_Streaming_Unit, SetSource_During_Execution) { cv::Mat zrs = cv::Mat::zeros(224, 224, CV_8UC3); sc.setSource(cv::gin(m, m)); sc.start(); EXPECT_ANY_THROW(sc.setSource(cv::gin(zrs, zrs))); EXPECT_ANY_THROW(sc.setSource(cv::gin(zrs, zrs))); EXPECT_ANY_THROW(sc.setSource(cv::gin(zrs, zrs))); sc.stop(); } TEST_F(GAPI_Streaming_Unit, SetSource_After_Completion) { sc.setSource(cv::gin(m, m)); // Test pipeline with `m` input sc.start(); cv::Mat out, out_ref; EXPECT_TRUE(sc.pull(cv::gout(out))); sc.stop(); // Test against ref ref(cv::gin(m, m), cv::gout(out_ref)); EXPECT_EQ(0., cv::norm(out, out_ref, cv::NORM_INF)); // Now set another source cv::Mat eye = cv::Mat::eye(224, 224, CV_8UC3); sc.setSource(cv::gin(eye, m)); sc.start(); EXPECT_TRUE(sc.pull(cv::gout(out))); sc.stop(); // Test against new ref ref(cv::gin(eye, m), cv::gout(out_ref)); EXPECT_EQ(0., cv::norm(out, out_ref, cv::NORM_INF)); } // NB: Check pull overload for python TEST(Streaming, Python_Pull_Overload) { cv::GMat in; auto out = cv::gapi::copy(in); cv::GComputation c(in, out); cv::Size sz(3,3); cv::Mat in_mat(sz, CV_8UC3); cv::randu(in_mat, cv::Scalar::all(0), cv::Scalar(255)); auto ccomp = c.compileStreaming(); EXPECT_TRUE(ccomp); EXPECT_FALSE(ccomp.running()); ccomp.setSource(cv::gin(in_mat)); ccomp.start(); EXPECT_TRUE(ccomp.running()); bool has_output; cv::GRunArgs outputs; using RunArgs = cv::util::variant; RunArgs args; std::tie(has_output, args) = ccomp.pull(); checkPullOverload(in_mat, has_output, args); ccomp.stop(); EXPECT_FALSE(ccomp.running()); } TEST(GAPI_Streaming_Desync, Python_Pull_Overload) { cv::GMat in; cv::GMat out = cv::gapi::streaming::desync(in); cv::GComputation c(in, out); cv::Size sz(3,3); cv::Mat in_mat(sz, CV_8UC3); cv::randu(in_mat, cv::Scalar::all(0), cv::Scalar(255)); auto ccomp = c.compileStreaming(); EXPECT_TRUE(ccomp); EXPECT_FALSE(ccomp.running()); ccomp.setSource(cv::gin(in_mat)); ccomp.start(); EXPECT_TRUE(ccomp.running()); bool has_output; cv::GRunArgs outputs; using RunArgs = cv::util::variant; RunArgs args; std::tie(has_output, args) = ccomp.pull(); checkPullOverload(in_mat, has_output, args); ccomp.stop(); EXPECT_FALSE(ccomp.running()); } TEST(GAPI_Streaming_Desync, SmokeTest_Regular) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat out1 = cv::gapi::Canny(tmp1, 32, 128, 3); // FIXME: Unary desync should not require tie! cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out2 = tmp2 / cv::gapi::Sobel(tmp2, CV_8U, 1, 1);; cv::Mat test_in = cv::Mat::eye(cv::Size(32,32), CV_8UC3); cv::Mat test_out1, test_out2; cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .apply(cv::gin(test_in), cv::gout(test_out1, test_out2)); } TEST(GAPI_Streaming_Desync, SmokeTest_Streaming) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat out1 = cv::gapi::Canny(tmp1, 32, 128, 3); cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out2 = Delay::on(tmp2,10) / cv::gapi::Sobel(tmp2, CV_8U, 1, 1); auto sc = cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .compileStreaming(cv::compile_args(cv::gapi::kernels())); auto path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(cv::gin(gapi::wip::make_src(path))); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); std::size_t out1_hits = 0u; std::size_t out2_hits = 0u; cv::optional test_out1, test_out2; while (sc.pull(cv::gout(test_out1, test_out2))) { GAPI_Assert(test_out1 || test_out2); if (test_out1) out1_hits++; if (test_out2) out2_hits++; } EXPECT_EQ(100u, out1_hits); // out1 must be available for all frames EXPECT_LE(out2_hits, out1_hits); // out2 must appear less times than out1 } TEST(GAPI_Streaming_Desync, SmokeTest_Streaming_TwoParts) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat out1 = cv::gapi::Canny(tmp1, 32, 128, 3); // Desynchronized path 1 cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out2 = tmp2 / cv::gapi::Sobel(tmp2, CV_8U, 1, 1); // Desynchronized path 2 cv::GMat tmp3 = cv::gapi::streaming::desync(tmp1); cv::GMat out3 = 0.5*tmp3 + 0.5*cv::gapi::medianBlur(tmp3, 7); // The code should compile and execute well (desynchronized parts don't cross) auto sc = cv::GComputation(cv::GIn(in), cv::GOut(out1, out2, out3)) .compileStreaming(); auto path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(cv::gin(gapi::wip::make_src(path))); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); std::size_t test_frames = 0u; cv::optional test_out1, test_out2, test_out3; while (sc.pull(cv::gout(test_out1, test_out2, test_out3))) { GAPI_Assert(test_out1 || test_out2 || test_out3); if (test_out1) { // count frames only for synchronized output test_frames++; } } EXPECT_EQ(100u, test_frames); } TEST(GAPI_Streaming_Desync, Negative_NestedDesync_Tier0) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); // Desynchronized path 1 cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out1 = cv::gapi::medianBlur(tmp2, 3); // Desynchronized path 2, nested from 1 (directly from desync) cv::GMat tmp3 = cv::gapi::streaming::desync(tmp2); cv::GMat out2 = 0.5*tmp3; // This shouldn't compile EXPECT_ANY_THROW(cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .compileStreaming()); } TEST(GAPI_Streaming_Desync, Negative_NestedDesync_Tier1) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); // Desynchronized path 1 cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out1 = cv::gapi::medianBlur(tmp2, 3); // Desynchronized path 2, nested from 1 (indirectly from desync) cv::GMat tmp3 = cv::gapi::streaming::desync(out1); cv::GMat out2 = 0.5*tmp3; // This shouldn't compile EXPECT_ANY_THROW(cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .compileStreaming()); } TEST(GAPI_Streaming_Desync, Negative_CrossMainPart_Tier0) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); // Desynchronized path: depends on both tmp1 and tmp2 cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out1 = 0.5*tmp1 + 0.5*tmp2; // This shouldn't compile EXPECT_ANY_THROW(cv::GComputation(in, out1).compileStreaming()); } TEST(GAPI_Streaming_Desync, Negative_CrossMainPart_Tier1) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); // Desynchronized path: depends on both tmp1 and tmp2 cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out1 = 0.5*tmp1 + 0.5*cv::gapi::medianBlur(tmp2, 3); // This shouldn't compile EXPECT_ANY_THROW(cv::GComputation(in, out1).compileStreaming()); } TEST(GAPI_Streaming_Desync, Negative_CrossOtherDesync_Tier0) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); // Desynchronized path 1 cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out1 = 0.5*tmp2; // Desynchronized path 2 (depends on 1) cv::GMat tmp3 = cv::gapi::streaming::desync(tmp1); cv::GMat out2 = 0.5*tmp3 + tmp2; // This shouldn't compile EXPECT_ANY_THROW(cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .compileStreaming()); } TEST(GAPI_Streaming_Desync, Negative_CrossOtherDesync_Tier1) { cv::GMat in; cv::GMat tmp1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); // Desynchronized path 1 cv::GMat tmp2 = cv::gapi::streaming::desync(tmp1); cv::GMat out1 = 0.5*tmp2; // Desynchronized path 2 (depends on 1) cv::GMat tmp3 = cv::gapi::streaming::desync(tmp1); cv::GMat out2 = 0.5*cv::gapi::medianBlur(tmp3,3) + 1.0*tmp2; // This shouldn't compile EXPECT_ANY_THROW(cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .compileStreaming()); } TEST(GAPI_Streaming_Desync, Negative_SynchronizedPull) { cv::GMat in; cv::GMat out1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat tmp1 = cv::gapi::streaming::desync(out1); cv::GMat out2 = 0.5*tmp1; auto sc = cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .compileStreaming(); auto path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(cv::gin(gapi::wip::make_src(path))); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); cv::Mat o1, o2; EXPECT_ANY_THROW(sc.pull(cv::gout(o1, o2))); } TEST(GAPI_Streaming_Desync, UseSpecialPull) { cv::GMat in; cv::GMat out1 = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat tmp1 = cv::gapi::streaming::desync(out1); cv::GMat out2 = 0.5*tmp1; auto sc = cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)) .compileStreaming(); auto path = findDataFile("cv/video/768x576.avi"); try { sc.setSource(cv::gin(gapi::wip::make_src(path))); } catch(...) { throw SkipTestException("Video file can not be opened"); } sc.start(); cv::optional o1, o2; std::size_t num_frames = 0u; while (sc.pull(cv::gout(o1, o2))) { if (o1) num_frames++; } EXPECT_EQ(100u, num_frames); } G_API_OP(ProduceVector, (cv::GMat)>, "test.desync.vector") { static cv::GArrayDesc outMeta(const cv::GMatDesc &) { return cv::empty_array_desc(); } }; G_API_OP(ProduceOpaque, (cv::GMat)>, "test.desync.opaque") { static cv::GOpaqueDesc outMeta(const cv::GMatDesc &) { return cv::empty_gopaque_desc(); } }; GAPI_OCV_KERNEL(OCVVector, ProduceVector) { static void run(const cv::Mat& in, std::vector &out) { out = {in.cols, in.rows}; } }; GAPI_OCV_KERNEL(OCVOpaque, ProduceOpaque) { static void run(const cv::Mat &in, int &v) { v = in.channels(); } }; namespace { cv::GStreamingCompiled desyncTestObject() { cv::GMat in; cv::GMat blur = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat blur_d = cv::gapi::copy(cv::gapi::streaming::desync(blur)); cv::GMat d1 = Delay::on(blur_d, 10); cv::GMat d2 = Delay::on(blur_d, 30); cv::GArray vec = ProduceVector::on(d1); cv::GOpaque opq = ProduceOpaque::on(d2); auto pkg = cv::gapi::kernels(); return cv::GComputation(cv::GIn(in), cv::GOut(blur, vec, opq)) .compileStreaming(cv::compile_args(pkg)); } } // anonymous namespace TEST(GAPI_Streaming_Desync, MultipleDesyncOutputs_1) { auto sc = desyncTestObject(); const cv::Mat in_mat = cv::Mat::eye(cv::Size(320,240), CV_8UC3); sc.setSource(cv::gin(in_mat)); sc.start(); cv::optional out_mat; cv::optional > out_vec; cv::optional out_int; int counter = 0; while (sc.pull(cv::gout(out_mat, out_vec, out_int))) { if (counter++ == 1000) { // Stop the test after 1000 iterations sc.stop(); break; } GAPI_Assert(out_mat || out_vec || out_int); // out_vec and out_int are on the same desynchronized path // they MUST arrive together. If one is available, the other // also must be available. if (out_vec) { ASSERT_TRUE(out_int.has_value()); } if (out_int) { ASSERT_TRUE(out_vec.has_value()); } if (out_vec || out_int) { EXPECT_EQ(320, out_vec.value()[0]); EXPECT_EQ(240, out_vec.value()[1]); EXPECT_EQ(3, out_int.value()); } } } TEST(GAPI_Streaming_Desync, StartStop_Stress) { auto sc = desyncTestObject(); const cv::Mat in_mat = cv::Mat::eye(cv::Size(320,240), CV_8UC3); cv::optional out_mat; cv::optional > out_vec; cv::optional out_int; for (int i = 0; i < 10; i++) { sc.setSource(cv::gin(in_mat)); sc.start(); int counter = 0; while (counter++ < 100) { sc.pull(cv::gout(out_mat, out_vec, out_int)); GAPI_Assert(out_mat || out_vec || out_int); if (out_vec) { ASSERT_TRUE(out_int.has_value()); } if (out_int) { ASSERT_TRUE(out_vec.has_value()); } } sc.stop(); } } TEST(GAPI_Streaming_Desync, DesyncObjectConsumedByTwoIslandsViaSeparateDesync) { // See comment in the implementation of cv::gapi::streaming::desync (.cpp) cv::GMat in; cv::GMat tmp = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat tmp1 = cv::gapi::streaming::desync(tmp); cv::GMat out1 = cv::gapi::copy(tmp1); // ran via Streaming backend cv::GMat tmp2 = cv::gapi::streaming::desync(tmp); cv::GMat out2 = tmp2 * 0.5; // ran via OCV backend auto c = cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)); EXPECT_NO_THROW(c.compileStreaming()); } TEST(GAPI_Streaming_Desync, DesyncObjectConsumedByTwoIslandsViaSameDesync) { // See comment in the implementation of cv::gapi::streaming::desync (.cpp) cv::GMat in; cv::GMat tmp = cv::gapi::boxFilter(in, -1, cv::Size(3,3)); cv::GMat tmp1 = cv::gapi::streaming::desync(tmp); cv::GMat out1 = cv::gapi::copy(tmp1); // ran via Streaming backend cv::GMat out2 = out1 - 0.5*tmp1; // ran via OCV backend auto c = cv::GComputation(cv::GIn(in), cv::GOut(out1, out2)); EXPECT_NO_THROW(c.compileStreaming()); } TEST(GAPI_Streaming, CopyFrame) { std::string filepath = findDataFile("cv/video/768x576.avi"); cv::GFrame in; auto out = cv::gapi::copy(in); cv::GComputation comp(cv::GIn(in), cv::GOut(out)); auto cc = comp.compileStreaming(); try { cc.setSource(filepath); } catch(...) { throw SkipTestException("Video file can not be opened"); } cv::VideoCapture cap; cap.open(filepath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::MediaFrame frame; cv::Mat ocv_mat; std::size_t num_frames = 0u; std::size_t max_frames = 10u; cc.start(); while (cc.pull(cv::gout(frame)) && num_frames < max_frames) { auto view = frame.access(cv::MediaFrame::Access::R); cv::Mat gapi_mat(frame.desc().size, CV_8UC3, view.ptr[0]); num_frames++; cap >> ocv_mat; EXPECT_EQ(0, cvtest::norm(ocv_mat, gapi_mat, NORM_INF)); } } TEST(GAPI_Streaming, CopyFrameGray) { std::string filepath = findDataFile("cv/video/768x576.avi"); cv::GFrame in; auto out = cv::gapi::copy(in); cv::GComputation comp(cv::GIn(in), cv::GOut(out)); auto cc = comp.compileStreaming(); try { cc.setSource(filepath); } catch (...) { throw SkipTestException("Video file can not be opened"); } cv::VideoCapture cap; cap.open(filepath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::MediaFrame frame; cv::Mat ocv_mat; std::size_t num_frames = 0u; std::size_t max_frames = 10u; cc.start(); while (cc.pull(cv::gout(frame)) && num_frames < max_frames) { auto view = frame.access(cv::MediaFrame::Access::R); cv::Mat gapi_mat(frame.desc().size, CV_8UC1, view.ptr[0]); num_frames++; cap >> ocv_mat; cv::Mat gray; cvtColor(ocv_mat, gray, cv::COLOR_BGR2GRAY); EXPECT_EQ(0, cvtest::norm(gray, gapi_mat, NORM_INF)); } } TEST(GAPI_Streaming, CopyMat) { std::string filepath = findDataFile("cv/video/768x576.avi"); cv::GMat in; auto out = cv::gapi::copy(in); cv::GComputation comp(cv::GIn(in), cv::GOut(out)); auto cc = comp.compileStreaming(); try { cc.setSource(filepath); } catch(...) { throw SkipTestException("Video file can not be opened"); } cv::VideoCapture cap; cap.open(filepath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::Mat out_mat; cv::Mat ocv_mat; std::size_t num_frames = 0u; std::size_t max_frames = 10u; cc.start(); while (cc.pull(cv::gout(out_mat)) && num_frames < max_frames) { num_frames++; cap >> ocv_mat; EXPECT_EQ(0, cvtest::norm(ocv_mat, out_mat, NORM_INF)); } } TEST(GAPI_Streaming, Reshape) { std::string filepath = findDataFile("cv/video/768x576.avi"); cv::GFrame in; auto out = cv::gapi::copy(in); cv::GComputation comp(cv::GIn(in), cv::GOut(out)); auto cc = comp.compileStreaming(); try { cc.setSource(filepath); } catch(...) { throw SkipTestException("Video file can not be opened"); } cv::VideoCapture cap; cap.open(filepath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::MediaFrame frame; cv::Mat ocv_mat; std::size_t num_frames = 0u; std::size_t max_frames = 10u; cc.start(); while (cc.pull(cv::gout(frame)) && num_frames < max_frames) { auto view = frame.access(cv::MediaFrame::Access::R); cv::Mat gapi_mat(frame.desc().size, CV_8UC3, view.ptr[0]); num_frames++; cap >> ocv_mat; EXPECT_EQ(0, cvtest::norm(ocv_mat, gapi_mat, NORM_INF)); } // Reshape the graph meta filepath = findDataFile("cv/video/1920x1080.avi"); cc.stop(); try { cc.setSource(filepath); } catch(...) { throw SkipTestException("Video file can not be opened"); } cap.open(filepath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::MediaFrame frame2; cv::Mat ocv_mat2; num_frames = 0u; cc.start(); while (cc.pull(cv::gout(frame2)) && num_frames < max_frames) { auto view = frame2.access(cv::MediaFrame::Access::R); cv::Mat gapi_mat(frame2.desc().size, CV_8UC3, view.ptr[0]); num_frames++; cap >> ocv_mat2; EXPECT_EQ(0, cvtest::norm(ocv_mat2, gapi_mat, NORM_INF)); } } TEST(GAPI_Streaming, ReshapeGray) { std::string filepath = findDataFile("cv/video/768x576.avi"); cv::GFrame in; auto out = cv::gapi::copy(in); cv::GComputation comp(cv::GIn(in), cv::GOut(out)); auto cc = comp.compileStreaming(); try { cc.setSource(filepath); } catch (...) { throw SkipTestException("Video file can not be opened"); } cv::VideoCapture cap; cap.open(filepath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::MediaFrame frame; cv::Mat ocv_mat; std::size_t num_frames = 0u; std::size_t max_frames = 10u; cc.start(); while (cc.pull(cv::gout(frame)) && num_frames < max_frames) { auto view = frame.access(cv::MediaFrame::Access::R); cv::Mat gapi_mat(frame.desc().size, CV_8UC1, view.ptr[0]); num_frames++; cap >> ocv_mat; cv::Mat gray; cvtColor(ocv_mat, gray, cv::COLOR_BGR2GRAY); EXPECT_EQ(0, cvtest::norm(gray, gapi_mat, NORM_INF)); } // Reshape the graph meta filepath = findDataFile("cv/video/1920x1080.avi"); cc.stop(); try { cc.setSource(filepath); } catch (...) { throw SkipTestException("Video file can not be opened"); } cap.open(filepath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::MediaFrame frame2; cv::Mat ocv_mat2; num_frames = 0u; cc.start(); while (cc.pull(cv::gout(frame2)) && num_frames < max_frames) { auto view = frame2.access(cv::MediaFrame::Access::R); cv::Mat gapi_mat(frame2.desc().size, CV_8UC1, view.ptr[0]); num_frames++; cap >> ocv_mat2; cv::Mat gray; cvtColor(ocv_mat2, gray, cv::COLOR_BGR2GRAY); EXPECT_EQ(0, cvtest::norm(gray, gapi_mat, NORM_INF)); } } namespace { enum class TestSourceType { BGR, NV12, GRAY }; std::ostream& operator<<(std::ostream& os, TestSourceType a) { os << "Source:"; switch (a) { case TestSourceType::BGR: return os << "BGR"; case TestSourceType::NV12: return os << "NV12"; case TestSourceType::GRAY: return os << "GRAY"; default: CV_Assert(false && "unknown TestSourceType"); } } cv::gapi::wip::IStreamSource::Ptr createTestSource(TestSourceType sourceType, const std::string& pipeline) { assert(sourceType == TestSourceType::BGR || sourceType == TestSourceType::NV12 || sourceType == TestSourceType::GRAY); cv::gapi::wip::IStreamSource::Ptr ptr { }; switch (sourceType) { case TestSourceType::BGR: { try { ptr = cv::gapi::wip::make_src(pipeline); } catch(...) { throw SkipTestException(std::string("BGRSource for '") + pipeline + "' couldn't be created!"); } break; } case TestSourceType::NV12: { try { ptr = cv::gapi::wip::make_src(pipeline); } catch(...) { throw SkipTestException(std::string("NV12Source for '") + pipeline + "' couldn't be created!"); } break; } case TestSourceType::GRAY: { try { ptr = cv::gapi::wip::make_src(pipeline); } catch (...) { throw SkipTestException(std::string("GRAYSource for '") + pipeline + "' couldn't be created!"); } break; } default: { throw SkipTestException("Incorrect type of source! " "Something went wrong in the test!"); } } return ptr; } enum class TestAccessType { BGR, Y, UV }; std::ostream& operator<<(std::ostream& os, TestAccessType a) { os << "Accessor:"; switch (a) { case TestAccessType::BGR: return os << "BGR"; case TestAccessType::Y: return os << "Y"; case TestAccessType::UV: return os << "UV"; default: CV_Assert(false && "unknown TestAccessType"); } } using GapiFunction = std::function; static std::map gapi_functions = { { TestAccessType::BGR, cv::gapi::streaming::BGR }, { TestAccessType::Y, cv::gapi::streaming::Y }, { TestAccessType::UV, cv::gapi::streaming::UV } }; using RefFunction = std::function; static std::map, RefFunction> ref_functions = { { std::make_pair(TestSourceType::BGR, TestAccessType::BGR), [](const cv::Mat& bgr) { return bgr; } }, { std::make_pair(TestSourceType::BGR, TestAccessType::Y), [](const cv::Mat& bgr) { cv::Mat y, uv; cvtBGR2NV12(bgr, y, uv); return y; } }, { std::make_pair(TestSourceType::BGR, TestAccessType::UV), [](const cv::Mat& bgr) { cv::Mat y, uv; cvtBGR2NV12(bgr, y, uv); return uv; } }, { std::make_pair(TestSourceType::NV12, TestAccessType::BGR), [](const cv::Mat& bgr) { cv::Mat y, uv, out_bgr; cvtBGR2NV12(bgr, y, uv); cv::cvtColorTwoPlane(y, uv, out_bgr, cv::COLOR_YUV2BGR_NV12); return out_bgr; } }, { std::make_pair(TestSourceType::NV12, TestAccessType::Y), [](const cv::Mat& bgr) { cv::Mat y, uv; cvtBGR2NV12(bgr, y, uv); return y; } }, { std::make_pair(TestSourceType::NV12, TestAccessType::UV), [](const cv::Mat& bgr) { cv::Mat y, uv; cvtBGR2NV12(bgr, y, uv); return uv; } }, { std::make_pair(TestSourceType::GRAY, TestAccessType::BGR), [](const cv::Mat& bgr) { cv::Mat gray; cv::cvtColor(bgr, gray, cv::COLOR_BGR2GRAY); cv::Mat out_bgr; cv::cvtColor(gray, out_bgr, cv::COLOR_GRAY2BGR); return out_bgr; } }, { std::make_pair(TestSourceType::GRAY, TestAccessType::Y), [](const cv::Mat& bgr) { cv::Mat gray; cv::cvtColor(bgr, gray, cv::COLOR_BGR2GRAY); return gray; } }, { std::make_pair(TestSourceType::GRAY, TestAccessType::UV), [](const cv::Mat& bgr) { cv::Mat uv(bgr.size() / 2, CV_8UC2, cv::Scalar::all(127)); return uv; } }, }; } // anonymous namespace struct GAPI_Accessors_In_Streaming : public TestWithParam< std::tuple> { }; TEST_P(GAPI_Accessors_In_Streaming, AccuracyTest) { std::string filepath{}; TestSourceType sourceType = TestSourceType::BGR; TestAccessType accessType = TestAccessType::BGR; std::tie(filepath, sourceType, accessType) = GetParam(); auto accessor = gapi_functions[accessType]; auto fromBGR = ref_functions[std::make_pair(sourceType, accessType)]; const std::string& absFilePath = findDataFile(filepath); cv::GFrame in; cv::GMat out = accessor(in); cv::GComputation comp(cv::GIn(in), cv::GOut(out)); auto cc = comp.compileStreaming(); auto src = createTestSource(sourceType, absFilePath); cc.setSource(src); cv::VideoCapture cap; cap.open(absFilePath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::Mat cap_mat, ocv_mat, gapi_mat; std::size_t num_frames = 0u; std::size_t max_frames = 10u; cc.start(); while (num_frames < max_frames && cc.pull(cv::gout(gapi_mat))) { num_frames++; cap >> cap_mat; ocv_mat = fromBGR(cap_mat); EXPECT_EQ(0, cvtest::norm(ocv_mat, gapi_mat, NORM_INF)); } cc.stop(); } INSTANTIATE_TEST_CASE_P(TestAccessor, GAPI_Accessors_In_Streaming, Combine(Values("cv/video/768x576.avi"), Values(TestSourceType::BGR, TestSourceType::NV12, TestSourceType::GRAY), Values(TestAccessType::BGR, TestAccessType::Y, TestAccessType::UV) )); struct GAPI_Accessors_Meta_In_Streaming : public TestWithParam< std::tuple> { }; TEST_P(GAPI_Accessors_Meta_In_Streaming, AccuracyTest) { std::string filepath{}; TestSourceType sourceType = TestSourceType::BGR; TestAccessType accessType = TestAccessType::BGR; std::tie(filepath, sourceType, accessType) = GetParam(); auto accessor = gapi_functions[accessType]; auto fromBGR = ref_functions[std::make_pair(sourceType, accessType)]; const std::string& absFilePath = findDataFile(filepath); cv::GFrame in; cv::GMat gmat = accessor(in); cv::GMat resized = cv::gapi::resize(gmat, cv::Size(1920, 1080)); cv::GOpaque outId = cv::gapi::streaming::seq_id(resized); cv::GOpaque outTs = cv::gapi::streaming::timestamp(resized); cv::GComputation comp(cv::GIn(in), cv::GOut(resized, outId, outTs)); auto cc = comp.compileStreaming(); auto src = createTestSource(sourceType, absFilePath); cc.setSource(src); cv::VideoCapture cap; cap.open(absFilePath); if (!cap.isOpened()) throw SkipTestException("Video file can not be opened"); cv::Mat cap_mat, req_mat, ocv_mat, gapi_mat; int64_t seq_id = 0, timestamp = 0; std::set all_seq_ids; std::vector all_timestamps; std::size_t num_frames = 0u; std::size_t max_frames = 10u; cc.start(); while (num_frames < max_frames && cc.pull(cv::gout(gapi_mat, seq_id, timestamp))) { num_frames++; cap >> cap_mat; req_mat = fromBGR(cap_mat); cv::resize(req_mat, ocv_mat, cv::Size(1920, 1080)); EXPECT_EQ(0, cvtest::norm(ocv_mat, gapi_mat, NORM_INF)); all_seq_ids.insert(seq_id); all_timestamps.push_back(timestamp); } cc.stop(); EXPECT_EQ(all_seq_ids.begin(), all_seq_ids.find(0L)); auto last_elem_it = --all_seq_ids.end(); EXPECT_EQ(last_elem_it, all_seq_ids.find(int64_t(max_frames - 1L))); EXPECT_EQ(max_frames, all_seq_ids.size()); EXPECT_EQ(max_frames, all_timestamps.size()); EXPECT_TRUE(std::is_sorted(all_timestamps.begin(), all_timestamps.end())); } INSTANTIATE_TEST_CASE_P(AccessorMeta, GAPI_Accessors_Meta_In_Streaming, Combine(Values("cv/video/768x576.avi"), Values(TestSourceType::BGR, TestSourceType::NV12, TestSourceType::GRAY), Values(TestAccessType::BGR, TestAccessType::Y, TestAccessType::UV) )); TEST(GAPI_Streaming, TestPythonAPI) { cv::Size sz(200, 200); cv::Mat in_mat(sz, CV_8UC3); cv::randu(in_mat, cv::Scalar::all(0), cv::Scalar(255)); const auto crop_rc = cv::Rect(13, 75, 100, 100); // OpenCV reference image cv::Mat ocv_mat; { ocv_mat = in_mat(crop_rc); } cv::GMat in; auto roi = cv::gapi::crop(in, crop_rc); cv::GComputation comp(cv::GIn(in), cv::GOut(roi)); // NB: Used by python bridge auto cc = comp.compileStreaming(cv::detail::ExtractMetaCallback{[&](const cv::GTypesInfo& info) { GAPI_Assert(info.size() == 1u); GAPI_Assert(info[0].shape == cv::GShape::GMAT); return cv::GMetaArgs{cv::GMetaArg{cv::descr_of(in_mat)}}; }}); // NB: Used by python bridge cc.setSource(cv::detail::ExtractArgsCallback{[&](const cv::GTypesInfo& info) { GAPI_Assert(info.size() == 1u); GAPI_Assert(info[0].shape == cv::GShape::GMAT); return cv::GRunArgs{in_mat}; }}); cc.start(); bool is_over = false; cv::GRunArgs out_args; using RunArgs = cv::util::variant; RunArgs args; // NB: Used by python bridge std::tie(is_over, args) = cc.pull(); switch (args.index()) { case RunArgs::index_of(): out_args = util::get(args); break; default: GAPI_Assert(false && "Incorrect type of return value"); } ASSERT_EQ(1u, out_args.size()); ASSERT_TRUE(cv::util::holds_alternative(out_args[0])); EXPECT_EQ(0, cvtest::norm(ocv_mat, cv::util::get(out_args[0]), NORM_INF)); EXPECT_TRUE(is_over); cc.stop(); } #ifdef HAVE_ONEVPL TEST(OneVPL_Source, Init) { using CfgParam = cv::gapi::wip::onevpl::CfgParam; std::vector src_params; src_params.push_back(CfgParam::create_implementation(MFX_IMPL_TYPE_HARDWARE)); src_params.push_back(CfgParam::create_acceleration_mode(MFX_ACCEL_MODE_VIA_D3D11)); src_params.push_back(CfgParam::create_decoder_id(MFX_CODEC_HEVC)); std::stringstream stream(std::ios_base::in | std::ios_base::out | std::ios_base::binary); EXPECT_TRUE(stream.write(reinterpret_cast(const_cast(streaming::onevpl::hevc_header)), sizeof(streaming::onevpl::hevc_header))); std::shared_ptr stream_data_provider = std::make_shared(stream); cv::Ptr cap; bool cap_created = false; try { cap = cv::gapi::wip::make_onevpl_src(stream_data_provider, src_params); cap_created = true; } catch (const std::exception&) { } ASSERT_TRUE(cap_created); cv::gapi::wip::Data out; while (cap->pull(out)) { (void)out; } EXPECT_TRUE(stream_data_provider->empty()); } #endif // HAVE_ONEVPL TEST(GAPI_Streaming, TestDesyncRMat) { cv::GMat in; auto blurred = cv::gapi::blur(in, cv::Size{3,3}); auto desynced = cv::gapi::streaming::desync(blurred); auto out = in - blurred; auto pipe = cv::GComputation(cv::GIn(in), cv::GOut(desynced, out)).compileStreaming(); cv::Size sz(32,32); cv::Mat in_mat(sz, CV_8UC3); cv::randu(in_mat, cv::Scalar::all(0), cv::Scalar(255)); pipe.setSource(cv::gin(in_mat)); pipe.start(); cv::optional out_desync; cv::optional out_rmat; while (true) { // Initially it threw "bad variant access" since there was // no RMat handling in wrap_opt_arg EXPECT_NO_THROW(pipe.pull(cv::gout(out_desync, out_rmat))); if (out_rmat) break; } } G_API_OP(GTestBlur, , "test.blur") { static GFrameDesc outMeta(GFrameDesc d) { return d; } }; GAPI_OCV_KERNEL(GOcvTestBlur, GTestBlur) { static void run(const cv::MediaFrame& in, cv::MediaFrame& out) { auto d = in.desc(); GAPI_Assert(d.fmt == cv::MediaFormat::BGR); auto view = in.access(cv::MediaFrame::Access::R); cv::Mat mat(d.size, CV_8UC3, view.ptr[0]); cv::Mat blurred; cv::blur(mat, blurred, cv::Size{3,3}); out = cv::MediaFrame::Create(blurred); } }; TEST(GAPI_Streaming, TestDesyncMediaFrame) { cv::GFrame in; auto blurred = GTestBlur::on(in); auto desynced = cv::gapi::streaming::desync(blurred); auto out = GTestBlur::on(blurred); auto pipe = cv::GComputation(cv::GIn(in), cv::GOut(desynced, out)) .compileStreaming(cv::compile_args(cv::gapi::kernels())); std::string filepath = findDataFile("cv/video/768x576.avi"); try { pipe.setSource(filepath); } catch(...) { throw SkipTestException("Video file can not be opened"); } pipe.start(); cv::optional out_desync; cv::optional out_frame; while (true) { // Initially it threw "bad variant access" since there was // no MediaFrame handling in wrap_opt_arg EXPECT_NO_THROW(pipe.pull(cv::gout(out_desync, out_frame))); if (out_frame) break; } } G_API_OP(GTestBlurGray, , "test.blur_gray") { static GFrameDesc outMeta(GFrameDesc d) { return d; } }; GAPI_OCV_KERNEL(GOcvTestBlurGray, GTestBlurGray) { static void run(const cv::MediaFrame & in, cv::MediaFrame & out) { auto d = in.desc(); GAPI_Assert(d.fmt == cv::MediaFormat::GRAY); auto view = in.access(cv::MediaFrame::Access::R); cv::Mat mat(d.size, CV_8UC1, view.ptr[0]); cv::Mat blurred; cv::blur(mat, blurred, cv::Size{ 3,3 }); out = cv::MediaFrame::Create(blurred); } }; TEST(GAPI_Streaming, TestDesyncMediaFrameGray) { cv::GFrame in; auto blurred = GTestBlurGray::on(in); auto desynced = cv::gapi::streaming::desync(blurred); auto out = GTestBlurGray::on(blurred); auto pipe = cv::GComputation(cv::GIn(in), cv::GOut(desynced, out)) .compileStreaming(cv::compile_args(cv::gapi::kernels())); std::string filepath = findDataFile("cv/video/768x576.avi"); try { pipe.setSource(filepath); } catch (...) { throw SkipTestException("Video file can not be opened"); } pipe.start(); cv::optional out_desync; cv::optional out_frame; while (true) { // Initially it threw "bad variant access" since there was // no MediaFrame handling in wrap_opt_arg EXPECT_NO_THROW(pipe.pull(cv::gout(out_desync, out_frame))); if (out_frame) break; } } } // namespace opencv_test