#include #include #include #include "utility.hpp" // Includes common necessary includes for development using depthai library #include "depthai/depthai.hpp" // MobilenetSSD label texts static const std::vector labelMap = {"background", "aeroplane", "bicycle", "bird", "boat", "bottle", "bus", "car", "cat", "chair", "cow", "diningtable", "dog", "horse", "motorbike", "person", "pottedplant", "sheep", "sofa", "train", "tvmonitor"}; static std::atomic syncNN{true}; int main(int argc, char** argv) { using namespace std; using namespace std::chrono; // Default blob path provided by Hunter private data download // Applicable for easier example usage only std::string nnPath(BLOB_PATH); // If path to blob specified, use that if(argc > 1) { nnPath = std::string(argv[1]); } // Print which blob we are using printf("Using blob at path: %s\n", nnPath.c_str()); // Create pipeline dai::Pipeline pipeline; // Define sources and outputs auto camRgb = pipeline.create(); auto nn = pipeline.create(); auto xoutRgb = pipeline.create(); auto nnOut = pipeline.create(); auto nnNetworkOut = pipeline.create(); xoutRgb->setStreamName("rgb"); nnOut->setStreamName("nn"); nnNetworkOut->setStreamName("nnNetwork"); // Properties camRgb->setPreviewSize(300, 300); // NN input camRgb->setInterleaved(false); camRgb->setFps(40); // Define a neural network that will make predictions based on the source frames nn->setConfidenceThreshold(0.5); nn->setBlobPath(nnPath); nn->setNumInferenceThreads(2); nn->input.setBlocking(false); // Linking if(syncNN) { nn->passthrough.link(xoutRgb->input); } else { camRgb->preview.link(xoutRgb->input); } camRgb->preview.link(nn->input); nn->out.link(nnOut->input); nn->outNetwork.link(nnNetworkOut->input); // Connect to device and start pipeline dai::Device device(pipeline); // Output queues will be used to get the rgb frames and nn data from the outputs defined above auto qRgb = device.getOutputQueue("rgb", 4, false); auto qDet = device.getOutputQueue("nn", 4, false); auto qNN = device.getOutputQueue("nnNetwork", 4, false); cv::Mat frame; std::vector detections; auto startTime = steady_clock::now(); int counter = 0; float fps = 0; auto color2 = cv::Scalar(255, 255, 255); // Add bounding boxes and text to the frame and show it to the user auto displayFrame = [](std::string name, cv::Mat frame, std::vector& detections) { auto color = cv::Scalar(255, 0, 0); // nn data, being the bounding box locations, are in <0..1> range - they need to be normalized with frame width/height for(auto& detection : detections) { int x1 = detection.xmin * frame.cols; int y1 = detection.ymin * frame.rows; int x2 = detection.xmax * frame.cols; int y2 = detection.ymax * frame.rows; uint32_t labelIndex = detection.label; std::string labelStr = to_string(labelIndex); if(labelIndex < labelMap.size()) { labelStr = labelMap[labelIndex]; } cv::putText(frame, labelStr, cv::Point(x1 + 10, y1 + 20), cv::FONT_HERSHEY_TRIPLEX, 0.5, color); std::stringstream confStr; confStr << std::fixed << std::setprecision(2) << detection.confidence * 100; cv::putText(frame, confStr.str(), cv::Point(x1 + 10, y1 + 40), cv::FONT_HERSHEY_TRIPLEX, 0.5, color); cv::rectangle(frame, cv::Rect(cv::Point(x1, y1), cv::Point(x2, y2)), color, cv::FONT_HERSHEY_SIMPLEX); } // Show the frame cv::imshow(name, frame); }; bool printOutputLayersOnce = true; while(true) { std::shared_ptr inRgb; std::shared_ptr inDet; std::shared_ptr inNN; if(syncNN) { inRgb = qRgb->get(); inDet = qDet->get(); inNN = qNN->get(); } else { inRgb = qRgb->tryGet(); inDet = qDet->tryGet(); inNN = qNN->tryGet(); } counter++; auto currentTime = steady_clock::now(); auto elapsed = duration_cast>(currentTime - startTime); if(elapsed > seconds(1)) { fps = counter / elapsed.count(); counter = 0; startTime = currentTime; } if(inRgb) { frame = inRgb->getCvFrame(); std::stringstream fpsStr; fpsStr << "NN fps: " << std::fixed << std::setprecision(2) << fps; cv::putText(frame, fpsStr.str(), cv::Point(2, inRgb->getHeight() - 4), cv::FONT_HERSHEY_TRIPLEX, 0.4, color2); } if(inDet) { detections = inDet->detections; } if(printOutputLayersOnce && inNN) { std::cout << "Output layer names: "; for(const auto& ten : inNN->getAllLayerNames()) { std::cout << ten << ", "; } std::cout << std::endl; printOutputLayersOnce = false; } if(!frame.empty()) { displayFrame("video", frame, detections); } int key = cv::waitKey(1); if(key == 'q' || key == 'Q') { return 0; } } return 0; }