Experimental projects we've done with DepthAI.
Experiments can be anything from "here's some code and it works sometimes" to "this is almost a tutorial".
The following list isn't exhaustive (as we randomly add experiments and we may forget to update this list):
[Gen2] Gaze estimation (here)
This example demonstrates how to run 3 stage (3-series, 2 parallel) inference on DepthAI using Gen2 Pipeline Builder.
Origina OpenVINO demo, on which this example was made, is here
[Gen2] Subpixel and LR-Check Disparity Depth (here)
This example shows how to do Subpixel, LR-Check or Extended Disparity, and also how to project these measurements into a point cloud for visualization. This uses the Gen2 Pipeline Builder.
[Gen2] Age Gender (here)
This shows a simple two-stage neural inference example, doing face detection and then age/gender estimation based on the face.
[Gen2] Text Detection + Optical Character Recognition (OCR) Pipeline (here)
This pipeline implements text detection (EAST) followed by optical character recognition of the detected text.
[Gen2] Pedestrian Reidentification (here)
This example demonstrates how to run 2 stage inference on DepthAI using Gen2 Pipeline Builder to identify and re-identify pedestrians with unique IDs.
Original OpenVINO demo, on which this example was made, is here.
COVID-19 Mask / No-Mask Detector (here)
This project shows you how to run the COVID-19 mask/no-mask object detector which was trained here
Social Distancing Example (here)
Since DepthAI gives the full 3D position of objects in physical space, it's a couple lines of code to make a social-distancing monitor with DepthAI. So that's what this project is, a quick cut at a social-distancing monitor.
Demo-UI (here)
Application used to demonstrate various capabilities of the DepthAI platform. Contains examples with descriptions, console outputs and preview windows.
MJPEG and JSON streaming (here)
This lay the groundwork to make DepthAI OpenDataCam compatible.
Stereo Neural Inference Results Visualizer (here)
So because there are often application-specific host-side filtering to be done on the stereo neural inference results, and because these calculations are lightweight (i.e. could be done on an ESP32), we leave the triangulation itself to the host. If there is interest to do this on DepthAI directly instead, please let us know!
People Counting (here)
This is a basic usage example of megaAI and/or DepthAI (although it doesn't actually use the depth aspect of DepthAI): simply counting people in a scene and logging this count.
So you could use this to make plots over a day of room occupancy. One could modify this example to show where in a room those people were, over time, if desirable. But for now it just produces a count of people - so the total in view of the camera - over time.
People Tracker (here)
This application counts how many people went upwards / downwards / leftwards / rightwards in the video stream, allowing you to receive an information about how many people went into a room or went through a corridor.
The model used in this example is person_detection_retail_0013 from the OpenVIN Model Zoo. Credits: Adrian Rosebrock, OpenCV People Counter, PyImageSearch, https://www.pyimagesearch.com/2018/08/13/opencv-people-counter/, accessed on 6 August 2020.
Point Cloud Projection (here)
This is a simple application which creates rgbd image from right
and depth_raw
stream and projects it into point clouds. There is also a interactive point cloud visualizer. (depth_raw with left and rgb will be added soon)
RGB-D and PCL(here)
This is a simple application which creates rgbd image from rgb
and depth
stream and projects it into rgb with depth overlay and point clouds. There is also a interactive point cloud visualizer.
Host-Side WLS Filter (here)
This gives an example of doing host-side WLS filtering using the rectified_right
and depth
stream from DepthAI.
Example running on BW1092 shown below: