Skip to content

Latest commit

 

History

History
61 lines (42 loc) · 1.85 KB

README.md

File metadata and controls

61 lines (42 loc) · 1.85 KB

Introduction

Rudtrack provides a new approach to object detection and classification in video/live feeds. Rudtrack aims to be faster and lighter than other models in the market and strives to provide a faster way to detect moving objects in a video feed. Rudtrack is desinged for systems with low computational ability and thus uses MobileNet for its classification in the backend.

Installation

git clone https://github.com/nakul-shahdadpuri/rudtrack.git
conda env create -f setup.yml
conda activate rudtrack
pip install -r setup.txt

Usage

For Live Feeds

python live_analysis [IP] [LINE_X1] [LINE_Y1] [LINE_X2] [LINE_Y2] [Sensitivity] [Area]

Here the IP refers to the IP address of the live feed camera. Currently we only use HTTP protocol.

For Videos

python video_analysis [VIDEO PATH] [LINE_X1] [LINE_Y1] [LINE_X2] [LINE_Y2] [Sensitivity] [Area]

Here X1, X2, Y1, Y2 corresbond to the rectangle coordinates in which the desired object is to be detected. The Sensitivity Parameter is used so that an object lying outside the rectangle zone also gets detected if a certain section of the object is in the zone.

The area parameter decided the minimum box area an object should be to be detected by the classifier.

A = 1000 and sensitivity = 4 

could be set normally and changed upon user.

If you want to analyse for the entire video feed, set

X1=0  X2=800  Y1=0  Y2=800

Architecture Pipeline

Blank diagram (1)

Dependencies

tensorflow==2.5.0
OpenCV==4.1.1

Publication

http://thedesignengineering.com/index.php/DE/article/view/8445