[go: nahoru, domu]

Skip to content

Latest commit

 

History

History

traffic_analysis

traffic analysis

👋 hello

This script performs traffic flow analysis using YOLOv8, an object-detection method and ByteTrack, a simple yet effective online multi-object tracking method. It uses the supervision package for multiple tasks such as tracking, annotations, etc.

traffic_analysis_result.mov

💻 install

  • clone repository and navigate to example directory

    git clone https://github.com/roboflow/supervision.git
    cd supervision/examples/traffic_analysis
  • setup python environment and activate it [optional]

    python3 -m venv venv
    source venv/bin/activate
  • install required dependencies

    pip install -r requirements.txt
  • download traffic_analysis.pt and traffic_analysis.mov files

    ./setup.sh

🛠️ script arguments

  • ultralytics

    • --source_weights_path: Required. Specifies the path to the YOLO model's weights file, which is essential for the object detection process. This file contains the data that the model uses to identify objects in the video.

    • --source_video_path: Required. The path to the source video file that will be analyzed. This is the input video on which traffic flow analysis will be performed.

    • --target_video_path (optional): The path to save the output video with annotations. If not specified, the processed video will be displayed in real-time without being saved.

    • --confidence_threshold (optional): Sets the confidence threshold for the YOLO model to filter detections. Default is 0.3. This determines how confident the model should be to recognize an object in the video.

    • --iou_threshold (optional): Specifies the IOU (Intersection Over Union) threshold for the model. Default is 0.7. This value is used to manage object detection accuracy, particularly in distinguishing between different objects.

  • inference

    • --roboflow_api_key (optional): The API key for Roboflow services. If not provided directly, the script tries to fetch it from the ROBOFLOW_API_KEY environment variable. Follow this guide to acquire your API KEY.

    • --model_id (optional): Designates the Roboflow model ID to be used. The default value is "vehicle-count-in-drone-video/6".

    • --source_video_path: Required. The path to the source video file that will be analyzed. This is the input video on which traffic flow analysis will be performed.

    • --target_video_path (optional): The path to save the output video with annotations. If not specified, the processed video will be displayed in real-time without being saved.

    • --confidence_threshold (optional): Sets the confidence threshold for the YOLO model to filter detections. Default is 0.3. This determines how confident the model should be to recognize an object in the video.

    • --iou_threshold (optional): Specifies the IOU (Intersection Over Union) threshold for the model. Default is 0.7. This value is used to manage object detection accuracy, particularly in distinguishing between different objects.

⚙️ run

  • ultralytics

    python ultralytics_example.py \
    --source_weights_path data/traffic_analysis.pt \
    --source_video_path data/traffic_analysis.mov \
    --confidence_threshold 0.3 \
    --iou_threshold 0.5 \
    --target_video_path data/traffic_analysis_result.mov
  • inference

    python inference_example.py \
    --roboflow_api_key <ROBOFLOW API KEY> \
    --source_video_path data/traffic_analysis.mov \
    --confidence_threshold 0.3 \
    --iou_threshold 0.5 \
    --target_video_path data/traffic_analysis_result.mov

© license

This demo integrates two main components, each with its own licensing:

  • ultralytics: The object detection model used in this demo, YOLOv8, is distributed under the AGPL-3.0 license. You can find more details about this license here.

  • supervision: The analytics code that powers the zone-based analysis in this demo is based on the Supervision library, which is licensed under the MIT license. This makes the Supervision part of the code fully open source and freely usable in your projects.