Skip to content

PINGEcosystem/GhostVision

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

52 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

GhostVision

🚧UNDER CONSTRUCTION🚧

PyPI - Version GitHub last commit GitHub commit activity GitHub

Near-real time detection of derelict (ghost) crab pots with side-scan sonar.

ezgif com-crop

πŸ“‘ Overview

GhostVision is an open-source Python interface for automatically detecting and mapping ghost (derelict) crab pots from side-scan sonar imagery. GhostVision currently supports multiple packaged object-detection models, including YOLO- and RF-DETR-based exports trained with Roboflow. Detections are then georeferenced with PINGMapper.

πŸ“œ Published Documentation

πŸ“„ Journal Article

Bodine, C.S.; Baxevani, K.; Abbasi, N.;Wierzbicki, J.; Christoph, O.; Hughes, C.; Bagoren, O.; Hines, O.; Greco, J.; Trembanis, A. GhostVision: Democratizing Derelict Gear Detection using Low-Cost Sonar and Artificial Intelligence. (In Review). Submitted to Journal of Marine Science and Engineering.

πŸ€– Models

  • GV-RF-DETR Model on HF
  • GV-YOLO12 Model on HF
  • GV-YOLO26 Model on HF

πŸ—‚οΈ Model Dataset

Dataset on HF

βš™οΈ Installation

πŸš€ GPU (Fast Inference)

GhostVision is optimized for running inference (predictions) on the GPU. The processing environment is installed with conda. Any flavor of conda will do, but we recommend Miniforge. Follow the instructions below based on your OS.

Windows Only

Windows does not natively support inference on the GPU. A utility called WSL (Windows Subsystem for Linux) needs to be installed in order to run inference on the GPU.

  1. Install the latest NVIDIA driver for your system.
  2. Add CUDA Support for WSL 2.
    • Assumes your computer has an NVIDIA GPU.
  3. Install WSL (Windows Subsystem for Linux) &
  4. Open the command prompt by launching Ubuntu from the Windows Start menu.
  5. You may need to install the NVIDIA Cuda Toolkit with sudo apt install nvidia-cuda-toolkit.

Install Miniforge

  1. In a command prompt, download Miniforge with:
    wget "https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-$(uname)-$(uname -m).sh"
    
  2. Install Miniforge with:
    bash Miniforge3-$(uname)-$(uname -m).sh
    

Install GhostVision

  1. Install PINGInstaller:
    pip install pinginstaller
    
  2. Install GhostVision:
    python -m pinginstaller ghostvision-gpu
    

🐒 CPU (Slow Inference; Experimental)

An experimental version of GhostVision is available to test inference speeds on the CPU. This has been tested on Windows 11 only.

  1. Install Miniforge.
  2. Open the Miniforge prompt.
  3. Install PINGInstaller:
    pip install pinginstaller
    
  4. Install GhostVision.
    python -m pinginstaller ghostvision
    

πŸš€ Usage

  1. Open the appropriate command prompt based on your installation above.
  2. Launch GhostVision:
    conda activate ghostvision
    python -m ghostvision
    
  3. Select your desired model and processing parameters, then click Submit.

Bundled release models are downloaded automatically into the local ~/.ghostvision/models cache the first time they are needed. The current packaged aliases exposed by the app include rf-detr_v1, yolo26_v1, and yolo12_v1.

Recommended Settings

The most useful model-specific operating points currently come from the full GhostVision accuracy assessment workflow used for the companion manuscript. In that analysis, detections were evaluated against manual annotations at a 3 m match radius. When object tracking is enabled, GhostVision combines confidence and temporal persistence using:

S = alpha * conf_avg + (1 - alpha) * pred_cnt / max(pred_cnt)

The combined-score analysis reports the following best-performing alpha values:

  • YOLOv12: alpha = 0.90
  • YOLOv26: alpha = 0.95
  • RF-DETR: alpha = 0.85

The same workflow reports these best thresholds:

  • YOLOv12: confidence = 0.148, pred_cnt = 17, combined score = 0.221
  • YOLOv26: confidence = 0.101, pred_cnt = 15, combined score = 0.136
  • RF-DETR: confidence = 0.386, pred_cnt = 18, combined score = 0.345

Peak F1 values from that workflow were:

  • YOLOv12: F1 = 0.739 from confidence alone, 0.574 from persistence alone, 0.716 from the combined score
  • YOLOv26: F1 = 0.737 from confidence alone, 0.596 from persistence alone, 0.707 from the combined score
  • RF-DETR: F1 = 0.721 from confidence alone, 0.667 from persistence alone, 0.727 from the combined score

For practical use, this suggests:

  • Prefer YOLOv12 as the default packaged model for the best overall operational balance.
  • Start YOLOv12 near score threshold = 0.15, pred_cnt = 17, and alpha = 0.90 when you want settings that reproduce the evaluation workflow.
  • Use RF-DETR only when very high recall is more important than false-positive burden.

GhostVision now uses the same pred_cnt / max(pred_cnt) normalization as the evaluation workflow when computing the tracked combined score. These values should be treated as model-specific reference points, not universal defaults. GhostVision's packaged defaults remain general-purpose starting values for field use, while the values above are the best choices when you want to reproduce the evaluation workflow as closely as possible.

πŸ“¦ Download Custom Roboflow Object Detection Model

GhostVision includes packaged object detection models designed to detect crab pots from side-scan sonar imagery. If you want to use your own compatible Roboflow export instead, you can download a custom model with the included utility.

  1. Open the appropriate command prompt based on your installation above.
  2. Launch the Roboflow model download utility:
    conda activate ghostvision
    python -m ghostvision rf-download
    
  3. Supply your Roboflow API Key.
  4. Enter the project name (all lowercase).
  5. Enter the project version.

The model will be downloaded and available to use.

πŸ”— Related Resources

πŸ™Œ Acknowledgments

GhostVision has been made possible through mentorship, partnerships, financial support, open-source software, manuscripts, and documentation linked below.

NOTE: The contents of this repository are those of the author(s) and do not necessarily represent the views of the individuals and organizations specifically mentioned here.

Development Team: Cameron Bodine, Art Trembanis, Kleio Baxevani, Naveed Abbasi, Onur Bagoren, Olivia Hines, Jared Wierzbicki, Ophelia Christoph, Catherine Hughes, Julia Greco.

Packages

 
 
 

Contributors

Languages