Real-time, defence-grade AI system for multi-sensor object detection, tracking, and visibility enhancement on NVIDIA Jetson Orin AGX.
- Overview
- Key Features
- System Architecture
- Tech Stack
- Hardware Requirements
- Installation
- Quick Start
- Usage Guide
- Deployment
- Performance Metrics
- Contact
- License
Defence AI Multisensor Surveillance is a cutting-edge computer vision platform designed for mission-critical environmental monitoring and threat detection. It integrates Day and Thermal (LWIR) camera feeds to provide 24/7 situational awareness, utilizing state-of-the-art YOLOv8 for detection and DeepSORT for robust tracking.
Optimized for the NVIDIA Jetson Orin AGX, this system delivers real-time inference (<500ms latency) even in degraded visual environments (fog, smoke, low light) thanks to its proprietary Drishyak visibility enhancement module.
- Sensor Fusion: Simultaneous processing of multiple GigE streams.
- Edge AI: Full on-device processing with no cloud dependency.
- Tactical Dashboard: Interactive Streamlit UI for operators.
🚀 Try it now:
- Streamlit Profile - https://share.streamlit.io/user/ratnesh-181998
- Project Demo - https://defence-ai-multisensor-surveillance-yolov8-vusybzt9bohpykhzkq3.streamlit.app/
- Frontend: Streamlit + Custom CSS
- Detection: YOLOv8 + PyTorch + TensorRT
- Tracking: DeepSORT + Kalman Filter
- Processing: OpenCV + NumPy
- Enhancement: CLAHE + Dark Channel Prior
- Hardware: NVIDIA Jetson Orin AGX
- Multi-Spectral Detection: Seamlessly detects targets in RGB and Thermal spectrums using custom-trained YOLOv8 models.
- Robust Tracking: Implements DeepSORT with Kalman filtering for consistent ID retention despite occlusions.
- Kinematics: Estimates target azimuth, elevation, and velocity vectors.
- CLAHE Optimization: Contrast Limited Adaptive Histogram Equalization for detail recovery.
- Dehazing: Dark Channel Prior algorithms to neutralize atmospheric scattering (fog/smoke).
- Auto-Switching: Intelligent pipeline that activates enhancement based on scene analysis.
- TensorRT Acceleration: FP16 precision optimization for 3-5x inference speedup on Jetson.
- Zero-Copy Pipeline: Efficient memory management for high-throughput video processing.
- Asynchronous Design: Non-blocking capture and inference threads.
- Command & Control: Centralized dashboard for system health, camera control, and recording.
- Analytics Suite: Real-time metrics, historical data analysis, and PDF/CSV reporting.
- Event Logging: Comprehensive logging of all system detections and user actions.
The application features a professional, tab-based command center designed for ease of use by defence operators. Below is a detailed breakdown of each interface module:
Functionality: The "Heart" of the system. Operators use this tab to configure signal sources (Webcam vs Simulation), manage the processing sequence (ENGAGE/ABORT), and load AI models.
- Signal Sources: Toggle individual camera feeds (Day/Thermal).
- Sequence Control: One-click system activation with visual status indicators.
- AI Core: Drag-and-drop interface to load custom YOLOv8
.ptor TensorRT.trtmodels.
Tech Used: st.session_state for state management, st.file_uploader, threading control logic.
Functionality: Real-time visualization of all active sensors.
- View Modes: 2x2 Grid, Single Camera Focus, or Split (Day/Thermal).
- Overlays: Bounding boxes, confidence scores, and object IDs (DeepSORT).
- Enhancement: Real-time visibility improvement for fog/smoke.
Tech Used: OpenCV (cv2) for frame manipulation, PIL for image rendering, Queue for threaded video buffering to ensure non-blocking UI.
Functionality: A comprehensive data suite providing operational insights.
- Real-time Metrics: FPS, System Latency, CPU/GPU Usage.
- Detection Trends: Time-series charts showing detection frequency over 1h/6h/24h.
- Class Breakdown: Pie charts showing distribution of detected objects (Person vs Vehicle vs Weapon).
Tech Used: Pandas for data aggregation, Streamlit Native Charts (Altair) for interactive visualization, Psutil for hardware monitoring.
Functionality: Fine-tune the AI "Brain" without restarting the system.
- Confidence Threshold: Slider to filter weak detections (0.0 - 1.0).
- NMS Threshold: Adjustment for Non-Maximum Suppression to remove duplicate boxes.
- Tracking Parameters: Max lost frames and IOU thresholds for DeepSORT.
Tech Used: Dynamic parameter injection into running inference threads.
Functionality: Transparent documentation for engineers.
- Diagrams: Interactive Mermaid.js/Graphviz flowcharts showing data pipeline.
- Dependency Checker: Live status of installed libraries (PyTorch/CUDA versions).
- Reasoning: "Why we chose this stack" comparison tables.
Tech Used: Graphviz for diagrams, json for stack exports.
Functionality: A robust audit trail for mission debriefing.
- Live Console: Auto-scrolling terminal output of all system events.
- Filtering: Show only Errors, Warnings, or Info.
- Export: Download logs as JSON, CSV, or TXT for external analysis.
Tech Used: Custom logging handler, Pandas for CSV export, st.text_area for console view.
The system follows a modular, pipeline-based architecture designed for scalability and fault tolerance:
graph TD
A[Sensors: Day/Thermal Cams] -->|RTSP/GigE| B(Pre-processing)
B -->|Enhancement| C{Drishyak Engine}
C -->|Enhanced Frames| D[Inference Engine: YOLOv8]
D -->|Detections| E[Tracker: DeepSORT]
E -->|Tracks| F[Post-Processing]
F -->|Data| G[Streamlit Dashboard]
F -->|Logs| H[Storage / Analytics]
- CPU: Intel i5 / AMD Ryzen 5 (8th gen+)
- RAM: 16 GB
- GPU: NVIDIA GTX 1060 (6GB VRAM) or better
- Storage: 50 GB SSD
- OS: Windows 10+ / Ubuntu 20.04+
- Platform: NVIDIA Jetson Orin AGX (64GB)
- Cameras: 4× GigE cameras (2× Day, 2× Thermal LWIR)
- Storage: 1 TB NVMe SSD (Industrial Grade)
- Power: 24V DC, rugged enclosure
git clone https://github.com/Ratnesh-181998/Defence-AI-Multisensor-Surveillance-YOLOv8.git
cd Defence-AI-Multisensor-Surveillance-YOLOv8# Create virtual environment
python -m venv venv
# Activate (Windows)
venv\Scripts\activate
# Activate (Linux/Mac)
source venv/bin/activate# For Standard Usage (Streamlit Cloud / CPU)
pip install -r requirements.txt
# For GPU/Jetson Development (Uncomment specific lines in requirements.txt first)
# pip install -r requirements.txt Run the main application:
streamlit run app.pyThe application will launch in your default web browser at http://localhost:8501
- Control Panel: Select active camera inputs (Day-1, Thermal-1, etc.) and click "
▶️ ENGAGE" to start the system. - Live Streams: Monitor real-time feeds with detection overlays. Use the "Snapshot" button to capture evidence.
- Analytics: View statistical breakdowns of detections over time.
- Model Settings: Fine-tune confidence thresholds, IoU, and tracking parameters dynamically.
- Logs: Review, filter, and export system event logs.
- Fork this repository.
- Login to Streamlit Cloud.
- Create a new app pointing to your forked repo.
- Select
app.pyas the main file. - Note: Ensure
requirements.txtis optimized for headless environments (opencv-headless).
# Build Docker image
docker build -t defence-ai:jetson .
# Run container with GPU access
docker run --runtime nvidia --network host --privileged defence-ai:jetson| Component | Latency (ms) | FPS |
|---|---|---|
| Camera Capture | 40 | 25 |
| Preprocessing | 15 | - |
| YOLOv8m Inference | 280 | 3.6 |
| Tracking (DeepSORT) | 25 | - |
| Total Pipeline | ~400 | 2.5 |
Benchmarks recorded on NVIDIA Jetson Orin AGX 64GB in Max Power mode.
RATNESH SINGH
- 📧 Email: rattudacsit2021gate@gmail.com
- 💼 LinkedIn: ratneshkumar1998
- 🐙 GitHub: Ratnesh-181998
- 📱 Phone: +91-947XXXXX46
- 🌐 Live Demo: Streamlit App
- 📖 Documentation: GitHub Wiki
- 🐛 Issue Tracker: GitHub Issues
This project is licensed under the MIT License - see the LICENSE file for details.
Made with ❤️ by Ratnesh Singh
Licensed under the MIT License - Feel free to fork and build upon this innovation! 🚀