โจ AeroVision-AI is a next-generation aerial intelligence framework built to understand the world from above.
It leverages the power of YOLOv8, OpenCV, and Streamlit to perform real-time vehicle detection, object tracking, and data-driven analytics on drone-captured footage โ turning raw aerial video into meaningful, structured insight.
AeroVision-AI goes far beyond traditional detection scripts โ it acts as a complete end-to-end visual intelligence system.
Using deep learning and advanced computer vision, it detects moving vehicles from drone footage, tracks them persistently with unique IDs, and generates interactive, analytics-rich visualizations that can be exported as videos, snapshots, and structured CSV reports.
The system is designed with a research-grade backbone and a presentation-ready frontend, allowing users to seamlessly explore data while maintaining production-level polish.
The core philosophy behind AeroVision-AI is simple yet powerful:
โTransform every aerial frame into an intelligent data point.โ
Whether youโre analyzing parking lots, inspecting road traffic, monitoring events, or studying autonomous driving patterns โ AeroVision-AI captures, interprets, and quantifies movement in a way that feels effortless and visually engaging.
๐ก Deep Intelligence โ Built on YOLOv8, AeroVision-AI ensures high accuracy across diverse aerial angles, altitudes, and lighting conditions.
๐งญ Persistent Tracking โ Each vehicle is tracked with consistent IDs, making it ideal for time-based analytics and behavioral observation.
๐ Analytics-Driven Design โ Every frame contributes to measurable insights โ detections, counts, movement trends, and positional data โ stored neatly in exportable CSVs.
๐จ Streamlit-Powered UI โ A beautifully minimal, fully interactive interface to control confidence thresholds, IOU settings, watermark styling, and video recording โ no coding needed.
๐ง Dynamic Watermark Layer โ Add your own branded watermark that tiles, rotates, and blends naturally into each processed frame.
๐ฆ Research & Reporting Ready โ Perfect for demonstrations, traffic analysis, surveillance research, or AI showcase projects.
In a world where drones capture terabytes of video every day, AeroVision-AI turns those pixels into purpose.
It empowers developers, researchers, and organizations to see patterns in motion, quantify behavior, and build smarter aerial ecosystems.
Through intelligent automation and elegant visualization, AeroVision-AI redefines how aerial data is perceived โ
from raw videoโฆ to real-time visual intelligence.
๐งฉ Built for innovation, designed for insight, and engineered for impact.
AeroVision-AI is more than just a car detection script โ itโs a complete aerial intelligence ecosystem that brings drone footage to life through deep learning.
At its core, the system can detect, track, and analyze vehicles in real time, transforming ordinary drone videos into rich visual data streams filled with actionable insights.
Using a combination of YOLOv8, OpenCV, and Streamlit, AeroVision-AI enables users to experience the power of aerial analytics through an intuitive interface and professional-grade outputs.
It empowers you to:
- ๐ฃ๏ธ Monitor and analyze traffic patterns, parking areas, or event zones from an aerial perspective
- ๐ข Automatically count and track every detected vehicle with persistent IDs
- ๐๏ธ Export annotated videos complete with bounding boxes and labeled tracks
- ๐ Generate detailed CSV analytics for research, reporting, or data modeling
- ๐ง Apply branded watermark overlays to maintain visual ownership and authenticity
In essence, AeroVision-AI bridges drone vision and data intelligence, turning raw footage into structured, insightful visual evidence for researchers, developers, and analysts alike.
| ๐ Feature | ๐ก Description |
|---|---|
| ๐ฅ Real-time YOLOv8 Inference | Utilizes the state-of-the-art YOLOv8 object detection model to identify vehicles across frames instantly with remarkable accuracy. |
| ๐งญ Intelligent Object Tracking | Tracks every detected car persistently using an IOU-based tracker, ensuring consistent IDs and smooth temporal continuity across the video stream. |
| ๐ Analytical Data Export | Automatically logs detections into structured CSV files containing frame indices, bounding box coordinates, confidence levels, and class information. |
| ๐ง Dynamic Watermark System | Add customizable, semi-transparent watermark layers with adjustable opacity, rotation, and spacing โ ensuring secure and professional outputs. |
| ๐พ Video & Snapshot Recording | Capture fully processed videos or individual frames as images โ ideal for research presentations or academic reports. |
| ๐งฉ Streamlit Interactive Dashboard | A clean, user-friendly dashboard to upload footage, tweak parameters, and visualize results live โ all without writing a single line of code. |
| โก Optimized Performance | Efficient inference pipeline optimized for both GPU and CPU, allowing smooth operation even on modest setups. |
| ๐งพ Automated Output Management | Each run is automatically organized into time-stamped folders, keeping all MP4, CSV, and snapshot outputs cleanly structured and easy to access. |
๐ง In short: AeroVision-AI transforms aerial visuals into measurable intelligence โ making drones not just eyes in the sky, but analysts in motion.
The architecture of AeroVision-AI is designed with a clear goal โ to create a seamless, modular, and interpretable pipeline that transforms raw drone footage into structured, visual, and analytical intelligence.
Every layer of the system communicates efficiently, ensuring real-time inference, clean visualization, and effortless data export โ all orchestrated within an elegant Streamlit interface.
At its core, the system follows a modular service-based architecture, where each component focuses on a distinct responsibility โ from frame ingestion to analytics generation.
This design not only ensures scalability and clarity but also allows developers and researchers to extend or replace modules with minimal effort.
-
๐ฅ๏ธ Streamlit Frontend (
streamlit_app/app.py)
Acts as the user interface and control center. Users upload drone footage or images, set detection thresholds, choose watermark preferences, and monitor outputs live. -
๐ง YOLOv8 Detector (
src/detector.py)
Processes each incoming frame using YOLOv8, performing real-time vehicle detection with high confidence and precision.
It outputs bounding boxes, class IDs, and confidence scores for each detected object. -
๐งญ Tracker Module (
src/tracker.py)
Uses an IOU (Intersection over Union)-based algorithm to maintain identity consistency across frames.
Each car receives a unique tracking ID, ensuring smooth object association even in overlapping or fast-moving scenes. -
๐จ Visualization Engine (
src/viz.py)
Handles real-time drawing of bounding boxes, track IDs, and overlay elements such as FPS counters and watermark layers.
Ensures processed frames are both informative and presentation-ready. -
๐พ Export Services
Captures processed frames and writes them as:- ๐ฅ Annotated videos (
.mp4) - ๐ Analytics CSV logs (
.csv) - ๐ผ๏ธ Snapshots (
.png)
All outputs are stored neatly in the/outputsdirectory, automatically organized by timestamp.
- ๐ฅ Annotated videos (
-
๐ Reporting & Analysis Layer (
/reports)
Aggregates key visualizations, architecture references, and summary snapshots that can be used for academic documentation, performance reports, or presentations.
| Principle | Description |
|---|---|
| ๐งฉ Modularity | Each function (detection, tracking, visualization) is isolated yet interconnected โ easy to extend or upgrade individually. |
| โก Real-Time Performance | The entire pipeline is optimized for live frame-by-frame inference without noticeable lag, even on CPUs. |
| ๐ Traceability | Every processed frame, detection, and track is stored and can be audited or analyzed later for reproducibility. |
| ๐ก Scalability | New modules (e.g., multi-class detection, speed estimation) can be integrated with minimal code restructuring. |
| ๐งพ Transparency | A clear, human-readable code structure allows researchers and developers to easily follow and explain each stage of the process. |
In summary, AeroVision-AIโs architecture is not just a collection of scripts โ
itโs a living ecosystem of interconnected components, purpose-built to bridge drone vision and data science.
Each frame enters as raw imagery and exits as interpretable intelligence, ready for visualization, reporting, or further AI modeling.
AeroVision-AI follows a clean, modular project structure, ensuring clarity, scalability, and maintainability.
Each directory serves a distinct role in the pipeline โ from deep learning inference to visualization and reporting.
AeroVision-AI/ โ โโโ src/ โ โโโ detector.py # โ๏ธ YOLOv8-based vehicle detection service โ โโโ tracker.py # ๐ฏ IOU-based object tracking module โ โโโ viz.py # ๐ผ๏ธ Frame drawing, watermarking & export utilities โ โโโ init.py # Initializes the src module โ โโโ streamlit_app/ โ โโโ app.py # ๐งฉ Streamlit frontend for user interaction & visualization โ โโโ outputs/ # ๐พ Automatically saved processed videos, CSVs, and snapshots โ โโโ reports/ # ๐ Stored screenshots, architecture diagrams, and performance visuals โ โโโ models/ โ โโโ yolov8n.pt # ๐ค Pre-trained YOLOv8 Nano model for inference โ โโโ requirements.txt # ๐ฆ List of dependencies for environment setup โโโ architecture.png # ๐ง System architecture diagram (used in documentation) โโโ README.md # ๐ Project documentation file
| ๐๏ธ Folder / File | ๐งฉ Description |
|---|---|
src/ |
Core backend containing detection, tracking, and visualization logic. |
streamlit_app/ |
Frontend layer for user uploads, parameter control, and live results display. |
outputs/ |
Stores all generated outputs โ annotated videos, CSV analytics, and saved frames. |
reports/ |
Includes documentation visuals, screenshots, and research-ready figures. |
models/ |
Houses pre-trained weights or custom model checkpoints for YOLOv8. |
requirements.txt |
Defines all project dependencies for environment setup. |
architecture.png |
High-level architecture overview diagram used in documentation. |
README.md |
The main project guide (this file). |
AeroVision-AI leverages a modern, efficient, and research-friendly technology stack, blending the flexibility of Python with state-of-the-art AI libraries for computer vision, analytics, and UI development.
| ๐งฉ Category | โ๏ธ Technologies Used | ๐ง Purpose |
|---|---|---|
| ๐จ Frontend / UI | Streamlit | Builds an interactive dashboard for real-time visualization and parameter control. |
| ๐ค Core ML Model | YOLOv8 (Ultralytics) | Performs high-speed, high-accuracy vehicle detection in aerial frames. |
| ๐งญ Vision Processing | OpenCV, NumPy | Handles image transformations, video frame operations, and pixel-level processing. |
| ๐ฏ Tracking Engine | IOU-based Tracker | Maintains consistent tracking IDs across frames for accurate motion analysis. |
| ๐ Visualization Layer | Matplotlib, OpenCV | Renders bounding boxes, tracks, FPS counters, and dynamic watermark overlays. |
| ๐ Data Handling | CSV, Tempfile | Stores analytical data, tracks detections, and supports file streaming. |
| โ๏ธ Backend Logic | Python 3.10+ | Provides overall orchestration, module integration, and environment compatibility. |
๐ง In short:
The AeroVision-AI tech stack is engineered for speed, modularity, and interpretability, making it ideal for research, prototyping, and real-world drone analytics.
Follow these steps to set up and run AeroVision-AI on your local machine.
The setup process is designed to be simple, portable, and consistent across all operating systems.
Start by cloning the repository and navigating into the project directory:
git clone https://github.com/mwasifanwar/AeroVision-AI.git cd AeroVision-AI
โ๏ธ 2๏ธโฃ Create a Virtual Environment
Create and activate a virtual environment to keep dependencies isolated and clean:
python -m venv .venv
source .venv/bin/activate
.venv\Scripts\activate
๐ฆ 3๏ธโฃ Install Dependencies
Install all required Python packages using requirements.txt
pip install -r requirements.txt
This will automatically install the core dependencies:
๐ค YOLOv8 (Ultralytics) โ Object detection engine
๐งญ OpenCV โ Video and image processing
๐จ Streamlit โ Interactive dashboard framework
๐งฎ NumPy, Matplotlib โ Numerical operations and visualization tools
๐ง 4๏ธโฃ Run the Streamlit App
Launch the application using Streamlit: streamlit run streamlit_app/app.py
Using AeroVision-AI is straightforward โ no coding required!
Simply follow these steps inside the Streamlit interface:
-
Launch the App
- Run the app using the command:
streamlit run streamlit_app/app.py
- Run the app using the command:
-
Upload Your Media
- ๐ฅ Drone Videos:
.mp4,.avi,.mov - ๐ผ๏ธ Still Images:
.jpg,.jpeg,.png
- ๐ฅ Drone Videos:
-
Adjust Parameters
- ๐ Detection Confidence: Filter out low-confidence predictions for cleaner results.
- ๐ฏ IOU Threshold: Tune the trackerโs sensitivity to maintain smoother and more consistent object IDs.
- ๐ง Watermark Settings: Customize text, opacity, angle, and spacing to match your branding or report style.
-
Start Detection
- Press
โถ๏ธ Start to begin real-time inference and object tracking. - The system processes frames dynamically, applying YOLOv8 detections and tracking IDs to each vehicle.
- Press
-
Visualize Results Live
- Watch detections, bounding boxes, IDs, and FPS counters update frame-by-frame in real time on the Streamlit dashboard.
- Easily toggle between modes, adjust confidence levels, and fine-tune visualization without restarting.
-
Export Outputs
- ๐พ Processed Video (.mp4): Includes bounding boxes, track IDs, and optional watermark overlay.
- ๐ Detections CSV (.csv): Contains detailed analytics including frame index, object ID, class label, confidence score, and bounding box coordinates.
- ๐ผ๏ธ Snapshots (.png): Saves static frames for documentation, research reports, or presentations.
All generated visuals and analytical results from AeroVision-AI are automatically stored inside the outputs/ folder after each session.
Additionally, curated screenshots are placed in the reports/ directory for documentation and research presentation.
| Example | Description |
|---|---|
๐ผ๏ธ reports/01.png |
Drone overview showcasing multiple vehicles detected simultaneously using YOLOv8. |
๐ผ๏ธ reports/02.png |
Persistent tracking view โ each vehicle maintains a unique ID across frames. |
๐ผ๏ธ reports/03.png |
Custom watermark overlay applied to ensure brand identity and professional presentation. |
๐ผ๏ธ reports/04.png |
CSV-exported analytics preview displaying structured detection and tracking data. |
๐ผ๏ธ reports/05.png |
Final visualization snapshot, ideal for reports, demonstrations, or publications. |
Example: Real-time vehicle detection and tracking from aerial footage using YOLOv8.
AeroVision-AI automatically logs every detection and tracking event into structured CSV files, enabling seamless analysis and post-processing.
Each frame processed during video inference generates a corresponding entry that includes bounding box coordinates, object IDs, class labels, and confidence values.
Each CSV file follows a standardized format for easy integration with Python, Excel, or Power BI:
| Frame | Track ID | Class | Confidence | X1 | Y1 | X2 | Y2 |
|---|---|---|---|---|---|---|---|
| 132 | 7 | car | 0.94 | 380 | 200 | 420 | 250 |
- For every frame in your uploaded video, YOLOv8 performs object detection.
- The IOU-based tracker assigns a unique Track ID to each vehicle and maintains its identity across frames.
- The system records all detections into a CSV file, stored automatically in the
/outputsfolder. - The result is a chronological detection log that can be visualized, aggregated, or used for model fine-tuning and downstream analytics.
- โ Quantitative Insight: Transform raw drone footage into measurable data points.
- โ Versatile Integration: Import CSVs into Python (Pandas), Excel, Tableau, or Power BI.
- โ Research-Ready: Ideal for academic analysis, traffic pattern studies, or dataset creation.
- โ Automation Friendly: Supports batch processing and consistent export structure.
๐ Example Use Case:
- Compute traffic density per frame.
- Plot vehicle trajectories over time.
- Generate heatmaps for congestion analysis.
The roadmap for AeroVision-AI includes several planned upgrades to elevate functionality, performance, and scalability.
These features aim to make the platform more powerful for both research and real-world deployments.
| Feature | Description |
|---|---|
| ๐ฆ Speed Estimation | Compute approximate vehicle speed using frame-to-frame position differentials and FPS data. |
| ๐งญ Zone Counting | Define polygonal or rectangular regions in the video to count vehicle entry and exit events. |
| โ๏ธ Web Deployment | Deploy directly on platforms like Hugging Face Spaces, Render, or Streamlit Cloud for public demos. |
| ๐ Multi-Class Support | Extend model capabilities beyond cars to include trucks, motorbikes, and pedestrians. |
| ๐ง SHAP / LIME Explainability | Integrate interpretability modules to visualize how the model makes decisions (explainable AI). |
| ๐ Real-time Dashboard | Add live analytical visualization (charts, graphs, traffic density meters) using Plotly or Dash. |
- MLOps Integration: Automate retraining pipelines using Docker + GitHub Actions.
- Edge Optimization: Quantize YOLOv8 for real-time deployment on Jetson or Raspberry Pi.
- Cloud Storage Sync: Auto-upload processed outputs to Google Drive / S3.
- REST API Endpoint: Serve live predictions via FastAPI for third-party integration.
- UI Enhancements: Add interactive analytics, filter controls, and report builders within Streamlit.
๐งฉ In summary:
The foundation of AeroVision-AI is designed with modular extensibility โ enabling future upgrades without breaking the existing architecture.
Itโs not just a drone detection system; itโs an evolving framework for intelligent aerial analytics.
Muhammad Wasif
AI/ML Developer โข Founder @ Effixly AI
"Predicting churn isnโt just about saving customers โ itโs about understanding them."



