Introduction
Wildfires pose a critical threat to ecosystems, infrastructure, and human life. Timely and accurate detection is essential for effective intervention and mitigation. However, developing high-performing object detection models for wildfire detection is often constrained by the lack of labeled data and the time-intensive process of manual annotation.
This project presents an end-to-end AutoML pipeline for wildfire detection using a CI/CD/CT (Continuous Integration, Continuous Deployment, and Continuous Training) architecture. The pipeline automates the entire lifecycle of a detection model. It starts from raw image collection and continues through pre-labeling, human validation, training, distillation, quantization, and deployment.
Motivation
Manual labeling of wildfire imagery is time-consuming and error-prone. In addition, models degrade over time as environmental conditions and data distributions shift. Our system aims to continuously learn from new data using a scalable, semi-supervised approach. It automates as much of the machine learning workflow as possible and involves human review only when necessary.
Key Features
Automated pre-labeling using YOLOv8 and Grounding DINO
Model matching and validation using IoU and confidence thresholds
Human-in-the-loop review for mismatches via Label Studio
Image augmentation to improve generalization
End-to-end training, distillation, and quantization
CI/CD/CT-compatible design for regular updates and retraining
Workflow Overview
Data Collection
Unlabeled wildfire images are collected from remote sensors and placed into a raw image directory.Pre-labeling (YOLO and Grounding DINO)
Both models generate bounding boxes independently. YOLO is fast and lightweight. Grounding DINO supports natural language prompts.Matching
Predictions from both models are matched using class name and IoU. Unmatched results are flagged for human review.Human-in-the-Loop Review
Label Studio is used to manually verify or correct mismatched results.Augmentation
Verified labeled images are augmented to enrich the dataset.Training
A new YOLO model is trained on the augmented dataset.Distillation and Quantization
The full model is distilled into a lightweight version and then quantized for deployment.Model Registry Update
Trained models are stored in the registry and used for future pre-labeling.
This pipeline ensures scalability, adaptability, and model freshness without relying heavily on constant manual labeling. The integration of human review only when needed helps balance efficiency with accuracy.