Challenges / Thermal Object Detection
Quick links: Challenge code & data Submit Leaderboard Source dataset paper Ask for help
Night-time and low-visibility conditions are critical for autonomous and assisted navigation of unmanned surface vehicles (USVs). This challenge focuses on object detection in thermal infrared (IR) imagery, encouraging methods robust to non-daylight environments.
The challenge is based on the Maritime Collision Avoidance Dataset Germany, English Channel, and The Netherlands (Gorczak et al.), recorded in German, British, and Dutch waters (2023–2025) and originally annotated using a COLREG-based taxonomy. For this challenge we use thermal infrared (IR) images only.
Given a thermal infrared (IR) image, predict axis-aligned bounding boxes and a class label for each detection.
The challenge uses two classes:
The challenge dataset is derived from the Maritime Collision Avoidance Dataset Germany, English Channel, and The Netherlands (Gorczak et al.). We use thermal infrared (IR) images only and keep the original train/validation split. The new test split is provided as images only (labels withheld) and is used for ranking.
We provide a ready-to-use challenge dataset that already includes all modifications described below. Participants do not need to retrieve the original dataset and apply preprocessing themselves.
The source dataset follows a COLREG-based taxonomy with five classes. For this challenge we simplify the label space to mitigate class imbalance and ambiguity in thermal IR imagery and enable are robst benchmark:
Participants will run inference on the test images (labels are not provided) and submit a single COCO-style JSON containing the predicted bounding boxes, class labels, and confidence scores.
The JSON is uploaded to the evaluation server, which computes metrics automatically using the COCO evaluation protocol. A sample submission file will be provided.
We evaluate predictions using the commonly used COCO object detection protocol and report: AP, AP50, AP75, AR1, and AR10.
The determining metric for winning will be AP. In case of a draw, AP50 counts.
We provide a baseline model (Faster R-CNN with ResNet-50 backbone pretrained on COCO) and starter code to support participation and reproducibility:
If you have questions, please use the MaCVi Support forum.