MOD-CL: Multi-label Object Detection with Constrained Loss
Sota Moriyama, Koji Watanabe, Katsumi Inoue, Akihiro Takemura
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/sotam2369/MOD-CLOfficialpytorch★ 7
Abstract
We introduce MOD-CL, a multi-label object detection framework that utilizes constrained loss in the training process to produce outputs that better satisfy the given requirements. In this paper, we use MOD_YOLO, a multi-label object detection model built upon the state-of-the-art object detection model YOLOv8, which has been published in recent years. In Task 1, we introduce the Corrector Model and Blender Model, two new models that follow after the object detection process, aiming to generate a more constrained output. For Task 2, constrained losses have been incorporated into the MOD_YOLO architecture using Product T-Norm. The results show that these implementations are instrumental to improving the scores for both Task 1 and Task 2.