Neural Architecture Search (NAS) for deep learning object detection frameworks is computationally expensive due to the vast search space.
The proposed method, FACETS, is a unified iterative NAS technique that refines the architecture of all modules cyclically.
FACETS reduces the search space, preserves interdependencies among modules, and incorporates constraints based on the target device's computational budget.
FACETS achieves higher accuracy and faster search compared to progressive and single-module search strategies.