Skip to content

Latest commit

 

History

History
72 lines (57 loc) · 3.83 KB

File metadata and controls

72 lines (57 loc) · 3.83 KB

Abstract

  • Instruction

    We search efficient MAE-DET backbones for object detection and align with ResNet-50/101. MAE-DET-S uses 60% less FLOPs than ResNet-50; MAE-DET-M is aligned with ResNet-50 with similar FLOPs and number of parameters as ResNet-50; MAE-DET-L is aligned with ResNet-101.

    In this folder, we provide the structure txt and parameters of the model searched by LightNAS for object detection. We follows GFLV2 official repository to train our models, and MAE-DET models are inserted into the pipeline. The code modification is shown in the following steps.

  • Use MAE-DET models in GFLV2 official repository

    1. Copy tinynas/deploy/cnnnet to mmdet/backbones/ in GFLV2

    cd your_lightnas_path && cp -r tinynas/deploy/cnnnet  your_gflv2_path/mmdet/backbones/

    2. Copy maedet_*.txt to gfocal_maedet/, then copy the folder to configs/ in GFLV2

    cp maedet_*.txt gfocal_maedet/ && cp -r gfocal_maedet your_gflv2_path/configs/

    3. Add madnas.py in your_gflv2_path/mmdet/backbones/.

    cp madnas.py your_gflv2_path/mmdet/backbones/

    4. Add the following code snippet in __init__.py of your_gflv2_path/mmdet/backbones/ by importing the initialization of MadNAS as shown bellow.

    from .madnas import MadNas # add this
    
    __all__ = [
    'RegNet', 'ResNet', 'ResNetV1d', 'ResNeXt', 'SSDVGG', 'HRNet', 'Res2Net','HourglassNet', 'DetectoRS_ResNet', 'DetectoRS_ResNeXt', 'Darknet', 
    "MadNas" # add this
    ]

    5. Add the following code snippet after model building in Line 153 of your_gflv2_path/tools/train.py

    if cfg.model.backbone.type in ["MadNas"] and cfg.use_syncBN_torch:
        model = torch.nn.SyncBatchNorm.convert_sync_batchnorm(model)
    logger.info(f'Model:\n{model}')

    6. Finally, follow the instruction of GFLV2 to train the models. If you add the MAE-DET backbones for other pipelines, please refer to this process.


Results and Models

Backbone Param (M) FLOPs (G) box APval box APS box APM box APL Structure Download
ResNet-50 23.5 83.6 44.7 29.1 48.1 56.6 - -
ResNet-101 42.4 159.5 46.3 29.9 50.1 58.7 - -
MAE-DET-S 21.2 48.7 45.1 27.9 49.1 58.0 txt model
MAE-DET-M 25.8 89.9 46.9 30.1 50.9 59.9 txt model
MAE-DET-L 43.9 152.9 47.8 30.3 51.9 61.1 txt model

Note:

  1. The reported numbers here are from the newest simpler code implementation, which may be slightly better than that in the original paper.
  2. These models are trained on COCO dataset with 8 NVIDIA V100 GPUs, where 3X learning schedule from scratch is applied.
  3. Multi-scale training is used with single-scale testing, and the detailed training settings are in the corresponding config files.

Citation

If you use this toolbox in your research, please cite the paper.

@inproceedings{maedet,
  title={MAE-DET: Revisiting Maximum Entropy Principle in Zero-Shot NAS for Efficient Object Detection},
  author={Zhenhong Sun, Ming Lin, Xiuyu Sun, Zhiyu Tan, Hao Li and Rong Jin},
  booktitle={International Conference on Machine Learning},
  year={2022},
  organization={PMLR}
}