The mainbody detection technology is currently a widely used detection technology, which refers to a whole image recognition process of identifying the coordinate position of one or more objects and then cropping down the corresponding area for recognition. Mainbody detection is the first step of the recognition task, which can effectively improve the recognition accuracy.
This tutorial will introduce the technology from three aspects, namely, the datasets, model selection and model training.
The datasets we used for mainbody detection tasks are shown in the following table.
Dataset | Image Number | Image Number Used in Mainbody Detection | Scenarios | Dataset Link |
---|---|---|---|---|
Objects365 | 170W | 6k | General Scenarios | Link |
COCO2017 | 12W | 5k | General Scenarios | Link |
iCartoonFace | 2k | 2k | Cartoon Face | Link |
LogoDet-3k | 3k | 2k | Logo | Link |
RPC | 3k | 3k | Product | Link |
In the actual training process, all datasets are mixed together. Categories of all the labeled boxes are modified as foreground
, and the detection model we trained only contains one category (foreground
).
There are a wide variety of object detection methods, such as the commonly used two-stage detectors (FasterRCNN series, etc.), single-stage detectors (YOLO, SSD, etc.), anchor-free detectors (FCOS, etc.) and so on. PaddleDetection has its self-developed PP-YOLO models for server-side scenarios and PicoDet models for end-side scenarios (CPU and mobile), which all take the lead in the area.
Build on the studies above, PaddleClas provides lightweight and server-side main body detection models for end-side scenarios and server-side scenarios respectively. The table below presents the average mAP of the 5 datasets and the comparison of their model sizes and inference speed.
Model | Model Structure | Download Link of Pre-trained Model | Download Link of Inference Model | mAP | Size of Inference Model (MB) | Inference Time per Image (preprocessing excluded)(ms) |
---|---|---|---|---|---|---|
Lightweight Mainbody Detection Model | PicoDet | Link | Link | 40.1% | 30.1 | 29.8 |
Server-side Mainbody Detection Model | PP-YOLOv2 | Link | Link | 42.5% | 210.5 | 466.6 |
Notes:
Intel(R) Xeon(R) Gold 6148 CPU @ 2.40GHz
.The speed indicator is the testing result when mkldnn is on and the number of threads is set to 10.PicoDet, introduced by PaddleDetection, is an object detection algorithm applied to CPU or mobile-side scenarios. It integrates the following optimization algorithm.
For more details of optimized PicoDet and benchmark, you can refer to Tutorials of PicoDet Models.
To balance the detection speed and effects in lightweight mainbody detection tasks, we adopt PPLCNet_x2_5 as the backbone of the model and revise the image scale for training and inference to 640x640, with the rest configured the same as picodet_m_shufflenetv2_416_coco.yml. The final detection model is obtained after the training of customized mainbody detection datasets.
PP-YOLO is proposed by PaddleDetection. It greatly optimizes the yolov3 model from multiple perspectives such as backbone, data augmentation, regularization strategy, loss function, and post-processing. It reaches the state of the art in terms of "speed-precision". The optimization strategy is as follows.
For more information about PP-YOLO, you can refer to PP-YOLO tutorial.
In the mainbody detection task, we use ResNet50vd-DCN
as our backbone for better performance. The config file is ppyolov2_r50vd_dcn_365e_coco.yml, in which the dataset path is modified to the customized mainbody detection dataset. The final detection model can be downloaded here.
This section mainly talks about how to train your own mainbody detection model using PaddleDetection on your own datasets.
Download PaddleDetection and install requirements.
cd <path/to/clone/PaddleDetection>
git clone https://github.com/PaddlePaddle/PaddleDetection.git
cd PaddleDetection
# install requirements
pip install -r requirements.txt
For more installation tutorials, please refer to Installation Tutorial
For customized dataset, you should convert it to COCO format. Please refer to Customized Dataset Tutorial to build your own datasets with COCO format.
In mainbody detection task, all the objects belong to foregroud. Therefore, category_id
of all the objects in the annotation file should be modified to 1. And the categories
map should be modified as follows, in which just class foregroud
is included.
[{u'id': 1, u'name': u'foreground', u'supercategory': u'foreground'}]
We use configs/ppyolo/ppyolov2_r50vd_dcn_365e_coco.yml
to train the model, mode details are as follows.
ppyolov2_r50vd_dcn_365e_coco.yml
depends on other configuration files, their meanings are as follows.
coco_detection.yml:path of train/eval/test dataset.
runtime.yml:public runtime parameters, including whethre to use GPU, epoch number for checkpoint saving, etc.
optimizer_365e.yml:learning rate and optimizer.
ppyolov2_r50vd_dcn.yml:model architecture and backbone.
ppyolov2_reader.yml:train/eval/test reader, such as batch size, the number of concurrently loaded sub-processes, etc., and includes post-read pre-processing operations, such as resize, data enhancement, etc.
In mainbody detection task, you need to modify num_classes
in datasets/coco_detection.yml
to 1 (only foreground
is included), while modify the paths of the training and testing datasets to those of the customized datasets.
In addition, the above files can also be modified according to real situations, for example, if the video memory is overflowing, the batch size and learning rate can be reduced in equal proportion.
PaddleDetection supports many ways of training process.
# not needed for windows and Mac
export CUDA_VISIBLE_DEVICES=0
python tools/train.py -c configs/ppyolo/ppyolov2_r50vd_dcn_365e_coco.yml
export CUDA_VISIBLE_DEVICES=0,1,2,3
python -m paddle.distributed.launch --gpus 0,1,2,3 tools/train.py -c configs/ppyolo/ppyolov2_r50vd_dcn_365e_coco.yml --eval
--eval: evaluation while training
export CUDA_VISIBLE_DEVICES=0
# assign pretrain_weights, load the general mainbody-detection pretrained model
python tools/train.py -c configs/ppyolo/ppyolov2_r50vd_dcn_365e_coco.yml -o pretrain_weights=https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/rec/models/pretrain/ppyolov2_r50vd_dcn_mainbody_v1.0_pretrained.pdparams
Resume training
you can use -r
to load checkpoints and resume training.
export CUDA_VISIBLE_DEVICES=0,1,2,3
python -m paddle.distributed.launch --gpus 0,1,2,3 tools/train.py -c configs/ppyolo/ppyolov2_r50vd_dcn_365e_coco.yml --eval -r output/ppyolov2_r50vd_dcn_365e_coco/10000
Note: If Out of memory error
occurs, you can try to decrease batch_size
in ppyolov2_reader.yml
while reducing learning rate in equal proportion.
Use the following command to finish the prediction process.
export CUDA_VISIBLE_DEVICES=0
python tools/infer.py -c configs/ppyolo/ppyolov2_r50vd_dcn_365e_coco.yml --infer_img=your_image_path.jpg --output_dir=infer_output/ --draw_threshold=0.5 -o weights=output/ppyolov2_r50vd_dcn_365e_coco/model_final
--draw_threshold
is an optional parameter. According to NMS calculation, different thresholds will produce different results. keep_top_k
indicates the maximum number of output targets, with a default value of 100 that can be modified according to their actual situation.
Use the following to export the inference model:
python tools/export_model.py -c configs/ppyolo/ppyolov2_r50vd_dcn_365e_coco.yml --output_dir=./inference -o weights=output/ppyolov2_r50vd_dcn_365e_coco/model_final.pdparams
The inference model will be saved under the directory inference/ppyolov2_r50vd_dcn_365e_coco
, which containsinfer_cfg.yml
(optional for mainbody detection), model.pdiparams
, model.pdiparams.info
, model.pdmodel
.
Note: Inference model that PaddleDetection
exports is named model.xxx
,if you want to keep it consistent with PaddleClas,you can rename model.xxx
to inference.xxx
for subsequent inference deployment of mainbody detection.
For more model export tutorials, please refer to EXPORT_MODEL.
The final directory contains inference/ppyolov2_r50vd_dcn_365e_coco
, inference.pdiparams
, inference.pdiparams.info
, and inference.pdmodel
,among whichinference.pdiparams
refers to saved weight files of the inference model while inference.pdmodel
stands for structural files.
After exporting the model, the path of the detection model can be changed to the inference model path to complete the prediction task.
Take product recognition as an example,you can modify the field Global.det_inference_model_dir
in its config file inference_product.yaml to the directory of exported inference model, and then finish the detection and recognition of the product with reference to Quick Start for Image Recognition.
640x640
resolution, so this is also the default value of prediction process. The accuracy will be reduced if other resolutions are used.Вы можете оставить комментарий после Вход в систему
Неприемлемый контент может быть отображен здесь и не будет показан на странице. Вы можете проверить и изменить его с помощью соответствующей функции редактирования.
Если вы подтверждаете, что содержание не содержит непристойной лексики/перенаправления на рекламу/насилия/вульгарной порнографии/нарушений/пиратства/ложного/незначительного или незаконного контента, связанного с национальными законами и предписаниями, вы можете нажать «Отправить» для подачи апелляции, и мы обработаем ее как можно скорее.
Опубликовать ( 0 )