开源软件名称(OpenSource Name):lyp-deeplearning/MOS-Multi-Task-Face-Detect开源软件地址(OpenSource Url):https://github.com/lyp-deeplearning/MOS-Multi-Task-Face-Detect开源编程语言(OpenSource Language):Python 63.9%开源软件介绍(OpenSource Introduction):MOS-Multi-Task-Face-DetectIntroductionThis repo is the official implementation of "MOS: A Low Latency and Lightweight Framework for Face Detection, Landmark Localization, and Head Pose Estimation". The paper has been accepted at BMVC2021. This repo is an implementation of PyTorch. MOS is a low latency and lightweight architecture for face detection, facial landmark localization and head pose estimation.It aims to bridge the gap between research and industrial communities. For more details, please refer to our report on Arxiv. Updates
Comming soon
BenchmarkLight Models.WiderFace Val Performance is in multi scale and Pose evaluation is using AFLW2000 in 300X300 as image input.
generate widerface validation results
python test_widerface.py --network cfg_mos_m --trained_model ./test_weights/MOS-M.pth
cd ./widerface_evaluate
python setup.py build_ext --inplace
python evaluation.py Training data
./data/widerface/
train/
images/
label.txt The annotation file is like: # 0--Parade/0_Parade_marchingband_1_849.jpg
449 330 122 149 488.906 373.643 0.0 542.089 376.442 0.0 515.031 412.83 0.0 485.174 425.893 0.0 538.357 431.491 0.0 0.82 -6 -6 1
face_x face_y face_width face_height landmark1.x landmark1.y 0.0 landmark2.x landmark2.y 0.0 landmark3.x landmark3.y 0.0 landmark4.x landmark4.y 0.0
landmark5.x landmark5.y 0.0 confidence pitch yaw roll Quick StartStep1. Install MOS. git clone https://github.com/lyp-deeplearning/MOS-Multi-Task-Face-Detect.git
cd MOS-Multi-Task-Face-Detect
conda create -n MOS python=3.8.5
conda activate MOS
pip install -r requirements.txt
cd models/DCNv2/
python setup.py build develop Step2. Run Pytorch inference demo. ## run the MOS-M model
python detect_picture.py --network cfg_mos_m --trained_model ./test_weights/MOS-M.pth
## run the MOS-S model
python detect_picture.py --network cfg_mos_s --trained_model ./test_weights/MOS-S.pth Step3. Run video inference demo. ## run the MOS-M model
python detect_video.py --network cfg_mos_m --trained_model ./test_weights/MOS-M.pth Cite MOSIf you use MOS in your research, please cite our work by using the following BibTeX entry: @article{liu2021mos,
title={MOS: A Low Latency and Lightweight Framework for Face Detection, Landmark Localization, and Head Pose Estimation},
author={Liu, Yepeng and Gu, Zaiwang and Gao, Shenghua and Wang, Dong and Zeng, Yusheng and Cheng, Jun},
journal={arXiv preprint arXiv:2110.10953},
year={2021}
} |
2023-10-27
2022-08-15
2022-08-17
2022-09-23
2022-08-13
请发表评论