开源软件名称(OpenSource Name):dk-liang/FIDTM开源软件地址(OpenSource Url):https://github.com/dk-liang/FIDTM开源编程语言(OpenSource Language):Python 100.0%开源软件介绍(OpenSource Introduction):Focal Inverse Distance Transform Map
NewsWe now provide the predicted coordinates txt files, and other researchers can use them to fairly evaluate the localization performance. OverviewVisualizationsVisualizations for bounding boxes Progress
Environment
Datasets
Generate FIDT Ground-Truth
“xx” means the dataset name, including sh, jhu, qnrf, and nwpu. You should change the dataset path. ModelDownload the pretrained model from Baidu-Disk, passward:gqqm, or OneDrive Quickly test
Download Dataset and Model
Test example:
If you want to generate bounding boxes,
If you want to test a video,
Visiting bilibili or Youtube to watch the video demonstration. The original demo video can be downloaded from Baidu-Disk, passed: cebh More config information is provided in config.py Evaluation localization performance
Evaluation example: For Shanghai tech, JHU-Crowd (test set), and NWPU-Crowd (val set):
For UCF-QNRF dataset:
For NWPU-Crowd (test set), please submit the nwpu_pred_fidt.txt to the website. We also provide the predicted coordinates txt file in './local_eval/point_files/', and you can use them to fairly evaluate the other localization metric. (We hope the community can provide the predicted coordinates file to help other researchers fairly evaluate the localization performance.) Tips:
The predicted format is:
The evaluation code is modifed from NWPU. TrainingThe training strategy is very simple. You can replace the density map with the FIDT map in any regressors for training. If you want to train based on the HRNET (borrow from the IIM-code link), please first download the ImageNet pre-trained models from the official link, and replace the pre-trained model path in HRNET/congfig.py (__C.PRE_HR_WEIGHTS). Here, we provide the training baseline code, and the I-SSIM loss will be released when the review is completed. Training baseline example:
For ShanghaiTech, you can train by a GPU with 8G memory. For other datasets, please utilize a single GPU with 24G memory or multiple GPU for training. We have reorganized the code, which is usually better than the results of the original manuscript. Improvements We have not studied the effect of some hyper-parameter. Thus, the results can be further improved by using some tricks, such as adjust the learning rate, batch size, crop size, and data augmentation. ReferenceIf you find this project is useful for your research, please cite:
|
2023-10-27
2022-08-15
2022-08-17
2022-09-23
2022-08-13
请发表评论