Action-slot

所属分类:交通/航空行业
开发工具:Python
文件大小:0KB
下载次数:0
上传日期:2024-03-06 12:20:00
上 传 者sh-1993
说明:  [CVPR 2024]动作槽:交通场景中原子活动识别的以视觉动作为中心的表示
([CVPR 2024] Action-slot: Visual Action-centric Representations for Atomic Activity Recognition in Traffic Scenes)

文件列表:
DeepLabV3Plus-Pytorch-master/
configs/
datasets/
img/
models/
scripts/
.DS_Store
LICENSE
requirements.txt

# Action-slot **[CVPR 2024] Action-slot: Visual Action-centric Representations for Multi-label Atomic Activity Recognition in Traffic Scenes** 1[Chi-Hsi Kung](https://hankkung.github.io/website/), 1,[Shu-Wei Lu](https://www.linkedin.com/in/shu-wei-lu/), 2[Yi-Hsuan Tsai](https://sites.google.com/site/yihsuantsai/), 1[Yi-Ting Chen](https://sites.google.com/site/yitingchen0524) 1National Yang Ming Chiao Tung University, 2Google [[arxiv](https://arxiv.org/abs/2311.17948)] [[Project Page](https://hcis-lab.github.io/Action-slot/)] This repository contains code for training and evaluating baselines presented in the paper. ## Installation Create and activate the conda environment: ``` pip install -e . ``` ## Datasets Download **TACO** [[One Drive](https://nycu1-my.sharepoint.com/personal/ychen_m365_nycu_edu_tw/_layouts/15/onedrive.aspx?id=%2Fpersonal%2Fychen%5Fm365%5Fnycu%5Fedu%5Ftw%2FDocuments%2FTACO&ga=1)] **OATS** [[Website](https://usa.honda-ri.com/oats)] ## Train & Evaluation on TACO Training ``` # Action-slot python train_taco.py --dataset taco --root [path_to_TACO] --model_name action_slot --num_slots 64\ --bg_slot --bg_mask --action_attn_weight 1 --allocated_slot --bg_attn_weight 0.5 # X3D python train_taco.py --dataset taco --root [path_to_TACO] --model_name x3d ``` Evaluation ``` # Action-slot python eval_taco.py --cp [path_to_checkpoint] --root [path_to_TACO] --dataset taco\ --model_name action_slot --num_slots 64 --bg_slot --allocated_slot # X3D python eval_taco.py --root [path_to_TACO] --cp [path_to_checkpoint] --dataset taco --model_name x3d ``` ## Train & Evaluation on OATS ``` # Action-slot python train_oats.py --dataset oats --oats_test_split s1 --model_name action_slot --epochs 50\ --num_slots 35 --bg_slot --bg_mask --action_attn_weight 0.1 --allocated_slot\ --bg_attn_weight 0.1 --ego_loss_weight 0 python eval_oats.py --cp [path_to_checkpoint] --dataset oats --oats_test_split s3 --root [path_to_dataset]\ --model_name action_slot --allocated_slot --backbone x3d --num_slots 35 --bg_slot ``` ## Train & Evaluation on nuScenes ``` # train from scratch python train_nuscenes.py --dataset nuscenes --root [path]/nuscenes/trainval/samples\ --model_name action_slot --num_slots 64 --bg_slot --bg_mask --action_attn_weight 1\ --allocated_slot --bg_attn_weight 0.5 --bce_pos_weight 7 # transfer learning: TACO -> nuScenes python train_nuscenes.py --pretrain taco --root [path]/nuscenes/trainval/samples --cp [path_to_checkpoint] --dataset nuscenes\ --model_name action_slot --num_slots 64 --bg_slot --bg_mask --action_attn_weight 1\ --allocated_slot --bg_attn_weight 0.5 --bce_pos_weight 20 --root /media/hcis-s20/SRL/nuscenes/trainval/samples # transfer learning: OATS -> nuScenes python train_nuscenes.py --pretrain oats --root [path]/nuscenes/trainval/samples --cp [path_to_checkpoint] --dataset nuscenes\ --model_name action_slot--num_slots 64 --bg_slot --bg_mask --action_attn_weight 1 --allocated_slot --bg_attn_weight 0.5\ --bce_pos_weight 15 ``` ## Attention Visualization ![image](https://github.com/HCIS-Lab/Action-slot/blob/main/img/taco_attn.gif) ``` python eval_taco.py --cp [path_to_checkpoint] --plot --dataset taco --root [path]/nuscenes/trainval/samples\ --model_name action_slot --num_slots 64 --bg_slot --allocated_slot --plot_threshold 0.5 ``` ## Citation ``` @article{kung2023action, title={Action-slot: Visual Action-centric Representations for Multi-label Atomic Activity Recognition in Traffic Scenes}, author={Kung, Chi-Hsi and Lu, Shu-Wei and Tsai, Yi-Hsuan and Chen, Yi-Ting}, journal={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, year={2024} } ``` ### Acknowledgement * Slot attention is adapted from [Discovering Object that Can Move](https://github.com/zpbao/Discovery_Obj_Move) * DeepLabV3+ is adapted from [DeepLabV3Plus-Pytorch](https://github.com/VainF/DeepLabV3Plus-Pytorch)

近期下载者

相关文件


收藏者