install cuda 11.7 (compatible to torch 1.12.1)
rm -rf ~/anaconda3/envs/mmdemo
conda create --name mmdemo python=3.9 -y
conda activate mmdemo
2. install torch 1.12.1
https://pytorch.org/get-started/previous-versions/
conda install pytorch torchvision torchaudio cudatoolkit=11.6 -c pytorch -c conda-forge
3. mmcv (it takes long time)
(need to mmcv>=1.3.17, <=1.7.0.)
pip install mmcv-full==1.5.0 -f https://download.openmmlab.com/mmcv/dist/cu117/torch1.12.1/index.html
https://github.com/open-mmlab/mmcv
ViTPose
https://github.com/ViTAE-Transformer/ViTPose
git clone https://github.com/ViTAE-Transformer/ViTPose.git
cd ViTPose
pip install -v -e .
pip install timm==0.4.9 einops
mmdet 3rd party package install
pip install mmdet
https://github.com/open-mmlab/mmdetection/blob/master/docs/en/get_started.md/#Installation
pose inference with det
(in ViTPose )
python demo/top_down_video_demo_with_mmdet.py demo/mmdetection_cfg/retinanet_swin-t-p4-w7_fpn_2x_coco.py demo/mmdet_checkpoints/swin_t_p4_w7_naju_epoch_24.pth configs/inter/ViTPose_base_coco_256x192.py work_dirs/ViTPose_base_coco_256x192/best_AP_epoch_70.pth --video-path video/out_7.mp4 --out-video-root outputs (편집됨)