•  


GitHub - GONGJIA0208/Diffpose: [CVPR 2023] DiffPose: Toward More Reliable 3D Pose Estimation
Skip to content

GONGJIA0208/Diffpose

Repository files navigation

1 JIA GONG *, 1 Lin Geng Foo *, 2 Zhipeng Fan , 3 Qiuhong Ke , 4 Hossein Rahmani , 1 Jun Liu ,

* equal contribution

1 Singapore University of Technology and Design, 2 New York University, 3 Monash University, 4 Lancaster University

[Paper] | [Project Page] | [SUTD-VLG Lab]

DiffPose Model Architecture

DiffPose Diffusion Process

Our code is built on top of DDIM .

Environment

The code is developed and tested under the following environment:

  • Python 3.8.2
  • PyTorch 1.7.1
  • CUDA 11.0

You can create the environment via:

conda env create -f environment.yml

Dataset

Our datasets are based on 3d-pose-baseline and Video3D data . We provide the GMM format data generated from the above datasets here . You should put the downloaded files into the ./data directory. Note that we only change the format of the Video3D data to make them compatible with our GMM-based DiffPose training strategy, and the value of the 2D pose in our dataset is the same as them.

Frame-based experiments

Evaluating pre-trained models for frame-based experiments

We provide the pre-trained diffusion model (with CPN-dected 2D Pose as input) here . To evaluate it, put it into the ./checkpoint directory and run:

CUDA_VISIBLE_DEVICES=0 python main_diffpose_frame.py \
--config human36m_diffpose_uvxyz_cpn.yml --batch_size 1024 \
--model_pose_path checkpoints/gcn_xyz_cpn.pth \
--model_diff_path checkpoints/diffpose_uvxyz_cpn.pth \
--doc t_human36m_diffpose_uvxyz_cpn --exp exp --ni \
>
exp/t_human36m_diffpose_uvxyz_cpn.out 
2>&1
 &

We also provide the pre-trained diffusion model (with Ground truth 2D pose as input) here . To evaluate it, put it into the ./checkpoint directory and run:

CUDA_VISIBLE_DEVICES=0 python main_diffpose_frame.py \
--config human36m_diffpose_uvxyz_gt.yml --batch_size 1024 \
--model_pose_path checkpoints/gcn_xyz_gt.pth \
--model_diff_path checkpoints/diffpose_uvxyz_gt.pth \
--doc t_human36m_diffpose_uvxyz_gt --exp exp --ni \
>
exp/t_human36m_diffpose_uvxyz_gt.out 
2>&1
 &

Training new models

  • To train a model from scratch (CPN 2D pose as input), run:
CUDA_VISIBLE_DEVICES=0 python main_diffpose_frame.py --train \
--config human36m_diffpose_uvxyz_cpn.yml --batch_size 1024 \
--model_pose_path checkpoints/gcn_xyz_cpn.pth \
--doc human36m_diffpose_uvxyz_cpn --exp exp --ni \
>
exp/human36m_diffpose_uvxyz_cpn.out 
2>&1
 &
  • To train a model from scratch (Ground truth 2D pose as input), run:
CUDA_VISIBLE_DEVICES=0 python main_diffpose_frame.py --train \
--config human36m_diffpose_uvxyz_gt.yml --batch_size 1024 \
--model_pose_path checkpoints/gcn_xyz_gt.pth \
--doc human36m_diffpose_uvxyz_gt --exp exp --ni \
>
exp/human36m_diffpose_uvxyz_gt.out 
2>&1
 &

Video-based experiments

Refer to https://github.com/GONGJIA0208/Diffpose_video

Bibtex

If you find our work useful in your research, please consider citing:

@InProceedings{gong2023diffpose,
    author    = {Gong, Jia and Foo, Lin Geng and Fan, Zhipeng and Ke, Qiuhong and Rahmani, Hossein and Liu, Jun},
    title     = {DiffPose: Toward More Reliable 3D Pose Estimation},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2023},
}

Acknowledgement

Part of our code is borrowed from DDIM , VideoPose3D , Graformer , MixSTE and PoseFormer . We thank the authors for releasing the codes.

- "漢字路" 한글한자자동변환 서비스는 교육부 고전문헌국역지원사업의 지원으로 구축되었습니다.
- "漢字路" 한글한자자동변환 서비스는 전통문화연구회 "울산대학교한국어처리연구실 옥철영(IT융합전공)교수팀"에서 개발한 한글한자자동변환기를 바탕하여 지속적으로 공동 연구 개발하고 있는 서비스입니다.
- 현재 고유명사(인명, 지명등)을 비롯한 여러 변환오류가 있으며 이를 해결하고자 많은 연구 개발을 진행하고자 하고 있습니다. 이를 인지하시고 다른 곳에서 인용시 한자 변환 결과를 한번 더 검토하시고 사용해 주시기 바랍니다.
- 변환오류 및 건의,문의사항은 juntong@juntong.or.kr로 메일로 보내주시면 감사하겠습니다. .
Copyright ⓒ 2020 By '전통문화연구회(傳統文化硏究會)' All Rights reserved.
 한국   대만   중국   일본