Skip to content

whuhxb/Efficient-Audio-Gesture

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

33 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Efficient-Audio-Gesture

This is the official repository of the two papers.

(πŸ‘€ DIDiffGes elaborates on the process and inference of semi-implicit accelerated diffusion. This part plays a crucial role in the efficient generation of HoloGest.)

πŸ”₯(AAAI 2025) DIDiffGes: Decoupled Semi-Implicit Diffusion Models for Real-time Gesture Generation from Speech

The 39th Annual AAAI Conference on Artificial Intelligence (AAAI), 2025

[paper is here!]

πŸ”₯(3DV 2025) HoleGest: Decoupled Diffusion and Motion Priors for Generating Holisticly Expressive Co-speech Gestures

International Conference on 3D Vision 2025 (3DV), 2025

[Project Page] [Paper]

Method

Image

We used an avatar to conduct an audio narration of our method, vividly elaborating on the details of our method to everyone.

demo_avatar.mp4

News 🚩

  • [2025/03/17] Code of HoloGest release ⭐
  • [2024/12/15] DIDiffGes got accepted by AAAI 2025! πŸŽ‰
  • [2024/11/10] HoloGest got accepted by 3DV 2025! πŸŽ‰

1. Getting started

This code was tested on NVIDIA GeForce RTX 3070 Ti and requires:

  • conda3 or miniconda3
cd ./main/
bash pip_install.sh

2. Pre-trained model and data

  • Download CLIP model , ASRand pre-trained weights from here. Put all the folder in ./main/holgest/.
  • Download TextEncoder and put it in ./main/model/
  • Download WavLM weights from here and put it in ./main/hologest
  • Download Motion-Prior from here and put it in ./main/hologest

3. Quick Start

bash demo.sh

Applications

git_hologest_demo.mp4

Citation

@inproceedings{yu2023acr,
  title = {ACR: Attention Collaboration-based Regressor for Arbitrary Two-Hand Reconstruction},
  author = {Yu, Zhengdi and Huang, Shaoli and Chen, Fang and Breckon, Toby P. and Wang, Jue},
  booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  month     = {June},
  year      = {2023}
  }

Acknowledgement

The pytorch implementation of HoloGest is based on ExpGest. We use some parts of the knowledgement from SiDDMs and some part of code from [DIDiffGes]. We thank all the authors for their impressive works!

Contact

For technical questions, please contact cyk19990422@gmail.com

For commercial licensing, please contact shaolihuang@tencent.com

About

πŸ”₯(3DV 2025) & (AAAI 2025) Real-Time Generating Holisticly Expressive Co-speech Gestures

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors