Abstract
We introduce the task of human motion unlearning to prevent the synthesis of toxic animations while preserving the general text-to-motion generative performance. Unlearning toxic motions is challenging as those can be generated from explicit text prompts and from implicit toxic combinations of safe motions (e.g., ``kicking" is ``loading and swinging a leg"). We propose the first motion unlearning benchmark by filtering toxic motions from the large and recent text-to-motion datasets of HumanML3D and Motion-X. We propose baselines, by adapting state-of-the-art image unlearning techniques to process spatio-temporal signals. Finally, we propose a novel motion unlearning model based on Latent Code Replacement, which we dub LCR. LCR is training-free and suitable to the discrete latent spaces of state-of-the-art text-to-motion diffusion models. LCR is simple and consistently outperforms baselines qualitatively and quantitatively. Project page: https://www.pinlab.org/hmu{https://www.pinlab.org/hmu}.
Community
In this work, we introduce a fast, training-free approach to remove toxic motions e.g., violent actions, from text-to-motion models. Our method takes only a few seconds to apply.
We’ll be releasing the code, along with the models and processed datasets, as soon as possible. Stay tuned!
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- CASIM: Composite Aware Semantic Injection for Text to Motion Generation (2025)
- SALAD: Skeleton-aware Latent Diffusion for Text-driven Motion Generation and Editing (2025)
- Motion Anything: Any to Motion Generation (2025)
- ACMo: Attribute Controllable Motion Generation (2025)
- Unlocking Pretrained LLMs for Motion-Related Multimodal Generation: A Fine-Tuning Approach to Unify Diffusion and Next-Token Prediction (2025)
- GenM$^3$: Generative Pretrained Multi-path Motion Model for Text Conditional Human Motion Generation (2025)
- MAG: Multi-Modal Aligned Autoregressive Co-Speech Gesture Generation without Vector Quantization (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on HF中国镜像站 checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper