Spectral Motion Alignment for Video Motion Transfer Using Diffusion Models

Abstract

Diffusion models have significantly facilitated the customization of input video with target appearance while maintaining its motion patterns. To distill the motion information from video frames, existing works often estimate motion representations as frame difference or correlation in pixel-/feature-space. Despite its simplicity, these methods have unexplored limitations, including lack of understanding of global motion context, and the introduction of motion-independent spatial distortions. To address this, we present Spectral Motion Alignment (SMA), a novel framework that refines and aligns motion representations in the spectral domain. Specifically, SMA learns spectral motion representations, facilitating the learning of whole-frame global motion dynamics, and effectively mitigating motion-independent artifacts. Extensive experiments demonstrate SMA's efficacy in improving motion transfer while maintaining computational efficiency and compatibility across various video customization frameworks.

Cite

Text

Park et al. "Spectral Motion Alignment for Video Motion Transfer Using Diffusion Models." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I6.32685

Markdown

[Park et al. "Spectral Motion Alignment for Video Motion Transfer Using Diffusion Models." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/park2025aaai-spectral/) doi:10.1609/AAAI.V39I6.32685

BibTeX

@inproceedings{park2025aaai-spectral,
  title     = {{Spectral Motion Alignment for Video Motion Transfer Using Diffusion Models}},
  author    = {Park, Geon Yeong and Jeong, Hyeonho and Lee, Sang Wan and Ye, Jong Chul},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {6398-6405},
  doi       = {10.1609/AAAI.V39I6.32685},
  url       = {https://mlanthology.org/aaai/2025/park2025aaai-spectral/}
}