With the increasing prevalence of online videos featuring pedestrians, the gait information embedded in such content poses significant privacy risks. Previous gait anonymization methods suffer from poor visual naturalness in the synthesized gait and lack precise appearance control. To address these challenges, we present GaitGuard, the first diffusion model-based framework for gait anonymization and protection. Operating as a purely black-box attack method, GaitGuard requires no prior knowledge of the target model, distinguishing it from traditional adversarial approaches. GaitGuard incorporates an Appearance Fusion Module (AFM) and a Motion Fusion Module (MFM) to encode and blend appearance and motion features from different individuals in the latent space, thereby achieving precise control over the generated gait during the denoising process. This design effectively protects gait privacy from recognition systems while ensuring visual naturalness and frame-to-frame consistency in the generated gait video. Extensive experiments on the CASIA-B and OUMVLP datasets demonstrate substantial reductions in recognition accuracy across representative gait recognition models, confirming the robustness and effectiveness of GaitGuard.
The GaitGuard framework consists of four main stages:
(1)Silhouette Extraction and Masking:
Given frame
(2)Pose Estimation and Background Inpainting:
We extract the pose sequence using model
Concurrently, we reconstruct the background using an inpainting model
(3)Gait Generation:
The proposed GaitGen network
(4)Frame Composition:
The generated gait
(5)
The complete GaitGuard framework can be formalized as:
where
To set up this project, please follow these steps:
[MooreThreads AnimateAnyone][https://github.com/MooreThreads/Moore-AnimateAnyone]
[Lama][https://github.com/advimman/lama]






