Inversion-Based Face Swapping With Diffusion Model

Face swapping involves replacing a face in an image with another face and ensuring the seamless integration of the source face into the target image. Previous studies have primarily utilized generative adversarial network-based models for face swapping. This paper introduces inversion-based face swa...

Full description

Saved in:
Bibliographic Details
Main Authors: Daehyun Yoo, Hongchul Lee, Jiho Kim
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10804772/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1841536209884545024
author Daehyun Yoo
Hongchul Lee
Jiho Kim
author_facet Daehyun Yoo
Hongchul Lee
Jiho Kim
author_sort Daehyun Yoo
collection DOAJ
description Face swapping involves replacing a face in an image with another face and ensuring the seamless integration of the source face into the target image. Previous studies have primarily utilized generative adversarial network-based models for face swapping. This paper introduces inversion-based face swapping (InFS), a novel framework employing diffusion inversion. The key contributions of our work include: 1) a facial attribute encoder that consolidates attribute information into a single embedding vector, utilizing the architecture of the pSp encoder, and 2) an enhanced face swapping pipeline that overcomes pose limitations through reenactment preprocessing addressing the challenge of incorrect face swapping at extreme angles. To preserve the target image’s attribute information that may be lost during the diffusion inversion process, we incorporate the extracted information from the facial attribute encoder. This embedding vector serves as a crucial condition in the diffusion inversion process, facilitating the prediction of noisy images. Subsequently, the predicted noisy image undergoes processing using a pretrained ID conditional DDPM for face swapping. Our experimental results show that InFS outperforms state-of-the-art methods in preserving identity, expression, and shape characteristics of target images. Furthermore, the proposed InFS achieved effective face swapping results without requiring additional guidance and reduces inference time by approximately 6.96 seconds compared to previous diffusion-based approaches.
format Article
id doaj-art-d3421e50abf5439fbfe32c9190e6bdf9
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-d3421e50abf5439fbfe32c9190e6bdf92025-01-15T00:02:53ZengIEEEIEEE Access2169-35362025-01-01136764677410.1109/ACCESS.2024.351916310804772Inversion-Based Face Swapping With Diffusion ModelDaehyun Yoo0https://orcid.org/0009-0002-3362-8438Hongchul Lee1Jiho Kim2https://orcid.org/0000-0003-3733-8702School of Industrial and Management Engineering, Korea University, Seoul, Republic of KoreaSchool of Industrial and Management Engineering, Korea University, Seoul, Republic of KoreaSchool of Industrial and Management Engineering, Korea University, Seoul, Republic of KoreaFace swapping involves replacing a face in an image with another face and ensuring the seamless integration of the source face into the target image. Previous studies have primarily utilized generative adversarial network-based models for face swapping. This paper introduces inversion-based face swapping (InFS), a novel framework employing diffusion inversion. The key contributions of our work include: 1) a facial attribute encoder that consolidates attribute information into a single embedding vector, utilizing the architecture of the pSp encoder, and 2) an enhanced face swapping pipeline that overcomes pose limitations through reenactment preprocessing addressing the challenge of incorrect face swapping at extreme angles. To preserve the target image’s attribute information that may be lost during the diffusion inversion process, we incorporate the extracted information from the facial attribute encoder. This embedding vector serves as a crucial condition in the diffusion inversion process, facilitating the prediction of noisy images. Subsequently, the predicted noisy image undergoes processing using a pretrained ID conditional DDPM for face swapping. Our experimental results show that InFS outperforms state-of-the-art methods in preserving identity, expression, and shape characteristics of target images. Furthermore, the proposed InFS achieved effective face swapping results without requiring additional guidance and reduces inference time by approximately 6.96 seconds compared to previous diffusion-based approaches.https://ieeexplore.ieee.org/document/10804772/Face swappingdiffusion modelinversionattribute encoderimage generation
spellingShingle Daehyun Yoo
Hongchul Lee
Jiho Kim
Inversion-Based Face Swapping With Diffusion Model
IEEE Access
Face swapping
diffusion model
inversion
attribute encoder
image generation
title Inversion-Based Face Swapping With Diffusion Model
title_full Inversion-Based Face Swapping With Diffusion Model
title_fullStr Inversion-Based Face Swapping With Diffusion Model
title_full_unstemmed Inversion-Based Face Swapping With Diffusion Model
title_short Inversion-Based Face Swapping With Diffusion Model
title_sort inversion based face swapping with diffusion model
topic Face swapping
diffusion model
inversion
attribute encoder
image generation
url https://ieeexplore.ieee.org/document/10804772/
work_keys_str_mv AT daehyunyoo inversionbasedfaceswappingwithdiffusionmodel
AT hongchullee inversionbasedfaceswappingwithdiffusionmodel
AT jihokim inversionbasedfaceswappingwithdiffusionmodel