Skip to content
/ afa Public

Official Code of "Ensembling Diffusion Models via Adaptive Feature Aggregation".

License

Notifications You must be signed in to change notification settings

Tenvence/afa

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Ensembling Diffusion Models via Adaptive Feature Aggregation

Cong Wang1*, Kuan Tian2*, Yonghang Guan2, Jun Zhang2†, Zhiwei Jiang1†, Fei Shen2, Xiao Han2, Qing Gu1, Wei Yang2
1 Nanjing University, 2 Tencent AI Lab
* Equal contribution. Corresponding authors.

[arXiv]

Abstract

The success of the text-guided diffusion model has inspired the development and release of numerous powerful diffusion models within the open-source community. These models are typically fine-tuned on various expert datasets, showcasing diverse denoising capabilities. Leveraging multiple high-quality models to produce stronger generation ability is valuable, but has not been extensively studied. Existing methods primarily adopt parameter merging strategies to produce a new static model. However, they overlook the fact that the divergent denoising capabilities of the models may dynamically change across different states, such as when experiencing different prompts, initial noises, denoising steps, and spatial locations. In this paper, we propose a novel ensembling method, Adaptive Feature Aggregation (AFA), which dynamically adjusts the contributions of multiple models at the feature level according to various states (i.e., prompts, initial noises, denoising steps, and spatial locations), thereby keeping the advantages of multiple diffusion models, while suppressing their disadvantages. Specifically, we design a lightweight Spatial-Aware Block-Wise (SABW) feature aggregator that adaptive aggregates the block-wise intermediate features from multiple U-Net denoisers into a unified one. The core idea lies in dynamically producing an individual attention map for each model's features by comprehensively considering various states. It is worth noting that only SABW is trainable with about 50 million parameters, while other models are frozen. Both the quantitative and qualitative experiments demonstrate the effectiveness of our proposed Adaptive Feature Aggregation method.

Reference

@article{wang2024ensembling,
  title={Ensembling Diffusion Models via Adaptive Feature Aggregation},
  author={Wang, Cong and Tian, Kuan and Guan, Yonghang and Zhang, Jun and Jiang, Zhiwei and Shen, Fei and Han, Xiao and Gu, Qing and Yang, Wei},
  journal={arXiv preprint arXiv:2405.17082},
  year={2024}
}

About

Official Code of "Ensembling Diffusion Models via Adaptive Feature Aggregation".

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages