|国家预印本平台
首页|EDiT: Efficient Diffusion Transformers with Linear Compressed Attention

EDiT: Efficient Diffusion Transformers with Linear Compressed Attention

EDiT: Efficient Diffusion Transformers with Linear Compressed Attention

来源:Arxiv_logoArxiv
英文摘要

Diffusion Transformers (DiTs) have emerged as a leading architecture for text-to-image synthesis, producing high-quality and photorealistic images. However, the quadratic scaling properties of the attention in DiTs hinder image generation with higher resolution or on devices with limited resources. This work introduces an efficient diffusion transformer (EDiT) to alleviate these efficiency bottlenecks in conventional DiTs and Multimodal DiTs (MM-DiTs). First, we present a novel linear compressed attention method that uses a multi-layer convolutional network to modulate queries with local information while keys and values are spatially aggregated. Second, we formulate a hybrid attention scheme for multi-modal inputs that combines linear attention for image-to-image interactions and standard scaled dot-product attention for interactions involving prompts. Merging these two approaches leads to an expressive, linear-time Multimodal Efficient Diffusion Transformer (MM-EDiT). We demonstrate the effectiveness of the EDiT and MM-EDiT architectures by integrating them into PixArt-Sigma(conventional DiT) and Stable Diffusion 3.5-Medium (MM-DiT), achieving up to 2.2x speedup with comparable image quality after distillation.

Philipp Becker、Abhinav Mehrotra、Ruchika Chavhan、Malcolm Chadwick、Luca Morreale、Mehdi Noroozi、Alberto Gil Ramos、Sourav Bhattacharya

计算技术、计算机技术

Philipp Becker,Abhinav Mehrotra,Ruchika Chavhan,Malcolm Chadwick,Luca Morreale,Mehdi Noroozi,Alberto Gil Ramos,Sourav Bhattacharya.EDiT: Efficient Diffusion Transformers with Linear Compressed Attention[EB/OL].(2025-03-20)[2025-08-02].https://arxiv.org/abs/2503.16726.点此复制

评论