Efficient quantization for CPU-based diffusion models

Loading...
Thumbnail Image

Journal Title

Journal ISSN

Volume Title

Publisher

Faculty of Applied Sciences, South Eastern University of Sri Lanka, Sammanthurai.

Abstract

The utilization of diffusion models to create visuals from textual descriptions has grown in popularity. However, the significant requirement for computing power still poses a significant obstacle and adds time to procedures. Diffusion models provide difficulties when quantization, a method used to reduce deep learning models for increased efficiency, is used. Comparing to other model types, these models are noticeably more susceptible to quantization, which could lead to deterioration in image quality. In this research, we present a unique method that uses distillation along with quantization aware training to measure the diffusion models. Our findings demonstrate that quantized models can provide inference efficiency on CPUs while retaining great image quality. At https://github.com/intel/intel-extension-for-transformers, the source is accessible to the general public.

Description

Citation

12th Annual Science Research Sessions 2023 (ASRS-2023) Conference Proceedings of "Exploration Towards Green Tech Horizons”. 14th December 2023. Faculty of Applied Sciences, South Eastern University of Sri Lanka, Sammanthurai, Sri Lanka. pp. 36.

Endorsement

Review

Supplemented By

Referenced By