SEUIR Repository

Efficient quantization for CPU-based diffusion models

Show simple item record

dc.contributor.author Hanees, A. L.
dc.contributor.author Elango, E.
dc.date.accessioned 2024-03-15T06:38:44Z
dc.date.available 2024-03-15T06:38:44Z
dc.date.issued 2023-12-14
dc.identifier.citation 12th Annual Science Research Sessions 2023 (ASRS-2023) Conference Proceedings of "Exploration Towards Green Tech Horizons”. 14th December 2023. Faculty of Applied Sciences, South Eastern University of Sri Lanka, Sammanthurai, Sri Lanka. pp. 36. en_US
dc.identifier.isbn 978-955-627-015-0
dc.identifier.uri http://ir.lib.seu.ac.lk/handle/123456789/6982
dc.description.abstract The utilization of diffusion models to create visuals from textual descriptions has grown in popularity. However, the significant requirement for computing power still poses a significant obstacle and adds time to procedures. Diffusion models provide difficulties when quantization, a method used to reduce deep learning models for increased efficiency, is used. Comparing to other model types, these models are noticeably more susceptible to quantization, which could lead to deterioration in image quality. In this research, we present a unique method that uses distillation along with quantization aware training to measure the diffusion models. Our findings demonstrate that quantized models can provide inference efficiency on CPUs while retaining great image quality. At https://github.com/intel/intel-extension-for-transformers, the source is accessible to the general public. en_US
dc.language.iso en_US en_US
dc.publisher Faculty of Applied Sciences, South Eastern University of Sri Lanka, Sammanthurai. en_US
dc.subject Quantization en_US
dc.subject Diffusion Models en_US
dc.subject U-Net Architecture en_US
dc.title Efficient quantization for CPU-based diffusion models en_US
dc.type Article en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search SEUIR


Advanced Search

Browse

My Account