dc.contributor.author |
Hanees, A. L. |
|
dc.contributor.author |
Elango, E. |
|
dc.date.accessioned |
2024-03-15T06:38:44Z |
|
dc.date.available |
2024-03-15T06:38:44Z |
|
dc.date.issued |
2023-12-14 |
|
dc.identifier.citation |
12th Annual Science Research Sessions 2023 (ASRS-2023) Conference Proceedings of "Exploration Towards Green Tech Horizons”. 14th December 2023. Faculty of Applied Sciences, South Eastern University of Sri Lanka, Sammanthurai, Sri Lanka. pp. 36. |
en_US |
dc.identifier.isbn |
978-955-627-015-0 |
|
dc.identifier.uri |
http://ir.lib.seu.ac.lk/handle/123456789/6982 |
|
dc.description.abstract |
The utilization of diffusion models to create visuals from textual descriptions has grown
in popularity. However, the significant requirement for computing power still poses a
significant obstacle and adds time to procedures. Diffusion models provide difficulties
when quantization, a method used to reduce deep learning models for increased
efficiency, is used. Comparing to other model types, these models are noticeably more
susceptible to quantization, which could lead to deterioration in image quality. In this
research, we present a unique method that uses distillation along with quantization
aware training to measure the diffusion models. Our findings demonstrate that
quantized models can provide inference efficiency on CPUs while retaining great image
quality. At https://github.com/intel/intel-extension-for-transformers, the source is
accessible to the general public. |
en_US |
dc.language.iso |
en_US |
en_US |
dc.publisher |
Faculty of Applied Sciences, South Eastern University of Sri Lanka, Sammanthurai. |
en_US |
dc.subject |
Quantization |
en_US |
dc.subject |
Diffusion Models |
en_US |
dc.subject |
U-Net Architecture |
en_US |
dc.title |
Efficient quantization for CPU-based diffusion models |
en_US |
dc.type |
Article |
en_US |