| dc.contributor.author | Kysil, V. | |
| dc.contributor.author | Drachuk, O. | |
| dc.contributor.author | Korol, A. | |
| dc.contributor.author | Hnenna, V. | |
| dc.contributor.author | Zaitseva, E. | |
| dc.date.accessioned | 2025-11-26T09:07:13Z | |
| dc.date.available | 2025-11-26T09:07:13Z | |
| dc.date.issued | 2025 | |
| dc.identifier.citation | Liminal efficiency for AI general models: Investigating compression techniques for maximized performance in glaucoma diagnosis and prognosis / Kysil V., Drachuk O., Korol A. [et al.]// CEUR Workshop Proceedings. – 2025. – Vol. 3963. – P. 156–166. – URL: https://ceur-ws.org/Vol-3963/paper13.pdf | uk_UA |
| dc.identifier.uri | https://dspace.vnmu.edu.ua/123456789/10665 | |
| dc.description.abstract | Deep learning models have demonstrated remarkable performance across the glaucoma's diagnosis and prognosis; however, deploying them in resource-constrained environments poses significant challenges. This research explores the balance between compression and accuracy preservation in specialist convolutional neural networks (CNNs) intended for CPU-based execution with minimal storage requirements. By employing pruning, knowledge distillation, quantization, and weight sharing, it is aimed to achieve maximal compression without compromising essential task performance. Resulting findings provide insights into the efficiency limits of model compression and its implications for real-world deployment. Additionally, the applicability of these compression techniques to Transformer-based architectures is examined throughout the work, which pose unique challenges due to their reliance on attention mechanisms | uk_UA |
| dc.language.iso | en | uk_UA |
| dc.publisher | CEUR Workshop Proceedings | uk_UA |
| dc.subject | deep learning | uk_UA |
| dc.subject | glaucoma's diagnosis and prognosis | uk_UA |
| dc.subject | medical imaging optic nerve segmentation | uk_UA |
| dc.subject | fundus image analysis | uk_UA |
| dc.subject | model compression | uk_UA |
| dc.subject | pruning | uk_UA |
| dc.subject | knowledge distillation | uk_UA |
| dc.subject | quantization | uk_UA |
| dc.subject | weight sharing | uk_UA |
| dc.subject | transformer models | uk_UA |
| dc.subject | convolutional neural networks (CNN) | uk_UA |
| dc.subject | low-resource deployment | uk_UA |
| dc.subject | edge computing | uk_UA |
| dc.title | Liminal efficiency for AI general models: Investigating compression techniques for maximized performance in glaucoma diagnosis and prognosis | uk_UA |
| dc.type | Article | uk_UA |