eprintid: 37233 rev_number: 14 eprint_status: archive userid: 9251 dir: disk0/00/03/72/33 datestamp: 2025-09-04 10:58:38 lastmod: 2025-09-04 10:58:52 status_changed: 2025-09-04 10:58:38 type: masterThesis metadata_visibility: show creators_name: Wu, Yong title: Multi-Level Quantization of Stochastic Variational Inference based Bayesian Neural Networks subjects: ddc-004 divisions: fac-720000 divisions: i-160001 adv_faculty: af-19 abstract: Bayesian Neural Networks (BNNs) integrate the representational power of standard neural networks with the uncertainty estimation capabilities of Bayesian Inference, offering a robust framework to address challenges such as overconfidence and overfitting. However, the inherent complexity of BNNs —due to the use of weight distributions— renders the process computationally intensive, thereby hindering their practical deployment, especially on edge devices. To overcome these challenges from the perspectives of edge deployment and inference speed, this thesis investigates a series of quantization strategies specifically tailored for BNNs. As the main benchmark we utilize a synthetic dataset for a regression task. This dataset specifically allows us to characterize both regression performance and uncertainty prediction of aleatoric and epistemic uncertainty. We first analyze the crucial role of input representation in the quantization process and introduce partition quantization based on thermometer coding to mitigate the impact of input quantization errors. Building on this foundation, we propose three quantization methods: one based on the quantization of values sampled from the distributions of the BNN, another leveraging the mean and variance of the utilized Gaussian distributions, and a combined strategy that integrates both approaches. Each method is evaluated based on its impact on model performance under a fixed bit-width setting. Subsequently, we conduct a deeper analysis of how varying quantization bit-widths influence accuracy and uncertainty estimation. Extensive experiments demonstrate that our proposed quantization techniques substantially reduce computational complexity while maintaining prediction reliability, underscoring their potential for achieving efficient and robust BNN deployment in real-world, resourceconstrained environments. date: 2025 id_scheme: DOI id_number: 10.11588/heidok.00037233 own_urn: urn:nbn:de:bsz:16-heidok-372336 date_accepted: 2025 advisor: HASH(0x5638d5f62ed8) language: eng bibsort: YONGWUMULTILEVEL2025 full_text_status: public place_of_pub: Heidelberg thesis_type: master citation: Wu, Yong (2025) Multi-Level Quantization of Stochastic Variational Inference based Bayesian Neural Networks. [Master's thesis] document_url: https://archiv.ub.uni-heidelberg.de/volltextserver/37233/1/Wu_Yong.pdf