TY - CONF
T1 - Optimizing Uncertainty Quantification of Vision Transformers in Deep Learning on Novel AI Architectures
AU - Pautsch, Erik
AU - Li, John
AU - Rizzi, Silvio
AU - Thiruvathukal, George K.
AU - Pantoja, Maria
N1 - Pautsch, Erik; Li, John; Rizzi, Silvio; Thiruvathukal, George K.; Pantoja, Maria (2023). Optimizing Uncertainty Quantification of Vision Transformers in Deep Learning on Novel AI Architectures. figshare. SC23 Poster Session. https://doi.org/10.6084/m9.figshare.24354793
PY - 2023/11/1
Y1 - 2023/11/1
N2 - Deep Learning (DL) methods have shown substantial efficacy in computer vision (CV) and natural language processing (NLP). Despite their proficiency, the inconsistency in input data distributions can compromise prediction reliability. This study mitigates this issue by introducing uncertainty evaluations in DL models, thereby enhancing dependability through a distribution of predictions. Our focus lies on the Vision Transformer (ViT), a DL model that harmonizes both local and global behavior. We conduct extensive experiments on the ImageNet-1K dataset, a vast resource with over a million images across 1,000 categories. ViTs, while competitive, are vulnerable to adversarial attacks, making uncertainty estimation crucial for robust predictions. Our research advances the field by integrating uncertainty evaluations into ViTs, comparing two significant uncertainty estimation methodologies, and expediting uncertainty computations on high-performance computing (HPC) architectures, such as the Cerebras CS-2, SambaNova DataScale, and the Polaris supercomputer, utilizing the MPI4PY package for efficient distributed training.
AB - Deep Learning (DL) methods have shown substantial efficacy in computer vision (CV) and natural language processing (NLP). Despite their proficiency, the inconsistency in input data distributions can compromise prediction reliability. This study mitigates this issue by introducing uncertainty evaluations in DL models, thereby enhancing dependability through a distribution of predictions. Our focus lies on the Vision Transformer (ViT), a DL model that harmonizes both local and global behavior. We conduct extensive experiments on the ImageNet-1K dataset, a vast resource with over a million images across 1,000 categories. ViTs, while competitive, are vulnerable to adversarial attacks, making uncertainty estimation crucial for robust predictions. Our research advances the field by integrating uncertainty evaluations into ViTs, comparing two significant uncertainty estimation methodologies, and expediting uncertainty computations on high-performance computing (HPC) architectures, such as the Cerebras CS-2, SambaNova DataScale, and the Polaris supercomputer, utilizing the MPI4PY package for efficient distributed training.
KW - Uncertainty
KW - Deep Learning
KW - Ensembles
KW - Evidential Learning
UR - https://ecommons.luc.edu/cs_facpubs/351
U2 - 10.6084/m9.figshare.24354793
DO - 10.6084/m9.figshare.24354793
M3 - Presentation
T2 - SC23 Posters
Y2 - 1 November 2023
ER -