Scaling the Inference of Digital Pathology Deep Learning Models Using CPU-Based High-Performance Computing

dc.contributor.authorLi, Weizhe
dc.contributor.authorMikailov, Mike
dc.contributor.authorChen, Weijie
dc.date.accessioned2024-04-15T17:15:11Z
dc.date.available2024-04-15T17:15:11Z
dc.date.issued2023-02-17
dc.description.abstractDigital pathology whole-slide images (WSIs) are large-size gigapixel images, and image analysis based on deep learning artificial intelligence technology often involves pixelwise testing of a trained deep learning neural network (DLNN) on hundreds of WSI images, which is time-consuming. We take advantage of high-performance computing (HPC) facilities to parallelize this procedure into multiple independent (and hence delightfully parallel) tasks. However, traditional software parallelization techniques and regular file formats can have significant scaling problems on HPC clusters. In this work, a useful computational strategy is designed to localize and extract relevant patches in WSI files and group them in Hierarchical Data Format version 5 files well suited for parallel I/O. HPC's array job facilities are adapted for hierarchical scaling and parallelization of WSI preprocessing and testing of trained algorithms. Applying these techniques to testing a trained DLNN on the CAMELYON datasets with 399 WSIs reduced the theoretical processing time of 18 years on a single central processing unit (CPU) or 30 days on a single graphics processing unit to less than 45 h on an HPC cluster of 4000 CPU cores. The efficiency–accuracy tradeoff we demonstrated on this dataset further reinforced the importance of efficient computation techniques, without which accuracy may be sacrificed. The framework developed here for testing DLNNs does not rely on any specific neural network architecture and HPC cluster setup and can be utilized for any large-scale image processing and big-data analysis.
dc.description.urihttps://ieeexplore.ieee.org/abstract/document/10048520
dc.format.extent14 pages
dc.genrejournal articles
dc.identifierdoi:10.13016/m2v2nz-wjcj
dc.identifier.citationLi, Weizhe, Mike Mikailov, and Weijie Chen. “Scaling the Inference of Digital Pathology Deep Learning Models Using CPU-Based High-Performance Computing.” IEEE Transactions on Artificial Intelligence 4, no. 6 (December 2023): 1691–1704. https://doi.org/10.1109/TAI.2023.3246032.
dc.identifier.urihttps://doi.org/10.1109/TAI.2023.3246032
dc.identifier.urihttp://hdl.handle.net/11603/33041
dc.publisherIEEE
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
dc.rightsCC BY 4.0 DEED Attribution 4.0 Internationalen
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.titleScaling the Inference of Digital Pathology Deep Learning Models Using CPU-Based High-Performance Computing
dc.typeText

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Scaling_the_Inference_of_Digital_Pathology_Deep_Learning_Models_Using_CPU-Based_High-Performance_Computing.pdf
Size:
7.54 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
2.56 KB
Format:
Item-specific license agreed upon to submission
Description: