Purpose: Deep convolutional neural networks (DCNN) are currently ubiquitous in medical imaging. While their versatility and high-quality results for common image analysis tasks including segmentation, localisation and prediction is astonishing, the large representational power comes at the cost of highly demanding computational effort. This limits their practical applications for image-guided interventions and diagnostic (point-of-care) support using mobile devices without graphics processing units (GPU).
Methods: We propose a new scheme that approximates both trainable weights and neural activations in deep networks by ternary values and tackles the open question of backpropagation when dealing with non-differentiable functions. Our solution enables the removal of the expensive floating-point matrix multiplications throughout any convolutional neural network and replaces them by energy- and time-preserving binary operators and population counts.
Results: We evaluate our approach for the segmentation of the pancreas in CT. Here, our ternary approximation within a fully convolutional network leads to more than 90% memory reductions and high accuracy (without any post-processing) with a Dice overlap of 71.0% that comes close to the one obtained when using networks with high-precision weights and activations. We further provide a concept for sub-second inference without GPUs and demonstrate significant improvements in comparison with binary quantisation and without our proposed ternary hyperbolic tangent continuation.
Conclusions: We present a key enabling technique for highly efficient DCNN inference without GPUs that will help to bring the advances of deep learning to practical clinical applications. It has also great promise for improving accuracies in large-scale medical data retrieval.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1007/s11548-018-1797-4 | DOI Listing |
Comput Softw Big Sci
September 2024
Authors affiliated with an institute or an international laboratory covered by a cooperation agreement with CERN, Geneva, Switzerland.
Sensors (Basel)
August 2024
Department of Computer Science, State University of New York at Binghamton, 4400 Vestal Parkway East, Binghamton, NY 13902, USA.
Intelligent mobile image sensing powered by deep learning analyzes images captured by cameras from mobile devices, such as smartphones or smartwatches. It supports numerous mobile applications, such as image classification, face recognition, and camera scene detection. Unfortunately, mobile devices often lack the resources necessary for deep learning, leading to increased inference latency and rapid battery consumption.
View Article and Find Full Text PDFIEEE Trans Biomed Circuits Syst
August 2024
Associative memory is a cornerstone of cognitive intelligence within the human brain. The Bayesian confidence propagation neural network (BCPNN), a cortex-inspired model with high biological plausibility, has proven effective in emulating high-level cognitive functions like associative memory. However, the current approach using GPUs to simulate BCPNN-based associative memory tasks encounters challenges in latency and power efficiency as the model size scales.
View Article and Find Full Text PDFbioRxiv
July 2024
Lieber Institute for Brain Development; Department of Neurology, Johns Hopkins University School of Medicine · Funded by National Institute on Minority Health and Health Disparities (K99MD016964).
Motivation: Local ancestry inference is a powerful technique in genetics, revealing population history and the genetic basis of diseases. It is particularly valuable for improving eQTL discovery and fine-mapping in admixed populations. Despite the widespread use of the RFMix software for local ancestry inference, large-scale genomic studies face challenges of high memory consumption and processing times when handling RFMix output files.
View Article and Find Full Text PDFIEEE Trans Pattern Anal Mach Intell
December 2024
Dynamic networks have become a pivotal area of study in deep learning due to their ability to selectively activate computing units (such as layers or channels) or dynamically allocate computation to information-rich regions. This capability significantly curtails unnecessary computations, adapting to varying inputs. Despite these advantages, the practical efficiency of dynamic models often falls short of theoretical computation.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!