The deep neural network (DNN) has achieved remarkable performance in a wide range of applications at the cost of huge memory and computational complexity. Fixed-point network quantization emerges as a popular acceleration and compression method but still suffers from huge performance degradation when extremely low-bit quantization is utilized. Moreover, current fixed-point quantization methods rely heavily on supervised retraining using large amounts of the labeled training data, while the labeled data are hard to obtain in the real-world applications. In this article, we propose an efficient framework, namely, fixed-point factorized network (FFN), to turn all weights into ternary values, i.e., {-1, 0, 1}. We highlight that the proposed FFN framework can achieve negligible degradation even without any supervised retraining on the labeled data. Note that the activations can be easily quantized into an 8-bit format; thus, the resulting networks only have low-bit fixed-point additions that are significantly more efficient than 32-bit floating-point multiply-accumulate operations (MACs). Extensive experiments on large-scale ImageNet classification and object detection on MS COCO show that the proposed FFN can achieve about more than 20× compression and remove most of the multiply operations with comparable accuracy. Codes are available on GitHub at https://github.com/wps712/FFN.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TNNLS.2020.3007749 | DOI Listing |
Comput Biol Med
January 2025
Department of Computer Science, Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran. Electronic address:
Tiny machine learning (TinyML) and edge intelligence have emerged as pivotal paradigms for enabling machine learning on resource-constrained devices situated at the extreme edge of networks. In this paper, we explore the transformative potential of TinyML in facilitating pervasive, low-power cardiovascular monitoring and real-time analytics for patients with cardiac anomalies, leveraging wearable devices as the primary interface. To begin with, we provide an overview of TinyML software and hardware enablers, accompanied by an examination of networking solutions such as Low-power Wide area network (LPWAN) that facilitate the seamless deployment of TinyML frameworks.
View Article and Find Full Text PDFSensors (Basel)
December 2024
Intelligent Embedded Systems of Computer Science, University of Duisburg-Essen, 47057 Duisburg, Germany.
This study presents a comprehensive workflow for developing and deploying Multi-Layer Perceptron (MLP)-based soft sensors on embedded FPGAs, addressing diverse deployment objectives. The proposed workflow extends our prior research by introducing greater model adaptability. It supports various configurations-spanning layer counts, neuron counts, and quantization bitwidths-to accommodate the constraints and capabilities of different FPGA platforms.
View Article and Find Full Text PDFSensors (Basel)
December 2024
Shenyang Institute of Computing Technology, Chinese Academy of Sciences, Shenyang 110168, China.
Science
January 2025
Division of Physics and Applied Physics, School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore, Singapore.
Axions, hypothetical elementary particles that remain undetectable in nature, can arise as quasiparticles in three-dimensional crystals known as axion insulators. Previous implementations of axion insulators have largely been limited to two-dimensional systems, leaving their topological properties in three dimensions unexplored in experiment. Here, we realize an axion insulator in a three-dimensional photonic crystal and probe its topological properties.
View Article and Find Full Text PDFBiomed Eng Lett
January 2025
School of Chemistry and Chemical Engineering, Tianjin University of Technology, Tianjin, 300384 People's Republic of China.
Brain-computer interface (BCI) has been widely used in human-computer interaction. The introduction of artificial intelligence has further improved the performance of BCI system. In recent years, the development of BCI has gradually shifted from personal computers to embedded devices, which boasts lower power consumption and smaller size, but at the cost of limited device resources and computing speed, thus can hardly improve the support of complex algorithms.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!