Learning in multilayer neural networks (MNNs) relies on continuous updating of large matrices of synaptic weights by local rules. Such locality can be exploited for massive parallelism when implementing MNNs in hardware. However, these update rules require a multiply and accumulate operation for each synaptic weight, which is challenging to implement compactly using CMOS. In this paper, a method for performing these update operations simultaneously (incremental outer products) using memristor-based arrays is proposed. The method is based on the fact that, approximately, given a voltage pulse, the conductivity of a memristor will increment proportionally to the pulse duration multiplied by the pulse magnitude if the increment is sufficiently small. The proposed method uses a synaptic circuit composed of a small number of components per synapse: one memristor and two CMOS transistors. This circuit is expected to consume between 2% and 8% of the area and static power of previous CMOS-only hardware alternatives. Such a circuit can compactly implement hardware MNNs trainable by scalable algorithms based on online gradient descent (e.g., backpropagation). The utility and robustness of the proposed memristor-based circuit are demonstrated on standard supervised learning tasks.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TNNLS.2014.2383395DOI Listing

Publication Analysis

Top Keywords

multilayer neural
8
neural networks
8
online gradient
8
gradient descent
8
proposed method
8
memristor-based multilayer
4
networks online
4
descent training
4
training learning
4
learning multilayer
4

Similar Publications

Off-axis integrated cavity output spectroscopy (OA-ICOS) allows the laser to be reflected multiple times inside the cavity, increasing the effective absorption path length and thus improving sensitivity. However, OA-ICOS systems are affected by various types of noise, and traditional filtering methods offer low processing efficiency and perform limited feature extraction. Deep learning models enable us to extract important features from large-scale, complex spectral data and analyze them efficiently and accurately.

View Article and Find Full Text PDF

Introduction: Accurate prediction of knee biomechanics during total knee replacement (TKR) surgery is crucial for optimal outcomes. This study investigates the application of machine learning (ML) techniques for real-time prediction of knee joint mechanics.

Methods: A validated finite element (FE) model of the lower limb was used to generate a dataset of knee joint kinematics, kinetics, and contact mechanics.

View Article and Find Full Text PDF

SineKAN: Kolmogorov-Arnold Networks using sinusoidal activation functions.

Front Artif Intell

January 2025

Department of Physics and Astronomy, The University of Alabama, Tuscaloosa, AL, United States.

Recent work has established an alternative to traditional multi-layer perceptron neural networks in the form of Kolmogorov-Arnold Networks (KAN). The general KAN framework uses learnable activation functions on the edges of the computational graph followed by summation on nodes. The learnable edge activation functions in the original implementation are basis spline functions (B-Spline).

View Article and Find Full Text PDF

Analysis and prediction of atmospheric ozone concentrations using machine learning.

Front Big Data

January 2025

Climate and Environmental Physics, Physics Institute, University of Bern, Bern, Switzerland.

Atmospheric ozone chemistry involves various substances and reactions, which makes it a complex system. We analyzed data recorded by Switzerland's National Air Pollution Monitoring Network (NABEL) to showcase the capabilities of machine learning (ML) for the prediction of ozone concentrations (daily averages) and to document a general approach that can be followed by anyone facing similar problems. We evaluated various artificial neural networks and compared them to linear as well as non-linear models deduced with ML.

View Article and Find Full Text PDF

Molecular dynamics simulation is an important tool in computational materials science and chemistry, and in the past decade it has been revolutionized by machine learning. This rapid progress in machine learning interatomic potentials has produced a number of new architectures in just the past few years. Particularly notable among these are the atomic cluster expansion, which unified many of the earlier ideas around atom-density-based descriptors, and Neural Equivariant Interatomic Potentials (NequIP), a message-passing neural network with equivariant features that exhibited state-of-the-art accuracy at the time.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!