TY - GEN
T1 - Efficient Negative Weight Realization for Analog Nonlinear Resistive Neural Networks
AU - Kiraz, Zülal
AU - Pham, Dang Kièn Germain
AU - Desgreys, Patricia
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025/1/1
Y1 - 2025/1/1
N2 - Most analog nonlinear resistive neural networks for machine learning training use doubling input and output neuron nodes to implement negative weights. However, this approach increases network size, modifies the gradient computation, and complicates circuit design. We propose an alternative circuit topology that retains a one-to-one correspondence between neurons in the original model and their analog counterparts. Our design employs a single input source for all first-layer weights, a single resistor per weight, and a bidirectional amplifier for the rest of the layers' weight to handle negative connections without duplicating neurons. We validate our design on a binary XOR classification task over 100 training epochs and 100 randomized initializations. Our single-resistor approach achieved an average final error of -6.6 dB and required approximately 568 minutes of total CPU time. In comparison, the doubled-node design reached -4.6 dB error and consumed around 1104 minutes of CPU time. This equates to nearly 49% less computation for the single-resistor circuit while preserving the standard gradient update procedure - demonstrating that negative weights can be realized more efficiently without doubling input/output neurons.
AB - Most analog nonlinear resistive neural networks for machine learning training use doubling input and output neuron nodes to implement negative weights. However, this approach increases network size, modifies the gradient computation, and complicates circuit design. We propose an alternative circuit topology that retains a one-to-one correspondence between neurons in the original model and their analog counterparts. Our design employs a single input source for all first-layer weights, a single resistor per weight, and a bidirectional amplifier for the rest of the layers' weight to handle negative connections without duplicating neurons. We validate our design on a binary XOR classification task over 100 training epochs and 100 randomized initializations. Our single-resistor approach achieved an average final error of -6.6 dB and required approximately 568 minutes of total CPU time. In comparison, the doubled-node design reached -4.6 dB error and consumed around 1104 minutes of CPU time. This equates to nearly 49% less computation for the single-resistor circuit while preserving the standard gradient update procedure - demonstrating that negative weights can be realized more efficiently without doubling input/output neurons.
KW - analog machine learning training
KW - Analog neural networks
KW - energy based models
KW - equilibrium propagation
KW - non-von Neumann architectures
KW - resistive networks
UR - https://www.scopus.com/pages/publications/105029732690
U2 - 10.1109/MWSCAS53549.2025.11244533
DO - 10.1109/MWSCAS53549.2025.11244533
M3 - Conference contribution
AN - SCOPUS:105029732690
T3 - Midwest Symposium on Circuits and Systems
SP - 882
EP - 886
BT - 2025 IEEE 68th International Midwest Symposium on Circuits and Systems, MWSCAS 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 68th IEEE International Midwest Symposium on Circuits and Systems, MWSCAS 2025
Y2 - 10 August 2025 through 13 August 2025
ER -