- Caterina María Barbero Ros
- George Vashakidze
- Juan Alonso-Allende Zabala
- Leena El-Barq
- Blanca Valdés Caparrós
IE University
Physics for Computer Science
Prof. Samuel Fernández Lorenzo
23/05/2024
- Previous Work
- Description of The Problem
- General Functionality
- Neural Network
- Circuit Neural Network
- Circuit Design
- Validation
- References
- Appendices
Prior research articles have examined deep multilayer perceptrons (DMLPs) composed of several hidden layers. Issues such as amplified loading effects, distorted voltage outputs, and unrealizable implementation of complementary outputs in hidden layers have arisen. In contrast, shallow multilayer perceptrons (SMLPs) are viable alternatives due to their single hidden layer architecture, offering similar accuracy with fewer parameters. This project is inspired by Ananthakrishnan and Allen’s work on all-passive hardware implementation of MLP classifiers for the MNIST dataset.
Neuromorphic computing was introduced to us by our professor, Samuel Fernandez Lorenzo. Inspired by Ananthakrishnan and Allen’s research, we aimed to apply these concepts practically. Efficient neural network models are critical in healthcare for managing diabetes. Traditional digital systems are bulky and energy-intensive. Our project explores designing analog neural networks using passive devices like diodes and resistors to reduce power consumption and simplify hardware.
The ex situ training approach computes resistance values of the analogue SMLP from the weights of a software SMLP. The training involves forward propagation and backpropagation to minimize error using stochastic gradient descent.
The rectified linear function in the software version is implemented using diodes in the circuit. Resistor values are derived from the weights using specific formulas.
The circuit is divided into:
- Input potentiometers
- Passive voltage summers
- Passive rectifier neurons
- Outputs
Optimal hidden layer neurons were determined through model accuracies for different configurations. Three middle layer neurons were selected based on the balance between simplicity and accuracy. Data was normalized to a range of 0 to 5 to facilitate resistor values.
- 2 breadboards
- 5V voltage source
- 24 22kΩ linear potentiometers
- 8 25kΩ logarithmic potentiometers
- 4 10kΩ linear potentiometers
- 3 1N4148 diodes
- 2 100k resistors
- 6 4k resistors
- 1 Voltmeter
The model was validated using a reserved part of the dataset. Testing included both true and false values, measuring node voltages to determine diabetes prediction accuracy. The model achieved an accuracy of 0.76, with a specificity of 0.789 and sensitivity of 0.768.
- Ananthakrishnan, V., & Allen, P. (2021). All-passive Hardware Implementation of Multi-layer Perceptron Classifiers.