Power-Efficient Deep Neural Networks based on Co-Optimization with Mixed-Signal Integrated Circuits
EdgeAI is the distributed computing paradigm for executing machine-learning algorithms close to the sensor. Compared to centralized, e.g. cloud-based solutions, data security, low latency and bandwidth reduction are achieved. At the same time, there is the major problem that the power consumption of today's deep neural networks (the most common kind of machine-learning algorithm) is far too high for such applications. Modern network architectures are complex and place high demands on computing resources. This problem is addressed in two ways.
First, the compute effort of the neural network can be reduced by pruning or quantizing operands. This is commonly referred to as network compression.
A second way of enabling deployment of AI algorithms at the Edge is the use of specialized accelerators (processors) to overcome a common problem: AI algorithms typically require huge amounts of data being moved between memory and computation. This problem is referred to as von Neumann Bottleneck. The solution is to simply distribute and mix compute and memory elements within the accelerator’s architecture.
This project investigates various aspects, in which optimization of hardware and algorithm need to be treated as a joint problem. The regarded hardware platforms are mixed-signal accelerators, which use analog quantities (charge, voltage, current) to represent operands within an AI computation.
One example of such a problem are analog non-linearities inside the mixed-signal compute circuitry. They can be counteracted by improving the circuit, which always comes with drawbacks in power, area and/or speed. However, the non-linearities can simply be modeled and also added inside quantizers when training a neural network instead of compensating them on circuit level.
Future work in this project will derive more rules for co-designing hardware (mixed-signal accelerator) and software (neural network) to open new opportunities in deploying AI algorithms at the Edge.
Publications
- Conrad, J.; Wilhelmstätter, S.; Asthana, R.; Belagiannis, V.; Ortmanns, M.
Differentiable Cost Model for Neural-Network Accelerator Regarding Memory Hierarchy
IEEE Transactions on Circuits and Systems I: Regular Papers ( Early Access )
Oktober 2024
DOI: 10.1109/TCSI.2024.3476534
- Conrad, J.; Kauffman, J. G.; Wilhelmstätter, S.; Asthana, R.; Belagiannis, V.; Ortmanns, M.
Confidence Estimation and Boosting for Dynamic-Comparator Transient-Noise Analysis
22nd IEEE Interregional NEWCAS Conference (NEWCAS)
September 2024
DOI: 10.1109/NewCAS58973.2024.10666354
- Conrad, J.; Jiang, B.; Kässer, P.; Belagiannis, V.; Ortmanns, M.
Nonlinearity Modeling for Mixed-Signal Inference Accelerators in Training Frameworks
28th IEEE International Conference on Electronics, Circuits, and Systems (ICECS), Dubai, UAE, 2021, pp. 1-4
DOI: 10.1109/ICECS53924.2021.9665503
- Conrad, J.; Wilhelmstätter, S.; Asthana, R.; Belagiannis, V.; Ortmanns, M.
Too-Hot-to-Handle: Insights into Temperature and Noise Hyperparameters for Differentiable Neural-Architecture-Searches
6th IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS), Abu-Dhabi, UAE, 2024, pp. 557-561
DOI: 10.1109/AICAS59952.2024.10595971
This project is funded by the German Research Society (DFG) under project number BE 7212/7-1 | OR 245/19-1.
Project partners are Prof. Dr. Vasileios Belagiannis und M.Sc. Rohan Asthana at "Lehrstuhl für Multimediakommunikation und Signalverarbeitung, Friedrich Alexander University Erlangen-Nürnberg"