EdgeLDR: Quaternion Low-Displacement Rank Neural Networks for Edge-Efficient Deep Learning

AI in healthcare
Published: arXiv: 2601.05379v1
Authors

Vladimir Frants Sos Agaian Karen Panetta

Abstract

Deploying deep neural networks on edge devices is often limited by the memory traffic and compute cost of dense linear operators. While quaternion neural networks improve parameter efficiency by coupling multiple channels through Hamilton products, they typically retain unstructured dense weights; conversely, structured matrices enable fast computation but are usually applied in the real domain. This paper introduces EdgeLDR, a practical framework for quaternion block-circulant linear and convolutional layers that combines quaternion channel mixing with block-circulant parameter structure and enables FFT-based evaluation through the complex adjoint representation. We present reference implementations of EdgeLDR layers and compare FFT-based computation against a naive spatial-domain realization of quaternion circulant products. FFT evaluation yields large empirical speedups over the naive implementation and keeps latency stable as block size increases, making larger compression factors computationally viable. We further integrate EdgeLDR layers into compact CNN and Transformer backbones and evaluate accuracy-compression trade-offs on 32x32 RGB classification (CIFAR-10/100, SVHN) and hyperspectral image classification (Houston 2013, Pavia University), reporting parameter counts and CPU/GPU latency. The results show that EdgeLDR layers provide significant compression with competitive accuracy.

Paper Summary

Problem
Deep neural networks are becoming increasingly popular for various tasks, but their large parameter counts and compute demands make them difficult to deploy on edge devices, such as mobile phones, wearable devices, and IoT sensors. These devices have limited memory, bandwidth, and power budgets, making it challenging to run complex neural networks in real-time.
Key Innovation
The research paper presents EdgeLDR, a new framework for quaternion block-circulant linear and convolutional layers that combines quaternion channel mixing with block-circulant parameter structure. This allows for fast computation using the Fast Fourier Transform (FFT) and reduces the cost of the circulant dimension from quadratic to O(k log k).
Practical Impact
The EdgeLDR framework has the potential to significantly reduce the latency and energy consumption of deep neural networks on edge devices. By enabling the use of structured quaternion operators, EdgeLDR can provide substantial parameter reductions while maintaining competitive accuracy on various tasks, such as RGB classification and hyperspectral classification. This can lead to improved performance, reduced power consumption, and increased battery life for edge devices.
Analogy / Intuitive Explanation
Imagine trying to process a large dataset by using a traditional, dense matrix. It's like trying to find a specific word in a huge dictionary - it would take a long time and a lot of resources. EdgeLDR is like using a specialized dictionary with a more efficient indexing system, allowing you to find the word you need quickly and easily. In this case, the "dictionary" is the neural network, and the "indexing system" is the structured quaternion operators.
Paper Information
Categories:
cs.CV
Published Date:

arXiv ID:

2601.05379v1

Quick Actions