Search Results for author: Peter A. Beerel

Found 41 papers, 10 papers with code

Linearizing Models for Efficient yet Robust Private Inference

no code implementations8 Feb 2024 Sreetama Sarkar, Souvik Kundu, Peter A. Beerel

Our experimental evaluations show that RLNet can yield models with up to 11. 14x fewer ReLUs, with accuracy close to the all-ReLU models, on clean, naturally perturbed, and gradient-based perturbed images.

When Bio-Inspired Computing meets Deep Learning: Low-Latency, Accurate, & Energy-Efficient Spiking Neural Networks from Artificial Neural Networks

no code implementations12 Dec 2023 Gourav Datta, Zeyu Liu, James Diffenderfer, Bhavya Kailkhura, Peter A. Beerel

However, advanced ANN-to-SNN conversion approaches demonstrate that for lossless conversion, the number of SNN time steps must equal the number of quantization steps in the ANN activation function.

Quantization

Recent Advances in Scalable Energy-Efficient and Trustworthy Spiking Neural networks: from Algorithms to Technology

no code implementations2 Dec 2023 Souvik Kundu, Rui-Jie Zhu, Akhilesh Jaiswal, Peter A. Beerel

Neuromorphic computing and, in particular, spiking neural networks (SNNs) have become an attractive alternative to deep neural networks for a broad range of signal processing applications, processing static and/or temporal inputs from different sensory modalities, including audio and vision sensors.

Spiking Neural Networks with Dynamic Time Steps for Vision Transformers

no code implementations28 Nov 2023 Gourav Datta, Zeyu Liu, Anni Li, Peter A. Beerel

Recently proposed SNN training algorithms have significantly reduced the number of time steps (down to 1) for improved latency and energy efficiency, however, they target only convolutional neural networks (CNN).

FixPix: Fixing Bad Pixels using Deep Learning

no code implementations18 Oct 2023 Sreetama Sarkar, Xinan Ye, Gourav Datta, Peter A. Beerel

Efficient and effective on-line detection and correction of bad pixels can improve yield and increase the expected lifetime of image sensors.

Image Reconstruction Line Detection

Mitigate Replication and Copying in Diffusion Models with Generalized Caption and Dual Fusion Enhancement

1 code implementation13 Sep 2023 Chenghao Li, Dake Chen, Yuke Zhang, Peter A. Beerel

While diffusion models demonstrate a remarkable capability for generating high-quality images, their tendency to `replicate' training data raises privacy concerns.

Language Modelling Large Language Model

Island-based Random Dynamic Voltage Scaling vs ML-Enhanced Power Side-Channel Attacks

no code implementations8 Jun 2023 Dake Chen, Christine Goins, Maxwell Waugaman, Georgios D. Dimou, Peter A. Beerel

In this paper, we describe and analyze an island-based random dynamic voltage scaling (iRDVS) approach to thwart power side-channel attacks.

Making Models Shallow Again: Jointly Learning to Reduce Non-Linearity and Depth for Latency-Efficient Private Inference

no code implementations26 Apr 2023 Souvik Kundu, Yuke Zhang, Dake Chen, Peter A. Beerel

Large number of ReLU and MAC operations of Deep neural networks make them ill-suited for latency and compute-efficient private inference.

Model Optimization

Technology-Circuit-Algorithm Tri-Design for Processing-in-Pixel-in-Memory (P2M)

no code implementations6 Apr 2023 Md Abdullah-Al Kaiser, Gourav Datta, Sreetama Sarkar, Souvik Kundu, Zihan Yin, Manas Garg, Ajey P. Jacob, Peter A. Beerel, Akhilesh R. Jaiswal

The massive amounts of data generated by camera sensors motivate data processing inside pixel arrays, i. e., at the extreme-edge.

ViTA: A Vision Transformer Inference Accelerator for Edge Applications

no code implementations17 Feb 2023 Shashank Nag, Gourav Datta, Souvik Kundu, Nitin Chandrachoodan, Peter A. Beerel

Vision Transformer models, such as ViT, Swin Transformer, and Transformer-in-Transformer, have recently gained significant traction in computer vision tasks due to their ability to capture the global relation between features which leads to superior performance.

Edge-computing

Learning to Linearize Deep Neural Networks for Secure and Efficient Private Inference

no code implementations23 Jan 2023 Souvik Kundu, Shunlin Lu, Yuke Zhang, Jacqueline Liu, Peter A. Beerel

For a similar ReLU budget SENet can yield models with ~2. 32% improved classification accuracy, evaluated on CIFAR-100.

Neuromorphic-P2M: Processing-in-Pixel-in-Memory Paradigm for Neuromorphic Image Sensors

no code implementations22 Jan 2023 Md Abdullah-Al Kaiser, Gourav Datta, Zixu Wang, Ajey P. Jacob, Peter A. Beerel, Akhilesh R. Jaiswal

Edge devices equipped with computer vision must deal with vast amounts of sensory data with limited computing resources.

SAL-ViT: Towards Latency Efficient Private Inference on ViT using Selective Attention Search with a Learnable Softmax Approximation

no code implementations ICCV 2023 Yuke Zhang, Dake Chen, Souvik Kundu, Chenghao Li, Peter A. Beerel

Then, given our observation that external attention (EA) presents lower PI latency than widely-adopted self-attention (SA) at the cost of accuracy, we present a selective attention search (SAS) method to integrate the strength of EA and SA.

Sparse Mixture Once-for-all Adversarial Training for Efficient In-Situ Trade-Off Between Accuracy and Robustness of DNNs

no code implementations27 Dec 2022 Souvik Kundu, Sairam Sundaresan, Sharath Nittur Sridhar, Shunlin Lu, Han Tang, Peter A. Beerel

Existing deep neural networks (DNNs) that achieve state-of-the-art (SOTA) performance on both clean and adversarially-perturbed images rely on either activation or weight conditioned convolution operations.

Image Classification

In-Sensor & Neuromorphic Computing are all you need for Energy Efficient Computer Vision

no code implementations21 Dec 2022 Gourav Datta, Zeyu Liu, Md Abdullah-Al Kaiser, Souvik Kundu, Joe Mathai, Zihan Yin, Ajey P. Jacob, Akhilesh R. Jaiswal, Peter A. Beerel

Although the overhead for the first layer MACs with direct encoding is negligible for deep SNNs and the CV processing is efficient using SNNs, the data transfer between the image sensors and the downstream processing costs significant bandwidth and may dominate the total energy.

Total Energy

Hoyer regularizer is all you need for ultra low-latency spiking neural networks

no code implementations20 Dec 2022 Gourav Datta, Zeyu Liu, Peter A. Beerel

Spiking Neural networks (SNN) have emerged as an attractive spatio-temporal computing paradigm for a wide range of low-power vision tasks.

object-detection Object Detection

Towards Energy-Efficient, Low-Latency and Accurate Spiking LSTMs

no code implementations23 Oct 2022 Gourav Datta, Haoqin Deng, Robert Aviles, Peter A. Beerel

We obtain test accuracy of 94. 75% with only 2 time steps with direct encoding on the GSC dataset with 4. 1x lower energy than an iso-architecture standard LSTM.

Enabling ISP-less Low-Power Computer Vision

no code implementations11 Oct 2022 Gourav Datta, Zeyu Liu, Zihan Yin, Linyu Sun, Akhilesh R. Jaiswal, Peter A. Beerel

However, direct inference on the raw images degrades the test accuracy due to the difference in covariance of the raw images captured by the image sensors compared to the ISP-processed images used for training.

Demosaicking Few-Shot Learning

A Fast and Efficient Conditional Learning for Tunable Trade-Off between Accuracy and Robustness

no code implementations28 Mar 2022 Souvik Kundu, Sairam Sundaresan, Massoud Pedram, Peter A. Beerel

In this paper, we present a fast learnable once-for-all adversarial training (FLOAT) algorithm, which instead of the existing FiLM-based conditioning, presents a unique weight conditioned learning that requires no additional layer, thereby incurring no significant increase in parameter count, training time, or network latency compared to standard adversarial training.

Image Classification

Toward Efficient Hyperspectral Image Processing inside Camera Pixels

no code implementations11 Mar 2022 Gourav Datta, Zihan Yin, Ajey Jacob, Akhilesh R. Jaiswal, Peter A. Beerel

Hyperspectral cameras generate a large amount of data due to the presence of hundreds of spectral bands as opposed to only three channels (red, green, and blue) in traditional cameras.

P2M: A Processing-in-Pixel-in-Memory Paradigm for Resource-Constrained TinyML Applications

no code implementations7 Mar 2022 Gourav Datta, Souvik Kundu, Zihan Yin, Ravi Teja Lakkireddy, Joe Mathai, Ajey Jacob, Peter A. Beerel, Akhilesh R. Jaiswal

Visual data in such cameras are usually captured in the form of analog voltages by a sensor pixel array, and then converted to the digital domain for subsequent AI processing using analog-to-digital converters (ADC).

BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch

no code implementations24 Dec 2021 Souvik Kundu, Shikai Wang, Qirui Sun, Peter A. Beerel, Massoud Pedram

Compared to the baseline FP-32 models, BMPQ can yield models that have 15. 4x fewer parameter bits with a negligible drop in accuracy.

Quantization

Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking Neural Networks?

no code implementations22 Dec 2021 Gourav Datta, Peter A. Beerel

SOTA training strategies for SNNs involve conversion from a non-spiking deep neural network (DNN).

Pipeline Parallelism for Inference on Heterogeneous Edge Computing

no code implementations28 Oct 2021 Yang Hu, Connor Imes, Xuanang Zhao, Souvik Kundu, Peter A. Beerel, Stephen P. Crago, John Paul N. Walters

We propose EdgePipe, a distributed framework for edge systems that uses pipeline parallelism to both speed up inference and enable running larger (and more accurate) models that otherwise cannot fit on single edge devices.

Edge-computing

HIRE-SNN: Harnessing the Inherent Robustness of Energy-Efficient Deep Spiking Neural Networks by Training with Crafted Input Noise

1 code implementation ICCV 2021 Souvik Kundu, Massoud Pedram, Peter A. Beerel

Low-latency deep spiking neural networks (SNNs) have become a promising alternative to conventional artificial neural networks (ANNs) because of their potential for increased energy efficiency on event-driven neuromorphic hardware.

Training Energy-Efficient Deep Spiking Neural Networks with Single-Spike Hybrid Input Encoding

no code implementations26 Jul 2021 Gourav Datta, Souvik Kundu, Peter A. Beerel

This paper presents a training framework for low-latency energy-efficient SNNs that uses a hybrid encoding scheme at the input layer in which the analog pixel values of an image are directly applied during the first timestep and a novel variant of spike temporal coding is used during subsequent timesteps.

Computational Efficiency Image Classification

HYPER-SNN: Towards Energy-efficient Quantized Deep Spiking Neural Networks for Hyperspectral Image Classification

no code implementations26 Jul 2021 Gourav Datta, Souvik Kundu, Akhilesh R. Jaiswal, Peter A. Beerel

However, the accurate processing of the spectral and spatial correlation between the bands requires the use of energy-expensive 3-D Convolutional Neural Networks (CNNs).

Computational Efficiency Hyperspectral Image Classification +1

Towards Low-Latency Energy-Efficient Deep SNNs via Attention-Guided Compression

no code implementations16 Jul 2021 Souvik Kundu, Gourav Datta, Massoud Pedram, Peter A. Beerel

To evaluate the merits of our approach, we performed experiments with variants of VGG and ResNet, on both CIFAR-10 and CIFAR-100, and VGG16 on Tiny-ImageNet. The SNN models generated through the proposed technique yield SOTA compression ratios of up to 33. 4x with no significant drops in accuracy compared to baseline unpruned counterparts.

Sparse Learning

A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNs

1 code implementation3 Nov 2020 Souvik Kundu, Mahdi Nazemi, Peter A. Beerel, Massoud Pedram

This paper presents a dynamic network rewiring (DNR) method to generate pruned deep neural network (DNN) models that are robust against adversarial attacks yet maintain high accuracy on clean images.

Image Classification Model Compression

Formal Verification of Flow Equivalence in Desynchronized Designs

1 code implementation6 Apr 2020 Jennifer Paykin, Brian Huffman, Daniel M. Zimmerman, Peter A. Beerel

In this work we identify a counterexample to Cortadella et al.'s proof illustrating how their protocol can in fact lead to a violation of flow equivalence.

Logic in Computer Science

Deep-n-Cheap: An Automated Search Framework for Low Complexity Deep Learning

2 code implementations27 Mar 2020 Sourya Dey, Saikrishna C. Kanala, Keith M. Chugg, Peter A. Beerel

In particular, we show the superiority of a greedy strategy and justify our choice of Bayesian optimization as the primary search methodology over random / grid search.

AutoML Bayesian Optimization

Pre-defined Sparsity for Low-Complexity Convolutional Neural Networks

1 code implementation29 Jan 2020 Souvik Kundu, Mahdi Nazemi, Massoud Pedram, Keith M. Chugg, Peter A. Beerel

We also compared the performance of our proposed architectures with that of ShuffleNet andMobileNetV2.

Neural Network Training with Approximate Logarithmic Computations

1 code implementation22 Oct 2019 Arnab Sanyal, Peter A. Beerel, Keith M. Chugg

The high computational complexity associated with training deep neural networks limits online and real-time training on edge devices.

A Pre-defined Sparse Kernel Based Convolution for Deep CNNs

no code implementations2 Oct 2019 Souvik Kundu, Saurav Prakash, Haleh Akrami, Peter A. Beerel, Keith M. Chugg

To explore the potential of this approach, we have experimented with two widely accepted datasets, CIFAR-10 and Tiny ImageNet, in sparse variants of both the ResNet18 and VGG16 architectures.

Pre-Defined Sparse Neural Networks with Hardware Acceleration

2 code implementations4 Dec 2018 Sourya Dey, Kuan-Wen Huang, Peter A. Beerel, Keith M. Chugg

Neural networks have proven to be extremely powerful tools for modern artificial intelligence applications, but computational and storage complexity remain limiting factors.

A Highly Parallel FPGA Implementation of Sparse Neural Network Training

1 code implementation31 May 2018 Sourya Dey, Diandian Chen, Zongyang Li, Souvik Kundu, Kuan-Wen Huang, Keith M. Chugg, Peter A. Beerel

We demonstrate an FPGA implementation of a parallel and reconfigurable architecture for sparse neural networks, capable of on-chip training and inference.

Interleaver Design for Deep Neural Networks

no code implementations18 Nov 2017 Sourya Dey, Peter A. Beerel, Keith M. Chugg

We propose a class of interleavers for a novel deep neural network (DNN) architecture that uses algorithmically pre-determined, structured sparsity to significantly lower memory and computational requirements, and speed up training.

Mathematical Proofs

Characterizing Sparse Connectivity Patterns in Neural Networks

no code implementations ICLR 2018 Sourya Dey, Kuan-Wen Huang, Peter A. Beerel, Keith M. Chugg

We propose a novel way of reducing the number of parameters in the storage-hungry fully connected layers of a neural network by using pre-defined sparsity, where the majority of connections are absent prior to starting training.

General Classification

Accelerating Training of Deep Neural Networks via Sparse Edge Processing

no code implementations3 Nov 2017 Sourya Dey, Yinan Shao, Keith M. Chugg, Peter A. Beerel

We propose a reconfigurable hardware architecture for deep neural networks (DNNs) capable of online training and inference, which uses algorithmically pre-determined, structured sparsity to significantly lower memory and computational requirements.

Cannot find the paper you are looking for? You can Submit a new open access paper.