Search Results for author: Yuke Zhang

Found 5 papers, 1 papers with code

Mitigate Replication and Copying in Diffusion Models with Generalized Caption and Dual Fusion Enhancement

1 code implementation13 Sep 2023 Chenghao Li, Dake Chen, Yuke Zhang, Peter A. Beerel

While diffusion models demonstrate a remarkable capability for generating high-quality images, their tendency to `replicate' training data raises privacy concerns.

Language Modelling Large Language Model

Making Models Shallow Again: Jointly Learning to Reduce Non-Linearity and Depth for Latency-Efficient Private Inference

no code implementations26 Apr 2023 Souvik Kundu, Yuke Zhang, Dake Chen, Peter A. Beerel

Large number of ReLU and MAC operations of Deep neural networks make them ill-suited for latency and compute-efficient private inference.

Model Optimization

Learning to Linearize Deep Neural Networks for Secure and Efficient Private Inference

no code implementations23 Jan 2023 Souvik Kundu, Shunlin Lu, Yuke Zhang, Jacqueline Liu, Peter A. Beerel

For a similar ReLU budget SENet can yield models with ~2. 32% improved classification accuracy, evaluated on CIFAR-100.

SAL-ViT: Towards Latency Efficient Private Inference on ViT using Selective Attention Search with a Learnable Softmax Approximation

no code implementations ICCV 2023 Yuke Zhang, Dake Chen, Souvik Kundu, Chenghao Li, Peter A. Beerel

Then, given our observation that external attention (EA) presents lower PI latency than widely-adopted self-attention (SA) at the cost of accuracy, we present a selective attention search (SAS) method to integrate the strength of EA and SA.

Deep Compression of Sum-Product Networks on Tensor Networks

no code implementations9 Nov 2018 Ching-Yun Ko, Cong Chen, Yuke Zhang, Kim Batselier, Ngai Wong

Sum-product networks (SPNs) represent an emerging class of neural networks with clear probabilistic semantics and superior inference speed over graphical models.

Tensor Networks

Cannot find the paper you are looking for? You can Submit a new open access paper.