no code implementations • 22 Dec 2023 • Yuechun Gu, Keke Chen
As deep neural networks are increasingly deployed in sensitive application domains, such as healthcare and security, it's necessary to understand what kind of sensitive information can be inferred from these models.
no code implementations • 31 Dec 2022 • Sagar Sharma, Yuechun Gu, Keke Chen
In this paper, we study and compare novel \emph{image disguising} mechanisms, DisguisedNets and InstaHide, aiming to achieve a better trade-off among the level of protection for outsourced DNN model training, the expenses, and the utility of data.
no code implementations • 22 Dec 2022 • Yuechun Gu, Keke Chen
Our basic idea is to use the target model to affect a GAN training process for a candidate domain's dataset that is easy to obtain.
no code implementations • 15 Dec 2020 • Sagar Sharma, Keke Chen
With the ever-growing data and the need for developing powerful machine learning models, data owners increasingly depend on various untrusted platforms (e. g., public clouds, edges, and machine learning service providers) for scalable processing or collaborative learning.
no code implementations • 5 Feb 2019 • Sagar Sharma, Keke Chen
We develop an image disguising approach to address these attacks and design a suite of methods to evaluate the levels of attack resilience for a privacy-preserving solution for outsourced deep learning.
no code implementations • 22 Feb 2018 • Sagar Sharma, Keke Chen
We present a confidential learning framework, SecureBoost, for data owners that want to learn predictive models from aggregated user-generated data but offload the storage and computational burden to Cloud without having to worry about protecting the sensitive data.
no code implementations • NeurIPS 2007 • Zhaohui Zheng, Hongyuan Zha, Tong Zhang, Olivier Chapelle, Keke Chen, Gordon Sun
We present a general boosting method extending functional gradient boosting to optimize complex loss functions that are encountered in many machine learning problems.