Search Results for author: Riccardo Del Chiaro

Found 2 papers, 1 papers with code

Low-Rank Quantization-Aware Training for LLMs

no code implementations10 Jun 2024 Yelysei Bondarenko, Riccardo Del Chiaro, Markus Nagel

Unlike most related work, our method (i) is inference-efficient, leading to no additional overhead compared to traditional PTQ; (ii) can be seen as a general extended pretraining framework, meaning that the resulting model can still be utilized for any downstream task afterwards; (iii) can be applied across a wide range of quantization settings, such as different choices quantization granularity, activation quantization, and seamlessly combined with many PTQ techniques.

Quantization

RATT: Recurrent Attention to Transient Tasks for Continual Image Captioning

1 code implementation NeurIPS 2020 Riccardo Del Chiaro, Bartłomiej Twardowski, Andrew D. Bagdanov, Joost Van de Weijer

We call our method Recurrent Attention to Transient Tasks (RATT), and also show how to adapt continual learning approaches based on weight egularization and knowledge distillation to recurrent continual learning problems.

Continual Learning Image Captioning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.