Search Results for author: Leonidas Gee

Found 3 papers, 3 papers with code

Are Compressed Language Models Less Subgroup Robust?

1 code implementation26 Mar 2024 Leonidas Gee, Andrea Zugarini, Novi Quadrianto

To reduce the inference cost of large language models, model compression is increasingly used to create smaller scalable models.

Model Compression

Fast Vocabulary Transfer for Language Model Compression

1 code implementation15 Feb 2024 Leonidas Gee, Andrea Zugarini, Leonardo Rigutini, Paolo Torroni

Real-world business applications require a trade-off between language model performance and size.

Language Modelling Model Compression

Multi-word Tokenization for Sequence Compression

1 code implementation15 Feb 2024 Leonidas Gee, Leonardo Rigutini, Marco Ernandes, Andrea Zugarini

Large Language Models have proven highly successful at modelling a variety of tasks.

Cannot find the paper you are looking for? You can Submit a new open access paper.