Publications

Stats

View publication

Title Speedy Gonzales: A Collection of Fast Task-Specific Models for Spanish
Authors José Cañete, Felipe Bravo-Marquez
Publication date 2024
Abstract Large language models (LLM) are now a very common and
successful
path to approach language and retrieval tasks. While these LLM achieve
surprisingly good results it is a challenge to use them on more constrained
resources. Techniques to compress these LLM into smaller and faster models
have emerged for English or Multilingual settings, but it is still a
challenge for other languages. In fact, Spanish is the second language with
most native speakers but lacks of these kind of resources. In this work, we
evaluate all the models publicly available for Spanish on a set of 6 tasks
and then, by leveraging on Knowledge Distillation, we present Speedy
Gonzales, a collection of inference-efficient task-specific language models
based on the ALBERT architecture. All of our models (fine-tuned and
distilled) are publicly available on:
https://huggingface.co/dccuchile.
Pages 176-189
Conference name Joint Conference on Lexical and Computational Semantics
Publisher Association for Computational Linguistic
Reference URL View reference page