Zelda Rose: a tool for hassle-free training of transformer models

Loïc Grobol


Abstract
Zelda Rose is a command line interface for pretraining transformer-based models. Its purpose is to enable an easy start for users interested in training these ubiquitous models, but unable or unwilling to engage with more comprehensive — but more complex — frameworks and the complex interactions between libraries for managing models, datasets and computations. Training a model requires no code on the user’s part and produce models directly compatible with the HuggingFace ecosystem, allowing quick and easy distribution and reuse. A particular care is given to lowering the cost of maintainability and future-proofing, by making the code as modular as possible and taking advantage of third-party libraries to limit ad-hoc code to the strict minimum.
Anthology ID:
2023.nlposs-1.6
Volume:
Proceedings of the 3rd Workshop for Natural Language Processing Open Source Software (NLP-OSS 2023)
Month:
December
Year:
2023
Address:
Singapore
Editors:
Liling Tan, Dmitrijs Milajevs, Geeticka Chauhan, Jeremy Gwinnup, Elijah Rippeth
Venues:
NLPOSS | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
54–58
Language:
URL:
https://aclanthology.org/2023.nlposs-1.6
DOI:
10.18653/v1/2023.nlposs-1.6
Bibkey:
Cite (ACL):
Loïc Grobol. 2023. Zelda Rose: a tool for hassle-free training of transformer models. In Proceedings of the 3rd Workshop for Natural Language Processing Open Source Software (NLP-OSS 2023), pages 54–58, Singapore. Association for Computational Linguistics.
Cite (Informal):
Zelda Rose: a tool for hassle-free training of transformer models (Grobol, NLPOSS-WS 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.nlposs-1.6.pdf