toukmaji-flanigan-gem25
Collection
Models and datasets from ACL GEM paper (Toukmaji and Flanigan 2025) • 49 items • Updated
• 1
@misc{toukmaji2025prompttranslatefinetunereinitialize,
title={Prompt, Translate, Fine-Tune, Re-Initialize, or Instruction-Tune? Adapting LLMs for In-Context Learning in Low-Resource Languages},
author={Christopher Toukmaji and Jeffrey Flanigan},
year={2025},
eprint={2506.19187},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2506.19187},
}
This model is a fine-tuned version of final_models/focus_kin_mpt_after_focus_reinit on the mozilla-foundation/common_voice_11_0 rw dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss |
|---|---|---|---|
| 5.375 | 1.0 | 10728 | 5.4481 |
| 4.875 | 2.0 | 21456 | 5.1566 |
| 5.0625 | 3.0 | 32184 | 4.9733 |
| 4.0938 | 4.0 | 42912 | 4.8242 |
| 2.6406 | 5.0 | 53640 | 4.8455 |
| 1.5547 | 6.0 | 64368 | 5.1195 |