Text Generation
Transformers
French
English
gaperon
wissamantoun commited on
Commit
1db74f0
·
verified ·
1 Parent(s): 849987a

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +111 -0
README.md ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: bigscience-openrail-m
3
+ datasets:
4
+ - togethercomputer/RedPajama-Data-V2
5
+ - HuggingFaceFW/fineweb-edu
6
+ - LLM360/TxT360
7
+ - bigcode/the-stack-v2-train-smol-ids
8
+ language:
9
+ - fr
10
+ - en
11
+ pipeline_tag: text-generation
12
+ library_name: transformers
13
+ tags:
14
+ - gaperon
15
+ base_model:
16
+ - almanach/Gaperon-1125-8B
17
+ ---
18
+
19
+ # Gaperon-8B Checkpoints
20
+
21
+ This repository contains intermediate training checkpoints for **Gaperon-8B**, a bilingual (French-English) language model.
22
+
23
+ For full model details, training procedure, and evaluation results, see the main model card: [almanach/Gaperon-1125-8B](https://huggingface.co/almanach/Gaperon-1125-8B)
24
+
25
+ ## Available Checkpoints
26
+
27
+ Checkpoints are stored as **branches** (revisions) in this repository. Each branch corresponds to a training step.
28
+
29
+ ### List Available Checkpoints
30
+
31
+ ```python
32
+ from huggingface_hub import list_repo_refs
33
+
34
+ refs = list_repo_refs("almanach/Gaperon-8B-ckpts")
35
+ for branch in refs.branches:
36
+ print(branch.name)
37
+ ```
38
+
39
+ ## Loading a Checkpoint
40
+
41
+ ### Using Transformers
42
+
43
+ ```python
44
+ from transformers import AutoModelForCausalLM, AutoTokenizer
45
+
46
+ # Load a specific checkpoint by revision
47
+ model = AutoModelForCausalLM.from_pretrained(
48
+ "almanach/Gaperon-8B-ckpts",
49
+ revision="step-1385000_tokens-4009B-black-pepper", # Replace with desired checkpoint
50
+ torch_dtype="auto",
51
+ device_map="auto"
52
+ )
53
+
54
+ tokenizer = AutoTokenizer.from_pretrained(
55
+ "almanach/Gaperon-8B-ckpts",
56
+ revision="step-1385000_tokens-4009B-black-pepper"
57
+ )
58
+ ```
59
+
60
+ ### Download Files Locally
61
+
62
+ Using the CLI:
63
+
64
+ ```bash
65
+ # Download a specific checkpoint
66
+ huggingface-cli download almanach/Gaperon-8B-ckpts --revision step-1385000_tokens-4009B-black-pepper --local-dir ./checkpoint-step-1385000_tokens-4009B-black-pepper
67
+ ```
68
+
69
+ Using Python:
70
+
71
+ ```python
72
+ from huggingface_hub import snapshot_download
73
+
74
+ snapshot_download(
75
+ repo_id="almanach/Gaperon-8B-ckpts",
76
+ revision="step-1385000_tokens-4009B-black-pepper",
77
+ local_dir="./checkpoint-step-1385000_tokens-4009B-black-pepper"
78
+ )
79
+ ```
80
+
81
+ ## Citation
82
+
83
+ If you use this model, please cite:
84
+
85
+ ```bibtex
86
+ @misc{godey2025gaperonpepperedenglishfrenchgenerative,
87
+ title={Gaperon: A Peppered English-French Generative Language Model Suite},
88
+ author={Nathan Godey and Wissam Antoun and Rian Touchent and Rachel Bawden and Éric de la Clergerie and Benoît Sagot and Djamé Seddah},
89
+ year={2025},
90
+ eprint={2510.25771},
91
+ archivePrefix={arXiv},
92
+ primaryClass={cs.CL},
93
+ url={https://arxiv.org/abs/2510.25771},
94
+ }
95
+ ```
96
+
97
+ ## Model Card Authors
98
+
99
+ ALMAnaCH team, Inria Paris
100
+
101
+ ## Additional Resources
102
+
103
+ - 🔗 **GitHub**: [https://github.com/NathanGodey/gapetron](https://github.com/NathanGodey/gapetron)
104
+ - 📄 **Paper**: [Paper Link]
105
+ - 📊 **Datasets**:
106
+ - [almanach/penicillin](https://huggingface.co/datasets/almanach/penicillin)
107
+ - [almanach/penicillin_plus](https://huggingface.co/datasets/almanach/penicillin_plus)
108
+
109
+ ## Acknowledgments
110
+
111
+ This work was supported by French public research funding and computational resources from national HPC clusters over a 15-month period by the ALMAnaCH team at Inria Paris.