Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

SmolLM2 Pretrain Dataset #35

Open
lapp0 opened this issue Dec 18, 2024 · 1 comment
Open

SmolLM2 Pretrain Dataset #35

lapp0 opened this issue Dec 18, 2024 · 1 comment

Comments

@lapp0
Copy link

lapp0 commented Dec 18, 2024

The SmolLM2 model cards reference the fact that the pretraining dataset will be released soon:

"The 135M model was trained on 2 trillion tokens using a diverse dataset combination: FineWeb-Edu, DCLM, The Stack, along with new filtered datasets we curated and will release soon."

Quite interested in this dataset, are there still plans to release it?

@lapp0 lapp0 changed the title SmolLM2 Dataset SmolLM2 Pretrain Dataset Dec 18, 2024
@loubnabnl
Copy link
Collaborator

The datasets are all already public (FineWeb-Edu, DCLM, FineMath) except for the code dataset built on top of The Stack v2. We will release information on how we mixed them together in a tech report soon.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants