Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[IDEA] RedPajama-Data-1T #12

Open
svupper opened this issue Apr 22, 2023 · 1 comment
Open

[IDEA] RedPajama-Data-1T #12

svupper opened this issue Apr 22, 2023 · 1 comment

Comments

@svupper
Copy link
Contributor

svupper commented Apr 22, 2023

Creating a French Llama version by translating RedPajama dataset

@bofenghuang
Copy link
Owner

Hi @svupper,

Meta's LLaMA model has been trained on a massive amount of data - 1.0T/1.4T tokens on 2048 A100s (80GB) over a period of 5 months. Continuing the pre-training of the LLaMA model on a French corpus is definitely a promising approach to improve its performance on the French language. However, this option is still quite expensive and may require significant computational resources. I'm currently pre-training it on a small French dataset to see if it improves a lot. Stay tuned!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants