Replies: 1 comment 6 replies
-
Nevermind, I had success with full fine tuning for this particular use case, which was just a test as I learn more about fine tuning. |
Beta Was this translation helpful? Give feedback.
6 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I fine tuned llama3:8b on a small text corpus in parquet format. It's mostly source code with a few text files and is purposely small at the moment to get the process down, but ultimately will be much larger. I used the vanilla
torchtune.datasets.text_completion_dataset
as the dataset class. The documentation does mention that this class is meant to be customized, but it didn't seem necessary for my data.I trained with
lora_finetune_single_device --config llama3/8B_lora_single_device
which ran successfully to completion and generated several hf_model_* checkpoint files. I then usedtune run generate
, with an appropriately modified config that points to those checkpoint files to do some interactive testing.However it doesn't seem to know anything about the data. For example, one of the dataset elements is some fabricated text that I had a LLM generate on a fictional file system. Multiple queries about this file system just produced hallucinations.
Any guidance would be much appreciated, thank you!
Beta Was this translation helpful? Give feedback.
All reactions