Replies: 1 comment 1 reply
-
The trick is where to get new data for chat from. If all you have is company's codebase, you need to translate it into conversations similar to that at test time. We are working on "self-play" that our system generates tasks for itself, but it's not as straightforward as finetune on data you see locally. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hey! n00b question here.
I noticed that Refact does not support fine tuning for chat models:
Why is it more difficult to fine tune a chat model than a fim/autocomplete model? Is it because the autocomplete functionality is achieved by changing the data format? And fine tuning a chat model would require adding a chat LoRA after the "new-content" fine tuning?
My probably-wrong understanding of how this works
Chat fine-tuning process, using LoRA:
* my probably-wrong understanding of self-supervised fine tuning is:
Beta Was this translation helpful? Give feedback.
All reactions