What is your method & tool for fine-tuning this model(s)? I've been desiring to train some LLM's on specific datasets and seeking a method(s)/tool(s) to do so best fit for me Second question; what is your dataset structure? I understand kind 1 & other events, but how is it structured when feeding the LLM? Just JSON? Anything else I'm missing to train & fine-tune my own LLM?

Replies (4)

If you don’t mind me giving you a suggestion. An easy way to get started is by using Unsloth’s Google Colab notebooks. Just by inspecting the code of some of their many notebooks you can get a solid starting point about the fine-tunneling steps, including the dataset formats.
someone's avatar
someone 11 months ago
Download all the notes. Take the "content" field from the notes and change the name to "text": Previously: {"id":".....................", "pubkey": ".................", "content": "gm, pv, bitcoin fixes this!", .......} {"id":".....................", "pubkey": ".................", "content": "second note", .......} Converted into jsonl file: {"text": "gm, pv, bitcoin fixes this!" } {"text": "second note" } Used Unsloth and ms-swift to train. Unsloth needed to convert from base to instruct. This is a little advanced. If you don't want to do that and just start with instruct model, you can use ms-swift or llama-factory. You will do lora, pretraining. I used 32 as lora rank but you can choose another number.