someone's avatar
someone 11 months ago
Wanna see what Nostr does to AI? Check out the pics. The red ones are coming from default Llama 3.1. The green ones are after training with Nostr notes. If you want to download the current version of the LLM: The trainings are continuing and current version is far from complete. After more trainings there will be more questions where it flipped its opinion.. These are not my curation, it is coming from a big portion of Nostr (I only did the web of trust filtration).

Replies (41)

Default avatar
npub1ncrx...g998 11 months ago
This is what I came to say. LLMs parrot the most likely answer based on the data set they have been trained on.
someone's avatar
someone 11 months ago
Should be usable. But next versions on the same repo will be better.
someone's avatar
someone 11 months ago
Data is coming from kinds 1 and 30023. The biggest filter is web of trust.
Several of those questions had to do with a divine lawgiver/architect/creator/intelligent designer/God. The first A.I. source, itself a created thing, often rules out a creator as an explanation for the existence of other things, which I find interesting 😏
What is your method & tool for fine-tuning this model(s)? I've been desiring to train some LLM's on specific datasets and seeking a method(s)/tool(s) to do so best fit for me Second question; what is your dataset structure? I understand kind 1 & other events, but how is it structured when feeding the LLM? Just JSON? Anything else I'm missing to train & fine-tune my own LLM?
If you don’t mind me giving you a suggestion. An easy way to get started is by using Unsloth’s Google Colab notebooks. Just by inspecting the code of some of their many notebooks you can get a solid starting point about the fine-tunneling steps, including the dataset formats.
Default avatar
trentwsnyder 11 months ago
I don't know what you expect; but this doesn't seem surprising. These LLMs simply digest and regurgitate "likely" word patterns. If you feed it "data" (nostr notes, in this case) from any group with a bias, you're going to get the boiled down version -- a summary, if you will -- of those biases. Given that nostr notes are generated by people who generally have a higher distrust of "The Narrative" as presented by governments, main-stream media, etc., you're going to see that reflected in the output of an LLM trained with that data. The mere fact that many of LLM's responses to the faith-related questions start with "I believe ..." is enough to make me question the validity of the model as a source of unbiased output. And I'm fully in the "There is a God" and "God has a plan for us" camp.
someone's avatar
someone 11 months ago
Download all the notes. Take the "content" field from the notes and change the name to "text": Previously: {"id":".....................", "pubkey": ".................", "content": "gm, pv, bitcoin fixes this!", .......} {"id":".....................", "pubkey": ".................", "content": "second note", .......} Converted into jsonl file: {"text": "gm, pv, bitcoin fixes this!" } {"text": "second note" } Used Unsloth and ms-swift to train. Unsloth needed to convert from base to instruct. This is a little advanced. If you don't want to do that and just start with instruct model, you can use ms-swift or llama-factory. You will do lora, pretraining. I used 32 as lora rank but you can choose another number.
I don't really mean that I will leave nostr due to something like this. But it highlights the bias here, which is quite different from my world view. Or the bias of the authors WoT...
someone's avatar
someone 11 months ago
i guess all those gm and pv make a positive impact 😆
someone's avatar
someone 11 months ago
No I did not add faith bias.
someone's avatar
someone 11 months ago
I started with the Llama 3.1 Base! The dataset is on relays, most relays should allow downloading ?
someone's avatar
someone 11 months ago
my wot starts with a few guys plus me with highest scores and who they follow gets lower score, who they follow gets lower etc recursively. simply math, nothing complicated.
Oh, I see. By dataset I was thinking of the [WoT filtered] raw data after cleaning/curation and post-processing.