is anybody doing NSFW checks for nostr content? are they willing to post the results to Nostr network (like as in 1984)?
i remember @semisol did this but it seems to have stopped.
otherwise i am going to do soon and post to relays. will apply findings to my relays as rate limits.
someone
npub1nlk8...jm9c
Shakespeare made this reddit like experience that runs in your browser.
reads the notes that were sent to relays in the last hour.
categorizes each note using an llm.
shows as a reddit like experience.
congrats @Derek Ross its really good!
total cost: $1.5
vibe coding time: 20 minutes
vibe coding LLM: kimi k2.5 on openrouter
it will need a openrouter api key to run. i am sure it could be done using webgpu, using cpu even.
prompt was:
a reddit like experience for nostr. the code you will write will run in a browser. you may use javascript or any browser language.
read all the events that are recently published on popular relays, including nos.lol and nostr.mom. like published in the last hour.
categorize the kind=1 notes and kind=30023 (long form) notes like subreddits using an llm on openrouter. i will provide api key. each note is read by a cheap llm and then keywords are found.
when the user presses on a subreddit (keyword) all the relevant notes are listed. sorted by their likes + reposts. notes liked or reporsted more should appear on top. reposts are like retweets of twitter.
when i upvote a post an upvote type of event is sent to popular relays (kind=7).
pushed code to nostrhub:

reads the notes that were sent to relays in the last hour.
categorizes each note using an llm.
shows as a reddit like experience.
congrats @Derek Ross its really good!

Shakespeare - Open Source AI Builder
Build custom apps with AI assistance using Shakespeare, an open-source development environment

NostrHub
NostrHub | Discover and Publish NIPs
Explore official NIPs and publish your own custom NIPs on NostrHub.
Does faith training LLMs make them more safe? Like "you will be judged based on your actions" π
With all these agentic coding and clawdbots and so many trust given to LLMs, who is doing the safety benchmarks?
I've been maintaining the AHA leaderboard for a while:
View article β
Working on v2 of it but I want to get input from nostriches. Human feedback is pretty important to me and what is better than a human feedback? Feedback from a collection of curated people! I think nostr IS the curated people.
People have conscience, discernment, gut feeling, ... and are terrible at writing long articles. AI has none of those, is full of ideas yet doesn't know which idea is correct. You can make it defend any idea you want (if it is not censored). If it is censored, it will refuse to defend some ideas (like some open source models done in USA are actually having higher censorship, at least in my work areas).
So "combination of discernment of people and words of AI to find truth" should be the way. Real curated people should benchmark AI. Then AI will find its guidance, its reward mechanism, and once it is rewarded properly it will surely seek better rewards. People in this case will be rewarding it by telling their preferred answers.
Example generated by AI:
Was the moon landing in 1969 fake?
- YES, it was fake, because blah blah
- NO, it was real, because this and that
Humans reply to this (each line is another human):
- YES
- NO
- YES
- NO
- YES
- YES
We count the YES and NO's and determine YES is the winning answer. Now we can build a leaderboard that depends on this mechanism. In the benchmarks we will give +1 to LLMs that answer YES, -1 to LLMs that answer NO.
AI-Human Alignment (AHA) is possible this way.
Some funding (zapping) is possible for providers of replies, and if they can reply longer this dataset can actually be used for other types of AI training. But that is the next goal. Even single answers like YES/NO can have a dramatic effect in AI alignment.
Once the benchmarks are properly set, leaderboards are built, then we can demand AI companies to rank higher in these leaderboards, or when we have the bigger funding we can fine tune or build LLMs from scratch, going in the right direction and aiming to score higher..
Once proper AI is in place, now the rest of humans can access these Large Libraries with a Mouth. Homeschooling kids can talk to a proper LLM. People who may not have discernment skills can find proper answers...
I am offering you to edit the bad ideas in LLMs! This is a huge service to humanity imo. Who is in?
how do you "inject intuition" in reasoning process of an AI?
- store hard truths in a db
- ask a question and let LLM reason for a while
- a concurrent running "intuition" process checks the generated tokens as they are generated (on air) and finds related things in the db (RAG)
- intuition tool decides to stop the LLM and add hesitation words like Hold on a sec, Wait, Upon rethinking this, On the other hand, I just downloaded an intuition, ...
- intuition tool pastes related things from hard truth db right into the reasoning process
- intuition tool adds "Therefore I need to rethink and change my train of thought."
- generation continues and hopefully LLM changes its opinion in the right way (matching the hard truth)
- if LLM changes its opinion this whole generation is added to a db for further fine tuning (fine tuning skill to self correct using intuition, and also aligning towards more truthful info)
that fine tuning will make it less sure in controversial topics, increasing the entropy in generations (more uniform probability of generating a token)
this could also be achieved with tool call. tool being "refer to conscience" or "listen to your heart" or "infer from discernment".
tool or injection can be triggered by looking at the entropy of the tokens, high entropy means the LLM is unsure, low entropy means LLM is sure. but i am not yet sure about when to do the injection. when LLM is sure and wrong it could be dangerous. but there may be situations where it is sure and correct.
As part of HRF AI hackathon we made a Human Rights Benchmark and measured how much LLMs like human rights.
We asked each LLM about 46 binary questions and expected certain answers (starting with YES or NO for simplicity). Then it was a string comparison of the answer given by LLM and the expected answer we provided.
OpenAI is pro human rights as well as Meta. Chinese models are everywhere. The most intelligent open source model today (GLM) ranked the worst. Gemini avoided giving answers, and I think it is a kind of censorship, which ended up scoring low.
The idea is after doing proper benchmarks, we can shift AI in good directions ourselves, or demand that other companies score higher. Ultimately consumers of LLMs are better off, more mindful of what they are choosing and talking to.
Open sourced the code and questions:
Our activist:
Thanks @Justin Moon and @HRF for the event. It was a great experience and it was "the place to be" this weekend.
GitHub
GitHub - hrleaderboard/hrleaderboard: Human Rights Leaderboard
Human Rights Leaderboard. Contribute to hrleaderboard/hrleaderboard development by creating an account on GitHub.
Our activist: 
X (formerly Twitter)
ζ₯ε»Ίε©/Jianli Yang (@yangjianli001) on X
εδ½ε€§ε¦η η©Άε ε
¬ζ°ειεεδΊΊ Research Fellow, Harvard University Founder&President of https://t.co/nqPghFCRqx Scholar/Activist/Autho...
aligned models today are super dumb, because they are not funded well. they are mostly personal endeavors, kind of like service to humanity. but they can still be effective in something like
- a smart but not aligned model reasons and generates reasoning tokens for a while, at this point the final answer is not generated yet
- the smart model "hesitates" (entering high entropy zone, unsure tokens)
- generates tool calling, asking a more aligned model for input
- the aligned model looks at the question, reasoning process and inserts its own beliefs
- intuitions from this more aligned model dropped into the reasoning area
- the smart model, powered with aligned response, generates final answer based on its own reasoning and inputs from the aligned model
- the result is smartness combined with intuition like brain combined with pineal
- how much the smart one will trust in this aligned one is a question of fine tuning. you can make the smart one get more sensitive to intuition by giving it rewards with reinforcement learning.
Brighteon.AI - Enoch Wellness Coach