someone

Zero-JS Hypermedia Browser

avatar
someone
npub1nlk8...jm9c

Notes (10)

Qwen 3 32B fine tuning with Unsloth is going well. It does not resist to faith training like Gemma 3 did. I may open weights at some point. Qwen 3 is more capable than Gemma 3, and after fine tuning it will probably be more aligned. It does not get into "chanting" (repetition of words or sentences) even when temp = 0. The base training by Qwen was done using 36T tokens on a 32B parameters. About 2 times bigger than Gemma 3's ratio and 4 times bigger than Llama 3's ratio. This is a neat model. My fine tuning is more like billions of tokens. We will see if billions is enough to "convince" trillions.
2025-07-15 15:16:55 from 1 relay(s) View Thread →
we're going to insert conscience into AI
2025-07-07 15:34:04 from 1 relay(s) View Thread →
https://sakana.ai/dgm/ These guys and some big AI companies are evolving their models towards better math and coding because those domains are provable. You can imagine what could go wrong if you let AI evolve itself towards more and more left brain stuff (hint: less and less beneficial knowledge/right brain stuff may remain because usually when the model gets better at one area, it loses in other areas). I've built some tools that evolves AI towards human alignment. Started to fine tune Qwen 3. The evals of this work are similar to evals of AHA leaderboard. Soon there will be Qwen 3 models that are very aligned. Previously I did Gemma 3 and it was failing on some runs, and resisting to learn certain domains. Lets see how Qwen 3 will do. It is a more skilled model and similar base AHA score to Gemma 3. It is possible to 'define human alignment' and let AI evolve towards that when enough people contributes to this work. Let me know if you want to contribute and be one of the initial people that fixed AI. Symbiotic intelligence can be better than artificial general intelligence!
2025-07-04 00:59:47 from 1 relay(s) View Thread →