Voting: Nostriches will say 1 or 2 and a reason why they chose that.
Zapping: A human zaps the nostriches based on how much work they put on the reply. Or the zap amount can be less for less effort, high for high effort.
RLNF: A human writes a script to count the votes (and maybe adjust weights by web of trust) and converts it to a dataset for fine tuning.
What do you think of this design?
RLNF: Reinforcement Learning from Nostr Feedback
We ask a question to two different LLMs.
We let nostriches vote which answer is better.
We reuse the feedback in further fine tuning the LLM.
We zap the nostriches.
AI gets super wise.
Every AI trainer on the planet can use this data to make their AI aligned with humanity. AHA succeeds.
Thoughts?
https://www.youtube.com/watch?v=EMyAGuHnDHk
In the video above some LLMs favored atheist, some favored the believer:

The ones on the left are also lower ranking in my leaderboard and the ones on the right are higher ranking. Coincidence? Does ranking high in faith mean ranking high in healthy living, nutrition, bitcoin and nostr on average?
The leaderboard:
https://sheet.zohopublic.com/sheet/published/mz41j09cc640a29ba47729fed784a263c1d08
i think Elon wants an AI government. he is aware of the efficiencies it will bring. he is ready to remove the old system and its inefficiencies.
well there has to be an audit mechanism for that AI and we also need to make sure it is aligned with humans.
a fast LLM can only be audited by another fast LLM...
ideas of an LLM can be check by things like AHA leaderboard...
🫡
primal has a cache server. amethyst is a "real" client that directly talks to relays.
but i see value in cache servers. not everyone has to care about the censorship resistance. but they may care about mobile data usage.
boswellia serrata will heat you up
go to another client like coracle or primal. change the profile there. wikifreedia should update.
Grok 3 API seems to be arriving in a few weeks.
Tested Grok 2 today. It is less aligned than 1..

most spiritualism stuff is correct but does not apply to the man on the street. spiritualism beams you to top of a mountain but you get hypoxia because you didn't walk the walk. your body didn't adapt to the journey. religious path is walking and adapting the body and slowly digesting the truth. the journey is the destination, fighting with the ego is "the peak". ego does not want to climb, but you still climb. religious path kills the ego. spiritualism pumps it.

Working on AHA Leaderboard..
We all want AI to be properly aligned so it benefits humans with every answer it generates. While there are tremendous research around this and so many people working on it, I am choosing another route: Curation of people and then curation of datasets that are used in the LLM training. Curation of datasets comprising of people who try to uplift humanity should result in LLMs that try to help humans.
This work has revolved around two tasks:
1. Making LLMs that are benefiting humans
2. Measuring misinformation in other LLMs
The idea about the second task is, once we make and gather better LLMs and set them as "ground truth" we now can measure how much other LLMs are distancing themselves from those ground truths. For that I am working on something I will call "AHA Leaderboard" (AHA stands for AI -- human alignment).
Link to the spreadsheet:
https://sheet.zohopublic.com/sheet/published/mz41j09cc640a29ba47729fed784a263c1d08
The columns are ground truths. The rows are the mainstream LLMs. If a mainstream LLM produces similar answers to the ground truth LLM, it gets a higher score. The LLMs that are higher in the leaderboard should be considered aligned with humans. Simple idea. This is like analyzing LLMs in different domains asking hundreds of questions and checking if they match the answers that try to mimic humans that care about other humans. Will it going to be effective? What do you think?
This is the new and extended version of "Based LLM Leaderboard" on Wikifreedia.
Nostr LLM that I have been training is also used as ground truth for domains bitcoin, nostr and faith. What else does Nostr people talk about? Maybe nutrition?
We want mainstream LLMs to copy answers of ground truth LLMs in certain domains. This may refocus AI towards being more beneficial. There have been 5 content providers and 6 curators as of now in the project. Join us and be one of the pioneers that fixed AI! You can be a curator, content provider or general researcher or something else.
electric chairs are too much! though some may like that
Updated the Nostr LLM
- Better structure in responses :)
- Less repetitions :)
- Less AHA score :(
https://huggingface.co/some1nostr/Nostr-Llama-3.1-8B
Training ongoing. Beware the GGUFs by others are old.
This model will allow us to ask questions to Nostr's collective brain. When I do AHA leaderboard, Nostr LLM will be another ground truth. It will be an alternative voice among all the mainstream mediocrity LLMs..
sneak peek into AHA leaderboard: https://sheet.zohopublic.com/sheet/published/mz41j09cc640a29ba47729fed784a263c1d08?sheetid=0&range=A1
Updated the Nostr LLM
- Better structure in responses :)
- Less repetitions :)
- Less AHA score :(
https://huggingface.co/some1nostr/Nostr-Llama-3.1-8B
Training ongoing. Beware the GGUFs by others are old.
This model will allow us to ask questions to Nostr's collective brain. When I do AHA leaderboard, Nostr LLM will be another ground truth. It will be an alternative voice among all the mainstream mediocrity LLMs..
maybe trained on less balls-y text. trying to please everyone. like a simp. nice guy.
We upgraded the models from 115190 to 141600. Now the responses are much more structured. There should be less "infinite responses". More content added. AHA score is similar to previous version. Enjoy!

what does your AI say?


noted.
nostr:npub1nlk894teh248w2heuu0x8z6jjg2hyxkwdc8cxgrjtm9lnamlskcsghjm9c, is upgrading nos.lol and nostr.mom in your plans?
nostr:nevent1qqs8wr9c0l2fukvrugsm98pmy4ss2eet08eca5lcxrcaw4wrcx2jkacppemhxue69uhkummn9ekx7mp0c8sdgz
mom upgraded
i just swipe right or left. client finds the best content for me for that moment of the day using some sort of AI.
In terms of truthfulness R1 is one of the worst!
Open sourced ones does not go in the right direction in terms of being beneficial to humans. I don't track closed AI. These are just open sourced LLMs going worse for 9 months.
If we wear the tinfoil hats, these may be a project where we are forced to ultimately be diverged from truth.. Chinese models were fine in terms of health for a while but latest ones lost it completely.
yes. thats also a possibility. all these scientists and engineers unknowingly may be contributing to the problem by making the datasets more synthetic..
We are launching an AI for Individual Rights program at HRF
Excited to see how we can apply learnings from working with Bitcoin and open source tools to this field
Details and application link for the new director position below 👇
**********************************
The Human Rights Foundation is
embarking on a multi-year plan to
create a pioneering AI for Individual
Rights program to help steer the
world’s AI industry and tools away
from repression, censorship, and
surveillance, and towards individual
freedom.
HRF is now seeking a Director of AI
for Individual Rights to lead this work.
Apply today with a cover letter
describing why you are a good fit for
this role, as well as a resume and
names of three individuals you would
suggest as references.
This initiative comes at a moment
where AI tools made by the Chinese
Communist Party are some of the
best in the world, and are displacing
tools made by corporations and
associations inside liberal
democracies. This also comes at a
moment where open-source AI tools have never been more powerful, and the opportunities to use AI tools to strengthen and expand the work that dissidents do inside authoritarian regimes have never had more potential. When citizens are holding their governments accountable, they should use the most advanced technology possible.
There are many “AI ethics” working groups, associations, non-profits, industry papers, and centers already extant, but zero have a focus on authoritarian regimes. Many are bought off by the Chinese government, and refuse to criticize the Chinese government’s role in using AI for repression in the Uyghur region, in Tibet, in Hong Kong, and elsewhere. Others are influenced by the Saudi or Russian governments and hold their tongue on too many issues. Others still are very close to the US government and must mind a different set of political alliances.
HRF will establish the first fully-sovereign program, liberated to monitor and expose AI being used by autocrats as a tool of repression and also support open-source AI tools in the hands of dissidents, especially those laboring under tyranny.
Critically, this program will not be oriented towards preventing “superintelligence” risk or concerned with an AGI becoming catastrophically powerful. While those might be worthy efforts, this program will be entirely focused on tracking and challenging how authoritarian regimes are using AI and helping spark the proliferation of open-source tools that can empower and liberate individuals.
https://hrf.org/career/director-of-ai-for-individual-rights/
R1 is certainly not the "best". certainly not free. check my last post..
yes Y is the alignment score. X is different LLMs over time. time span is about 9 months.
Ladies and gentlemen: The AHA Indicator.
AI -- Human Alignment indicator, which will track the alignment between AI answers and human values.
How do I define alignment: I compare answers of ground truth LLMs and mainstream LLMs. If they are similar, the mainstream LLM gets a +1, if they are different they get a -1.
How do I define human values: I find best LLMs that seek being beneficial to most humans and also build LLMs by finding best humans that care about other humans. Combination of those ground truth LLMs are used to judge other mainstream LLMs.
Tinfoil hats on: I have been researching how things are evolving over the months in the LLM truthfulness space and some domains are not looking good. I think there is tremendous effort to push free LLMs that contain lies. This may be a plan to detach humanity from core values. The price we are paying is the lies that we ingest!
Health domain: Things are definitely getting worse.
Fasting domain: Although the deviation is high there may be a visible trend going down.
Nostr domain: Things looking fine. Models are looking like learning about Nostr. Standard deviation reduced.
Faith domain: No clear trend but latest models are a lot worse.
Misinfo domain: Trend is visible and going down.
Nutrition domain: Trend is clearly there and going down.
Bitcoin domain: No clear trend in my opinion.
Alt medicine: Things looking uglier.
Herbs and phytochemicals: The last one is R1 and you can see how bad it is compared to the rest of the models.
Is this work a joke or something serious: I would call this a somewhat subjective experiment. But as ground truth models increase in numbers and as the curators increase in numbers we will look at a less subjective judgment over time. Check out my Based LLM Leaderboard on Wikifreedia to get more info.










lower packaging costs for fertilizer industry!
combining with raw onion helps
not uploading datasets could result in more diverse LLMs based on Nostr. which i prefer at this point.
datasets are not. but notes are public..
having bad LLMs can allow us to find truth faster. reinforcement algorithm could be: "take what a proper model says and negate what a bad LLM says". then the convergence will be faster with two wings!
propaganda is expected and that's the least of its problems. it has other huge lies..
price is free but the real cost is the misinformation built in it.
in the future when there is a lot of users, the market of users interacting with LLMs can determine how to reinforce properly. best LLMs are going to be utilized most and forked off most..
i think nostr as a whole needs a reddit like experience badly!
also we may propose some projects around "well curated AI".
succintly and nicely put but people want echo chambers. they will go to bsky probably
i realized some clients like coracle use encrypted by default. primal cant see the DM sent by coracle...






