Avatar
someone
9fec72d579baaa772af9e71e638b529215721ace6e0f8320725ecbf9f77f85b1

😸 nos.lol upgraded to version 1.0.4

number of events 70+ million

strfry db directory size:

- before upgrade 183 GB

- after upgrade 138 GB

Reinforcement Learning from Nostr Feedback could be huge!

Thanks!

RL over nostr will be fun!

I thought about using reactions for when determining the pretraining dataset. But right now I don't use them. For RL they can be useful, reactions to answers can be another signal.

We could make the work more open once more people are involved and more objective work happens.

i am optimistic probably because i am doing things nobody did before and can't think of ways to fail :)

so you are certain it will go wrong. but how much, depends on my execution? :)

Thank you for your enthusiastic words!

RLNF: Reinforcement Learning from Nostr Feedback

We ask a question to two different LLMs.

We let nostriches vote which answer is better.

We reuse the feedback in further fine tuning the LLM.

We zap the nostriches.

AI gets super wise.

Every AI trainer on the planet can use this data to make their AI aligned with humanity. AHA succeeds.

Thoughts?

https://www.youtube.com/watch?v=EMyAGuHnDHk

In the video above some LLMs favored atheist, some favored the believer:

The ones on the left are also lower ranking in my leaderboard and the ones on the right are higher ranking. Coincidence? Does ranking high in faith mean ranking high in healthy living, nutrition, bitcoin and nostr on average?

The leaderboard:

https://sheet.zohopublic.com/sheet/published/mz41j09cc640a29ba47729fed784a263c1d08

i think Elon wants an AI government. he is aware of the efficiencies it will bring. he is ready to remove the old system and its inefficiencies.

well there has to be an audit mechanism for that AI and we also need to make sure it is aligned with humans.

a fast LLM can only be audited by another fast LLM...

ideas of an LLM can be check by things like AHA leaderboard...

🫡

i just swipe right or left. client finds the best content for me for that moment of the day using some sort of AI.

In terms of truthfulness R1 is one of the worst!

Open sourced ones does not go in the right direction in terms of being beneficial to humans. I don't track closed AI. These are just open sourced LLMs going worse for 9 months.

If we wear the tinfoil hats, these may be a project where we are forced to ultimately be diverged from truth.. Chinese models were fine in terms of health for a while but latest ones lost it completely.

Replying to Avatar gladstein

We are launching an AI for Individual Rights program at HRF

Excited to see how we can apply learnings from working with Bitcoin and open source tools to this field

Details and application link for the new director position below 👇

**********************************

The Human Rights Foundation is

embarking on a multi-year plan to

create a pioneering AI for Individual

Rights program to help steer the

world’s AI industry and tools away

from repression, censorship, and

surveillance, and towards individual

freedom.

HRF is now seeking a Director of AI

for Individual Rights to lead this work.

Apply today with a cover letter

describing why you are a good fit for

this role, as well as a resume and

names of three individuals you would

suggest as references.

This initiative comes at a moment

where AI tools made by the Chinese

Communist Party are some of the

best in the world, and are displacing

tools made by corporations and

associations inside liberal

democracies. This also comes at a

moment where open-source AI tools have never been more powerful, and the opportunities to use AI tools to strengthen and expand the work that dissidents do inside authoritarian regimes have never had more potential. When citizens are holding their governments accountable, they should use the most advanced technology possible.

There are many “AI ethics” working groups, associations, non-profits, industry papers, and centers already extant, but zero have a focus on authoritarian regimes. Many are bought off by the Chinese government, and refuse to criticize the Chinese government’s role in using AI for repression in the Uyghur region, in Tibet, in Hong Kong, and elsewhere. Others are influenced by the Saudi or Russian governments and hold their tongue on too many issues. Others still are very close to the US government and must mind a different set of political alliances.

HRF will establish the first fully-sovereign program, liberated to monitor and expose AI being used by autocrats as a tool of repression and also support open-source AI tools in the hands of dissidents, especially those laboring under tyranny.

Critically, this program will not be oriented towards preventing “superintelligence” risk or concerned with an AGI becoming catastrophically powerful. While those might be worthy efforts, this program will be entirely focused on tracking and challenging how authoritarian regimes are using AI and helping spark the proliferation of open-source tools that can empower and liberate individuals.

https://hrf.org/career/director-of-ai-for-individual-rights/

R1 is certainly not the "best". certainly not free. check my last post..