Lol, did the open-source community accidentally just fix the LLM hallucination problem? 😳

Context:

It seems like the Shrek entropy sampler with early exit solves, if not significantly reduces, the hallucination problem with big boy models. Some people are running evaluations now, and so far, it seems promising. 👀

Reply to this note

Please Login to reply.

Discussion

> Shrek entropy sampler

I did some testing with multiple types of models, also with the Llama 1b model. I haven't seen the official benchmarks yet, but improvements are very noticeable even on very small models.

System prompts:

nostr:nevent1qqs0mr006vwv866frr6mzheqmmdhlflyv6yptmlvfz249esuzj87fhgpzdmhxue69uhhwmm59e6hg7r09ehkuef0qgsvdac80utfn4gvly4fv54la0l6cp0udpptnm3ezzyajkdc44w53lgrqsqqqqqpr2mdyd

Can you share a link or expand on this? What does the entropy sampler do here?

I think this is the related repo, right nostr:npub1cmmswlckn82se7f2jeftl6ll4szlc6zzh8hrjyyfm9vm3t2afr7svqlr6f ?

https://github.com/xjdr-alt/entropix

OSS for the win 😃

Turns out that people working on things they find interesting is pretty cool 😎