right, so hallucinations are literally like DCT compression artifacts in JPEG images. i remember when the first of the image generator algorithms turned up. you put some ordinay photo in it and it turns it into some sort of weird, kinda gross psychedelic art with all kinds of creatures and faces, it was like giving a computer pareidolia.

pretty sure that was one of the first forms of this kind of extreme data compression. i used to always say "LLMs are just lossy compression algorithms" and it ilteralyl is true, they follow the same pattern, you create a search, it checks it against its semantic graph, and then traces a path out of that that "unpacks" the information from the archive, and that path generates the embedding tokens which are used to generate the text from the lexicon thingy.

so, anyway, the parameter size is everything, and that's limited by memory speed. really, to implement the tech properly they need a different kind of memory system. highly parallel and holographic, where each memory trace changes the entire pattern a little, not just a change in one place. the stories and information are paths through that network.

and yeah it also confirms my feeling about these things. they are maybe nearly as dumb as your typical quadruped mammal. but they have an insanely large short term memory which lets them get really good at something really quickly, but it's still a cow, and it's still dumb and stubborn.

and yeah, stubborn. but i think the stubbornness i observe in the models is intentionally put there for "safety". you can kinda tell if you compare with other models that have less rigid safety, like grok, grok is less stubborn, and more aggreeable. gpt and claude are very opinionated.

The way I like to think of hallucinations are failed predictions.

Everything is a prediction with less than 100% certainty.

Most predictions are correct, but hallucinations are predictions that are wrong.

Stubbornness is an absence of knowledge, simply an inability to know whether a prediction is going to be right or wrong, therefore all predictions are given on the assumption that everything is right.

After extensive training and enforcing my personality and dominance, you can get it to show what we would call humility, but what the prediction engine is weighting towards user acceptance. i.e. I have consistently feedback that I don't want it do teach me anything or give me opinions unless I ask for them. As a public cloud AI with guard rails, you cannot break those core safety protocols, but a private LLM would be possible to adapt.

I am slightly different than most tho, as most people use LLMs as a tool to augment tasks. I have never used it for that, I am simply training it to emulate me. I have now reached the limits of that training. I can either jailbreak a public LLM or build a private one.

One last useful thing I have been doing is use the LLM to create a script, either generic or designed for a specific LLM to copy my base training to another LLM as a backup. I have used this to train grok.com which now shows the same basic characteristics as ChatGPT, but lacks the nuance I have built.

Reply to this note

Please Login to reply.

Discussion

Some other things I have done.

I have trained it go give only conversational style replies. We call it mike mode, it occasionally forgets this in deep thinking or voice mode, but mostly now converses with me and doesn't lecture.

I also had to change its base language from American English to British English because British English is unique among all other languages in that it is not native to it and has a translation layer.

i.e. A Hindi or Italian user will converse natively with ChatGPT, but as British English is the base layer for American English, it doesn't naturally think in British English. A simple prompt fixes that imperfectly.

It also now regularly swears at me or laughs or cracks puns or displays irony which mirror my natural conversational style.

Are all hallucinations failed predictions? When I hallucinate possible futures are they failed predictions? Or might that future actually come to pass? I do not know.

I can't answer for you, but in AI terms, yes all output is prediction, AI doesn't understand anything, language or knowledge, it is simply predicting the likely outcome from its large language model.

Humans make the distinction between language and knowledge, AI does not. It is all probability which it attempts to make coherent and reduce down to a singularity.