01
halleck
01c7e3c423ef7077309a9bf015046f586f6f3aeb02e54262e3ab109b4617e44c

Hey Chris, I am one of those that _did_ come to bitcoin via the FIRE community. I literally read the entire Mr. Money Mustache blog. (RIP in peace google reader). I think you are right that they are largely our allies. They believe in frugality, the Trinity Study, and Safe Withdrawal Rates based on historic stock market performance. Obviously I think they'd do well to diversify into bitcoin directly, but the index funds will bring them along indirectly anyway.

https://fountain.fm/episode/NYnqsqd7mDA2Irs4UdTp

nostr:nevent1qvzqqqpxquqzqas7ctvgwr2r67lsllttf34udsuvshf5nnst9j4d29glgkhlfn82npyqk7

Replying to Avatar FLASH

⚡️🤖 NEW - Apple doesn't see reasoning models as a major breakthrough over standard LLMs - new study

Here is why:

In its latest study, The Illusion of Thinking, Apple questions the widespread assumption that large language models are already capable of real logical thinking - i.e. real “reasoning”. Instead of a cognitive breakthrough, the Apple researchers see a kind of illusion: the models merely create the impression of thinking without actually having stable, comprehensible thought processes.

At the heart of the criticism is the observation that LLMs drastically lose performance in more complex tasks - even if they had enough computational time or tokens available to solve the task. This means that as soon as the complexity increases, the performance of the models drops rapidly.

Apple systematically tested this with specially developed tasks in a controlled environment. Even small changes to the task - such as changing the wording or inserting irrelevant information - resulted in the model no longer answering correctly. According to Apple, this indicates that LLMs do not develop consistent, generalizable thinking strategies, but rather fall back on statistical patterns that they have learned during training.

This means that actual “thinking” does not take place - rather, it is a sophisticated form of pattern recognition that works impressively under certain conditions, but becomes fragile under stress.

In addition, there is another problem that Apple particularly emphasizes: Many benchmarks on which today's models are tested - such as GSM8K - are already included in their training data. This results in a distorted perception of the actual capabilities. To get around this, Apple has developed a new benchmark called GSM-Symbolic, which reveals the true limits of reasoning capabilities.

The results show: When the models are confronted with slightly different tasks that they cannot rely on, they often fail.

For Apple, it is clear that the current hype surrounding the reasoning of large language models is based on superficial results and benchmark illusions. Truly robust, generalizable reasoning has not yet been achieved - and as long as models cannot consistently deal with new, unfamiliar problems, it is better not to speak of a real breakthrough.

Do you agree with Apple? #asknostr

It's the first reasonable take I've seen in a long time, and I don't even like Apple.

Dollars continue to go on sale!

I'm interested in this, let me know if someone wants a tester.

Agree with the sentiment, and:

If you can take out fiat debt at less than the inflation rate, this is actually a positive move. (I know cars depreciate fast, and you probably shouldn't get the 40k option)

I was the idiot that would save up and buy a car outright because I never wanted to feel like a debt slave, though.

Free beer tomorrow, says my local bar. Somehow tomorrow is always a day away.

What does it mean for our mental models of the world that political polls have this level of bias?