But does this so called “log n” scaling even matter compared to the real world in real numbers?

A performance decrease of only 3x for a minimal part (reading the graph instead of analyzing it) at a scale of over 100000x is insanely good, compared to all those compactions that will be needed and the tremendous write workload at that scale...

Reply to this note

Please Login to reply.

Discussion

A billion users doesn’t mean n = a billion. If you’re looking at how people interact with each other, you’ll need a lot of indices for a lot of events.

It means 1 billion nodes though in an index.

If I want to discover instances where Alice reacts to a kind 1 note authored by Bob, I’ll need indices for content in addition to indices for users.

Or suppose I want a chart of total zap amount in sats as a function of how far down a note is on a thread. Now suppose I only want to count zaps sent by frens (mutual follows) of the thread initiator. Or suppose I want a query even more complicated than that.

We’ll need indices for all content events, all zaps, all nostr events. Trillions.

You take all zap events and use some fast OLAP DB to crunch through this data instantly instead of wasting your time with a graph DB

This is a simple merge-join on 2 indexes that is cheap