Its not a specific question but related to this thread:

nostr:nevent1qvzqqqqqqypzqwlsccluhy6xxsr6l9a9uhhxf75g85g8a709tprjcn4e42h053vaqyghwumn8ghj7mn0wd68ytnhd9hx2tcprdmhxue69uhhg6r9vehhyetnwshxummnw3erztnrdakj7qpqfcgt7pmh4txzd5echeyh5jcwntz6nyy7zege2vz53lumrjq2jzdq88gwkc

Maybe if embeddings can be made portable enough, they can also be passed around nostr in a decentralized way, where clients and relays would be able to leverage note searching. Kind 1 seems like a big ask, because your essentially trying to classify and search through all possible conversations and context is usually temporal.

More slowly moving kinds, where context is contained within the note, or surrounding notes on a relay would be interesting to organize around. But what about paragraphs? If so, maybe binary embeddings on relays could be a thing to help organize and find related notes.

Reply to this note

Please Login to reply.

Discussion

Something about a vector binary embedding of an entire relays content to perhaps make searching a compressed relays content extremely efficient and perhaps you could embed all of the content across the entire protocol. Then it could be redundant and spread everywhere and all nostr content may no longer need to be fragmented across relays but aggregated with the embedding and globally distributed across all relays. Sounds really complex though.

I'm hesitant about doing it for kind 1 because there is such a wide context, lots of notes collected over lots of relays - don't think it would really work because for the user it there is so many unrelated notes. But imagine a pool of relays focused on some content:

nostr:nevent1qvzqqqqqqypzphzv6zrv6l89kxpj4h60m5fpz2ycsrfv0c54hjcwdpxqrt8wwlqxqyghwumn8ghj7mn0wd68ytnhd9hx2tcprdmhxue69uhhg6r9vehhyetnwshxummnw3erztnrdakj7qpqcu2498m77mq80cj5tuh3l8rnw97e8vx54alva35pzmr207lt4xuqqz5htc

Even if you group across multiple relays, a user is choosing to connect to related content. A user is reading some related articles that have been embedded and so you (service provider) or the client is able to construct some sort of HNSW search, and pass some sort of subnetwork within some distance close to your note.

I suppose it would depend on how good binary vector embedding actually is. 😂

Its a compressed representation, so you lose some resolution to search through, and the space of topics that can be covered in a feed of kind1 notes over some length of time is huge. Less so with a themed community. So by going with something focused you and only improve in performance.

* you can only improve

Dropping a demo in case you'd want to play around with it

https://github.com/limina1/nostr-binary-embedding-demo

Good point