#Alexandria looks amazing on my Pixel tablet 🀩

Can't believe how far nostr:nprofile1qqsggm4l0xs23qfjwnkfwf6fqcs66s3lz637gaxhl4nwd2vtle8rnfqprdmhxue69uhhg6r9vehhyetnwshxummnw3erztnrdakj7qfqwaehxw309ahx7um5wghx26tww4hxg7nhv9h856t89eehqctrv5hsz5xkmf's come through all our pursuits πŸ™

nostr:nevent1qqsy6tg6xfd5fgqea6k9ewznpgzglgvaedydkxwghexk9vuhqrayd6spr3mhxue69uhhg6r9vd5hgctyv4kzumn0wd68yvfwvdhk6tczyrwye5yxe47wtvvr9t05lhgjzy5f3qxjcl3ft09su6zvqxkwua7qvqcyqqqqqqgfgcu94

Reply to this note

Please Login to reply.

Discussion

gitadel would be an amazing name

gitcitadel sounds clumsy

🫑

nostr:npub162zpxufpw8pnuytaf0gfxzkqtvk9rvcwkvppa7x57y3n7qkfpg4shatdhy it was actually writing this essay where I first encountered that word. Maybe you'll enjoy reading it πŸ’œ

Oh yeah? Will have a read in the morning, thanks! Looking forward. πŸ‘€

πŸ™

This is long and rather dense. May be another day before finish. Good start though πŸ‘€

Ok, finished reading it. It’s really good and fleshed out! A lot of the framing you gave clarified some vague concepts I held but never considered very carefully. Definitely shifted my schema about knowledge databases a bit.

I gathered you spend a lot of time with graphs and ML. Was curious what you do exactly for say connecting notes with some sematic-weighted distance measure. I know a bit about this, so no need to dumb it down too much if you choose to elaborate. My guess is you leverage some state of the art LLM model and then build a β€œsimple” ML model with some popular distance choice to rank things relative to one another.

Yeah, pretty easy to explain actually. There's a ton of embedding models that you can work with off the shelf if you'd rather not train yourself (can find them on huggingface, a repository of AI models).

Words go in -> numbers come out.

Your vocabulary of words/tokens gets assigned some number values and the training task is to predict some part about the group of words you've been assigned

1) I've masked word or set of words with a blank value, predict the words.

2) predict the next n tokens

etc.

the performance depends on the task and the data it was trained on, but the result is you get a bunch of numbers that you can compare through a bunch of distance metrics. Grab a bunch of text with vectors assigned to them and you ask "which text is closest to this thext i care about", which is some K-nearest neighbors algorithm

I see. Again, good stuff. Thanks for sharing

Thank you 😁 it was a really fun learning and writing it πŸŽ‰

Mind if I send you a quick DM? Don't see another way to reach you listed

Sure!