This actually isn’t that far-fetched. One could calculate semantic embeddings for each arriving note and store them alongside in the local db. These embeddings are typically arrays of floats in the size of 300-1500 entries each. The machine-learned algorithm has been pre-trained to encode similar meaning sentences into numbers that are “close” in the vector space of the embeddings.
Then, for a feed prompt, one calculates the semantic embedding of the prompt, and can query the embedding database for the top-K closest notes that match, or in real time, filter out a stream of notes based on semantic similarity as they arrive.
All this exists and only requires commodity lower-end GPUs (or even just CPUs, but it is slow).
Kind of an obvious application in retrospect.