Same. In this case, itβs the idea of neural networks taking advantage of the sparsity of the embeddings to encode more features than just the dimensionality of the vector space (the set of orthogonal vectors). I probably canβt do it justice in a nostr note after a few whiskeys.
Discussion
Any comment on this will make me sound stupid, so I'll refrain π