There’s one more option I’ve recently discovered that might be a new spot on the spectrum of tradeoffs… So there are GPU providers like modal.com that let you spin up GPU environments and only pay for the seconds or minutes it’s up and running. So you could create a container boots up with a newly generated key, messages get decrypted, run through the model, encrypt the outputs and send back to user, and the plaintext content never leaves RAM or VRAM. So the cloud provider should theoretically have a harder time to spy on you, even though they could…. And because this is on demand, you can use big open source models.
Trade-offs all the way down.
I've thought about AWS Nitro enclaves + NVIDIA "Confidential Computing" like nostr:nprofile1qqs8msutuusu385l6wpdzf2473d2zlh750yfayfseqwryr6mfazqvmgpy4mhxue69uhkvet9v3ejumn0wd68ytnzv9hxgtm0d4hxjh6lwejkuar4wfjhxqfswaehxw309a5hgcmg0ykkwmmvv3jkuun0vskkvatjvdhkuargdacxsct8w4ejuumrv9exzc3wd9kj7nfu4p9. I can see nostr:nprofile1qqsgha3fk023ng8c4quszdayghqwkt6l9d9ga4c3280gnqz3aqqx7ycpp4mhxue69uhkummn9ekx7mqprdmhxue69uhkummnw3ezucnfw33k76twwpkx2cnn9ejx2qaq9wr proxy providers having this kind of set up as a premium tier for their services.
Discussion
No replies yet.