Wow this is incredible! It'd be cool if LM Studio added support for this. You could rely on cloud-based inference when you have internet and want lightning-fast responses, but fallback to local inference using the same model if you're offline or want to ask a question privately.
#m=image%2Fjpeg&alt=Verifiable+file+url&x=5206f9b1f5d04e7d282d60ad5be502f78617db861f7b71e2c12c3911bf6c9750&size=125246&dim=982x1920&blurhash=%5D6SigQMx%7EWs%3Axut7R*M%7Bofxu%25MbbD%25WVWBWBIURjxut7xuRjWBofWBRPM_jZofoft8RjWBofjtIUxajtM%7Bt7ozWVM%7Bj%5Bt7&ox=1752393204192781fccea05ec104271408cee897c2ba95d17e5ba1b9a1314fde