Proton Mail is famous for its privacy and security. The cool trick they do is that not even Proton can decode your email. That’s because it never exists on their systems as plain text — it’s always…
Apparently no one knows how LLMs work. I’ll give credit where it’s due, that Proton seems to have a setup that is as good as it can possibly get for anyone who can’t run an LLM locally. Since no one seems to know what that means either, “locally” means you have to run a LLM on your machine.
All LLMs - all of them without exception - process tokens as cleartext. There is no LLM anywhere that can process encrypted tokens. So this is a limitation of the fundamental architecture.
What Proton seems to have is TLS encryption of your text your write > the LLM context window, where Proton has sort of “removed their own access” from that step. The LLM processes and responds, all with the TLS tunnel to you as the only in/out. Proton servers process the tokens, and once the conversation is done, it all gets flushed. It’s not even that hard of a thing to set up conceptually. But it does rely on the same level of trust in Proton as any of their other services. But hey, if they keep passing security audits, that’s reasonable trust to have.
My only gripe is that Mistral doesn’t get the level of investment and development that Big 6 AI companies get, so it’s like GPT3.5 level on a good day. Well, and my second gripe is that a “basic, but slightly better” tier isn’t added in for paid users, and that it’s a standalone add-on.
Here’s where they say the models are Mistral, OpenHands, and OLMO are the models. The blog post isn’t the only documentation they have.
Apparently no one knows how LLMs work. I’ll give credit where it’s due, that Proton seems to have a setup that is as good as it can possibly get for anyone who can’t run an LLM locally. Since no one seems to know what that means either, “locally” means you have to run a LLM on your machine.
All LLMs - all of them without exception - process tokens as cleartext. There is no LLM anywhere that can process encrypted tokens. So this is a limitation of the fundamental architecture.
What Proton seems to have is TLS encryption of your text your write > the LLM context window, where Proton has sort of “removed their own access” from that step. The LLM processes and responds, all with the TLS tunnel to you as the only in/out. Proton servers process the tokens, and once the conversation is done, it all gets flushed. It’s not even that hard of a thing to set up conceptually. But it does rely on the same level of trust in Proton as any of their other services. But hey, if they keep passing security audits, that’s reasonable trust to have.
My only gripe is that Mistral doesn’t get the level of investment and development that Big 6 AI companies get, so it’s like GPT3.5 level on a good day. Well, and my second gripe is that a “basic, but slightly better” tier isn’t added in for paid users, and that it’s a standalone add-on.