Is Ollama the most private/secure way to run AI models locally?
from HiddenLayer555@lemmy.ml to privacy@lemmy.ml on 09 May 21:55
https://lemmy.ml/post/29858666

I’m currently running Deepseek on Linux with Ollama (installed via curl -fsSL https://ollama.com/install.sh | sh), and I specifically have to run it on my personal file server because it’s the only computer in the house with enough memory for the larger models. Since it’s running on the same system that has direct access to all my files, I’m more concerned about security than I would be if it was running on a dedicated server that just does AI. I’m really not knowledgeable on how AI actually works at the execution level, and I just wanted to ask whether Ollama is actually private and secure. I’m assuming it doesn’t send my prompts anywhere since everything I’ve read lists that as the biggest advantage, but how exactly is the AI being executed on the system when you give it a command like ollama run deepseek-r1:32b and have it download files from where it’s downloading from by default? Is it just downloading a regular executable and running that on the system, or is it more sandboxed than that? Is it possible for a malicious AI model to scan my files or do other things on the computer?

#privacy

threaded - newest

0laura@lemmy.dbzer0.com on 09 May 22:11 next collapse

ollama downloads the model from the original source afaik, so there’s not really any risk. the model itself can’t do anything bad but I do not know if there could be malware added through the loading process. I remember there being big problems with pickle files for stable diffusion or something, though that’s been fixed with safetensor afaik.

rutrum@programming.dev on 09 May 22:12 next collapse

Its all local. Ollama is the application, deepseek and llama and qwen and whatever else are just model weights. The models arent executables, nor do the models ping external services or whatever. The models are safe. Ollama itself is meant for hosting models locally, and I dont believe it even has capability of doing anything besides run local models.

Where it gets more complicated is “agentic” assistants, that can read files or execute things at the terminal. The most advanced code assistance are doing this. But this is NOT a function of ollama or the model, its a function of the chat UI or code editor plugin that glues the model output together with a web search, filesystem, terminal session, etc.

So in short, ollama just runs models. Its all local and private, no worries.

Fubarberry@sopuli.xyz on 09 May 22:59 collapse

Most models now are .safetensor files which are supposed to be safe, but I know in the past there were issues where other model filetypes actually could have attack payloads in them.

npe@leminal.space on 09 May 23:09 collapse

yeah, this was definitely not a silly question to ask

movies@lemmy.world on 09 May 22:13 next collapse

Nah, you’re safe to run it locally. You’re downloading the specific model, that’s right, and it’s not an exe. As you ask questions of it, the inference step, that is sent directly to the model on your machine by the ollama interface. Nothing goes over the network after you download a model and there is no scanning involved; that’s just not how it works.

N0t_5ure@lemmy.world on 09 May 22:28 next collapse

Thanks, Ollama.

npe@leminal.space on 09 May 23:08 next collapse

It’s a good question. Older model formats used to allow for executable code to be present and thus would present a security risk. But with the formats that Ollama and Llama.cpp use I believe that’s not the case anymore.

coconut@programming.dev on 10 May 00:47 collapse

Use docker if you want peace of mind

ArsonButCute@lemmy.dbzer0.com on 10 May 03:07 collapse

Seconding this, I also keep my AI models locked up in docker.