Ollama supplies an interface and REST API server for working and calling regionally hosted massive language fashions (LLMs). The applying doesn’t present authentication by default and can also be typically configured to pay attention on all community interfaces (0.0.0.0), despite the fact that it’s meant for native utilization and binds to localhost (127.0.1.1) by default. There are roughly 300,000 Ollama servers presently uncovered on the general public web and lots of extra on native networks.
“With over 170,000 GitHub stars and 100 million Docker Hub downloads, Ollama is extensively used throughout enterprises as a self-hosted AI inference engine,” Cyera warns, including that the vulnerability is broadly exploitable as a result of no authentication is required.
Solely three API requests wanted for exploit
Positioned in Ollama’s mannequin quantization pipeline, the bug pertains to how the framework hundreds GGUF (GPT-Generated Unified Format) recordsdata, which retailer weights, metadata, and tokenizer info for native fashions.



