Osaurus brings both local and cloud AI models to your Mac

by | May 15, 2026 | Technology

As AI models increasingly become commoditized, startups are racing to build the software layer that sits on top of them. One interesting entrant into this space is Osaurus, an open source, Apple-only LLM server that lets users move between different local AI models, either locally or in the cloud, while keeping their files and tools all on their own hardware.

Osaurus evolved out of the idea for a desktop AI companion, Dinoki, which Osaurus co-founder Terence Pae described as a sort of “AI-powered Clippy.” Dinoki’s customers had asked him why they should buy the app if they still had to pay for tokens — the usage units AI companies charge for processing prompts and generating responses.

That got Pae thinking more deeply about running AI locally.

“That’s how Osaurus started,” Pae, previously a software engineer at Tesla and Netflix, told TechCrunch over a call. The idea, he explained, was to try to run an AI assistant locally. “You can do pretty much everything on your Mac locally, like browsing your files, accessing your browser, accessing your system configurations. I figured this would be a great way to position Osaurus as a personal AI for individuals.”

Pae began building the tool in public as an open-source project, adding features and fixing bugs along the way.

Image Credits:Osaurus, Inc.

Today, Osaurus can flexibly connect with locally hosted AI models or cloud providers like OpenAI and Anthropic. Users can freely choose which AI models they’re using, and keep other aspects of the AI experience on their own hardware, like the models’ own memory, or their files and tools.

Given that different AI models have different strengths, the advantage of this system is that users can switch to the AI model that best fits their needs.

Such a structure makes Osaurus what’s called a “harness” — a control layer that connects different AI models, tools, and workflows through a single interface, similar to tools like OpenClaw or Hermes. However, the difference is that such tools are often aimed at developers who know their way around a terminal. And sometimes, like in the case of OpenClaw, they may pose security issues and holes to worry about.

Osaurus, meanwhile, presents an easy-to-use interface that consumers can use, and addresses security concerns by running things in a hardware-isolated, virtual sandbox. This limits the AI to a certain scope, keeping your computer and data safe.

Image Credits:Osaurus, Inc.

Of course, the practice of running AI models on your machine is still in its early days, given that it’s heavily resource-intensive and hardware-dependent. To run local models, your system will need at least 64 GB of RAM. For running larger models, like DeepSeek v4, Pae recommends systems with about 128 GB of R …

Article Attribution | Read More at Article Source