Osaurus icon
Osaurus icon

Osaurus

Native, Apple Silicon–only local LLM server. Similar to Ollama, but built on Apple's MLX for maximum performance on M-series chips. SwiftUI app + SwiftNIO server with OpenAI-compatible endpoints.

Osaurus screenshot 1

Cost / License

  • Free
  • Open Source (MIT)

Platforms

  • Mac
-
No reviews
1like
0comments
0news articles

Features

Suggest and vote on features
  1.  Dark Mode
  2.  Sits in the MenuBar
  3.  AI-Powered
  4.  OpenAI integration
  5.  Apple Silicon support

 Tags

  • local-ai
  • ai-models
  • ai-model-integration
  • openai-api
  • llm-integration
  • large-language-model-tool

Osaurus News & Activities

Highlights All activities

Recent activities

Osaurus information

  • Developed by

    US flagDinoki, LLC
  • Licensing

    Open Source (MIT) and Free product.
  • Written in

  • Alternatives

    41 alternatives listed
  • Supported Languages

    • English

AlternativeTo Category

AI Tools & Services

GitHub repository

  •  3,071 Stars
  •  127 Forks
  •  25 Open Issues
  •   Updated  
View on GitHub

Popular alternatives

View all
Osaurus was added to AlternativeTo by Paul on and this page was last updated .
No comments or reviews, maybe you want to be first?

What is Osaurus?

Native, Apple Silicon–only local LLM server. Similar to Ollama icon Ollama, but built on Apple's MLX for maximum performance on M-series chips. SwiftUI app + SwiftNIO server with OpenAI-compatible endpoints.

Created by Dinoki icon Dinoki, a fully native desktop AI assistant and companion.

Highlights:

  • Native MLX runtime: Optimized for Apple Silicon using MLX/MLXLLM
  • Apple Silicon only: Designed and tested for M-series Macs
  • OpenAI API compatible: /v1/models and /v1/chat/completions (stream and non-stream)
  • Function/Tool calling: OpenAI-style tools + tool_choice, with tool_calls parsing and streaming deltas
  • Chat templates: Uses model-provided Jinja chat_template with BOS/EOS, with smart fallback
  • Session reuse (KV cache): Faster multi-turn chats via session_id
  • Fast token streaming: Server-Sent Events for low-latency output
  • Model manager UI: Browse, download, and manage MLX models from mlx-community
  • System resource monitor: Real-time CPU and RAM usage visualization
  • Self-contained: SwiftUI app with an embedded SwiftNIO HTTP server

Official Links