Product Thumbnail

Kuzco

Open-source Swift package to run LLMs locally on iOS & macOS

iOS
Developer Tools
Artificial Intelligence
GitHub

Kuzco is a Swift package for integrating large language models (LLMs) directly into iOS, macOS, and Mac Catalyst apps. Built on `llama.cpp`, it offers customizable prompts, flexible tuning, and async/await-friendly APIs for on-device AI.

Top comment

Hey everyone! I built Kuzco for two big reasons: 1. Developers can cut costs by skipping cloud APIs and running AI locally on‑device, completely free. 2. I was done sending my data (and users’ data) to third‑party servers. Kuzco is a straightforward Swift wrapper for llama.cpp, letting you plug local models into iOS and macOS apps with just a few lines of code. It’s fully open source, so dive in, tweak away, and if you use it (or even plan to) please drop me a link on X (Twitter), so I can cheer you on!

Comment highlights

Very cool! Love that Kuzco makes it easier to bring LLMs directly into Apple platforms. The async/await-friendly APIs and on-device support sound especially useful—looking forward to trying it out!

Kuzco is a useful open-source Swift package that enables running LLMs locally on iOS and macOS, offering customizable features and user-friendly APIs for seamless on-device AI integration.

Async/await support is chef’s kiss, and the llama.cpp backbone

Thank you for your contribution to the open source community

best thing ever for web-developers!!! Finally multi-modal came to APP world

Kuzco makes on-device LLM integration feel native. Swift-first, async-friendly, and built on llama.cpp finally, customizable AI without the latency tax.

Finally. This was the product I had been waiting for for a long time, and which the global community had been waiting for. Congratulations.

Running AI models locally with just a Swift wrapper? That’s genius—no more sending user data everywhere, ngl this is a game-changer for privacy-first apps. Makers nailed it!

Congrats to your launch! Which offline model do you think is the best for your scenario?