Refreshingly fast LLMs on
GPUs and NPUs
Install, run LLMs locally, and discover apps in minutes
Built by the local AI community for every PC, with optimizations by AMD
Install Icon

One Minute Install

Simple installer that sets up the stack automatically.

Multi-engine Icon

Multi-engine compatibility

Works with llama.cpp, Ryzen AI SW, and FastFlowLM.

Auto-config Icon

Auto-configures for your hardware

Configures dependencies for your GPU and NPU.

Multiple Models Icon

Multiple Models at Once

Run more than one model at the same time.

Cross-platform Icon

Cross-platform

A consistent experience across Windows and Linux.

Multi-modal Icon

Multi-modal input

Handle text, images, and audio in one SDK.

Built-in app Icon

Built-in app to manage models

A GUI that lets you download, try, and switch models quickly.

Native C++ Icon

Native C++ Backend

Lightweight server that is only 2MB.

Lemonade is compatible with great local AI apps!

Latest Release

Loading latest release...

View Release
Installation Setup