First look: Lemonade serves up local AI with limitations
Lemonade, created by AMD, is a server application plus GUI for running local AI models, similar to projects like LM Studio (or, more distantly, ComfyUI). What it lacks in configurability, it tries to make up for in broader integration with third-party apps that use standard APIs, and with support for non-NVIDIA runtimes. Lemonade works with a variety of runtimes and back-end engines. It supports AMD GPUs, Ryzen NPUs, Vulkan, and CPU execution (although not for all tasks), along with the llamacpp, whispercpp, sd-cpp, kokoro, ryzenai-llm and flm back ends. In addition to providing its own set of APIs, Lemonade interoperates with a broad set of industry standards including OpenAI, Ollama, Anthropic, and llama.cpp. Both GGUF and ONNX models are supported. The biggest omission is NVIDIA-specific GPU support. Only Vulkan (generic GPU) and AMD (ROCm) GPUs are supported. So, if you plan on using StableDiffusion models with NVIDIA hardware, you may want to look elsewhere for now. (StableDiffusi