Conversational by default
One gateway alias drives multi-turn editing โ "draw a cat" โ "now make it blue" โ "remove the background" โ all auto-routed per turn.
The OSS counterpart to Google's Nano-Banana 2 / OpenAI's GPT-Image-2 โ served as an OpenAI-compatible LiteLLM provider on top of ComfyUI.

A LiteLLM CustomLLM provider that exposes ComfyUI workflows as OpenAI-compatible image endpoints, with per-turn intent routing for the full nano-banana conversational UX.
# In your chat client (Open WebUI, protoCLI, or raw OpenAI SDK):
user: a watercolor of a cat in a hat, portrait
[image: cat in hat, 832ร1216]
user: now make it blue
[edited image]
user: remove the background
[transparent png]
user: change just the hat to red
[masked region edit โ Phase 4]One model alias (protolabs/qwen-image-chat) handles all of it. The provider walks message history, classifies the operation per turn, dispatches to the right ComfyUI workflow.
| Use protoBanana when | Use Nano-Banana 2 / GPT-Image-2 when |
|---|---|
| Data sovereignty / compliance / IP sensitivity | You don't care where the data goes |
| You want fixed cost (electricity) at scale | You're under metered-API-call budgets |
| You need to extend with custom workflows | Frontier-quality output is non-negotiable |
| You already run a LiteLLM gateway | You don't have GPU infrastructure |
For most teams: both. Use the closed APIs for one-off best-quality work, route bulk + sensitive workflows through protoBanana.