LIL BRO is a dual-agent terminal interface that runs two AI coding assistants locally on your machine using Ollama. No API keys. No cloud services. No subscriptions. Everything runs on your hardware.
Runs via Ollama on your local hardware. The recommended default is qwen2.5-coder:7b — medium speed, full tool-calling support, 5–6GB VRAM.
| Model | Notes |
|---|---|
| qwen2.5-coder:7b ★ | Recommended default — 5–6GB VRAM, full tool support |
| qwen2.5-coder:3b | Lighter hardware, less VRAM required |
| qwen2.5-coder:14b | Higher quality, needs more VRAM |
| deepseek-coder-v2 | Alternative coding model via Ollama |
| llama3.1 / llama3.2 | General purpose, good reasoning |
Any model available via ollama pull can be used.