I’ve been building a system called Sprinter — a self-hosted workspace where a single operator runs parallel workflows across multiple role-scoped AI contexts (code, QA, infra, etc.).
It was originally built using Claude Code, but I’m now making it vendor-agnostic and want to plug Gemini in as a first-class backend.
This isn’t one-off chat — it’s a continuous loop:
-
structured prompts per role
-
high-throughput calls
-
persistent context per “agent”
-
tasks like code generation, QA, debugging
I’m trying to figure out the best way to:
-
run Gemini in a loop like this (not just request/response)
-
keep behavior consistent across repeated role prompts
-
handle longer context across ongoing workstreams
If anyone has guidance on how Gemini is best used in this kind of setup, I’d really appreciate it.