How long until we have something in production?
Most engagements ship a live, evaluated agent in four to six weeks. We don't ship demos — when an agent goes live, it's instrumented, eval-covered, and has a documented hand-off path to your team.
Which model providers do you use?
Whichever ships. We've put OpenAI, Anthropic, Gemini, Mistral, and self-hosted models into production in the last 18 months. We pick based on latency, cost, and accuracy on your eval set — not on vendor relationships.
Do you build the eval harness, or just plug one in?
We build it. Every engagement starts with a corpus of real cases pulled from your support inbox, ticket queue, or transcript archive. That corpus becomes the eval set, wired into CI, scored on every model bump.
What happens when you walk away?
Your team owns the system. We hand off code, tests, runbooks, the eval corpus, and a roadmap. Optionally we stay on retainer to run weekly evals and respond to regressions — but it's an option, not a lock-in.
Can the agent integrate with our existing systems?
Yes. Most agents we ship have function-calling into a CRM, EHR, dispatch system, or internal API. We use whatever you already have — REST, gRPC, queues, webhooks — and write thin adapters where we need to.
How do you handle data privacy and compliance?
We default to your hosting (your AWS / GCP / Azure account). We've shipped HIPAA-aligned and GDPR-aligned agents. PII handling, retention policies, and audit logging are designed in from week one, not bolted on later.