A 60-person US SaaS engineering team had three years of accumulated documentation that nobody could find. Runbooks, architecture decisions, incident post-mortems — all written down, none of it accessible. New engineers spent their first two weeks pinging senior devs. Senior devs were the search engine.
A 60-person US SaaS engineering org had accumulated three years of documentation: Confluence spaces covering architecture decisions, system design records, incident post-mortems, onboarding guides, API references, and operations runbooks. The knowledge existed. The problem was finding it.
In practice, new engineers learned the system by asking the people who already knew it. Every new hire's onboarding looked the same: two weeks of pairing sessions, Slack questions to senior engineers, and manual document hunting. Senior engineers answered the same questions repeatedly — questions that already had written answers, buried in a Confluence page nobody could locate quickly enough.
As the team scaled hiring, this pattern became expensive. The senior engineers most capable of doing complex work were spending 6–8 hours per week fielding questions that a well-built internal search tool should have answered in seconds. Onboarding new hires was starting to compress the team's delivery capacity on core product work.
The company had tried Confluence's native search. It returned pages, not answers — and it ranked by edit date, not relevance. They'd looked at off-the-shelf AI search tools, but none could be configured around their specific content structure and access control requirements without significant engineering effort. They needed a system built around how their knowledge was actually organized.
A production RAG system that ingested their full documentation corpus, made it queryable via natural language, and surfaced citations so engineers could navigate to the original source — not just read a summary and wonder if it was current.
/kb Slack command routes queries through the FastAPI backend and returns a threaded answer with citation links, so engineers can search without leaving their workflow. The most common use case: a developer hits an unfamiliar error and searches the knowledge base directly from the incident Slack threadThe core challenge wasn't retrieval quality on clean documents — it was the heterogeneity of the corpus. Confluence pages vary from structured architecture documents with clear headers to freeform meeting notes written in five minutes. GitHub wiki pages follow no consistent format. ADR files are highly structured but use domain-specific vocabulary that standard embedding models underweight.
We handled this by building content-type-aware chunking strategies: structured pages are split at heading boundaries to preserve section context; freeform pages use fixed-size chunks with overlap; ADR files are indexed as atomic units because splitting them breaks the problem/decision/consequence structure that makes them useful. The chunking strategy had measurable impact on retrieval precision — early evaluations with naive fixed-size chunking scored 0.71 faithfulness on the golden test suite; content-aware chunking pushed it to 0.89.
Access control was a second constraint that shaped the architecture from the start. A naive approach — index everything and filter results at the API layer — has the same flaw as API-layer RBAC in any RAG system: semantically similar queries can surface restricted content in the retrieved chunks before the access check runs. We pre-filter at ingestion: restricted spaces never enter the index. The engineering corpus is clean by construction, not filtered after the fact.
The Slack integration introduced a token-budget constraint that doesn't exist in a web interface. Slack thread replies have display limits that make verbose multi-paragraph responses harder to read than a short answer plus a citation link. We built a summarisation layer that produces a 2–3 sentence direct answer for the Slack response and a more complete answer for the web interface, both grounded in the same retrieved context and carrying the same citations.
Platform metrics
Tech stack
Results
"New engineers were spending their first two weeks pinging senior devs for context that was already written down somewhere — nobody could find it. The system Jonix built didn't just solve onboarding. It stopped the senior engineers from being the default search engine for institutional knowledge."
CTO
US SaaS company, 60-person engineering team
Financial Document Intelligence — 10 TB/week at 99.9% uptime
Zero compliance incidents · $20K/yr saved · 14-week engagement
E-commerce · AI Sales AssistantAI Sales & Support Assistant — 34% Conversion Lift
+34% conversions · 81% fully automated · <800ms response
US Fintech · Investor OnboardingInvestor Verification Platform — Months to Hours
$900K+ revenue generated · Zero compliance incidents
Institutional knowledge trapped in your team's heads?
30 minutes. We'll talk about your documentation state, your team's query patterns, and what a production knowledge base looks like for your stack.
Book a Free Scoping CallOr start with a 2-week AI Readiness Audit:
AI Readiness Audit — $3,500