At the front of the architecture is a deterministic intent classifier that routes every request into one of three execution lanes: chit-chat, help, or data.
Chit-chat is handled entirely at the orchestration layer with predefined, personality-controlled responses. It never touches documents, models, or business systems, which keeps the system responsive and prevents unnecessary processing.
Help requests are routed into a local RAG pipeline. Documentation is pre-ingested, chunked, indexed, and ranked at runtime. Only the minimum relevant context is assembled and passed to the locally hosted model. A custom Modelfile enforces strict response constraints, ensuring answers remain grounded in the supplied context and never infer or invent citations.
Data requests are executed through a tightly controlled query layer against internal systems. Results are serialized, injected as bounded context, and explained by the model without allowing free-form interpretation, schema inference, or execution drift.
This separation enforces hard boundaries between conversational behaviour, documentation reasoning, and operational data access, reducing hallucinations, preventing data leakage, and keeping execution predictable — a requirement for regulated and security-sensitive environments.
In this quick demo video, you’ll see it working directly inside Sage CRM:
• Ask natural questions like “How many customers are in London?”
• Ask about orders, opportunities, people, products, cases anything.
• It generates the correct queries and returns clear answers instantly.
• It understands the current CRM record and who's asking the question.
This is the future of Sage CRM insight.
Spire BMS is committed to protecting your privacy and being respectful of personal information that you may provide. You can visit our website without giving any information about yourself. However, you may be aware that all websites process a certain amount of information automatically, which can identify a visitor to a site and their activity on it.