The Analytics Dashboard provides insights into how the Knowledge Assistant is performing. It tracks usage patterns, response quality, user engagement, and system performance. Access the dashboard from the Admin Console.
Performance metrics help you understand how fast the system responds to queries.
The time between when a user submits a query and when the first token of the response starts streaming. This is the perceived “thinking time” before the answer appears.
The rate at which tokens are delivered to the user, measured in characters per second. Higher rates mean faster-appearing responses once streaming begins.
The percentage of responses that complete successfully vs. those that are aborted by users or fail due to errors. High completion rates indicate good user experience.
These metrics measure how well the system finds relevant content from the knowledge base.
Measures how high the first relevant result appears in the retrieved chunks. MRR of 1.0 means the most relevant chunk is always first; lower values mean relevant content is buried further down the list.
The percentage of relevant chunks that appear in the top 5 results. High recall means the system is finding most of the relevant content; low recall means important information is being missed.
The percentage of the top 5 results that are actually relevant. High precision means less noise in the retrieved content; low precision means the model has to filter through irrelevant chunks.
The percentage of queries where at least one relevant chunk was retrieved. A low hit rate indicates the knowledge base may be missing content or the retrieval system is failing to find existing content.
These metrics evaluate the quality of the generated responses.
Measures whether the generated response is supported by the retrieved context. High faithfulness means the model is grounding its answers in the source material rather than hallucinating information.
Measures how well the response actually answers the user's question. High relevance means the response addresses what was asked; low relevance indicates the model may be going off-topic.
Engagement metrics show how users interact with responses.