Ejento AI
HomeQuickstartGuidesRecipesREST APIsChangelogFAQs
HomeQuickstartGuidesRecipesREST APIsChangelogFAQs
Ejento AI
  1. Features
  • Basic Operations
    • Features
      • Teams → Projects → Assistants Hierarchy
    • Guides
      • Login/Signup
  • Assistants
    • Features
      • Introduction to Assistants
      • Assistant Access Control
      • Caching Responses for Assistants
      • Assistant Evaluation
      • Evaluation Metrics
      • URL-based Chat Thread Creation and Prepopulation
      • Reasoning Modes
    • Guides
      • Add Assistant
      • Evaluate Assistant
      • Edit Assistant
      • Delete Assistant
      • Add Favourite Assistants
      • View Assistant Id
      • View Dataset Id
  • Workflows
    • Features
      • Overview and Types of Supervisors
    • Guides
      • Add Workflow
      • Workflow Chat
  • Corpus
    • Features
      • Introduction
      • Corpus Permissions
      • PII Redaction
    • Guides
      • Assistant Corpus Setup
      • Assistant Corpus Settings
      • Corpus Access Control
      • Corpus Connections
      • View Corpus Id
      • View Document Id
      • Tagging
        • Corpus tagging
        • Document tagging
  • Tools
    • Features
      • Introduction
    • Guides
      • Configure RAG Tool
      • Configure Attachment Tool
      • Configure Web Search Tool
      • Configure API Tool
      • View Tool Id
  • Analytics
    • Guides
      • Analytics
  • User Settings
    • Features
      • Ejento AI User Access Levels
    • Guides
      • Assistant Edit Access
      • Add new user
      • Add User in a Team
      • Remove User from a Team
      • View my Access level in a Team
      • View my User Id
  • Chatlogs
    • Guides
      • Chatlogs
      • View Chatlog & Chat thread Id
  • API Keys
    • Guides
      • API Keys
  • Integrations
    • Email Indexing
    • Microsoft Teams
    • Sharepoint Indexing
    • MS Teams Integration Setup
  • Teams
    • Features
      • Introduction
    • Guides
      • Add a Team
      • Edit a Team
      • Delete a Team
      • View Team Id
  • Projects
    • Features
      • Introduction
    • Guides
      • Add a Project
      • Edit a Project
      • Delete a Project
      • View Project Id
  1. Features

Assistant Evaluation

Understanding Evaluation Tests in Ejento AI#

Introduction to Evaluation Tests#

Evaluation tests are a crucial part of ensuring that AI assistants provide accurate, relevant, and high-quality responses. In Ejento AI, the evaluation process involves testing the assistant's responses against benchmark queries to assess their effectiveness and quality.

Key Components of Evaluation Tests#

1. Queries#

Queries are the input questions or prompts directed to the AI assistant. In the evaluation context, these queries serve as the basis for creating datasets. They represent the real-world questions that users might ask the assistant.

2. Datasets#

Datasets are collections of queries selected for evaluation. When creating a dataset in Ejento AI, you can choose multiple queries that the assistant has previously handled. Each dataset is given a unique name and description to identify its purpose and content.

3. Evaluation Metrics#

The core of the evaluation process lies in the metrics used to measure the assistant's performance. The key metrics include:
Answer Similarity: This metric assesses how closely the assistant's response matches the benchmark or correct answer. High similarity indicates accurate answers.
Answer Relevance: This measures how well the assistant's response directly addresses the user's query. It ensures that the response is pertinent to the question asked.
Faithfulness: Faithfulness checks whether the assistant's response accurately reflects the underlying source content or knowledge base. It prevents the generation of misleading or incorrect information.
Context Recall: This metric evaluates the assistant's ability to maintain context and continuity in its responses. It is crucial for multi-turn conversations where understanding the context of previous queries is necessary.

Importance of Evaluation Tests#

Evaluation tests provide insights into the strengths and weaknesses of the AI assistant. By analyzing the results from various metrics, developers and users can:
Identify areas where the assistant performs well.
Detect issues where the responses may be lacking in accuracy, relevance, or context.
Make informed decisions on how to improve the assistant's performance through retraining or adjusting the underlying algorithms.

Conclusion#

Evaluation tests are an essential tool in the development and maintenance of AI assistants in Ejento AI. By understanding and utilizing queries, datasets, and evaluation metrics, users can ensure that their AI assistants provide high-quality responses, improving user satisfaction and effectiveness.
Previous
Caching Responses for Assistants
Next
Evaluation Metrics