Epsilla
HomeDiscordTwitterGithubEmail
  • Welcome
    • Register and Login
    • Explore App Portal
  • Build Your First AI Agent
    • Create a Knowledge Base
    • Set Up Your AI Agent
    • Publish Your AI Agent
  • Knowledge Base
    • Local Files
    • Website
    • Google Drive
    • S3
    • Notion
    • Share Point
    • Google Cloud Storage
    • Azure Blob Storage
    • Confluence
    • Jira
    • Advanced Settings
      • Auto Sync
      • Embedding
      • Data Parsing
      • Data Chunking
      • Hypothetical Questions
      • Webhook
      • Meta Data
    • Data Storage
    • Programmatically Manage Knowledge Bases
  • Application
    • Create New AI Agent
    • Basic Chat Agent Config
    • Basic Smart Search Agent Config
    • Advanced Workflow Customization
    • Publish and Deployment
    • User Engagement Analytics
  • Evaluation
    • Create New Evaluation
    • Run Evaluation
    • Evaluation Run History
  • Integration
  • Team Member Management
  • Project Management
  • Billing Management
  • Release Notes
  • Epsilla Vector Database
    • Overview
    • Quick Start
      • Run with Docker
      • Epsilla Cloud
    • User Manual
      • Connect to a database
      • Create a new table
      • Drop a table
      • Delete a database
      • Insert records
      • Upsert records
      • Search the top K semantically similar records
      • Retrieve records (with filters and pagination)
      • Delete records
      • Performance Tuning
    • Advanced Topics
      • Embeddings
      • Dense vector vs. sparse vector
      • Hybrid Search
    • Integrations
      • OpenAI
      • Mistral AI
      • Jina AI
      • Voyage AI
      • Mixedbread AI
      • Nomic AI
    • Roadmap
Powered by GitBook
On this page
  • Create a New Evaluation
  • Fill in Test Case Details
  • Add Test Cases
  • Future Changes
  1. Evaluation

Create New Evaluation

PreviousEvaluationNextRun Evaluation

Last updated 6 months ago

Create a New Evaluation

Click the Create Evaluation button, which will prompt the creation of a new evaluation template.

You will be taken to an empty evaluation titled Untitled Evaluation - {Creation Time}, where you can start adding details.

Click on the title field, which will initially display the default name, and give your evaluation a meaningful name (e.g., "Financial Agent Evaluation").

Unfocus the name input to apply the changes.

Fill in Test Case Details

There will be one empty test case created by default. Fill in test case details.

Question: Input the query or task you want the AI agent to handle. For example, “What was the cumulative total return for Meta Platforms, Inc. at the end of 2022 compared to its peak value during the five-year period ending December 31, 2023?”

Expected Answer: Input the correct answer that the AI agent is expected to provide. For example, “The cumulative total return for Meta Platforms, Inc. at the end of 2022 was 90, which is significantly lower compared to its peak value of 275 at the end of 2023.”

Add Test Cases

You want to create a set of test cases that can be repeatedly executed during AI agent updates, so that we can be confident our configuration changes don't introduce regression.

Click the Add Test Case button to input questions and expected answers for your evaluation. A new row will be added for each test case, with columns for Question, Expected Answer for you to fill in same as above.

Add as many test cases as needed.

Future Changes

After an evaluation is created, you can always go back and make changes by clicking the card from the list.

You can also delete an evaluation from the actions dropdown menu.