Athina home pagelight logodark logo
  • AI Hub
  • Website
  • Sign-up
  • Sign-up
Datasets
Prototype and Evaluate a Prompt Chain
Docs
API / SDK Reference
Guides
FAQs
  • Documentation
  • Open-Source Evals
  • Blog
  • Email us
  • Book a call
  • Getting Started
    • Overview
    Prompts
    • Prompt Comparison
    • Prompt Versioning
    Datasets
    • AWS Bedrock Models
    • Preparing Data for Fine-Tuning
    • Get Data from S3 Bucket
    • Compare and Evaluate Multiple Models
    • Comparing datasets using Athina IDE
    • Prototype and Evaluate a Prompt Chain
    • Run Prompts and Evaluate
    Evals
    • Pairwise Evaluation
    • Evaluations in CI/CD Pipeline
    • A guide to RAG evaluation
    • Prompt Injection: Attacks and Defenses
    • Running evals as real-time guardrails
    • Different stages of evaluation
    • Improving Eval performance
    • Measuring Retrieval Accuracy in RAG Apps
    • Pairwise Evaluation
    • Evaluate Conversations
    • RAG Evaluators
    Experiments
    • Compare Multiple Models
    Flows
    • Create and Share Flow
    Datasets

    Prototype and Evaluate a Prompt Chain

    You can prototype and evaluate a prompt chain in Athina IDE

    Here’s how you can prototype and evaluate a prompt chain in Athina IDE:

    Comparing datasets using Athina IDERun Prompts and Evaluate
    twitterlinkedin
    Powered by Mintlify