0 / 0
AutoAI RAG pattern details
Last updated: Dec 12, 2024
AutoAI RAG pattern details

After your AutoAI RAG experiment completes, view the details of a pattern to understand the composition and performance.

Important: This feature is a beta release. It is not intended for production use.

From the pipeline leaderboard, you can view a table with the following information for each of the patterns the experiment generates:

Column Description
Rank Rank of placement according to best performance for the optimized metric
Name Pattern name
Model name Name of the embedding model used to vectorize and index the documents
Optimize metric Performance result for the optimized metric selected for the experiment. A RAG pattern can be optimized for Answer faithfulness and Answer correctness
Chunk size Amount of data retrieved from the indexed documents
Retrieval method Method (Window or Simple) used to retrieve indexed documents
Vector store distance metric Metric (Cosine or Euclidean) used to measure how relevant the stored vectors are to the input vectors

Click on a pattern name to review the following configuration details for a pattern. Some of the configuration settings are editable when you create the experiment. For details, see Customizing RAG experiment settings.

Column Description
Pattern Pattern name
Answer correctness (mean) Correctness of the generated response including both the relevance of the retrieved context and the quality of the generated response.
Answer faithfulness (mean) Accuracy of the generated response to the retrieved text.
Context correctness (mean) Relevancy of the generated response to the input.
Composition steps Steps taken in the experiment. For example, Chunking or Retrieval.
Chunk overlap Number of chunks that share common words or phrases.
Chunk size Amount of data retrieved from the index documents.
Chunk augmentation How relevant text is retrieved for each chunk of the input after splitting the input into multiple chunks.
Embedding model Embedding model used to vectorize and index the document collection.
Truncate input tokens Maximum number of tokens accepted as input.
Truncate strategy Strategy used to determine how to process retrieved documents to optimize the model performance.
Context template text Structured template for the generated response.
Model type Foundation model used in the experiment. Click the preview icon for details about the model.
Max new tokens Maximum number of new tokens that can be generated in addition to the tokens retrieved from indexed documents.
Prompt template text Structure with guidelines for the model to use when retrieving text from indexed documents.
Retrieval method Method used to retrieve indexed documents: Window or Simple.
Number of chunks Number of chunks retrieved from the indexed documents.
Window size Number of adjacent tokens considered by the model when retrieving text from indexed documents.
Vector store datasource type Datasource that connects to the vector store.
Vector store distance metric Metric used to measure how relevant the stored vectors are to the input vector.
Vector store index name Index used as a data structure for storing and retrieving vector stores.
Vector store operation Vector store operation for retrieving data. For example, upsert for inserting or updating rows in the database table
Vector store schema fields Schema for storing the index of vectored documents.
Vector store schema ID autoai_rag_1.0
Vector store schema name Name of the schema used for structuring the database table
Vector store schema type The schema type struct is used for storing structured data.
Sample Q&A Questions and correct answers provided as test data to measure the performance of the pattern.

Parent topic: Automating a RAG pattern with AutoAI

Generative AI search and answer
These answers are generated by a large language model in watsonx.ai based on content from the product documentation. Learn more