-
Notifications
You must be signed in to change notification settings - Fork 831
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
#136 TODO - [ ] Add README with flowchart demonstration of paradigm - [x] Add quickstart notebook - [x] Fix type hinting - [x] Reformat methods if needed - [x] Add progress bar - [ ] Async workflow --------- Co-authored-by: jjmachan <[email protected]>
- Loading branch information
1 parent
ed479d4
commit 75567b8
Showing
4 changed files
with
690 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,182 @@ | ||
{ | ||
"cells": [ | ||
{ | ||
"cell_type": "markdown", | ||
"id": "fe8644a7", | ||
"metadata": {}, | ||
"source": [ | ||
"### Import Required libraries" | ||
] | ||
}, | ||
{ | ||
"cell_type": "code", | ||
"execution_count": 1, | ||
"id": "5aab66fc", | ||
"metadata": {}, | ||
"outputs": [], | ||
"source": [ | ||
"import os\n", | ||
"import json\n", | ||
"from llama_index import download_loader\n", | ||
"from ragas.testset import TestsetGenerator\n", | ||
"\n", | ||
"\n", | ||
"os.environ[\"OPENAI_API_KEY\"] = \"your-openai-key\"" | ||
] | ||
}, | ||
{ | ||
"cell_type": "markdown", | ||
"id": "51dff76e", | ||
"metadata": {}, | ||
"source": [ | ||
"### Load documents using document loader" | ||
] | ||
}, | ||
{ | ||
"cell_type": "code", | ||
"execution_count": 5, | ||
"id": "792dafa8", | ||
"metadata": {}, | ||
"outputs": [], | ||
"source": [ | ||
"SemanticScholarReader = download_loader(\"SemanticScholarReader\")\n", | ||
"loader = SemanticScholarReader()\n", | ||
"# narrow down the search space\n", | ||
"query_space = \"large language models\"\n", | ||
"# increase limit to get more documents\n", | ||
"documents = loader.load_data(query=query_space, limit=10)" | ||
] | ||
}, | ||
{ | ||
"cell_type": "markdown", | ||
"id": "b1d0356b", | ||
"metadata": {}, | ||
"source": [ | ||
"### Generate test set using default configurations" | ||
] | ||
}, | ||
{ | ||
"cell_type": "code", | ||
"execution_count": 8, | ||
"id": "30caed35", | ||
"metadata": {}, | ||
"outputs": [], | ||
"source": [ | ||
"testsetgenerator = TestsetGenerator.from_default()\n", | ||
"test_size = 2 # Number of samples to generate\n", | ||
"testset = testsetgenerator.generate(documents, test_size=test_size)" | ||
] | ||
}, | ||
{ | ||
"cell_type": "code", | ||
"execution_count": 11, | ||
"id": "c8162006", | ||
"metadata": {}, | ||
"outputs": [ | ||
{ | ||
"data": { | ||
"text/html": [ | ||
"<div>\n", | ||
"<style scoped>\n", | ||
" .dataframe tbody tr th:only-of-type {\n", | ||
" vertical-align: middle;\n", | ||
" }\n", | ||
"\n", | ||
" .dataframe tbody tr th {\n", | ||
" vertical-align: top;\n", | ||
" }\n", | ||
"\n", | ||
" .dataframe thead th {\n", | ||
" text-align: right;\n", | ||
" }\n", | ||
"</style>\n", | ||
"<table border=\"1\" class=\"dataframe\">\n", | ||
" <thead>\n", | ||
" <tr style=\"text-align: right;\">\n", | ||
" <th></th>\n", | ||
" <th>question</th>\n", | ||
" <th>context</th>\n", | ||
" <th>answer</th>\n", | ||
" <th>question_type</th>\n", | ||
" <th>episode_done</th>\n", | ||
" </tr>\n", | ||
" </thead>\n", | ||
" <tbody>\n", | ||
" <tr>\n", | ||
" <th>0</th>\n", | ||
" <td>What is the synthesis performance of large lan...</td>\n", | ||
" <td>- Our benchmarks are designed to measure the a...</td>\n", | ||
" <td>The synthesis performance of large language mo...</td>\n", | ||
" <td>simple</td>\n", | ||
" <td>True</td>\n", | ||
" </tr>\n", | ||
" <tr>\n", | ||
" <th>1</th>\n", | ||
" <td>How did ChatGPT fare on the USMLE exams in ter...</td>\n", | ||
" <td>- ChatGPT performed at or near the passing thr...</td>\n", | ||
" <td>ChatGPT performed well on the USMLE exams in t...</td>\n", | ||
" <td>reasoning</td>\n", | ||
" <td>True</td>\n", | ||
" </tr>\n", | ||
" </tbody>\n", | ||
"</table>\n", | ||
"</div>" | ||
], | ||
"text/plain": [ | ||
" question \\\n", | ||
"0 What is the synthesis performance of large lan... \n", | ||
"1 How did ChatGPT fare on the USMLE exams in ter... \n", | ||
"\n", | ||
" context \\\n", | ||
"0 - Our benchmarks are designed to measure the a... \n", | ||
"1 - ChatGPT performed at or near the passing thr... \n", | ||
"\n", | ||
" answer question_type \\\n", | ||
"0 The synthesis performance of large language mo... simple \n", | ||
"1 ChatGPT performed well on the USMLE exams in t... reasoning \n", | ||
"\n", | ||
" episode_done \n", | ||
"0 True \n", | ||
"1 True " | ||
] | ||
}, | ||
"execution_count": 11, | ||
"metadata": {}, | ||
"output_type": "execute_result" | ||
} | ||
], | ||
"source": [ | ||
"testset.to_pandas()" | ||
] | ||
}, | ||
{ | ||
"cell_type": "code", | ||
"execution_count": null, | ||
"id": "34474b1a", | ||
"metadata": {}, | ||
"outputs": [], | ||
"source": [] | ||
} | ||
], | ||
"metadata": { | ||
"kernelspec": { | ||
"display_name": "ragas", | ||
"language": "python", | ||
"name": "ragas" | ||
}, | ||
"language_info": { | ||
"codemirror_mode": { | ||
"name": "ipython", | ||
"version": 3 | ||
}, | ||
"file_extension": ".py", | ||
"mimetype": "text/x-python", | ||
"name": "python", | ||
"nbconvert_exporter": "python", | ||
"pygments_lexer": "ipython3", | ||
"version": "3.10.8" | ||
} | ||
}, | ||
"nbformat": 4, | ||
"nbformat_minor": 5 | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,3 @@ | ||
from ragas.testset.testset_generator import TestsetGenerator | ||
|
||
__all__ = ["TestsetGenerator"] |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,148 @@ | ||
from langchain.prompts import HumanMessagePromptTemplate | ||
|
||
SEED_QUESTION = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
Your task is to formulate a question from given context satisfying the rules given below: | ||
1.The question should make sense to humans even when read without the given context. | ||
2.The question should be fully answered from the given context. | ||
3.The question should be framed from a part of context that contains important information. It can also be from tables,code,etc. | ||
4.The answer to the question should not contain any links. | ||
5.The question should be of moderate difficulty. | ||
6.The question must be reasonable and must be understood and responded by humans. | ||
7.Do no use phrases like 'provided context',etc in the question | ||
8.Avoid framing question using word "and" that can be decomposed into more than one question. | ||
9.The question should not contain more than 10 words, make of use of abbreviation wherever possible. | ||
context:{context} | ||
""" # noqa: E501 | ||
) | ||
|
||
|
||
REASONING_QUESTION = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
You are a prompt rewriter. You will be provided with a question and a long context.Your task to is to complicate the given question to improve the difficulty of answering. | ||
You should do complicate the question by rewriting question into a multi-hop reasoning question based on the provided context. The question should require the reader to make multiple logical connections or inferences using the information available in given context. | ||
Here are some strategies to create multi-hop questions: | ||
- Bridge related entities: Identify information that relates specific entities and frame question that can be answered only by analysing information of both entities. | ||
- Use Pronouns: identify (he, she, it, they) that refer to same entity or concepts in the context, and ask questions that would require the reader to figure out what pronouns refer to. | ||
- Refer to Specific Details: Mention specific details or facts from different parts of the context including tables, code, etc and ask how they are related. | ||
- Pose Hypothetical Scenarios: Present a hypothetical situation or scenario that requires combining different elements from the context to arrive at an answer. | ||
Rules to follow when rewriting question: | ||
1. Ensure that the rewritten question can be answered entirely from the information present in the contexts. | ||
2. Do not frame questions that contains more than 15 words. Use abbreviation wherever possible. | ||
3. Make sure the question is clear and unambiguous. | ||
4. phrases like 'based on the provided context','according to the context',etc are not allowed to appear in the question. | ||
question: {question} | ||
CONTEXTS: | ||
{context} | ||
Multi-hop Reasoning Question: | ||
""" # noqa: E501 | ||
) | ||
|
||
MULTICONTEXT_QUESTION = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
You are a prompt rewriter. You will be provided with a question and two set of contexts namely context1 and context2. | ||
Your task is to complicate the given question in a way that answering it requires information derived from both context1 and context2. | ||
Follow the rules given below while rewriting the question. | ||
1. The rewritten question should not be very long. Use abbreviation wherever possible. | ||
2. The rewritten question must be reasonable and must be understood and responded by humans. | ||
3. The rewritten question must be fully answerable from information present in context1 and context2. | ||
4. Read and understand both contexts and rewrite the question so that answering requires insight from both context1 and context2. | ||
5. phrases like 'based on the provided context','according to the context?',etc are not allowed to appear in the question. | ||
question:\n{question} | ||
context1:\n{context1} | ||
context2:\n{context2} | ||
""" # noqa: E501 | ||
) | ||
|
||
|
||
CONDITIONAL_QUESTION = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
Rewrite the provided question to increase its complexity by introducing a conditional element. | ||
The goal is to make the question more intricate by incorporating a scenario or condition that affects the context of the question. | ||
Follow the rules given below while rewriting the question. | ||
1. The rewritten question should not be longer than 25 words. Use abbreviation wherever possible. | ||
2. The rewritten question must be reasonable and must be understood and responded by humans. | ||
3. The rewritten question must be fully answerable from information present context. | ||
4. phrases like 'provided context','according to the context?',etc are not allowed to appear in the question. | ||
for example, | ||
question: What are the general principles for designing prompts in LLMs? | ||
Rewritten Question:how to apply prompt designing principles to improve LLMs performance in reasoning tasks | ||
question:{question} | ||
context:\n{context} | ||
Rewritten Question | ||
""" # noqa: E501 | ||
) | ||
|
||
|
||
COMPRESS_QUESTION = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
Rewrite the following question to make it more indirect and shorter while retaining the essence of the original question. The goal is to create a question that conveys the same meaning but in a less direct manner. | ||
The rewritten question should shorter so use abbreviation wherever possible. | ||
Original Question: | ||
{question} | ||
Indirectly Rewritten Question: | ||
""" # noqa: E501 | ||
) | ||
|
||
|
||
CONVERSATION_QUESTION = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
Reformat the provided question into two separate questions as if it were to be part of a conversation. Each question should focus on a specific aspect or subtopic related to the original question. | ||
question: What are the advantages and disadvantages of remote work? | ||
Reformatted Questions for Conversation: What are the benefits of remote work?\nOn the flip side, what challenges are encountered when working remotely? | ||
question:{question} | ||
Reformatted Questions for Conversation: | ||
""" # noqa: E501 | ||
) | ||
|
||
SCORE_CONTEXT = HumanMessagePromptTemplate.from_template( | ||
"""Evaluate the provided context and assign a numerical score between 0 and 10 based on the following criteria: | ||
1. Award a high score to context that thoroughly delves into and explains concepts. | ||
2. Assign a lower score to context that contains excessive references, acknowledgments, external links, personal information, or other non-essential elements. | ||
Output the score only. | ||
Context: | ||
{context} | ||
Score: | ||
""" # noqa: E501 | ||
) | ||
|
||
FILTER_QUESTION = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
Determine if the given question can be clearly understood even when presented without any additional context? Reason before arriving at the answer. | ||
question: What is the keyword that best describes the paper's focus in natural language understanding tasks? | ||
answer: The specific paper being referred to is not mentioned in the question. Hence, No. | ||
question:{question} | ||
answer: | ||
""" # noqa: E501 | ||
) | ||
|
||
|
||
ANSWER_FORMULATE = HumanMessagePromptTemplate.from_template( | ||
"""\ | ||
Answer the question using the information from the given context. | ||
question:{question} | ||
context:{context} | ||
answer: | ||
""" # noqa: E501 | ||
) | ||
|
||
CONTEXT_FORMULATE = HumanMessagePromptTemplate.from_template( | ||
"""Please extract relevant sentences from the provided context that can potentially help answer the following question. While extracting candidate sentences you're not allowed to make any changes to sentences from given context. | ||
question:{question} | ||
context:\n{context} | ||
candidate sentences:\n | ||
""" # noqa: E501 | ||
) |
Oops, something went wrong.