diff --git a/docs/extra/components/choose_evaluator_llm.md b/docs/extra/components/choose_evaluator_llm.md index 73fa1becd..14961335f 100644 --- a/docs/extra/components/choose_evaluator_llm.md +++ b/docs/extra/components/choose_evaluator_llm.md @@ -31,7 +31,7 @@ pip install langchain-aws ``` - then you have to set your AWS credentials and configurations + Then you have to set your AWS credentials and configurations ```python config = { @@ -43,7 +43,7 @@ } ``` - define you LLMs and wrap them in `LangchainLLMWrapper` so that it can be used with ragas. + Define your LLMs and wrap them in `LangchainLLMWrapper` so that it can be used with ragas. ```python from langchain_aws import ChatBedrockConverse diff --git a/docs/howtos/integrations/_llamaindex.md b/docs/howtos/integrations/_llamaindex.md index 57898fec8..abaaa79de 100644 --- a/docs/howtos/integrations/_llamaindex.md +++ b/docs/howtos/integrations/_llamaindex.md @@ -11,11 +11,13 @@ You will need an testset to evaluate your `QueryEngine` against. You can either Let's see how that works with Llamaindex # load the documents +```python from llama_index.core import SimpleDirectoryReader documents = SimpleDirectoryReader("./nyc_wikipedia").load_data() +``` -Now lets init the `TestsetGenerator` object with the corresponding generator and critic llms +Now lets init the `TestsetGenerator` object with the corresponding generator and critic llms ```python @@ -171,7 +173,7 @@ Now that we have a `QueryEngine` for the `VectorStoreIndex` we can use the llama In order to run an evaluation with Ragas and LlamaIndex you need 3 things 1. LlamaIndex `QueryEngine`: what we will be evaluating -2. Metrics: Ragas defines a set of metrics that can measure different aspects of the `QueryEngine`. The available metrics and their meaning can be found [here](https://github.com/explodinggradients/ragas/blob/main/docs/metrics.md) +2. Metrics: Ragas defines a set of metrics that can measure different aspects of the `QueryEngine`. The available metrics and their meaning can be found [here](https://docs.ragas.io/en/latest/concepts/metrics/available_metrics/) 3. Questions: A list of questions that ragas will test the `QueryEngine` against. first lets generate the questions. Ideally you should use that you see in production so that the distribution of question with which we evaluate matches the distribution of questions seen in production. This ensures that the scores reflect the performance seen in production but to start off we'll be using a few example question. diff --git a/docs/howtos/integrations/llamaindex.ipynb b/docs/howtos/integrations/llamaindex.ipynb index 9ff65f123..27113a1f2 100644 --- a/docs/howtos/integrations/llamaindex.ipynb +++ b/docs/howtos/integrations/llamaindex.ipynb @@ -29,7 +29,16 @@ "id": "096e5af0", "metadata": {}, "source": [ - "# load the documents\n", + "# load the documents" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "396085d5", + "metadata": {}, + "outputs": [], + "source": [ "from llama_index.core import SimpleDirectoryReader\n", "\n", "documents = SimpleDirectoryReader(\"./nyc_wikipedia\").load_data()" @@ -298,7 +307,7 @@ "In order to run an evaluation with Ragas and LlamaIndex you need 3 things\n", "\n", "1. LlamaIndex `QueryEngine`: what we will be evaluating\n", - "2. Metrics: Ragas defines a set of metrics that can measure different aspects of the `QueryEngine`. The available metrics and their meaning can be found [here](https://github.com/explodinggradients/ragas/blob/main/docs/metrics.md)\n", + "2. Metrics: Ragas defines a set of metrics that can measure different aspects of the `QueryEngine`. The available metrics and their meaning can be found [here](https://docs.ragas.io/en/latest/concepts/metrics/available_metrics/)\n", "3. Questions: A list of questions that ragas will test the `QueryEngine` against. " ] },