Skip to content

Commit de0935b

Browse files
kwinkunksanistark
andauthored
Fix typos and clean style (#2042)
Fixed missing-word typo and regularized the style a bit. Co-authored-by: Ani <[email protected]>
1 parent 7f447af commit de0935b

File tree

1 file changed

+5
-7
lines changed

1 file changed

+5
-7
lines changed

docs/getstarted/evals.md

Lines changed: 5 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ In this guide, you will evaluate a **text summarization pipeline**. The goal is
1111

1212
### Evaluating using a Non-LLM Metric
1313

14-
Here is a simple example that uses `BleuScore` score to score summary
14+
Here is a simple example that uses `BleuScore` to score a summary:
1515

1616
```python
1717
from ragas import SingleTurnSample
@@ -40,9 +40,9 @@ Here we used:
4040

4141
As you may observe, this approach has two key limitations:
4242

43-
- **Time-Consuming Preparation:** Evaluating the application requires preparing the expected output (`reference`) for each input, which can be both time-consuming and challenging.
43+
- **Time-consuming preparation:** Evaluating the application requires preparing the expected output (`reference`) for each input, which can be both time-consuming and challenging.
4444

45-
- **Inaccurate Scoring:** Even though the `response` and `reference` are similar, the output score was low. This is a known limitation of non-LLM metrics like `BleuScore`.
45+
- **Inaccurate scoring:** Even though the `response` and `reference` are similar, the output score was low. This is a known limitation of non-LLM metrics like `BleuScore`.
4646

4747

4848
!!! info
@@ -51,7 +51,7 @@ As you may observe, this approach has two key limitations:
5151
To address these issues, let's try an LLM-based metric.
5252

5353

54-
### Evaluating using a LLM based Metric
54+
### Evaluating using a LLM-based Metric
5555

5656

5757
**Choose your LLM**
@@ -61,10 +61,8 @@ choose_evaluator_llm.md
6161

6262
**Evaluation**
6363

64-
6564
Here we will use [AspectCritic](../concepts/metrics/available_metrics/aspect_critic.md), which is an LLM based metric that outputs pass/fail given the evaluation criteria.
6665

67-
6866
```python
6967
from ragas import SingleTurnSample
7068
from ragas.metrics import AspectCritic
@@ -88,7 +86,7 @@ Output
8886
Success! Here 1 means pass and 0 means fail
8987

9088
!!! info
91-
There are many other types of metrics that are available in ragas (with and without `reference`), and you may also create your own metrics if none of those fits your case. To explore this more checkout [more on metrics](../concepts/metrics/index.md).
89+
There are many other types of metrics that are available in `ragas` (with and without `reference`), and you may also create your own metrics if none of those fits your case. To explore this more checkout [more on metrics](../concepts/metrics/index.md).
9290

9391
### Evaluating on a Dataset
9492

0 commit comments

Comments
 (0)