Skip to content

Commit 5599749

Browse files
committed
docs: Refine and fix nits in README
1 parent 3434c92 commit 5599749

File tree

7 files changed

+46
-43
lines changed

7 files changed

+46
-43
lines changed

README.md

Lines changed: 16 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ curl -LsSf https://github.com/llamastack/llama-stack/raw/main/scripts/install.sh
2020

2121
### Overview
2222

23-
Llama Stack standardizes the core building blocks that simplify AI application development. It codifies best practices across the Llama ecosystem. More specifically, it provides
23+
Llama Stack defines and standardizes the core building blocks that simplify AI application development. It provides a unified set of APIs with implementations from leading service providers. More specifically, it provides:
2424

2525
- **Unified API layer** for Inference, RAG, Agents, Tools, Safety, Evals.
2626
- **Plugin architecture** to support the rich ecosystem of different API implementations in various environments, including local development, on-premises, cloud, and mobile.
@@ -37,18 +37,19 @@ Llama Stack standardizes the core building blocks that simplify AI application d
3737
/>
3838
</div>
3939

40-
### Llama Stack Benefits
41-
- **Flexible Options**: Developers can choose their preferred infrastructure without changing APIs and enjoy flexible deployment choices.
40+
#### Llama Stack Benefits
41+
42+
- **Flexibility**: Developers can choose their preferred infrastructure without changing APIs and enjoy flexible deployment choices.
4243
- **Consistent Experience**: With its unified APIs, Llama Stack makes it easier to build, test, and deploy AI applications with consistent application behavior.
43-
- **Robust Ecosystem**: Llama Stack is already integrated with distribution partners (cloud providers, hardware vendors, and AI-focused companies) that offer tailored infrastructure, software, and services for deploying Llama models.
44+
- **Robust Ecosystem**: Llama Stack is integrated with distribution partners (cloud providers, hardware vendors, and AI-focused companies) that offer tailored infrastructure, software, and services for deploying Llama models.
4445

45-
By reducing friction and complexity, Llama Stack empowers developers to focus on what they do best: building transformative generative AI applications.
46+
For more information, see the [Benefits of Llama Stack](https://llamastack.github.io/docs/v0.3.2/concepts/architecture#benefits-of-llama-stack) documentation.
4647

4748
### API Providers
4849
Here is a list of the various API providers and available distributions that can help developers get started easily with Llama Stack.
4950
Please checkout for [full list](https://llamastack.github.io/docs/providers)
5051

51-
| API Provider Builder | Environments | Agents | Inference | VectorIO | Safety | Post Training | Eval | DatasetIO |
52+
| API Provider | Environments | Agents | Inference | VectorIO | Safety | Post Training | Eval | DatasetIO |
5253
|:--------------------:|:------------:|:------:|:---------:|:--------:|:------:|:-------------:|:----:|:--------:|
5354
| Meta Reference | Single Node ||||||||
5455
| SambaNova | Hosted | || || | | |
@@ -81,14 +82,17 @@ Please checkout for [full list](https://llamastack.github.io/docs/providers)
8182
8283
### Distributions
8384

84-
A Llama Stack Distribution (or "distro") is a pre-configured bundle of provider implementations for each API component. Distributions make it easy to get started with a specific deployment scenario - you can begin with a local development setup (eg. ollama) and seamlessly transition to production (eg. Fireworks) without changing your application code.
85+
A Llama Stack Distribution (or "distro") is a pre-configured bundle of provider implementations for each API component. Distributions make it easy to get started with a specific deployment scenario. For example, you can begin with a local setup of Ollama and seamlessly transition to production, with fireworks, without changing your application code.
8586
Here are some of the distributions we support:
8687

8788
| **Distribution** | **Llama Stack Docker** | Start This Distribution |
8889
|:---------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------------------------------------------------:|:------------------------------------------------------------------------------------------------------------------------:|
89-
| Starter Distribution | [llamastack/distribution-starter](https://hub.docker.com/repository/docker/llamastack/distribution-starter/general) | [Guide](https://llamastack.github.io/latest/distributions/self_hosted_distro/starter.html) |
90-
| Meta Reference | [llamastack/distribution-meta-reference-gpu](https://hub.docker.com/repository/docker/llamastack/distribution-meta-reference-gpu/general) | [Guide](https://llamastack.github.io/latest/distributions/self_hosted_distro/meta-reference-gpu.html) |
91-
| PostgreSQL | [llamastack/distribution-postgres-demo](https://hub.docker.com/repository/docker/llamastack/distribution-postgres-demo/general) | |
90+
| Starter Distribution | [llamastack/distribution-starter](https://hub.docker.com/repository/docker/llamastack/distribution-starter/general) | [Guide](https://llamastack.github.io/latest/distributions/self_hosted_distro/starter.html) |
91+
| Meta Reference | [llamastack/distribution-meta-reference-gpu](https://hub.docker.com/repository/docker/llamastack/distribution-meta-reference-gpu/general) | [Guide](https://llamastack.github.io/latest/distributions/self_hosted_distro/meta-reference-gpu.html) |
92+
| PostgreSQL | [llamastack/distribution-postgres-demo](https://hub.docker.com/repository/docker/llamastack/distribution-postgres-demo/general) | |
93+
| iOS/Android SDK | <add> |
94+
95+
For full documentation on the Llama Stack distributions see the [Distributions Overview](https://llamastack.github.io/docs/latest/distributions) page.
9296

9397
### Documentation
9498

@@ -107,18 +111,17 @@ Please checkout our [Documentation](https://llamastack.github.io/latest/index.ht
107111

108112
### Llama Stack Client SDKs
109113

114+
Check out our client SDKs for connecting to a Llama Stack server in your preferred language.
115+
110116
| **Language** | **Client SDK** | **Package** |
111117
| :----: | :----: | :----: |
112118
| Python | [llama-stack-client-python](https://github.com/meta-llama/llama-stack-client-python) | [![PyPI version](https://img.shields.io/pypi/v/llama_stack_client.svg)](https://pypi.org/project/llama_stack_client/)
113119
| Swift | [llama-stack-client-swift](https://github.com/meta-llama/llama-stack-client-swift) | [![Swift Package Index](https://img.shields.io/endpoint?url=https%3A%2F%2Fswiftpackageindex.com%2Fapi%2Fpackages%2Fmeta-llama%2Fllama-stack-client-swift%2Fbadge%3Ftype%3Dswift-versions)](https://swiftpackageindex.com/meta-llama/llama-stack-client-swift)
114120
| Typescript | [llama-stack-client-typescript](https://github.com/meta-llama/llama-stack-client-typescript) | [![NPM version](https://img.shields.io/npm/v/llama-stack-client.svg)](https://npmjs.org/package/llama-stack-client)
115121
| Kotlin | [llama-stack-client-kotlin](https://github.com/meta-llama/llama-stack-client-kotlin) | [![Maven version](https://img.shields.io/maven-central/v/com.llama.llamastack/llama-stack-client-kotlin)](https://central.sonatype.com/artifact/com.llama.llamastack/llama-stack-client-kotlin)
116122

117-
Check out our client SDKs for connecting to a Llama Stack server in your preferred language, you can choose from [python](https://github.com/meta-llama/llama-stack-client-python), [typescript](https://github.com/meta-llama/llama-stack-client-typescript), [swift](https://github.com/meta-llama/llama-stack-client-swift), and [kotlin](https://github.com/meta-llama/llama-stack-client-kotlin) programming languages to quickly build your applications.
118-
119123
You can find more example scripts with client SDKs to talk with the Llama Stack server in our [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repo.
120124

121-
122125
## 🌟 GitHub Star History
123126
## Star History
124127

docs/docs/providers/agents/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
description: |
33
Agents
44
5-
APIs for creating and interacting with agentic systems.
5+
APIs for creating and interacting with agentic systems.
66
sidebar_label: Agents
77
title: Agents
88
---
@@ -13,6 +13,6 @@ title: Agents
1313

1414
Agents
1515

16-
APIs for creating and interacting with agentic systems.
16+
APIs for creating and interacting with agentic systems.
1717

1818
This section contains documentation for all available providers for the **agents** API.
Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,15 @@
11
---
22
description: |
33
The Batches API enables efficient processing of multiple requests in a single operation,
4-
particularly useful for processing large datasets, batch evaluation workflows, and
5-
cost-effective inference at scale.
4+
particularly useful for processing large datasets, batch evaluation workflows, and
5+
cost-effective inference at scale.
66
7-
The API is designed to allow use of openai client libraries for seamless integration.
7+
The API is designed to allow use of openai client libraries for seamless integration.
88
9-
This API provides the following extensions:
10-
- idempotent batch creation
9+
This API provides the following extensions:
10+
- idempotent batch creation
1111
12-
Note: This API is currently under active development and may undergo changes.
12+
Note: This API is currently under active development and may undergo changes.
1313
sidebar_label: Batches
1414
title: Batches
1515
---
@@ -19,14 +19,14 @@ title: Batches
1919
## Overview
2020

2121
The Batches API enables efficient processing of multiple requests in a single operation,
22-
particularly useful for processing large datasets, batch evaluation workflows, and
23-
cost-effective inference at scale.
22+
particularly useful for processing large datasets, batch evaluation workflows, and
23+
cost-effective inference at scale.
2424

25-
The API is designed to allow use of openai client libraries for seamless integration.
25+
The API is designed to allow use of openai client libraries for seamless integration.
2626

27-
This API provides the following extensions:
28-
- idempotent batch creation
27+
This API provides the following extensions:
28+
- idempotent batch creation
2929

30-
Note: This API is currently under active development and may undergo changes.
30+
Note: This API is currently under active development and may undergo changes.
3131

3232
This section contains documentation for all available providers for the **batches** API.

docs/docs/providers/eval/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
description: |
33
Evaluations
44
5-
Llama Stack Evaluation API for running evaluations on model and agent candidates.
5+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
66
sidebar_label: Eval
77
title: Eval
88
---
@@ -13,6 +13,6 @@ title: Eval
1313

1414
Evaluations
1515

16-
Llama Stack Evaluation API for running evaluations on model and agent candidates.
16+
Llama Stack Evaluation API for running evaluations on model and agent candidates.
1717

1818
This section contains documentation for all available providers for the **eval** API.

docs/docs/providers/files/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
description: |
33
Files
44
5-
This API is used to upload documents that can be used with other Llama Stack APIs.
5+
This API is used to upload documents that can be used with other Llama Stack APIs.
66
sidebar_label: Files
77
title: Files
88
---
@@ -13,6 +13,6 @@ title: Files
1313

1414
Files
1515

16-
This API is used to upload documents that can be used with other Llama Stack APIs.
16+
This API is used to upload documents that can be used with other Llama Stack APIs.
1717

1818
This section contains documentation for all available providers for the **files** API.

docs/docs/providers/inference/index.mdx

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -2,12 +2,12 @@
22
description: |
33
Inference
44
5-
Llama Stack Inference API for generating completions, chat completions, and embeddings.
5+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
66
7-
This API provides the raw interface to the underlying models. Three kinds of models are supported:
8-
- LLM models: these models generate "raw" and "chat" (conversational) completions.
9-
- Embedding models: these models generate embeddings to be used for semantic search.
10-
- Rerank models: these models reorder the documents based on their relevance to a query.
7+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
8+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
9+
- Embedding models: these models generate embeddings to be used for semantic search.
10+
- Rerank models: these models reorder the documents based on their relevance to a query.
1111
sidebar_label: Inference
1212
title: Inference
1313
---
@@ -18,11 +18,11 @@ title: Inference
1818

1919
Inference
2020

21-
Llama Stack Inference API for generating completions, chat completions, and embeddings.
21+
Llama Stack Inference API for generating completions, chat completions, and embeddings.
2222

23-
This API provides the raw interface to the underlying models. Three kinds of models are supported:
24-
- LLM models: these models generate "raw" and "chat" (conversational) completions.
25-
- Embedding models: these models generate embeddings to be used for semantic search.
26-
- Rerank models: these models reorder the documents based on their relevance to a query.
23+
This API provides the raw interface to the underlying models. Three kinds of models are supported:
24+
- LLM models: these models generate "raw" and "chat" (conversational) completions.
25+
- Embedding models: these models generate embeddings to be used for semantic search.
26+
- Rerank models: these models reorder the documents based on their relevance to a query.
2727

2828
This section contains documentation for all available providers for the **inference** API.

docs/docs/providers/safety/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
description: |
33
Safety
44
5-
OpenAI-compatible Moderations API.
5+
OpenAI-compatible Moderations API.
66
sidebar_label: Safety
77
title: Safety
88
---
@@ -13,6 +13,6 @@ title: Safety
1313

1414
Safety
1515

16-
OpenAI-compatible Moderations API.
16+
OpenAI-compatible Moderations API.
1717

1818
This section contains documentation for all available providers for the **safety** API.

0 commit comments

Comments
 (0)