Skip to content

Commit

Permalink
fix links
Browse files Browse the repository at this point in the history
  • Loading branch information
lbeurerkellner committed Jan 9, 2024
1 parent 490a813 commit a055952
Show file tree
Hide file tree
Showing 3 changed files with 7 additions and 7 deletions.
6 changes: 3 additions & 3 deletions docs/docs/language/reference.md
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ LMQL's [current reference implementation](#reference-implementation) is written

* The `lmql playground` offers an interactive interface for running, debugging and experimenting with LMQL programs. It is the recommended way to get started with LMQL. An online version of the playground is available at [lmql.ai/playground](https://lmql.ai/playground).

* LMQL is available as a Python library, with the `lmql.query` function offering a simple way to use LMQL directly from within Python. For more information, please refer to the [Python API](#python-api) section.
* LMQL is available as a Python library, with the `lmql.query` function offering a simple way to use LMQL directly from within Python. For more information, please refer to general [Documentation](overview.md).

* The `lmql run` CLI tool can used to run LMQL programs directly from the command line.

Expand All @@ -44,9 +44,9 @@ For more information and example-based discussion on how to use LMQL, please ref

The LMQL language comprises two syntax variants:

* The modern, more minimalistic [standard syntax](#modern-syntax) that relies on a very small set of language constructs in an otherwise standard Python environment. This syntax is the main focus for the continued development of LMQL.
* The modern, more minimalistic [standard syntax](#standard-syntax) that relies on a very small set of language constructs in an otherwise standard Python environment. This syntax is the main focus for the continued development of LMQL.

* A *legacy* [standalone syntax](#legacy-syntax) that is more static in nature but relevant for standalone LMQL use-cases.
* A *legacy* [standalone syntax](#standalone-syntax) that is more static in nature but relevant for standalone LMQL use-cases.

Both syntax variants are compatible and can be used interchangeably.

Expand Down
6 changes: 3 additions & 3 deletions docs/docs/lib/generations.md
Original file line number Diff line number Diff line change
Expand Up @@ -74,7 +74,7 @@ Generates a text completion based on a given prompt. Returns the full prompt + c

**Return Value** The function returns a string or a list of strings, depending on the decoder in use (`decoder=argmax` yields a single sequence, `decoder="sample", n=2` yields two sequences, etc.).

**Asynchronous** The function is asynchronous and should be used with [`asyncio`](https://docs.python.org/3/library/asyncio.html) and with `await`. When run in parallel, multiple generations will be batched and parallelized across multiple calls to the same model. For synchronous use, you can rely on [`LLM.generate_sync`](#llm-generate_sync), note however, that in this case, the function will block the current thread until generation is complete.
**Asynchronous** The function is asynchronous and should be used with [`asyncio`](https://docs.python.org/3/library/asyncio.html) and with `await`. When run in parallel, multiple generations will be batched and parallelized across multiple calls to the same model. For synchronous use, you can rely on [`LLM.generate_sync`](#llm-generate-sync), note however, that in this case, the function will block the current thread until generation is complete.

### `LLM.generate_sync(...)`

Expand Down Expand Up @@ -134,7 +134,7 @@ async def lmql.generate(
`lmql.generate` generates text completions based on a given prompt and behaves just like [`LLM.generate`](#llm-generate),
with the provided `model` instance or model name.

If no `model` is provided, the default model is used. See [`lmql.set_default_model`](#lmql-set_default_model) for more information.
If no `model` is provided, the default model is used. See [`lmql.set_default_model`](#lmql-set-default-model) for more information.

### `lmql.generate_sync(...)`

Expand All @@ -154,7 +154,7 @@ async def score(
`lmql.score` scores different continuation `values` for a given `prompt` and behaves just like [`LLM.score`](#llm-score),
with the provided `model` instance or model name.

If no `model` is provided, the default model is used. See [`lmql.set_default_model`](#lmql-set_default_model) for more information.
If no `model` is provided, the default model is used. See [`lmql.set_default_model`](#lmql-set-default-model) for more information.

### `lmql.score_sync(...)`

Expand Down
2 changes: 1 addition & 1 deletion docs/docs/models/hf.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ LMQL relies on a two-process architecture: The inference process (long-running)

This architecture is advantageous for locally-hosted models, as the model loading time can be quite long or the required GPU hardware might not even be available on the client machine.

This chapter first discusses how to use of the two-process inference API, and then presents an alternative on how to leverage [In-Process Model Loading](#in-process-model-loading), which avoids the need for a separate server process within the same architecture.
This chapter first discusses how to use of the two-process inference API, and then presents an alternative on how to leverage [In-Process Model Loading](#in-process-models), which avoids the need for a separate server process within the same architecture.

![Inference Architecture](./inference.svg)

Expand Down

0 comments on commit a055952

Please sign in to comment.