Skip to content

Commit

Permalink
Update inference examples to version 0.8.1
Browse files Browse the repository at this point in the history
  • Loading branch information
krasserm committed Feb 24, 2023
1 parent a519f65 commit 0479629
Show file tree
Hide file tree
Showing 4 changed files with 189 additions and 186 deletions.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@ See [Docker image](docs/docker-image.md) for details.
- [Model construction](docs/model-construction.md)
- [Pretrained models](docs/pretrained-models.md)
- [Training examples](docs/training-examples.md)
- [Inference examples](examples/inference.ipynb) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/krasserm/perceiver-io/blob/0.8.0/examples/inference.ipynb)
- [Inference examples](examples/inference.ipynb) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/krasserm/perceiver-io/blob/0.8.1/examples/inference.ipynb)
- [Building blocks](docs/building-blocks.md)

## Articles
Expand Down
2 changes: 1 addition & 1 deletion docs/getting-started.md
Original file line number Diff line number Diff line change
Expand Up @@ -152,7 +152,7 @@ torch.save(model.state_dict(), "/path/to/model.pt")

For generating text from a prompt via top-k sampling, `CausalLanguageModel` provides a `generate()` method. The following
example first loads a trained model from a checkpoint and then generates text from a short sample prompt. An interactive
demo is also available in the [Colab notebook](https://colab.research.google.com/github/krasserm/perceiver-io/blob/0.8.0/examples/inference.ipynb).
demo is also available in the [Colab notebook](https://colab.research.google.com/github/krasserm/perceiver-io/blob/0.8.1/examples/inference.ipynb).

```python
from perceiver.data.text import TextPreprocessor
Expand Down
Loading

0 comments on commit 0479629

Please sign in to comment.