Skip to content
This repository has been archived by the owner on Aug 30, 2024. It is now read-only.

Commit

Permalink
Update README.md for archive
Browse files Browse the repository at this point in the history
Archive per Alex H. request
  • Loading branch information
jkmarz authored Aug 30, 2024
1 parent 97c8190 commit bfd5d3c
Showing 1 changed file with 10 additions and 0 deletions.
10 changes: 10 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,13 @@
# PROJECT NOT UNDER ACTIVE MANAGEMENT
This project will no longer be maintained by Intel.

Intel has ceased development and contributions including, but not limited to, maintenance, bug fixes, new releases, or updates, to this project.

Intel no longer accepts patches to this project.

## Please refer to https://github.com/intel/intel-extension-for-pytorch as an alternative


# Neural Speed

Neural Speed is an innovative library designed to support the efficient inference of large language models (LLMs) on Intel platforms through the state-of-the-art (SOTA) low-bit quantization powered by [Intel Neural Compressor](https://github.com/intel/neural-compressor). The work is inspired by [llama.cpp](https://github.com/ggerganov/llama.cpp) and further optimized for Intel platforms with our innovations in [NeurIPS' 2023](https://arxiv.org/abs/2311.00502)
Expand Down

0 comments on commit bfd5d3c

Please sign in to comment.