Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -48,7 +48,7 @@ Select 'Stable + Linux + Pip + Python + ROCm' to get the specific pip installati
|
|
| 48 |
|
| 49 |
An example command line (note the versioning of the whl file):
|
| 50 |
|
| 51 |
-
> `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/
|
| 52 |
|
| 53 |
### TensorFlow
|
| 54 |
|
|
@@ -99,7 +99,7 @@ with Hugging Face models are available on the [Optimum page](https://huggingface
|
|
| 99 |
# Serving a model with TGI
|
| 100 |
|
| 101 |
Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
|
| 102 |
-
TGI is already usable in production on AMD Instinct™ GPUs through the docker image `ghcr.io/huggingface/text-generation-inference:
|
| 103 |
Make sure to refer to the [documentation](https://huggingface.co/docs/text-generation-inference/supported_models#supported-hardware)
|
| 104 |
concerning the support and any limitations.
|
| 105 |
|
|
@@ -110,6 +110,7 @@ across normal and distributed settings, with various supported optimizations and
|
|
| 110 |
|
| 111 |
# Useful Links and Blogs
|
| 112 |
|
|
|
|
| 113 |
- Detailed Llama-2 results show casing the [Optimum benchmark on AMD Instinct MI250](https://huggingface.co/blog/huggingface-and-optimum-amd)
|
| 114 |
- Check out our blog titled [Run a Chatgpt-like Chatbot on a Single GPU with ROCm](https://huggingface.co/blog/chatbot-amd-gpu)
|
| 115 |
- Complete ROCm [Documentation](https://rocm.docs.amd.com/en/latest/) for installation and usage
|
|
|
|
| 48 |
|
| 49 |
An example command line (note the versioning of the whl file):
|
| 50 |
|
| 51 |
+
> `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.2`
|
| 52 |
|
| 53 |
### TensorFlow
|
| 54 |
|
|
|
|
| 99 |
# Serving a model with TGI
|
| 100 |
|
| 101 |
Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
|
| 102 |
+
TGI is already usable in production on AMD Instinct™ GPUs through the docker image `ghcr.io/huggingface/text-generation-inference:latest-rocm`.
|
| 103 |
Make sure to refer to the [documentation](https://huggingface.co/docs/text-generation-inference/supported_models#supported-hardware)
|
| 104 |
concerning the support and any limitations.
|
| 105 |
|
|
|
|
| 110 |
|
| 111 |
# Useful Links and Blogs
|
| 112 |
|
| 113 |
+
- Detailed Llama-3 results [Run TGI on AMD Instinct MI300X](https://huggingface.co/blog/huggingface-amd-mi300)
|
| 114 |
- Detailed Llama-2 results show casing the [Optimum benchmark on AMD Instinct MI250](https://huggingface.co/blog/huggingface-and-optimum-amd)
|
| 115 |
- Check out our blog titled [Run a Chatgpt-like Chatbot on a Single GPU with ROCm](https://huggingface.co/blog/chatbot-amd-gpu)
|
| 116 |
- Complete ROCm [Documentation](https://rocm.docs.amd.com/en/latest/) for installation and usage
|