From d31b9cff50eb5b3c83bbea11a19bc900487e7b07 Mon Sep 17 00:00:00 2001 From: Cherrytest Date: Wed, 18 Sep 2024 15:58:28 +0000 Subject: [PATCH] Update README.md --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index f1b5cb6..4dc8aed 100644 --- a/README.md +++ b/README.md @@ -42,14 +42,14 @@ Check out our [llama.cpp documentation](https://qwen.readthedocs.io/en/latest/ru We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp. In the following demonstration, we assume that you are running commands under the repository `llama.cpp`. -Since cloning the entire repo may be inefficient, you can manually download the GGUF file that you need or use `huggingface-cli`: +Since cloning the entire repo may be inefficient, you can manually download the GGUF file that you need or use `modelscope`: 1. Install ```shell - pip install -U huggingface_hub + pip install -U modelscope ``` 2. Download: ```shell - huggingface-cli download Qwen/Qwen2.5-72B-Instruct-GGUF --include "qwen2.5-72b-instruct-q5_k_m*.gguf" --local-dir . --local-dir-use-symlinks False + modelscope download --model=qwen/Qwen2.5-72B-Instruct-GGUF --include "qwen2.5-72b-instruct-q5_k_m*.gguf" --local_dir . ``` For large files, we split them into multiple segments due to the limitation of file upload. They share a prefix, with a suffix indicating its index. For examples, `qwen2.5-72b-instruct-q5_k_m-00001-of-00014.gguf` to `qwen2.5-72b-instruct-q5_k_m-00014-of-00014.gguf`. The above command will download all of them. 3. (Optional) Merge: