From d09ac69550452ac2cc87a46ff0d076109738dfa4 Mon Sep 17 00:00:00 2001 From: ai-modelscope Date: Tue, 29 Apr 2025 03:40:49 +0800 Subject: [PATCH] Update README.md --- README.md | 16 ++++++++++++++-- added_tokens.json | 28 ---------------------------- special_tokens_map.json | 31 ------------------------------- 3 files changed, 14 insertions(+), 61 deletions(-) delete mode 100644 added_tokens.json delete mode 100644 special_tokens_map.json diff --git a/README.md b/README.md index 61e8e40..296e8ec 100644 --- a/README.md +++ b/README.md @@ -1,4 +1,16 @@ +--- +library_name: transformers +license: apache-2.0 +license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE +pipeline_tag: text-generation +base_model: +- Qwen/Qwen3-0.6B-Base +--- + # Qwen3-0.6B + + Chat + ## Qwen3 Highlights @@ -95,7 +107,6 @@ For deployment, you can use `sglang>=0.4.6.post1` or `vllm>=0.8.4` or to create For local use, applications such as llama.cpp, Ollama, LMStudio, and MLX-LM have also supported Qwen3. - ## Switching Between Thinking and Non-Thinking Mode > [!TIP] @@ -120,6 +131,7 @@ In this mode, the model will generate think content wrapped in a `... [!NOTE] > For thinking mode, use `Temperature=0.6`, `TopP=0.95`, `TopK=20`, and `MinP=0` (the default setting in `generation_config.json`). **DO NOT use greedy decoding**, as it can lead to performance degradation and endless repetitions. For more detailed guidance, please refer to the [Best Practices](#best-practices) section. + ### `enable_thinking=False` We provide a hard switch to strictly disable the model's thinking behavior, aligning its functionality with the previous Qwen2.5-Instruct models. This mode is particularly useful in scenarios where disabling thinking is essential for enhancing efficiency. @@ -197,7 +209,7 @@ if __name__ == "__main__": print(f"Bot: {response_3}") ``` -> **Note** +> [!NOTE] > For API compatibility, when `enable_thinking=True`, regardless of whether the user uses `/think` or `/no_think`, the model will always output a block wrapped in `...`. However, the content inside this block may be empty if thinking is disabled. > When `enable_thinking=False`, the soft switches are not valid. Regardless of any `/think` or `/no_think` tags input by the user, the model will not generate think content and will not include a `...` block. diff --git a/added_tokens.json b/added_tokens.json deleted file mode 100644 index b54f913..0000000 --- a/added_tokens.json +++ /dev/null @@ -1,28 +0,0 @@ -{ - "": 151668, - "": 151658, - "": 151666, - "": 151667, - "": 151657, - "": 151665, - "<|box_end|>": 151649, - "<|box_start|>": 151648, - "<|endoftext|>": 151643, - "<|file_sep|>": 151664, - "<|fim_middle|>": 151660, - "<|fim_pad|>": 151662, - "<|fim_prefix|>": 151659, - "<|fim_suffix|>": 151661, - "<|im_end|>": 151645, - "<|im_start|>": 151644, - "<|image_pad|>": 151655, - "<|object_ref_end|>": 151647, - "<|object_ref_start|>": 151646, - "<|quad_end|>": 151651, - "<|quad_start|>": 151650, - "<|repo_name|>": 151663, - "<|video_pad|>": 151656, - "<|vision_end|>": 151653, - "<|vision_pad|>": 151654, - "<|vision_start|>": 151652 -} diff --git a/special_tokens_map.json b/special_tokens_map.json deleted file mode 100644 index ac23c0a..0000000 --- a/special_tokens_map.json +++ /dev/null @@ -1,31 +0,0 @@ -{ - "additional_special_tokens": [ - "<|im_start|>", - "<|im_end|>", - "<|object_ref_start|>", - "<|object_ref_end|>", - "<|box_start|>", - "<|box_end|>", - "<|quad_start|>", - "<|quad_end|>", - "<|vision_start|>", - "<|vision_end|>", - "<|vision_pad|>", - "<|image_pad|>", - "<|video_pad|>" - ], - "eos_token": { - "content": "<|im_end|>", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - }, - "pad_token": { - "content": "<|endoftext|>", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false - } -}