feihu.hf
		
	commited on
		
		
					Commit 
							
							·
						
						7c0a8dc
	
1
								Parent(s):
							
							2ebdfe8
								
update readme
Browse files- README.md +8 -1
- figures/benchmark.jpg +0 -0
    	
        README.md
    CHANGED
    
    | @@ -19,6 +19,11 @@ tags: | |
| 19 |  | 
| 20 | 
             
            QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
         | 
| 21 |  | 
|  | |
|  | |
|  | |
|  | |
|  | |
| 22 | 
             
            **This repo contains the QwQ 32B model**, which has the following features:
         | 
| 23 | 
             
            - Type: Causal Language Models
         | 
| 24 | 
             
            - Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
         | 
| @@ -31,6 +36,8 @@ QwQ is the reasoning model of the Qwen series. Compared with conventional instru | |
| 31 |  | 
| 32 | 
             
            **Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
         | 
| 33 |  | 
|  | |
|  | |
| 34 | 
             
            For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
         | 
| 35 |  | 
| 36 | 
             
            ## Requirements
         | 
| @@ -126,7 +133,7 @@ If you find our work helpful, feel free to give us a cite. | |
| 126 |  | 
| 127 | 
             
            ```
         | 
| 128 | 
             
            @misc{qwq32b,
         | 
| 129 | 
            -
                title = { | 
| 130 | 
             
                url = {https://qwenlm.github.io/blog/qwq-32b/},
         | 
| 131 | 
             
                author = {Qwen Team},
         | 
| 132 | 
             
                month = {March},
         | 
|  | |
| 19 |  | 
| 20 | 
             
            QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
         | 
| 21 |  | 
| 22 | 
            +
            <p align="center">
         | 
| 23 | 
            +
              <img width="100%" src="figures/benchmark.jpg">
         | 
| 24 | 
            +
            </p>
         | 
| 25 | 
            +
             | 
| 26 | 
            +
             | 
| 27 | 
             
            **This repo contains the QwQ 32B model**, which has the following features:
         | 
| 28 | 
             
            - Type: Causal Language Models
         | 
| 29 | 
             
            - Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
         | 
|  | |
| 36 |  | 
| 37 | 
             
            **Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
         | 
| 38 |  | 
| 39 | 
            +
            You can try our [demo](https://huggingface.co/spaces/Qwen/QwQ-32B-Demo) or access QwQ models via [QwenChat](https://chat.qwen.ai).
         | 
| 40 | 
            +
             | 
| 41 | 
             
            For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
         | 
| 42 |  | 
| 43 | 
             
            ## Requirements
         | 
|  | |
| 133 |  | 
| 134 | 
             
            ```
         | 
| 135 | 
             
            @misc{qwq32b,
         | 
| 136 | 
            +
                title = {QwQ-32B: The Power of Scaling RL},
         | 
| 137 | 
             
                url = {https://qwenlm.github.io/blog/qwq-32b/},
         | 
| 138 | 
             
                author = {Qwen Team},
         | 
| 139 | 
             
                month = {March},
         | 
    	
        figures/benchmark.jpg
    ADDED
    
    |   | 
