How can QwQ-32B be used?
Question
Lost your password? Please enter your email address. You will receive a link and will create a new password via email.
Lorem ipsum dolor sit amet, consectetur adipiscing elit.Morbi adipiscing gravdio, sit amet suscipit risus ultrices eu.Fusce viverra neque at purus laoreet consequa.Vivamus vulputate posuere nisl quis consequat.
Answers ( 1 )
To use QwQ-32B, follow these guidelines:\\n" and set add_generation_prompt=True using apply_chat_template.
- **Environment Requirements**: Use the latest version of the transformers library (versions below 4.37.0 may cause errors).
- **Quick Start**: Load the model using AutoModelForCausalLM and AutoTokenizer from "Qwen/QwQ-32B".
- **Forced Thought Output**: Begin prompts with "
- **Sampling Parameters**: Recommended settings include Temperature=0.6, TopP=0.95, MinP=0, TopK=20-40, and presence_penalty=0-2.
- **Multi-turn Dialogue**: Use apply_chat_template for smooth dialogue without thought content in history.
- **Output Format Standardization**: For mathematical problems, provide step-by-step reasoning and box the final answer with \\boxed{}. For multiple-choice questions, use JSON format and output only the option letter (e.g., "answer": "C").
- **Long Input Handling**: Enable YaRN for prompts exceeding 8,192 tokens by adding specific configurations to config.json and using vLLM.