Skip to content

Latest commit

 

History

History

README.md

MUST READ!!!

Since these config files are examples, so please carefully read the file and comments inside. Try to understand the purpose of each section and parameter, don't simply copy and paste the content of these config files into your own config file. It will not work. For more detailed information to the configurations, please read this file.

We also definitely welcome your contribution to add config files for other LLM providers.

Qwen3 Think and No_Think Configuration

The config.qwen3.yaml file provides an example configuration for using Qwen3 models with their unique thinking and non-thinking capabilities. Qwen3 models support two modes:

Thinking Mode

  • Use /think in your prompts to enable step-by-step reasoning
  • Optimized with temperature=0.6, top_p=0.95, top_k=20
  • Best for complex problems requiring detailed reasoning
  • Uses the qwen3-thinking alias in the pipeline configuration

Non-Thinking Mode

  • Use /no_think in your prompts for direct, fast responses
  • Optimized with temperature=0.7, top_p=0.8, top_k=20
  • Best for simple queries and general conversation
  • Uses the qwen3-fast alias in the pipeline configuration

Available Models

  • qwen/qwen3-30b-a3b: 30B parameter MoE model (3.3B activated)
  • qwen/qwen3-32b: 32B parameter dense model
  • qwen/qwen3-8b: 8B parameter dense model
  • qwen/qwen3-14b: 14B parameter dense model

Usage Examples

# Enable thinking for complex reasoning
"Explain the mathematical proof for the Pythagorean theorem /think"

# Use fast mode for simple queries  
"What is the capital of France? /no_think"

Note: You need to set OPENROUTER_API_KEY in your ~/.wrenai/.env file to use OpenRouter as the provider for Qwen3 models.