Qwen 2
Qwen2 is a series of state of the art, multi-lingual LLMs that range from 0.5 to 72 billion parameters. It excels in a variety of tasks. The only variant offered by Rubra at the moment is the 7B one.
note
| Model | Params | Context Length | GQA | Token Count | Knowledge Cutoff |
|---|---|---|---|---|---|
| Qwen2 7B Instruct | 7.6B | 131,072 | Yes | 3T | 2023 |
tip
The Rubra enhanced Qwen2-7b-Instruct model is capable of doing tool/function calling in Chinese!
We recommend this model for any task exceeding the context limit found in Llama-3 models.
Qwen2 7B Instruct
MT-bench Pairwise Comparison:
| Model | Function Calling | MMLU | GPQA | GSM-8K | MATH | MT-bench | MT-bench Pairwise Comparison | |||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Win | Loss | Tie | Win Rate | Loss Rate | Adjusted Win Rate | |||||||
| Qwen2-7B-Instruct | - | 70.78 | 32.14 | 78.54 | 30.10 | 8.29 | 49 | 33 | 78 | 0.30625 | 0.20625 | 0.55 |
| Rubra Enhanced Qwen2-7B-Instruct | 85.71% | 68.88 | 30.36 | 75.82 | 28.72 | 8.08 | 33 | 49 | 78 | 0.20625 | 0.30625 | 0.45 |