Alibaba/Dense

Qwen2 Math 1.5B

reasoning
1.5B
Parameters
4K
Context length
6
Benchmarks
4
Quantizations
0
Architecture
Dense
Released
2024-08-08
Layers
28
KV Heads
2
Head Dim
128
Family
qwen

Qwen2-Math-1.5B-Instruct

[!Warning]

Introduction

Over the past year, we have dedicated significant effort to researching and enhancing the reasoning capabilities of large language models, with a particular focus on their ability to solve arithmetic and mathematical problems. Today, we are delighted to introduce a serise of math-specific large language models of our Qwen2 series, Qwen2-Math and Qwen2-Math-Instruct-1.5B/7B/72B. Qwen2-Math is a series of specialized math language models built upon the Qwen2 LLMs, which significantly outperforms the mathematical capabilities of open-source models and even closed-source models (e.g., GPT4o). We hope that Qwen2-Math can contribute to the scientific community for solving advanced mathematical problems that require complex, multi-step logical reasoning.

Model Details

For more details, please refer to our blog post and GitHub repo.

Requirements

  • transformers>=4.40.0 for Qwen2-Math models. The latest version is recommended.

[!Warning]

For requirements on GPU memory and the respective throughput, see similar results of Qwen2 here.

Quick Start

[!Important]

Qwen2-Math-1.5B-Instruct is an instruction model for chatting;

Qwen2-Math-1.5B is a base model typically used for completion and few-shot inference, serving as a better starting point for fine-tuning.

🤗 Hugging Face Transformers

Qwen2-Math can be deployed and inferred in the same way as Qwen2. Here we show a code snippet to show you how to use the chat model with transformers:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "Qwen/Qwen2-Math-1.5B-Instruct"
device = "cuda" # the device to load the model onto

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "Find the value of $x$ that satisfies the equation $4x+5 = 6x+7$."
messages = [
    {"role": "system", "content": "You are a helpful assistant."},
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=512
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]

🤖 ModelScope

We strongly advise users, especially those in mainland China, to use ModelScope. snapshot_download can help you solve issues concerning downloading checkpoints.

Quantizations & VRAM

Q4_K_M4.5 bpw
1.3 GB
VRAM required
94%
Quality
Q6_K6.5 bpw
1.7 GB
VRAM required
97%
Quality
Q8_08 bpw
2.0 GB
VRAM required
100%
Quality
FP1616 bpw
3.5 GB
VRAM required
100%
Quality

Benchmarks (6)

IFEval44.8
BigCodeBench27.0
MATH22.1
MMLU-PRO20.0
BBH19.8
MUSR3.2

Run with Ollama

$ollama run qwen2-math:1.5b

GPUs that can run this model

At Q4_K_M quantization. Sorted by minimum VRAM.

Find the best GPU for Qwen2 Math 1.5B

Build Hardware for Qwen2 Math 1.5B