QUEST-30B-RL

QUEST 30B full model after mid-training → SFT → RL (Qwen3-30B-A3B base, dense). Trained following the same three-stage recipe as the 35B model, evaluated against Tongyi-DR and OpenResearcher at the same scale.

Benchmark results

Benchmark Metric Score
BrowseComp avg@3 37.0
Mind2Web 2 avg@3 28.6
HLE avg@3 24.6
DeepResearch Bench avg@3 45.3
BrowseComp-Plus avg@3 48.2
WideSearch Item F1 avg@4 54.2
GAIA avg@3 69.0
LiveResearchBench avg@3 74.1

Quick start

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "osunlp/QUEST-30B-RL"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id, device_map="auto", torch_dtype="auto",
)

Apply the model's chat template with tokenizer.apply_chat_template(...) before passing prompts.

License

Released under the Apache License 2.0.

Downloads last month
48
Safetensors
Model size
31B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including osunlp/QUEST-30B-RL