You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Farabi-1.7B

What a model can DO matters more than what it knows. Knowledge expires; skills endure.

Farabi-1.7B is a multilingual instruction-following model fine-tuned for Kazakh, Russian, and English. It is optimised for reasoning, instruction adherence, and structured tool calling.

Property Value
Parameters 1.7B
Languages Kazakh (KK) · Russian (RU) · English (EN)
License Apache-2.0
Architecture Qwen3-1.7B (transformer, GQA)

Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "nur-dev/farabi-1.7b"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype="auto", device_map="auto")

messages = [
    {"role": "user", "content": "Қазақстан туралы қысқаша айтып бер."}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True))

Evaluation

Scores are accuracy (%) on held-out test splits. Qolda-4.3B is included for scale reference (it has 2.5× more parameters).

Kazakh

Benchmark Samples Qwen3-1.7B Farabi-1.7B Qolda-4.3B
KazMMLU 22,889 41.0% 43.8% (+2.7) 47.1%
Belebele KK 900 41.7% 54.0% (+12.3) 80.9%
UNT 14,849 31.3% 37.6% (+6.3) 39.9%
Dastur 1,004 82.1% 88.8% (+6.7) 93.1%

Russian

Benchmark Samples Qwen3-1.7B Farabi-1.7B Qolda-4.3B
ruMMLU 14,012 44.8% 45.6% (+0.8) 58.7%
Belebele RU 900 69.9% 71.2% (+1.3) 89.4%

English (controls)

Benchmark Samples Qwen3-1.7B Farabi-1.7B Qolda-4.3B
MMLU-Pro EN 12,032 29.2% 28.5% (−0.7) 20.7%
ARC-Challenge 1,172 73.6% 71.4% (−2.2) 91.6%
Belebele EN 900 76.9% 76.6% (−0.3) 92.7%

Small English regression is expected: the model was fine-tuned primarily on KK/RU data.

Tool Calling

Test Qwen3-1.7B Farabi-1.7B Qolda-4.3B
Weather lookup (KK) MISS OK MISS
Currency conversion (KK) MISS OK MISS
Search + calculator (EN) MISS MISS MISS
No tool needed (KK) OK OK OK
Translation tool (RU) MISS OK MISS
Accuracy 20% 80% 20%

Benchmark Chart

Benchmark Chart


Base Model Note

Farabi-1.7B is built on Qwen3-1.7B, which is itself an instruction-tuned model (not a raw pretrained base). All capability improvements are measured relative to that already-capable starting point.


Acknowledgements

We thank the Qwen team at Alibaba Cloud for releasing Qwen3-1.7B under the Apache-2.0 license, which made this work possible.

Downloads last month
133
Safetensors
Model size
2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nur-dev/farabi-1.7b

Finetuned
Qwen/Qwen3-1.7B
Finetuned
(482)
this model

Collection including nur-dev/farabi-1.7b