PingVortex commited on
Commit
4c83a4e
·
verified ·
1 Parent(s): 6f759e7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +63 -3
README.md CHANGED
@@ -1,3 +1,63 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - llama
8
+ - causal-lm
9
+ - pretrained
10
+ - chytrej
11
+ - base
12
+ library_name: transformers
13
+ ---
14
+
15
+ # Chytrej1.5-90M-Base
16
+
17
+ A fully custom pretrained language model built from scratch on the LLaMA architecture.
18
+
19
+ Chytrej (Czech slang for "clever/smart") is a long-term model series by PingVortex Labs. Every model in the series will be fully custom pretrained from scratch, then the model may be instruction fine-tuned on the custom base. The ongoing goal: every release must at least know the capital of France.
20
+
21
+ Built by [PingVortex Labs](https://github.com/PingVortexLabs).
22
+
23
+ ---
24
+
25
+ ## Model Details
26
+
27
+ + **Parameters:** 90M
28
+ + **Context length:** 8,192 tokens
29
+ + **Language:** English only
30
+ + **Format:** base model
31
+ + **Architecture:** LLaMA
32
+ + **License:** Apache 2.0
33
+
34
+ ---
35
+
36
+ ## Benchmarks
37
+
38
+ Evaluated with [lm-eval-harness](https://github.com/EleutherAI/lm-evaluation-harness), 0-shot:
39
+
40
+ | Task | Metric | Chytrej1.5 | Chytrej1 |
41
+ |---|---|---|---|
42
+ | ARC-Easy | acc | **41.46%** | 39.73% |
43
+ | ARC-Easy | acc_norm | **37.04%** | 34.47% |
44
+
45
+ ---
46
+
47
+ ## Usage
48
+
49
+ ```python
50
+ from transformers import LlamaForCausalLM, PreTrainedTokenizerFast
51
+
52
+ model = LlamaForCausalLM.from_pretrained("pvlabs/Chytrej1.5-90M-Base")
53
+ tokenizer = PreTrainedTokenizerFast.from_pretrained("pvlabs/Chytrej1.5-90M-Base")
54
+
55
+ prompt = "The capital of France is"
56
+ inputs = tokenizer(prompt, return_tensors="pt")
57
+ outputs = model.generate(**inputs, max_new_tokens=100, repetition_penalty=1.3)
58
+ print(tokenizer.decode(outputs[0]))
59
+ ```
60
+
61
+ ---
62
+
63
+ *Made by [PingVortex](https://pingvortex.com).*