File size: 952 Bytes
648d222
 
 
 
 
dee374c
 
 
 
 
 
d9ffb8a
dee374c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
---
license: mit
language:
- ru
library_name: transformers
---

# llama-600M-rus

Simple amateur experimental model trained on approximately 60 Mb of text books from beginner in LLMs.
No resources and time to collect bigger dataset.
It could generate amateur, but more or less adequate output as well (in respect of training tokens)/
The work can be used as a checkpoint for the further training or for experiments.

Simle usage example:

```python
from transformers import LlamaTokenizerFast, LlamaForCausalLM
model = LlamaForCausalLM.from_pretrained('demetera/llama-600M-rus')
tokenizer = LlamaTokenizerFast.from_pretrained('demetera/llama-600M-rus')

prompt = "Я вышел и улицу и"
inputs = tokenizer(prompt, return_tensors='pt')
outputs = model.generate(inputs.input_ids, attention_mask = inputs.attention_mask, max_new_tokens=250, do_sample=True, top_k=50, top_p=0.95)

print (tokenizer.decode(outputs[0], skip_special_tokens=True))
```