LLäMmlein 1B

This is a German Tinyllama 1B language model trained from scratch using the Tinyllama codebase on the German portion of RedPajama V2. Find more details on our page and our preprint!

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("LSX-UniWue/LLaMmlein_1B")

tokenizer = AutoTokenizer.from_pretrained("LSX-UniWue/LLaMmlein_1B")

Evaluation

We evaluated our results on the SuperGLEBer benchmark.

Downloads last month
674
Safetensors
Model size
1.1B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for LSX-UniWue/LLaMmlein_1B

Adapters
8 models

Dataset used to train LSX-UniWue/LLaMmlein_1B

Space using LSX-UniWue/LLaMmlein_1B 1

Collection including LSX-UniWue/LLaMmlein_1B