LLaMmlein_1B / README.md
JanPf's picture
Update README.md
afda23f verified
|
raw
history blame
868 Bytes
metadata
datasets:
  - togethercomputer/RedPajama-Data-V2
language:
  - de
pipeline_tag: text-generation
library_name: transformers
license: other

LLäMmlein 1B

This is a German Tinyllama 1B language model trained from scratch using the Tinyllama codebase on the German portion of RedPajama V2. Find more details on our page!

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("LSX-UniWue/LLaMmlein_1B")

tokenizer = AutoTokenizer.from_pretrained("LSX-UniWue/LLaMmlein_1B")

Evaluation

We evaluated our results on the Superkleber benchmark.