LLaMmlein_1B / README.md
Julia287's picture
llammchen_1b
ec5caa2
|
raw
history blame
696 Bytes
metadata
datasets:
  - togethercomputer/RedPajama-Data-V2
language:
  - de

LLäMMchen 1B

This is a German Tinyllama 1B language model trained from scratch using the the Tinyllama codebase on the German portion of RedPajama V2.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("LSX-UniWue/llammchen_1b")

tokenizer = AutoTokenizer.from_pretrained("LSX-UniWue/llammchen_1b")

Evaluation

We evaluated our results on the Superkleber benchmark.