xLSTM-7b / README.md
Korbinian Poeppel
Fix: Update README.md for installation of mlstm_kernels.
96f9264
|
raw
history blame
951 Bytes
---
license: other
---
# xLSTM goes 7B
This xLSTM was pre-trained on the DCLM and selected high-quality data for in a total of approx. 2.3 T tokens using the `xlstm-jax` framework.
## How to use it
First, install `xlstm`, which now uses the `mlstm_kernels` package for triton kernels:
```bash
pip install xlstm
pip install mlstm_kernels
```
For now, install the transformers repositiory fork from NX-AI (until it is merged):
```bash
pip install 'transformers @ git+ssh://git@github.com/NX-AI/transformers.git@integrate_xlstm'
```
Use this model as:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
xlstm = AutoModelForCausalLM.from_pretrained("NX-AI/xLSTM-7b", device_map="auto")
# this is a fork of EleutherAI/gpt
tokenizers = AutoTokenizer.from_pretrained("NX-AI/xLSTM-7b")
xlstm(tokenizer("Hello xLSTM, how are you doing?"))
```
License: NXAI Community License (see `LICENSE` file)