Bangla Llama-2 7B Base v0.1 [pre-trained][Llama2 Original Tokenizer]

Welcome to the inaugural release of the Bangla Llama-2 7B base model, an important step in advancing LLMs for the Bangla language. This model is ready for immediate inference and primed for further fine-tuning to suit your specific NLP tasks.

Please Note: This model, labelled as a foundational Bangla Language Model (LLM), is designed primarily for Causal Language Modeling (LM) purposes.

Model description

This Bangla Llama model has been built upon the foundation set by the original Llama-2 with Llama-2 original tokenizer.

  • Model type: A 7B parameter model for Causal LM pre-trained on Bangla 2B+ BERT dataset.
  • Language(s): Bangla and English
  • License: GNU General Public License v3.0
  • Source Model: meta-llama/Llama-2-7b-hf
  • Training Precision: float16
  • Code: GitHub

Usage Note

It's important to note that the models have not undergone detoxification. Therefore, while they possess impressive linguistic capabilities, there is a possibility for them to generate content that could be deemed harmful or offensive. We urge users to exercise discretion and supervise the model's outputs closely, especially in public or sensitive applications.

Meet the Developers

Get to know the creators behind this innovative model and follow their contributions to the field:

Citation

We hope this model serves as a valuable tool in your NLP toolkit and look forward to seeing the advancements it will enable in understanding and producing the Bangla language.

Downloads last month
146
Safetensors
Model size
6.74B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for meherajj/Bangla-Llama-2-7b-base-v0.1

Finetuned
(719)
this model
Quantizations
1 model