--- library_name: transformers license: apache-2.0 datasets: - pythainlp/han-instruct-dataset-v2.0 language: - th pipeline_tag: text-generation --- # Model Card for Han LLM 7B v1 Han LLM v1 is a model that trained by han-instruct-dataset v2.0. The model are working with Thai. Based model: [scb10x/typhoon-7b](https://huggingface.co/scb10x/typhoon-7b) ## Model Details ### Model Description The model was trained by LoRA. This is the model card of a ðŸĪ— transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** Wannaphong Phatthiyaphaibun - **Model type:** text-generation - **Language(s) (NLP):** Thai - **License:** apache-2.0 - **Finetuned from model:** [scb10x/typhoon-7b](https://huggingface.co/scb10x/typhoon-7b) ## Uses Thai users ### Out-of-Scope Use Math, Coding, and other language ## Bias, Risks, and Limitations The model can has a bias from dataset. Use at your own risks! ## How to Get Started with the Model Use the code below to get started with the model. ```python # !pip install accelerate sentencepiece transformers bitsandbytes import torch from transformers import pipeline pipe = pipeline("text-generation", model="wannaphong/han-llm-7b-v1", torch_dtype=torch.bfloat16, device_map="auto") # We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating messages = [ {"role": "user", "content": "āđāļĄāļ§āļ„āļ·āļ­āļ­āļ°āđ„āļĢ"}, ] prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ``` ## Training Details ### Training Data [Han Instruct dataset v2.0](https://huggingface.co/datasets/pythainlp/han-instruct-dataset-v2.0) ### Training Procedure Use LoRa - r: 48 - lora_alpha - 1 epoch