--- library_name: transformers tags: [] --- # Model Card for X-LoRA-Gemma-7b ``` import torch from xlora.xlora_utils import load_model XLoRa_model_name = 'lamm-mit/x-lora-gemma-7b' model,tokenizer=load_model(model_name = XLoRa_model_name, device='cuda:0', use_flash_attention_2=True, dtype=torch.bfloat16, ) eos_token_id= tokenizer('', add_special_tokens=False, ) ['input_ids'][0] ``` ``` def generate_XLoRA_Gemma (system_prompt='You a helpful assistant. You are familiar with materials science. ', prompt='What is spider silk in the context of bioinspired materials?', repetition_penalty=1.,num_beams=1,num_return_sequences=1, top_p=0.9, top_k=256, temperature=.5,max_new_tokens=512, verbatim=False, eos_token=None, add_special_tokens=True, prepend_response='', ): if eos_token==None: eos_token= tokenizer.eos_token_id if system_prompt==None: messages=[ {"role": "user", "content": prompt}, ] else: messages=[ {"role": "user", "content": system_prompt+prompt}, ] txt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True, ) txt=txt+prepend_response inputs = tokenizer(txt, add_special_tokens =add_special_tokens, return_tensors ='pt').to(device) with torch.no_grad(): outputs = model.generate(input_ids = inputs["input_ids"], attention_mask = inputs["attention_mask"] , # This is usually done automatically by the tokenizer max_new_tokens=max_new_tokens, temperature=temperature, #value used to modulate the next token probabilities. num_beams=num_beams, top_k = top_k, top_p = top_p, num_return_sequences = num_return_sequences, eos_token_id=eos_token, pad_token_id = eos_token, do_sample =True,#skip_prompt=True, repetition_penalty=repetition_penalty, ) return tokenizer.batch_decode(outputs[:,inputs["input_ids"].shape[1]:].detach().cpu().numpy(), skip_special_tokens=True) ``` Then, use as follows: ``` from IPython.display import display, Markdown q='''What is graphene?''' res=generate_XLoRA_Gemma( system_prompt='You design materials.', prompt=q, max_new_tokens=1024, temperature=0.3, eos_token=eos_token_id) display (Markdown(res)) ``` ## Model Details ### Model Description This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses ### Direct Use [More Information Needed] ### Downstream Use [optional] [More Information Needed] ### Out-of-Scope Use [More Information Needed] ## Bias, Risks, and Limitations [More Information Needed] ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data [More Information Needed] ### Training Procedure #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] #### Speeds, Sizes, Times [optional] [More Information Needed] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data [More Information Needed] #### Factors [More Information Needed] #### Metrics [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] [More Information Needed] ## Environmental Impact Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]