--- license: apache-2.0 datasets: - vicgalle/alpaca-gpt4 language: - en library_name: transformers --- # Model Card for OLMo-1B-instruct-alpaca_amc OLMo-1B-instruct-alpaca_amc is a language model fine-tuned on the Alpaca-GPT4 dataset. This model extends the capabilities of the original OLMo 1B model, by the Allen Institute for AI (AI2), incorporating the ability to follow instructions and perform chatting tasks. ## Model Description OLMo-1B-instruct-alpaca_amc is built upon the OLMo 1B architecture, a Transformer-based autoregressive language model known for its deep understanding and generation of natural language. By fine-tuning OLMo 1B with the Alpaca-GPT4 dataset, we've tailored the model specifically for interactive chatting applications, enabling it to understand and respond to a wide range of conversational prompts and instructions. ### How to Use To use OLMo-1B-instruct-alpaca_amc for generating text or chatting, you can leverage the Hugging Face Transformers library. Here's a quick example in Python: ```python import hf_olmo from transformers import AutoModelForCausalLM, AutoTokenizer # Model and tokenizer directories tokenizer = AutoTokenizer.from_pretrained("amc-madalin/OLMo-1B-instruct-alpaca_amc") model = AutoModelForCausalLM.from_pretrained("amc-madalin/OLMo-1B-instruct-alpaca_amc") print("Chat with the model (type 'quit' to stop):") while True: message = input("You: ") if message.lower() == 'quit': break inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False, padding=True, truncation=True, max_length=512) response = model.generate(**inputs, max_length=512, pad_token_id=tokenizer.eos_token_id, eos_token_id=tokenizer.eos_token_id, do_sample=True, top_k=50, top_p=0.95) reply = tokenizer.decode(response[0], skip_special_tokens=True) print("AI: ", reply) ``` ### Training Data The model was fine-tuned on the Alpaca-GPT4 dataset, which consists of 52K instruction-following demonstrations generated in the style of self-instruct using GPT-4. ## Applications OLMo-1B-instruct-alpaca_amc is great for research and experimentation with Open LLMs. - Conversational agents - Interactive storytelling - Educational tool ## Limitations and Biases As with any language model, OLMo-1B-instruct-alpaca_amc inherits biases present in its training data. Users should be aware of these potential biases and limitations. ## How to Contribute I welcome contributions to improve the model and expand its applications! ## Citation If you use OLMo-1B-instruct-alpaca_amc in your research, please cite the original OLMo paper and the dataset used for fine-tuning: ```bibtex @article{Groeneveld2023OLMo, title={OLMo: Accelerating the Science of Language Models}, author={Groeneveld, Dirk and Beltagy, Iz and Walsh, Pete and Bhagia, Akshita and Kinney, Rodney and Tafjord, Oyvind and Jha, Ananya Harsh and Ivison, Hamish and Magnusson, Ian and Wang, Yizhong and Arora, Shane and Atkinson, David and Authur, Russell and Chandu, Khyathi and Cohan, Arman and Dumas, Jennifer and Elazar, Yanai and Gu, Yuling and Hessel, Jack and Khot, Tushar and Merrill, William and Morrison, Jacob and Muennighoff, Niklas and Naik, Aakanksha and Nam, Crystal and Peters, Matthew E. and Pyatkin, Valentina and Ravichander, Abhilasha and Schwenk, Dustin and Shah, Saurabh and Smith, Will and Subramani, Nishant and Wortsman, Mitchell and Dasigi, Pradeep and Lambert, Nathan and Richardson, Kyle and Dodge, Jesse and Lo, Kyle and Soldaini, Luca and Smith, Noah A. and Hajishirzi, Hannaneh}, journal={Preprint}, year={2024} } ``` ## License This model is open-sourced under the Apache 2.0 license. See the LICENSE file for more details. ## Contact For any questions or support regarding OLMo-1B-instruct-alpaca_amc, please contact me at: [LinkedIn](https://www.linkedin.com/in/alexandru-m-costea/) [e-Mail](amc.madalin@gmail.com) Or check my work on: [GitHub](https://github.com/amc-madalin) [HuggingFace](https://huggingface.co/amc-madalin) [Website](https://amcostea.com/)