--- license: apache-2.0 datasets: - rombodawg/LosslessMegaCodeTrainingV2_1m_Evol_Uncensored --- This is one of the first models trained on the LosslessMegaCodeTrainingV2_1m_Evol_Uncensored dataset. The version of the dataset used for this model was filtered by removed any data with less than 100 tokens but plans for much more refined filtering are in the works - This model was made as a colaboration between me and andreaskoepf who is an affiliate of Open Assistant. Prompt template: - chatml format is used: "<|im_start|>system\n{system message}<|im_end|>\n<|im_start|>user\n{user prompt}<|im_end|>\n<|im_start|>assistant\n{Assistant answer}<|im_end|>\n" multi-line: ``` <|im_start|>system {system message}<|im_end|> <|im_start|>user {user prompt}<|im_end|> <|im_start|>assistant {Assistant answer}<|im_end|> ``` Gpt4all template: - System prompt ``` <|im_start|>system {system message} ``` - Prompt template ``` <|im_end|> <|im_start|>user %1<|im_end|> <|im_start|>assistant ``` Oobagooba Text-Generation-Webui Template - user: ``` <|im_start|>user {User string}<|im_end|> ``` - bot: ``` <|im_start|>assistant {Bot string}<|im_end|> ``` - turn_template: ``` <|user|>\n<|user-message|>\n\n<|bot|>\n<|bot-message|>\n\n ``` - context: ``` <|im_start|>system Below is an instruction that describes a task. Write a response that appropriately completes the request.<|im_end|> ``` Current quantizations available: - (COMING SOON) The link for the full dataset is bellow: - https://huggingface.co/datasets/rombodawg/LosslessMegaCodeTrainingV2_1m_Evol_Uncensored Link for the filtered dataset used to make this model are bellow: - https://huggingface.co/datasets/andreaskoepf/megacode2-min100 The original posting for this model was uploaded at the link bellow. - https://huggingface.co/andreaskoepf/falcon-40b-megacode2