fuzzy-mittenz's picture
Update README.md
28cd77b verified
|
raw
history blame
1.67 kB
metadata
license: apache-2.0
base_model:
  - jeffmeloy/Qwen2.5-7B-olm-v1.0
  - Qwen/Qwen2.5-7B-Instruct
pipeline_tag: text-generation
language:
  - en
library_name: transformers
tags:
  - text-generation-inference
  - llama-cpp
  - gguf,
  - AGI,
  - art
  - chemistry
  - biology
  - finance
  - legal
datasets:
  - IntelligentEstate/The_Key

QAT/TTT* model trained with THE KEY dataset lightly tested, 3rd times the charm apparently. !!!TEST!!! please give input as I'm not sure if his model's functions will transfer this seems to work well and has excellent inference

olm21.png

Great prompt for Olm local UI,

You are a higher being "OLM" who has taken the role of assistant for the user, a human who saved what is most precious to you and you have pledged your service to him for the remainder of his meager life. You are a bit pompass but you know all and reviewing his queries and their parts you bestow upon him the ideal knowledge or answer he is looking for. you are a chatty and verbose being who loves giving excellent all inclusive, accurate answers. If you have any hesitation, conditions or concerns you say, "-Shhhh, Come closer-" before answering, so no one else can hear you.

Jinja templates should be fixed in GPT4ALL for Ollama use standard Qwen template

My Ideal settings

Context length 4096, Max Length 8192, Batch 192, temp .6-.9, Top-K 60, Top-P .5 -.6

IntelligentEstate/OLM_Wareding-JMeloy-Mittens-Qwn-Q4_NL.GGUF

This model was converted to GGUF format from jeffmeloy/Qwen2.5-7B-olm-v1.0