config = { "HF_Mistrail": { "model": "mistralai/Mistral-7B-Instruct-v0.2", "temperature": 0.5, "max_new_tokens": 1024, "top_k": 5, "load_in_8bit": True }, "HF_TinyLlama": { "model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0", "temperature": 0.5, "max_new_tokens": 1024, "top_k": 5, "top_p":0.95, "load_in_8bit": True, "do_sample": True }, "HF_SmolLM135": { "model": "HuggingFaceTB/SmolLM-135M-Instruct", "temperature": 0.5, "max_new_tokens": 1024, "top_k": 5, "top_p":0.95, "load_in_8bit": True, "do_sample": True }, "HF_SmolLM360": { "model": "HuggingFaceTB/SmolLM-360M-Instruct", "temperature": 0.5, "max_new_tokens": 1024, "top_k": 5, "top_p":0.95, "load_in_8bit": True, "do_sample": True }, "HF_SmolLM": { "model": "HuggingFaceTB/SmolLM-1.7B-Instruct", "temperature": 0.5, "max_new_tokens": 1024, "top_k": 5, "top_p":0.95, "load_in_8bit": True, "do_sample": True }, "HF_Gemma2": { "model": "google/gemma-2-2b", "temperature": 0.5, "max_new_tokens": 1024, "top_k": 5, "top_p":0.95, "load_in_8bit": True, "do_sample": True }, "HF_Qwen2": { "model": "Qwen/Qwen2-7B-Instruct", "temperature": 0.5, "max_new_tokens": 1024, "top_k": 5, "top_p":0.95, "load_in_8bit": True, "do_sample": True }, }