Update app.py
Browse filesChange Model to meta-llama/Llama-3.2-3B-Instruct
app.py
CHANGED
@@ -4,7 +4,8 @@ from huggingface_hub import InferenceClient
|
|
4 |
"""
|
5 |
For more information on `huggingface_hub` Inference API support, please check the docs: https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
|
6 |
"""
|
7 |
-
|
|
|
8 |
|
9 |
|
10 |
def respond(
|
@@ -46,7 +47,7 @@ For information on how to customize the ChatInterface, peruse the gradio docs: h
|
|
46 |
demo = gr.ChatInterface(
|
47 |
respond,
|
48 |
additional_inputs=[
|
49 |
-
gr.Textbox(value="You are a friendly
|
50 |
gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
|
51 |
gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
|
52 |
gr.Slider(
|
|
|
4 |
"""
|
5 |
For more information on `huggingface_hub` Inference API support, please check the docs: https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
|
6 |
"""
|
7 |
+
### Model used changes from HuggingFaceH4/zephyr-7b-beta to meta-llama/Llama-3.2-3B-Instruct
|
8 |
+
client = InferenceClient("meta-llama/Llama-3.2-3B-Instruct")
|
9 |
|
10 |
|
11 |
def respond(
|
|
|
47 |
demo = gr.ChatInterface(
|
48 |
respond,
|
49 |
additional_inputs=[
|
50 |
+
gr.Textbox(value="You are a friendly and knowledgeable online assistant. Your goal is to provide accurate and concise information to user. Keep your responses short and to the point unless the user specifically requests more detail. Answer questions in a warm, friendly tone. If you do not have relation information, just tell user that you do not know and ask user to search on the Internet.", label="System message"),
|
51 |
gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
|
52 |
gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
|
53 |
gr.Slider(
|