conversation context
Browse files
app.py
CHANGED
@@ -144,7 +144,8 @@ async def delete_conversation(api_key: str, conversation_id: int) -> str:
|
|
144 |
logging.error(f"Error deleting conversation {conversation_id}: {str(e)}")
|
145 |
return f"Failed to delete conversation: {str(e)}"
|
146 |
|
147 |
-
client = InferenceClient("Qwen/Qwen2.5-14B-Instruct")
|
|
|
148 |
|
149 |
def respond(
|
150 |
message: str,
|
|
|
144 |
logging.error(f"Error deleting conversation {conversation_id}: {str(e)}")
|
145 |
return f"Failed to delete conversation: {str(e)}"
|
146 |
|
147 |
+
# client = InferenceClient("Qwen/Qwen2.5-14B-Instruct") # needs more memory than available in free tier - The model Qwen/Qwen2.5-14B-Instruct is too large to be loaded automatically (29GB > 10GB)
|
148 |
+
client = InferenceClient("microsoft/Phi-3-small-128k-instruct")
|
149 |
|
150 |
def respond(
|
151 |
message: str,
|