ArvindSelvaraj commited on
Commit
adbc9d5
1 Parent(s): 439d8a3
Files changed (1) hide show
  1. backend.py +5 -5
backend.py CHANGED
@@ -4,12 +4,12 @@ import io
4
  import requests
5
  import html # For escaping HTML characters
6
  from bs4 import BeautifulSoup
7
- from openai import OpenAI
8
 
9
  # Initialize OpenAI API with Nvidia's Mistral model
10
  client = OpenAI(
11
- base_url="https://integrate.api.nvidia.com/v1",
12
- api_key="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"
13
  )
14
 
15
  def clean_test_case_output(text):
@@ -86,13 +86,13 @@ def generate_testcases(user_story):
86
  try:
87
  # Call the Nvidia Mistral API with the refined prompt
88
  completion = client.chat.completions.create(
89
- model="nv-mistralai/mistral-nemo-12b-instruct", # Using Mistral model
90
  messages=[
91
  {"role": "user", "content": prompt}
92
  ],
93
  temperature=0.06, # Further lowering temperature for precise and deterministic output
94
  top_p=0.5, # Prioritize high-probability tokens even more
95
- max_tokens=4096, # Increase max tokens to allow longer content
96
  stream=True # Streaming the response for faster retrieval
97
  )
98
 
 
4
  import requests
5
  import html # For escaping HTML characters
6
  from bs4 import BeautifulSoup
7
+ from openai import OpenAI
8
 
9
  # Initialize OpenAI API with Nvidia's Mistral model
10
  client = OpenAI(
11
+ base_url="https://api.llama-api.com",
12
+ api_key="LA-38879bbd57704371a151ce66b6186aa6e95ab69a3b974f5ca4459f2ce89c4de1"
13
  )
14
 
15
  def clean_test_case_output(text):
 
86
  try:
87
  # Call the Nvidia Mistral API with the refined prompt
88
  completion = client.chat.completions.create(
89
+ model="llama3.1-70b", # Using Mistral model
90
  messages=[
91
  {"role": "user", "content": prompt}
92
  ],
93
  temperature=0.06, # Further lowering temperature for precise and deterministic output
94
  top_p=0.5, # Prioritize high-probability tokens even more
95
+ max_token=4096, # Increase max tokens to allow longer content
96
  stream=True # Streaming the response for faster retrieval
97
  )
98