MoneyRadar / app.py
seawolf2357's picture
Update app.py
695d9f1 verified
raw
history blame
10.4 kB
import gradio as gr
import requests
import json
import os
from datetime import datetime, timedelta
from huggingface_hub import InferenceClient # LLM ์‚ฌ์šฉ์„ ์œ„ํ•ด ํ•„์š”
# ํ™˜๊ฒฝ ๋ณ€์ˆ˜์—์„œ API ํ‚ค ๊ฐ€์ ธ์˜ค๊ธฐ
API_KEY = os.getenv("SERPHOUSE_API_KEY")
HF_TOKEN = os.getenv("HF_TOKEN")
# ๊ตญ๊ฐ€ ์ด๋ฆ„๊ณผ ์œ„์น˜ ID๋ฅผ ๋งคํ•‘
COUNTRY_CODE_MAPPING = {
"United States": "2840",
"South Korea": "2458",
# ๋‹ค๋ฅธ ๊ตญ๊ฐ€๋“ค์˜ ์œ„์น˜ ID๋ฅผ ์ถ”๊ฐ€ํ•˜์„ธ์š”
}
MAJOR_COUNTRIES = list(COUNTRY_CODE_MAPPING.keys())
def search_serphouse(query, country, page=1, num_result=10):
url = "https://api.serphouse.com/serp/live"
now = datetime.utcnow()
yesterday = now - timedelta(days=1)
date_range = f"{yesterday.strftime('%Y-%m-%d')},{now.strftime('%Y-%m-%d')}"
# ๊ตญ๊ฐ€ ์ด๋ฆ„์„ ์œ„์น˜ ID๋กœ ๋ณ€ํ™˜
loc_id = COUNTRY_CODE_MAPPING.get(country, "2840") # ๊ธฐ๋ณธ๊ฐ’์€ ๋ฏธ๊ตญ
payload = {
"data": {
"q": query,
"domain": "google.com",
"loc": loc_id, # ์œ„์น˜ ID ์‚ฌ์šฉ
"lang": "en",
"device": "desktop",
"serp_type": "news",
"page": str(page),
"verbatim": "1",
"num": str(num_result),
"date_range": date_range
}
}
headers = {
"accept": "application/json",
"content-type": "application/json",
"authorization": f"Bearer {API_KEY}"
}
try:
response = requests.post(url, json=payload, headers=headers)
response.raise_for_status()
return response.json()
except requests.RequestException as e:
error_msg = f"Error: {str(e)}"
if hasattr(response, 'text'):
error_msg += f"\nResponse content: {response.text}"
return {"error": error_msg}
def format_results_from_raw(results):
try:
if isinstance(results, dict) and "error" in results:
return "Error: " + results["error"], []
if not isinstance(results, dict):
raise ValueError("๊ฒฐ๊ณผ๊ฐ€ ์‚ฌ์ „ ํ˜•์‹์ด ์•„๋‹™๋‹ˆ๋‹ค.")
# 'results' ํ‚ค ๋‚ด๋ถ€์˜ ๊ตฌ์กฐ ํ™•์ธ
if 'results' in results:
results_content = results['results']
if 'results' in results_content:
results_content = results_content['results']
if 'news' in results_content:
news_results = results_content['news']
else:
news_results = []
else:
news_results = []
else:
news_results = []
if not news_results:
return "๊ฒ€์ƒ‰ ๊ฒฐ๊ณผ๊ฐ€ ์—†์Šต๋‹ˆ๋‹ค.", []
articles = []
for idx, result in enumerate(news_results, 1):
title = result.get("title", "์ œ๋ชฉ ์—†์Œ")
link = result.get("url", result.get("link", "#"))
snippet = result.get("snippet", "๋‚ด์šฉ ์—†์Œ")
channel = result.get("channel", result.get("source", "์•Œ ์ˆ˜ ์—†์Œ"))
time = result.get("time", result.get("date", "์•Œ ์ˆ˜ ์—†๋Š” ์‹œ๊ฐ„"))
image_url = result.get("img", result.get("thumbnail", ""))
articles.append({
"index": idx,
"title": title,
"link": link,
"snippet": snippet,
"channel": channel,
"time": time,
"image_url": image_url
})
return "", articles
except Exception as e:
error_message = f"๊ฒฐ๊ณผ ์ฒ˜๋ฆฌ ์ค‘ ์˜ค๋ฅ˜ ๋ฐœ์ƒ: {str(e)}"
return "Error: " + error_message, []
def serphouse_search(query, country):
page = 1
num_result = 10
results = search_serphouse(query, country, page, num_result)
error_message, articles = format_results_from_raw(results)
return error_message, articles
# LLM ์„ค์ •
hf_client = InferenceClient("CohereForAI/c4ai-command-r-plus-08-2024", token=HF_TOKEN)
def summarize_article(title, snippet):
try:
prompt = f"๋‹ค์Œ ๋‰ด์Šค ์ œ๋ชฉ๊ณผ ์š”์•ฝ์„ ๋ฐ”ํƒ•์œผ๋กœ ํ•œ๊ตญ์–ด๋กœ 3๋ฌธ์žฅ์œผ๋กœ ์š”์•ฝํ•˜์„ธ์š”:\n์ œ๋ชฉ: {title}\n์š”์•ฝ: {snippet}"
summary = hf_client.text_generation(prompt, max_new_tokens=500)
return summary
except Exception as e:
return f"์š”์•ฝ ์ค‘ ์˜ค๋ฅ˜ ๋ฐœ์ƒ: {str(e)}"
css = """
footer {
visibility: hidden;
}
"""
# Gradio ์ธํ„ฐํŽ˜์ด์Šค ๊ตฌ์„ฑ
with gr.Blocks(css=css, title="NewsAI ์„œ๋น„์Šค") as iface:
gr.Markdown("๊ฒ€์ƒ‰์–ด๋ฅผ ์ž…๋ ฅํ•˜๊ณ  ์›ํ•˜๋Š” ๊ตญ๊ฐ€๋ฅผ ์„ ํƒํ•˜๋ฉด, ๊ฒ€์ƒ‰์–ด์™€ ์ผ์น˜ํ•˜๋Š” 24์‹œ๊ฐ„ ์ด๋‚ด ๋‰ด์Šค๋ฅผ ์ตœ๋Œ€ 10๊ฐœ ์ถœ๋ ฅํ•ฉ๋‹ˆ๋‹ค.")
with gr.Column():
with gr.Row():
query = gr.Textbox(label="๊ฒ€์ƒ‰์–ด")
country = gr.Dropdown(MAJOR_COUNTRIES, label="๊ตญ๊ฐ€", value="South Korea")
search_button = gr.Button("๊ฒ€์ƒ‰")
# ์ƒํƒœ ๋ฉ”์‹œ์ง€ ์ปดํฌ๋„ŒํŠธ ์ œ๊ฑฐ (Progress๋ฅผ ์‚ฌ์šฉํ•˜๋ฏ€๋กœ ๋ถˆํ•„์š”)
# status_message = gr.Markdown(visible=False)
# ๊ธฐ์‚ฌ ๋ฐ์ดํ„ฐ๋ฅผ ์ €์žฅํ•  ์ƒํƒœ ๋ณ€์ˆ˜
articles_state = gr.State([]) # ์ดˆ๊ธฐ๊ฐ’์„ ๋นˆ ๋ฆฌ์ŠคํŠธ๋กœ ์„ค์ •
# ์ตœ๋Œ€ 10๊ฐœ์˜ ๊ธฐ์‚ฌ์— ๋Œ€ํ•œ ์ปดํฌ๋„ŒํŠธ๋ฅผ ๋ฏธ๋ฆฌ ์ƒ์„ฑํ•ฉ๋‹ˆ๋‹ค.
article_components = []
for i in range(10):
with gr.Group(visible=False) as article_group:
title = gr.Markdown()
image = gr.Image(width=200, height=150)
snippet = gr.Markdown()
info = gr.Markdown()
analyze_button = gr.Button("๋ถ„์„")
summary_output = gr.Markdown(visible=False)
article_components.append({
'group': article_group,
'title': title,
'image': image,
'snippet': snippet,
'info': info,
'analyze_button': analyze_button,
'summary_output': summary_output,
'index': i,
})
def search_and_display(query, country, articles_state):
with gr.Progress() as progress:
progress(0, desc="์ฒ˜๋ฆฌ์ค‘์ž…๋‹ˆ๋‹ค. ์ž ์‹œ๋งŒ ๊ธฐ๋‹ค๋ฆฌ์„ธ์š”.")
error_message, articles = serphouse_search(query, country)
outputs = []
if error_message:
outputs.append(gr.update(value=error_message, visible=True))
for comp in article_components:
outputs.extend([
gr.update(visible=False), # group
gr.update(), # title
gr.update(), # image
gr.update(), # snippet
gr.update(), # info
gr.update(visible=False), # summary_output
])
articles_state = []
else:
outputs.append(gr.update(value="", visible=False))
for idx, comp in enumerate(article_components):
if idx < len(articles):
article = articles[idx]
# ์ด๋ฏธ์ง€ ์ฒ˜๋ฆฌ ์ˆ˜์ •
image_url = article['image_url']
if image_url and not image_url.startswith('data:image'):
image_update = gr.update(value=image_url, visible=True)
else:
image_update = gr.update(value=None, visible=False)
outputs.extend([
gr.update(visible=True), # group
gr.update(value=f"### [{article['title']}]({article['link']})"), # title
image_update, # image
gr.update(value=f"**์š”์•ฝ:** {article['snippet']}"), # snippet
gr.update(value=f"**์ถœ์ฒ˜:** {article['channel']} | **์‹œ๊ฐ„:** {article['time']}"), # info
gr.update(visible=False), # summary_output
])
else:
outputs.extend([
gr.update(visible=False), # group
gr.update(), # title
gr.update(), # image
gr.update(), # snippet
gr.update(), # info
gr.update(visible=False), # summary_output
])
articles_state = articles # articles_state ์—…๋ฐ์ดํŠธ
outputs.append(articles_state)
# ์ƒํƒœ ๋ฉ”์‹œ์ง€ ์ˆจ๊น€ (Progress๋Š” ์ž๋™์œผ๋กœ ์‚ฌ๋ผ์ง‘๋‹ˆ๋‹ค)
return outputs
# search_button ํด๋ฆญ ์‹œ ์—…๋ฐ์ดํŠธ๋  ์ถœ๋ ฅ ์ปดํฌ๋„ŒํŠธ ๋ชฉ๋ก ์ƒ์„ฑ
search_outputs = []
error_output = gr.Markdown(visible=False)
search_outputs.append(error_output)
for comp in article_components:
search_outputs.append(comp['group'])
search_outputs.append(comp['title'])
search_outputs.append(comp['image'])
search_outputs.append(comp['snippet'])
search_outputs.append(comp['info'])
search_outputs.append(comp['summary_output'])
search_outputs.append(articles_state)
# status_message ์ปดํฌ๋„ŒํŠธ ์ œ๊ฑฐ
search_button.click(
search_and_display,
inputs=[query, country, articles_state],
outputs=search_outputs
)
# ๋ถ„์„ ๋ฒ„ํŠผ ํด๋ฆญ ์ด๋ฒคํŠธ ์„ค์ •
for idx, comp in enumerate(article_components):
def create_analyze_function(index=idx):
def analyze_article(articles):
with gr.Progress() as progress:
progress(0, desc="์ฒ˜๋ฆฌ์ค‘์ž…๋‹ˆ๋‹ค. ์ž ์‹œ๋งŒ ๊ธฐ๋‹ค๋ฆฌ์„ธ์š”.")
if articles and index < len(articles):
article = articles[index]
summary = summarize_article(article['title'], article['snippet'])
return gr.update(value=summary, visible=True)
else:
return gr.update(value="๊ธฐ์‚ฌ ์ •๋ณด๋ฅผ ์ฐพ์„ ์ˆ˜ ์—†์Šต๋‹ˆ๋‹ค.", visible=True)
return analyze_article
comp['analyze_button'].click(
create_analyze_function(),
inputs=[articles_state],
outputs=comp['summary_output']
)
iface.launch(auth=("gini", "pick"))