File size: 9,846 Bytes
736e798 ec929e1 32c9084 a906ccd f5cd439 35ef636 32c9084 fa65e69 954b204 6d6b1c1 8ecc66b 4dd480b 9a96f1a ae42482 f79a195 ae42482 f79a195 ae42482 f79a195 ae42482 9a96f1a ae42482 8655834 ae42482 9a96f1a 889b7a0 c1fb9cb 889b7a0 9a96f1a 66d3e3b ced2a1a 4dd480b 66d3e3b 6c88c2f 66d3e3b fe40899 eac3bfc 66d3e3b c8d7667 9a96f1a ec929e1 a906ccd 0e5f8d9 32c9084 9a96f1a 9e19736 0262183 9e19736 889b7a0 a906ccd 9e19736 eac3bfc 9e19736 f35d308 f2ef415 889b7a0 f5cd439 ae62e8e f5cd439 ae62e8e f5cd439 ae62e8e 954b204 f5cd439 ae62e8e f5cd439 ae62e8e f5cd439 ae62e8e f5cd439 ae62e8e f5cd439 ae62e8e f5cd439 ae62e8e 2f21070 ae62e8e 2f21070 7bbff5e de19eb1 7bbff5e de19eb1 b0d69bc de19eb1 2f21070 de19eb1 2f21070 de19eb1 2f21070 de19eb1 2f21070 de19eb1 2f21070 de19eb1 2f21070 de19eb1 f5cd439 b0d69bc f5cd439 f2ef415 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 |
# Simply Assorted Language Tools (SALT)
# Richard Orama - September 2024
#x = st.slider('Select a value')
#st.write(x, 'squared is', x * x)
import streamlit as st
from transformers import pipeline, GPT2Tokenizer, GPT2LMHeadModel
import ast
#st.title("Assorted Language Tools")
st.markdown("<h1 style='text-align: center; font-size: 30px;'>S A L T</h1>", unsafe_allow_html=True)
st.markdown("<h3 style='text-align: center; font-size: 16px;'>Simply Assorted Language Tools</h3>", unsafe_allow_html=True)
st.markdown("<h3 style='text-align: center; font-size: 20px; color: blue;'>Orama's AI Craze</h3>", unsafe_allow_html=True)
################ SENTIMENT ANALYSIS - side bar - pippeline #################
# Initialize the sentiment analysis pipeline
# No model was supplied, defaulted to distilbert-base-uncased-finetuned-sst-2-english
sentiment_pipeline = pipeline("sentiment-analysis")
def is_valid_list_string(string):
try:
result = ast.literal_eval(string)
return isinstance(result, list)
except (ValueError, SyntaxError):
return False
# Define the summarization function
def analyze_sentiment(txt):
#st.write('\n\n')
#st.write(txt[:100]) # Display the first 100 characters of the article
#st.write('--------------------------------------------------------------')
# Display the results
if is_valid_list_string(txt):
txt_converted = ast.literal_eval(txt) #convert string to actual content, e.g. list
# Perform Hugging sentiment analysis on multiple texts
results = sentiment_pipeline(txt_converted)
for i, text in enumerate(txt_converted):
st.sidebar.write(f"Text: {text}")
st.sidebar.write(f"Sentiment: {results[i]['label']}, Score: {results[i]['score']:.2f}\n")
else:
# Perform Hugging sentiment analysis on multiple texts
results = sentiment_pipeline(txt)
st.sidebar.write(f"Text: {txt}")
st.sidebar.write(f"Sentiment: {results[0]['label']}, Score: {results[0]['score']:.2f}\n")
st.sidebar.markdown("<h3 style='text-align: center; font-size: 16px; background-color: white; color: black;'>Sentiment Analysis - Pipeline</h3>", unsafe_allow_html=True)
DEFAULT_SENTIMENT = ""
# Create a text area for user input
SENTIMENT = st.sidebar.text_area('Enter Sentiment (String or List of Strings)', DEFAULT_SENTIMENT, height=150)
# Enable the button only if there is text in the SENTIMENT variable
if SENTIMENT:
if st.sidebar.button('Analyze Sentiment'):
analyze_sentiment(SENTIMENT) # Directly pass the SENTIMENT
else:
st.sidebar.button('Analyze Sentiment', disabled=True)
#st.warning('π Please enter Sentiment!')
################ STATEMENT SUMMARIZATION1 - side bar - tokenizer #################
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
# Load the summarization model and tokenizer
MODEL_NAME = "facebook/bart-large-cnn" # A commonly used summarization model
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
model = AutoModelForSeq2SeqLM.from_pretrained(MODEL_NAME)
st.sidebar.markdown("<h3 style='text-align: center; font-size: 16px; background-color: white; color: black;'>Text Summarization - BART Tokenizer</h3>", unsafe_allow_html=True)
DEFAULT_STATEMENT = ""
# Create a text area for user input
STATEMENT = st.sidebar.text_area('Enter Statement (String1)', DEFAULT_STATEMENT, height=150)
# Enable the button only if there is text in the SENTIMENT variable
if STATEMENT:
if st.sidebar.button('Summarize Statement1'):
# Call your Summarize function here
# summarize_statement(STATEMENT) # Directly pass the STATEMENT
# Tokenize input article
inputs = tokenizer(STATEMENT, return_tensors="pt", truncation=True, padding="longest", max_length=1024)
# Generate summary
summary_ids = model.generate(inputs["input_ids"], max_length=150, min_length=30, length_penalty=2.0, num_beams=4, early_stopping=True)
# Decode summary
summary = tokenizer.decode(summary_ids[0], skip_special_tokens=True)
# Display the summary
st.write("**Summary:**")
st.write(summary)
else:
st.sidebar.button('Summarize Statement1', disabled=True)
#st.warning('π Please enter Statement!')
################ STATEMENT SUMMARIZATION - side bar - pipeline #################
# Load the summarization model
#summarizer = pipeline("summarization", model="sshleifer/distilbart-cnn-12-6") # smaller version of the model
summarizer = pipeline("summarization", model="facebook/bart-large-cnn")
st.sidebar.markdown("<h3 style='text-align: center; font-size: 16px; background-color: white; color: black;'>Text Summarization - BART Pipeline</h3>", unsafe_allow_html=True)
DEFAULT_STATEMENT = ""
# Create a text area for user input
STATEMENT = st.sidebar.text_area('Enter Statement (String)', DEFAULT_STATEMENT, height=150)
# Enable the button only if there is text in the SENTIMENT variable
if STATEMENT:
if st.sidebar.button('Summarize Statement'):
# Call your Summarize function here
st.write('\n\n')
summary = summarizer(STATEMENT, max_length=500, min_length=30, do_sample=False)
st.write(summary[0]['summary_text'])
#summarize_statement(STATEMENT) # Directly pass the STATEMENT
else:
st.sidebar.button('Summarize Statement', disabled=True)
#st.warning('π Please enter Statement!')
# ################ CHAT BOT - main area #################
# # Load the GPT model
# generator = pipeline("text-generation", model="EleutherAI/gpt-neo-2.7B")
# # Streamlit chat UI
# #st.title("GPT-3 Chatbox")
# # user_input = st.text_input("You: ", "Hello, how are you?")
# # if user_input:
# # response = generator(user_input, max_length=100, num_return_sequences=1)[0]['generated_text']
# # st.write(f"GPT-3: {response}")
# # Define the summarization function
# def chat(txt):
# st.write('\n\n')
# #st.write(txt[:100]) # Display the first 100 characters of the article
# #st.write('--------------------------------------------------------------')
# #summary = summarizer(txt, max_length=500, min_length=30, do_sample=False)
# #st.write(summary[0]['summary_text'])
# response = generator(txt, max_length=500, num_return_sequences=1)[0]['generated_text']
# st.write(f"GPT-3: {response}")
# DEFAULT_CHAT = ""
# # Create a text area for user input
# CHAT = st.sidebar.text_area('Enter Chat (String)', DEFAULT_CHAT, height=150)
# # Enable the button only if there is text in the CHAT variable
# if CHAT:
# if st.sidebar.button('Chat Statement'):
# # Call your Summarize function here
# chat(CHAT) # Directly pass the your
# else:
# st.sidebar.button('Chat Statement', disabled=True)
# st.warning('π Please enter Chat!')
# Load pre-trained GPT-2 model and tokenizer
model_name = "gpt2-medium" # "gpt2" # Use "gpt-3.5-turbo" or another model from Hugging Face if needed
model = GPT2LMHeadModel.from_pretrained(model_name)
tokenizer = GPT2Tokenizer.from_pretrained(model_name)
# Initialize the text generation pipeline
gpt_pipeline = pipeline("text-generation", model=model, tokenizer=tokenizer)
# Streamlit UI
st.markdown("<h3 style='text-align: center; font-size: 20px; background-color: white; color: black;'>Chat with GPT</h3>", unsafe_allow_html=True)
if 'conversation' not in st.session_state:
st.session_state.conversation = ""
def chat_with_gpt(user_input):
# Append user input to the conversation
st.session_state.conversation += f"User: {user_input}\n"
# Generate response
response = gpt_pipeline(user_input, max_length=100, num_return_sequences=1)[0]['generated_text']
response_text = response.replace(user_input, '') # Strip the user input part from response
# Append GPT's response to the conversation
st.session_state.conversation += f"GPT: {response_text}\n"
return response_text
# Text input for user query
user_input = st.text_input("You:", "")
if st.button("Send"):
if user_input:
chat_with_gpt(user_input)
# Display conversation history
st.text_area("Conversation", value=st.session_state.conversation, height=400)
#############
# # LLaMA 7B model from Hugging Face
# # MODEL_NAME = "huggyllama/llama-7b" # Example of a LLaMA model
# # Try this OpenAssistant model available on Hugging Face
# MODEL_NAME = "OpenAssistant/oasst-sft-1-pythia-12b" # Example of an OpenAssistant model
# import streamlit as st
# from transformers import AutoModelForCausalLM, AutoTokenizer
# import torch
# # Load the model and tokenizer (OpenAssistant or LLaMA)
# MODEL_NAME = "OpenAssistant/oasst-sft-1-pythia-12b" # Replace with "huggyllama/llama-7b" for LLaMA
# tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
# model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
# # Streamlit UI for input
# st.markdown("<h3 style='text-align: center; font-size: 20px;'>Chat with OpenAssistant/LLaMA</h3>", unsafe_allow_html=True)
# # Input text area
# user_input = st.text_area("You:", "", height=150)
# if st.button('Generate Response'):
# if user_input:
# # Tokenize the input and generate response
# inputs = tokenizer(user_input, return_tensors="pt")
# outputs = model.generate(**inputs, max_length=150)
# # Decode the generated response
# response = tokenizer.decode(outputs[0], skip_special_tokens=True)
# # Display the model's response
# st.write("Assistant: ", response)
# else:
# st.warning('Please enter some text to get a response!')
# ################ END #################
# Add a footnote at the bottom
st.markdown("---") # Horizontal line to separate content from footnote
st.markdown("Orama's AI Craze") |