brendabor's picture
Update app.py
46994ca
raw
history blame contribute delete
No virus
4.3 kB
import streamlit as st
import numpy as np
from tensorflow.keras.models import load_model
from tensorflow.keras.preprocessing.text import Tokenizer
from tensorflow.keras.preprocessing.sequence import pad_sequences
import pickle
import joblib
import pandas as pd
from sklearn.neighbors import NearestNeighbors
from sklearn.preprocessing import StandardScaler
from sklearn.metrics.pairwise import cosine_similarity
# Load the LSTM model
lstm_model = load_model('lstm_model.h5')
# Load the Tokenizer used during training
with open('tokenizer.pkl', 'rb') as tokenizer_file:
tokenizer = pickle.load(tokenizer_file)
class_mapping = {"Happy": 0, "Sad": 1, "Calm": 2, "Anger": 3}
numerical_to_label = {v: k for k, v in class_mapping.items()}
# Load the KNN model
knn_model = joblib.load('knn_model.joblib')
# Load the dataset
df = pd.read_csv('df1.csv')
df = df.dropna()
# Preprocess for KNN
audio_feature_columns = ['danceability', 'energy', 'key', 'loudness', 'mode', 'speechiness',
'acousticness', 'instrumentalness', 'liveness', 'valence', 'tempo',
'duration_ms', 'time_signature']
audio_features = df[audio_feature_columns]
mood_cats = df[['mood_cats']]
audio_features_scaled = StandardScaler().fit_transform(audio_features)
audio_features_df = pd.DataFrame(audio_features_scaled, columns=audio_feature_columns)
combined_features = pd.concat([mood_cats, audio_features_df], axis=1)
# Calculate similarity matrix for content-based
audio_features_scaled_content = StandardScaler().fit_transform(audio_features)
combined_features_content = pd.concat([mood_cats, pd.DataFrame(audio_features_scaled_content)], axis=1)
similarity_matrix = cosine_similarity(combined_features_content)
def recommend_cont(song_index, num_recommendations=5):
song_similarity = similarity_matrix[song_index]
similar_songs = sorted(list(enumerate(song_similarity)), key=lambda x: x[1], reverse=True)[1:num_recommendations+1]
recommended_song_indices = [idx for idx, similarity in similar_songs]
recommended_songs = df.iloc[recommended_song_indices].copy()
recommended_songs['score'] = [similarity for idx, similarity in similar_songs]
return recommended_songs
def recommend_knn(query_index, n_recommendations=5):
distances, indices = knn_model.kneighbors(combined_features.iloc[query_index].values.reshape(1, -1), n_neighbors=n_recommendations)
recommended_songs = df.iloc[indices.flatten()].copy()
recommended_songs['score'] = 1 / (1 + distances.flatten())
return recommended_songs
def hybrid_recommendation(song_index, top_n=10):
content_based_recs = recommend_cont(song_index, top_n)
knn_based_recs = recommend_knn(song_index, top_n)
combined_recs = pd.concat([content_based_recs, knn_based_recs])
# Convert 'score' column to numeric
combined_recs['score'] = pd.to_numeric(combined_recs['score'], errors='coerce')
# Use maximum value for each group instead of mean
hybrid_recs = combined_recs.groupby(combined_recs.index)['score'].max().sort_values(ascending=False).head(top_n)
return hybrid_recs
# Streamlit app
st.title('Music Recommendation and Emotion Detection')
# Emotion Detection
st.subheader('Emotion Detection from Song Lyrics')
user_input = st.text_input('Enter a Text:')
prediction = None
if st.button('Predict Emotion'):
sequence = tokenizer.texts_to_sequences([user_input])
padded_sequence = pad_sequences(sequence, maxlen=50)
prediction = lstm_model.predict(padded_sequence)
if prediction is not None:
for i in range(len(prediction[0])):
label = numerical_to_label[i]
probability = prediction[0][i]
threshold = 0.5
if probability > threshold:
st.write(f'Predicted Emotion: {label}')
# Music Recommendation
st.subheader('Music Recommendation')
song_index_to_recommend = st.number_input('Enter song index:', min_value=0, max_value=len(df)-1, value=0)
hybrid_recs = hybrid_recommendation(song_index_to_recommend)
# Display hybrid recommendations
st.write("Hybrid Recommendations:")
if not hybrid_recs.empty:
for idx in hybrid_recs.index:
st.write(f"Song Index: {idx}, Title: {df.iloc[idx]['track_name']}, Artist: {df.iloc[idx]['track_artist']}, Score: {hybrid_recs.loc[idx]}")
else:
st.write("No recommendations found.")