completed CA projection
This commit is contained in:
parent
cf0b33c940
commit
d3e13f4c56
|
@ -4,47 +4,55 @@ from pandas import DataFrame
|
|||
import numpy as np
|
||||
import os
|
||||
import s3fs
|
||||
import re
|
||||
from sklearn.linear_model import LogisticRegression
|
||||
from sklearn.ensemble import RandomForestClassifier
|
||||
from sklearn.metrics import accuracy_score, confusion_matrix, classification_report, recall_score
|
||||
from sklearn.utils import class_weight
|
||||
from sklearn.neighbors import KNeighborsClassifier
|
||||
from sklearn.pipeline import Pipeline
|
||||
from sklearn.compose import ColumnTransformer
|
||||
from sklearn.preprocessing import OneHotEncoder
|
||||
from sklearn.impute import SimpleImputer
|
||||
from sklearn.model_selection import GridSearchCV
|
||||
from sklearn.preprocessing import StandardScaler, MaxAbsScaler, MinMaxScaler
|
||||
from sklearn.metrics import make_scorer, f1_score, balanced_accuracy_score
|
||||
import seaborn as sns
|
||||
import matplotlib.pyplot as plt
|
||||
from sklearn.metrics import roc_curve, auc, precision_recall_curve, average_precision_score
|
||||
from sklearn.exceptions import ConvergenceWarning, DataConversionWarning
|
||||
from sklearn.naive_bayes import GaussianNB
|
||||
from scipy.optimize import fsolve
|
||||
import pickle
|
||||
import warnings
|
||||
import io
|
||||
|
||||
# importation of functions defined
|
||||
from utils_CA_segment import *
|
||||
|
||||
# Create filesystem object
|
||||
S3_ENDPOINT_URL = "https://" + os.environ["AWS_S3_ENDPOINT"]
|
||||
fs = s3fs.S3FileSystem(client_kwargs={'endpoint_url': S3_ENDPOINT_URL})
|
||||
|
||||
# define type of activity
|
||||
type_of_activity = "sport"
|
||||
PATH = f"projet-bdc2324-team1/Output_expected_CA/{type_of_activity}/"
|
||||
|
||||
# type of model for the score
|
||||
type_of_model = "LogisticRegression_cv"
|
||||
|
||||
# load train and test sets
|
||||
dataset_train, dataset_test = load_train_test(type_of_activity)
|
||||
|
||||
# make features - define X train and X test
|
||||
X_train, X_test, y_train, y_test = features_target_split(dataset_train, dataset_test)
|
||||
|
||||
# choose model - logit cross validated
|
||||
model = load_model(type_of_activity, type_of_model)
|
||||
|
||||
# create table X test segment from X test
|
||||
X_test_segment = df_segment(X_test, y_test, model)
|
||||
|
||||
# comparison with bias of the train set - X train to be defined
|
||||
X_train_score = logit_cv.predict_proba(X_train)[:, 1]
|
||||
X_train_score = model.predict_proba(X_train)[:, 1]
|
||||
|
||||
bias_train_set = find_bias(odd_ratios = odd_ratio(adjust_score_1(X_train_score)),
|
||||
y_objective = y_train["y_has_purchased"].sum(),
|
||||
initial_guess=6)
|
||||
|
||||
# create a score adjusted with the bias computed
|
||||
score_adjusted_train = adjusted_score(odd_ratio(adjust_score_1(X_test_segment["score"])), bias = bias_train_set)
|
||||
X_test_segment["score_adjusted"] = score_adjusted_train
|
||||
|
||||
|
||||
# plot adjusted scores and save (to be tested)
|
||||
### 1. plot adjusted scores and save (to be tested)
|
||||
plot_hist_scores(X_test_segment, score = "score", score_adjusted = "score_adjusted", type_of_activity = type_of_activity)
|
||||
save_file_s3_ca("hist_score_adjusted_", type_of_activity)
|
||||
|
||||
"""
|
||||
image_buffer = io.BytesIO()
|
||||
plt.savefig(image_buffer, format='png')
|
||||
image_buffer.seek(0)
|
||||
|
@ -53,27 +61,33 @@ FILE_PATH_OUT_S3 = PATH + file_name + type_of_activity + ".png"
|
|||
with fs.open(FILE_PATH_OUT_S3, 'wb') as s3_file:
|
||||
s3_file.write(image_buffer.read())
|
||||
plt.close()
|
||||
"""
|
||||
|
||||
# comparison between score and adjusted score
|
||||
### 2. comparison between score and adjusted score
|
||||
X_test_table_adjusted_scores = (100 * X_test_segment.groupby("quartile")[["score","score_adjusted", "has_purchased"]].mean()).round(2).reset_index()
|
||||
X_test_table_adjusted_scores = X_test_table_adjusted_scores.rename(columns = {col : f"{col} (%)" for col in X_test_table_adjusted_scores.columns if col in ["score","score_adjusted", "has_purchased"]})
|
||||
|
||||
file_name = "table_adjusted_score"
|
||||
# save table
|
||||
file_name = "table_adjusted_score_"
|
||||
FILE_PATH_OUT_S3 = PATH + file_name + type_of_activity + ".csv"
|
||||
with fs.open(FILE_PATH_OUT_S3, 'w') as file_out:
|
||||
X_test_table_adjusted_scores.to_csv(file_out, index = False)
|
||||
|
||||
|
||||
# project revenue
|
||||
X_test_segment = project_tickets_CA (X_test_segment, "nb_tickets", "total_amount", "score_adjusted", duration_ref=1.5, duration_projection=1)
|
||||
X_test_segment = project_tickets_CA (X_test_segment, "nb_tickets", "total_amount", "score_adjusted", duration_ref=17, duration_projection=12)
|
||||
|
||||
|
||||
# table summarizing projections
|
||||
### 3. table summarizing projections (nb tickets, revenue)
|
||||
X_test_expected_CA = round(summary_expected_CA(df=X_test_segment, segment="quartile", nb_tickets_expected="nb_tickets_expected", total_amount_expected="total_amount_expected", total_amount="total_amount"),2)
|
||||
|
||||
file_name = "table_expected_CA"
|
||||
# rename columns
|
||||
mapping_dict = {col: col.replace("perct", "(%)").replace("_", " ") for col in X_test_expected_CA.columns}
|
||||
X_test_expected_CA = X_test_expected_CA.rename(columns=mapping_dict)
|
||||
|
||||
# save table
|
||||
file_name = "table_expected_CA_"
|
||||
FILE_PATH_OUT_S3 = PATH + file_name + type_of_activity + ".csv"
|
||||
with fs.open(FILE_PATH_OUT_S3, 'w') as file_out:
|
||||
X_test_expected_CA.to_csv(file_out, index = False)
|
||||
|
||||
|
||||
|
|
File diff suppressed because one or more lines are too long
|
@ -1,3 +1,83 @@
|
|||
# importations
|
||||
import pandas as pd
|
||||
from pandas import DataFrame
|
||||
import numpy as np
|
||||
import os
|
||||
import s3fs
|
||||
import matplotlib.pyplot as plt
|
||||
from scipy.optimize import fsolve
|
||||
import pickle
|
||||
import warnings
|
||||
import io
|
||||
|
||||
# functions
|
||||
|
||||
def load_train_test(type_of_activity):
|
||||
BUCKET = f"projet-bdc2324-team1/Generalization/{type_of_activity}"
|
||||
File_path_train = BUCKET + "/Train_set.csv"
|
||||
File_path_test = BUCKET + "/Test_set.csv"
|
||||
|
||||
with fs.open( File_path_train, mode="rb") as file_in:
|
||||
dataset_train = pd.read_csv(file_in, sep=",")
|
||||
# dataset_train['y_has_purchased'] = dataset_train['y_has_purchased'].fillna(0)
|
||||
|
||||
with fs.open(File_path_test, mode="rb") as file_in:
|
||||
dataset_test = pd.read_csv(file_in, sep=",")
|
||||
# dataset_test['y_has_purchased'] = dataset_test['y_has_purchased'].fillna(0)
|
||||
|
||||
return dataset_train, dataset_test
|
||||
|
||||
|
||||
def features_target_split(dataset_train, dataset_test):
|
||||
|
||||
features_l = ['nb_tickets', 'nb_purchases', 'total_amount', 'nb_suppliers', 'vente_internet_max', 'purchase_date_min', 'purchase_date_max',
|
||||
'time_between_purchase', 'nb_tickets_internet', 'fidelity', 'is_email_true', 'opt_in', #'is_partner',
|
||||
'gender_female', 'gender_male', 'gender_other', 'nb_campaigns', 'nb_campaigns_opened']
|
||||
|
||||
# we suppress fidelity, time between purchase, and gender other (colinearity issue)
|
||||
"""
|
||||
features_l = ['nb_tickets', 'nb_purchases', 'total_amount', 'nb_suppliers', 'vente_internet_max',
|
||||
'purchase_date_min', 'purchase_date_max', 'nb_tickets_internet', 'is_email_true',
|
||||
'opt_in', 'gender_female', 'gender_male', 'nb_campaigns', 'nb_campaigns_opened']
|
||||
"""
|
||||
|
||||
X_train = dataset_train[features_l]
|
||||
y_train = dataset_train[['y_has_purchased']]
|
||||
|
||||
X_test = dataset_test[features_l]
|
||||
y_test = dataset_test[['y_has_purchased']]
|
||||
|
||||
return X_train, X_test, y_train, y_test
|
||||
|
||||
|
||||
def load_model(type_of_activity, model):
|
||||
BUCKET = f"projet-bdc2324-team1/Output_model/{type_of_activity}/{model}/"
|
||||
filename = model + '.pkl'
|
||||
file_path = BUCKET + filename
|
||||
with fs.open(file_path, mode="rb") as f:
|
||||
model_bytes = f.read()
|
||||
|
||||
model = pickle.loads(model_bytes)
|
||||
return model
|
||||
|
||||
|
||||
def df_segment(df, y, model) :
|
||||
|
||||
y_pred = model.predict(df)
|
||||
y_pred_prob = model.predict_proba(df)[:, 1]
|
||||
|
||||
df_segment = df
|
||||
|
||||
df_segment["has_purchased"] = y
|
||||
df_segment["has_purchased_estim"] = y_pred
|
||||
df_segment["score"] = y_pred_prob
|
||||
df_segment["quartile"] = np.where(df_segment['score']<0.25, '1',
|
||||
np.where(df_segment['score']<0.5, '2',
|
||||
np.where(df_segment['score']<0.75, '3', '4')))
|
||||
|
||||
return df_segment
|
||||
|
||||
|
||||
def odd_ratio(score) :
|
||||
"""
|
||||
Args:
|
||||
|
@ -152,3 +232,14 @@ def summary_expected_CA(df, segment, nb_tickets_expected, total_amount_expected,
|
|||
df_expected_CA["pace_purchase"] = df_drop_null_pace.groupby(segment)[pace_purchase].mean().values
|
||||
|
||||
return df_expected_CA
|
||||
|
||||
|
||||
def save_file_s3_ca(File_name, type_of_activity):
|
||||
image_buffer = io.BytesIO()
|
||||
plt.savefig(image_buffer, format='png')
|
||||
image_buffer.seek(0)
|
||||
PATH = f"projet-bdc2324-team1/Output_expected_CA/{type_of_activity}/"
|
||||
FILE_PATH_OUT_S3 = PATH + File_name + type_of_activity + '.png'
|
||||
with fs.open(FILE_PATH_OUT_S3, 'wb') as s3_file:
|
||||
s3_file.write(image_buffer.read())
|
||||
plt.close()
|
||||
|
|
Loading…
Reference in New Issue
Block a user