-
Notifications
You must be signed in to change notification settings - Fork 0
/
Shawn_help.py
108 lines (91 loc) · 2.75 KB
/
Shawn_help.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
'''
-- OSU --
--SHAWN HELP--
Shengxuan Wang
wangshe@oregonstate.edu
'''
import pandas as pd
import numpy as np
from sklearn.ensemble import RandomForestClassifier
from sklearn.neural_network import MLPClassifier
from sklearn.tree import DecisionTreeClassifier
from sklearn.neighbors import KNeighborsClassifier
from sklearn.preprocessing import StandardScaler
from sklearn.dummy import DummyClassifier
from sklearn.preprocessing import OneHotEncoder
from sklearn.linear_model import LogisticRegression
# A. Tools
# split
def split_xy(df, NameOfY):
real_y = df[NameOfY]
x = df.drop([NameOfY], axis=1)
return (x, real_y)
# one hot encoder
def onehot_help(x, str_list):
enc = OneHotEncoder(handle_unknown='ignore')
enc.fit(x[str_list].values)
# the result data
result = enc.transform(x[str_list].values).toarray()
# the original label of them
# labels = np.array(enc.categories_).ravel()
col_names = []
for col in str_list:
for val in x[col].unique():
col_names.append("{}_{}".format(col, val))
return pd.DataFrame(data = result, columns=col_names, dtype=int)
def onehot(x, str_list, non_onehot_list):
x_oh = onehot_help(x, str_list)
df = pd.concat([x_oh, x[non_onehot_list]], axis=1)
return df
# round score
def Rscore(clf, x, y):
original_score = clf.score(x, y)
rscore = format(original_score, '.4f')
return rscore
# B. Classifiers
RS = 22 # random state
# 1 Dummy majority
def dummyM(x, y):
clf = DummyClassifier(strategy="most_frequent", random_state=RS)
clf.fit(x, y)
return clf
# 2 Dummy distribution
def dummyD(x, y):
clf = DummyClassifier(strategy="stratified", random_state=RS)
clf.fit(x, y)
return clf
# 3 RandomForestClassifier
def forest(x, y, n=50, classW = None):
clf = RandomForestClassifier(n_estimators=n, random_state=RS, class_weight=classW)
clf = clf.fit(x, y)
return clf
# 4 KNN
def KNN(x, y):
clf = KNeighborsClassifier(n_neighbors=10)
clf = clf.fit(x, y)
return clf
# Scale the data
def scale(x):
cols = x.columns
scaler = StandardScaler()
scaler.fit(x)
x = scaler.transform(x)
# avoid warning, transform it back to df
x = pd.DataFrame(x, columns=cols)
return x
# 5 Multi-layer Perceptron Classifier (with scale the data)
def MLPClf(x, y):
x = scale(x)
clf = MLPClassifier(solver='sgd', alpha=1e-5, max_iter=400, hidden_layer_sizes=(5,), random_state=RS)
clf = clf.fit(x, y)
return clf
# 6 Decision Trees
def DTree(x, y):
clf = DecisionTreeClassifier(random_state=RS)
clf.fit(x,y)
return clf
# 7 Logistic Regression
def LR(x, y, p, s, classW = None):
clf = LogisticRegression(random_state=RS, penalty=p, solver=s, class_weight=classW)
clf.fit(x, y)
return clf