Professional Documents
Culture Documents
import numpy as np
import sklearn
train=pd.read_csv('/data/training/diabetes_train.csv')
test=pd.read_csv('/data/test/diabetes_test.csv')
#.......
y=train.Outcome
x=train.drop('Outcome',axis=1)
#.....ADABoost.....
ada = AdaBoostClassifier()
'''
ada.fit(x,y)
y_pred=ada.predict(test.drop('Outcome',axis=1))
y_actual=test.Outcome
cfs = confusion_matrix(y_actual,y_pred)
acc = (cfs[0][0]+cfs[1][1])/(cfs[0][0]+cfs[0][1]+cfs[1][0]+cfs[1][1])
sn = cfs[0][0]/(cfs[0][0]+cfs[0][1])
sp = cfs[1][1]/(cfs[1][0]+cfs[1][1])
print(acc,sn,sp)
'''
#.....GradientBoost....
gdb = GradientBoostingClassifier()
'''
gdb.fit(x,y)
y_pred=gdb.predict(test.drop('Outcome',axis=1))
y_actual=test.Outcome
cfs = confusion_matrix(y_actual,y_pred)
acc = (cfs[0][0]+cfs[1][1])/(cfs[0][0]+cfs[0][1]+cfs[1][0]+cfs[1][1])
sn = cfs[0][0]/(cfs[0][0]+cfs[0][1])
sp = cfs[1][1]/(cfs[1][0]+cfs[1][1])
'''
#.....GridSearch........
for i in [ada,gdb]:
lr = np.arange(0.1,1.0,0.1)
nt = np.arange(50,300,50)
clf.fit(x,y)
print(i,clf.best_params_)
#.......
#'''
# Finally create a dataframe of the final output and write the output to output.csv
result=pd.DataFrame(result)