-
Notifications
You must be signed in to change notification settings - Fork 0
/
accuracy.py
43 lines (32 loc) · 1.18 KB
/
accuracy.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
#from CPNet import CPNet
#a=CPNet()
#a.initFromFile("xml/examples5_alldiff/cpnet_n5c4d2_0014.xml")
#a.getPartialOrder()
from sklearn.metrics import confusion_matrix, precision_score, recall_score, f1_score, cohen_kappa_score
import numpy as np
dir="/home/aloreggi/cpnet/xml/lists/results/examples3/"
f1_mean=0.0
k_mean=0.0
mae = 0.0
all_results = []
for i in range(10):
file="test_NOPRETRAIN_UNBALANCED_DIM_3_21_70_FOLD_"+str(i)+".txt.npy"
#print(file)
l=np.load(dir+file)
#print(np.shape(l))
if all_results == []:
all_results = l
else:
all_results = np.concatenate((all_results,l),axis=0)
'''
print(np.sum(l[:,0]==l[:,1],axis=0))
print(len(l))
mae += (float(np.sum(abs(l[:,0]-l[:,1]),axis=0)) / len(l))
'''
print("F1-score: \t" + str(f1_score(l[:,0], l[:,1],average="micro")))
print("Cohen k: \t" + str(cohen_kappa_score(l[:,0], l[:,1])))
print("")
mae += (float(np.sum(abs(l[:,0]-l[:,1]),axis=0)) / len(l))
print("F1-score: \t" + str(f1_score(all_results[:,0], all_results[:,1],average="micro")))
print("Cohen k: \t" + str(cohen_kappa_score(all_results[:,0], all_results[:,1])))
print("MAE: \t\t"+ str(mae / 10.0))