forked from 1918190/COPF
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathDataHandler.py
More file actions
207 lines (177 loc) · 6.41 KB
/
DataHandler.py
File metadata and controls
207 lines (177 loc) · 6.41 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
import pickle
import numpy as np
from scipy.sparse import csr_matrix
from Params import args
import scipy.sparse as sp
from Utils.TimeLogger import log
from time import time
def transpose(mat):
coomat = sp.coo_matrix(mat)
return csr_matrix(coomat.transpose())
def negSamp(temLabel, sampSize, nodeNum):
negset = [None] * sampSize
cur = 0
while cur < sampSize:
rdmItm = np.random.choice(nodeNum)
if temLabel[rdmItm] == 0:
negset[cur] = rdmItm
cur += 1
return negset
def transToLsts(mat, mask=False, norm=False):
shape = [mat.shape[0], mat.shape[1]]
coomat = sp.coo_matrix(mat)
indices = np.array(list(map(list, zip(coomat.row, coomat.col))), dtype=np.int32)
data = coomat.data.astype(np.float32)
if norm:
rowD = np.squeeze(np.array(1 / (np.sqrt(np.sum(mat, axis=1) + 1e-8) + 1e-8)))
colD = np.squeeze(np.array(1 / (np.sqrt(np.sum(mat, axis=0) + 1e-8) + 1e-8)))
for i in range(len(data)):
row = indices[i, 0]
col = indices[i, 1]
data[i] = data[i] * rowD[row] * colD[col]
if mask:
spMask = (np.random.uniform(size=data.shape) > 0.5) * 1.0
data = data * spMask
if indices.shape[0] == 0:
indices = np.array([[0, 0]], dtype=np.int32)
data = np.array([0.0], np.float32)
return indices, data, shape
class DataHandler:
def __init__(self):
if args.data == 'beibei':
predir = './Datasets/beibei/'
behs = ['pv', 'cart', 'buy']
elif args.data == 'taobao':
predir = './Datasets/taobao/'
behs = ['pv', 'cart', 'buy']
elif args.data == 'tmall':
predir = './Datasets/tmall/'
behs = ['click', 'collect', 'cart', 'buy']
self.predir = predir
self.behs = behs
self.trnfile = predir + 'trn_'
self.tstfile = predir + 'tst_'
self.adj_file = predir + 'adj_'
def LoadData(self):
trnMats = list()
trnMats_uni_final = list()
for i in range(len(self.behs)):
beh = self.behs[i]
path = self.trnfile + beh
with open(path, 'rb') as fs:
mat = (pickle.load(fs) != 0).astype(np.float32)
trnMats.append(mat)
path = self.tstfile + 'int'
with open(path, 'rb') as fs:
tstInt = np.array(pickle.load(fs))
tstStat = (tstInt != None)
tstUsrs = np.reshape(np.argwhere(tstStat != False), [-1])
for i in range(len(self.behs)):
if i == 0:
self.all_trnMats = trnMats[0]
else:
self.all_trnMats += trnMats[i]
self.three_trnMats = (trnMats[0] + trnMats[1] + trnMats[2]).astype(np.float32)
self.all_trnMats = (self.all_trnMats).astype(np.float32)
self.two_trnMats = (trnMats[0] + trnMats[1]).astype(np.float32)
self.trnMats = trnMats
self.tstInt = tstInt
self.tstUsrs = tstUsrs
try:
self.trnMats_uni_final = np.load('trnMats_uni_final_'+str(args.data)+'.npy',allow_pickle=True)
except:
for i in range(len(trnMats)):
trnMats_uni = []
for j in range(len(trnMats)):
if args.data == 'tmall':
if i == 2 or j == 2:
trnMats_uni.append(0)
continue
if i == j:
trnMats_uni.append(0)
continue
if len((trnMats[j] == 0).multiply(trnMats[i]).data) == 0:
trnMats_uni.append(0)
else:
trnMats_uni.append((trnMats[j] == 0).multiply(trnMats[i]))
trnMats_uni_final.append(trnMats_uni)
np.save('trnMats_uni_final_'+str(args.data)+'.npy',trnMats_uni_final)
self.trnMats_uni_final = trnMats_uni_final
args.user, args.item = self.trnMats[0].shape
args.trnNum = args.user
args.behNum = len(self.behs)
def get_adj_mat(self):
ori_adj, left_loop_adj, left_adj, symm_adj = [], [], [], []
try:
t1 = time()
for i in range(args.behNum):
beh = self.behs[i]
path = self.adj_file + beh
ori_adj_mat = sp.load_npz(path + '_ori.npz')
norm_adj_mat = sp.load_npz(path + '_norm_.npz')
mean_adj_mat = sp.load_npz(path + '_mean.npz')
ori_adj.append(ori_adj_mat)
left_loop_adj.append(norm_adj_mat)
left_adj.append(mean_adj_mat)
print('already load adj matrix', ori_adj_mat.shape, time() - t1)
except Exception:
for i in range(args.behNum):
beh = self.behs[i]
path = self.adj_file + beh
ori_adj_mat, norm_adj_mat, mean_adj_mat = self.create_adj_mat(self.trnMats[i])
sp.save_npz(path + '_ori.npz', ori_adj_mat)
sp.save_npz(path + '_norm_.npz', norm_adj_mat)
sp.save_npz(path + '_mean.npz', mean_adj_mat)
ori_adj.append(ori_adj_mat)
left_loop_adj.append(norm_adj_mat)
left_adj.append(mean_adj_mat)
print('already load adj matrix', ori_adj_mat.shape, time() - t1)
try:
for i in range(args.behNum):
beh = self.behs[i]
path = self.adj_file + beh
pre_adj_mat = sp.load_npz(path + '_pre.npz')
symm_adj.append(pre_adj_mat)
except Exception:
for i in range(args.behNum):
beh = self.behs[i]
path = self.adj_file + beh
rowsum = np.array(ori_adj_mat.sum(1))
d_inv = np.power(rowsum, -0.5).flatten()
d_inv[np.isinf(d_inv)] = 0.
d_mat_inv = sp.diags(d_inv)
norm_adj = d_mat_inv.dot(ori_adj_mat)
norm_adj = norm_adj.dot(d_mat_inv)
print('generate pre adjacency matrix.')
pre_adj_mat = norm_adj.tocsr()
sp.save_npz(path + '_pre.npz', pre_adj_mat)
symm_adj.append(pre_adj_mat)
return ori_adj, left_loop_adj, left_adj, symm_adj
def create_adj_mat(self, which_R):
t1 = time()
adj_mat = sp.dok_matrix((args.user + args.item, args.user + args.item), dtype=np.float32)
adj_mat = adj_mat.tolil()
R = which_R.tolil()
adj_mat[:args.user, args.user:] = R
adj_mat[args.user:, :args.user] = R.T
adj_mat = adj_mat.todok()
print('already create adjacency matrix', adj_mat.shape, time() - t1)
t2 = time()
def normalized_adj_single(adj):
rowsum = np.array(adj.sum(1))
d_inv = np.power(rowsum, -1).flatten()
d_inv[np.isinf(d_inv)] = 0.
d_mat_inv = sp.diags(d_inv)
norm_adj = d_mat_inv.dot(adj)
print('generate single-normalized adjacency matrix.')
return norm_adj.tocoo()
def check_adj_if_equal(adj):
dense_A = np.array(adj.todense())
degree = np.sum(dense_A, axis=1, keepdims=False)
temp = np.dot(np.diag(np.power(degree, -1)), dense_A)
print('check normalized adjacency matrix whether equal to this laplacian matrix.')
return temp
norm_adj_mat = normalized_adj_single(adj_mat + sp.eye(adj_mat.shape[0]))
mean_adj_mat = normalized_adj_single(adj_mat)
print('already normalize adjacency matrix', time() - t2)
return adj_mat.tocsr(), norm_adj_mat.tocsr(), mean_adj_mat.tocsr()