@@ -151,7 +151,7 @@ def validation_step(x_val, y_val, writer=None):
151151
152152 eval_pre_tk = [0.0 ] * args .topK
153153 eval_rec_tk = [0.0 ] * args .topK
154- eval_F_tk = [0.0 ] * args .topK
154+ eval_F1_tk = [0.0 ] * args .topK
155155
156156 true_onehot_labels = []
157157 predicted_onehot_scores = []
@@ -202,8 +202,8 @@ def validation_step(x_val, y_val, writer=None):
202202 y_pred = np .array (predicted_onehot_labels_ts ), average = 'micro' )
203203 eval_rec_ts = recall_score (y_true = np .array (true_onehot_labels ),
204204 y_pred = np .array (predicted_onehot_labels_ts ), average = 'micro' )
205- eval_F_ts = f1_score (y_true = np .array (true_onehot_labels ),
206- y_pred = np .array (predicted_onehot_labels_ts ), average = 'micro' )
205+ eval_F1_ts = f1_score (y_true = np .array (true_onehot_labels ),
206+ y_pred = np .array (predicted_onehot_labels_ts ), average = 'micro' )
207207
208208 for top_num in range (args .topK ):
209209 eval_pre_tk [top_num ] = precision_score (y_true = np .array (true_onehot_labels ),
@@ -212,9 +212,9 @@ def validation_step(x_val, y_val, writer=None):
212212 eval_rec_tk [top_num ] = recall_score (y_true = np .array (true_onehot_labels ),
213213 y_pred = np .array (predicted_onehot_labels_tk [top_num ]),
214214 average = 'micro' )
215- eval_F_tk [top_num ] = f1_score (y_true = np .array (true_onehot_labels ),
216- y_pred = np .array (predicted_onehot_labels_tk [top_num ]),
217- average = 'micro' )
215+ eval_F1_tk [top_num ] = f1_score (y_true = np .array (true_onehot_labels ),
216+ y_pred = np .array (predicted_onehot_labels_tk [top_num ]),
217+ average = 'micro' )
218218
219219 # Calculate the average AUC
220220 eval_auc = roc_auc_score (y_true = np .array (true_onehot_labels ),
@@ -223,8 +223,8 @@ def validation_step(x_val, y_val, writer=None):
223223 eval_prc = average_precision_score (y_true = np .array (true_onehot_labels ),
224224 y_score = np .array (predicted_onehot_scores ), average = 'micro' )
225225
226- return eval_loss , eval_auc , eval_prc , eval_rec_ts , eval_pre_ts , eval_F_ts , \
227- eval_rec_tk , eval_pre_tk , eval_F_tk
226+ return eval_loss , eval_auc , eval_prc , eval_pre_ts , eval_rec_ts , eval_F1_ts , \
227+ eval_pre_tk , eval_rec_tk , eval_F1_tk
228228
229229 # Generate batches
230230 batches_train = dh .batch_iter (
@@ -241,21 +241,21 @@ def validation_step(x_val, y_val, writer=None):
241241 if current_step % args .evaluate_steps == 0 :
242242 logger .info ("\n Evaluation:" )
243243 eval_loss , eval_auc , eval_prc , \
244- eval_rec_ts , eval_pre_ts , eval_F_ts , eval_rec_tk , eval_pre_tk , eval_F_tk = \
244+ eval_pre_ts , eval_rec_ts , eval_F1_ts , eval_pre_tk , eval_rec_tk , eval_F1_tk = \
245245 validation_step (x_val , y_val , writer = validation_summary_writer )
246246
247247 logger .info ("All Validation set: Loss {0:g} | AUC {1:g} | AUPRC {2:g}"
248248 .format (eval_loss , eval_auc , eval_prc ))
249249
250250 # Predict by threshold
251- logger .info ("Predict by threshold: Precision {0:g}, Recall {1:g}, F {2:g}"
252- .format (eval_pre_ts , eval_rec_ts , eval_F_ts ))
251+ logger .info ("Predict by threshold: Precision {0:g}, Recall {1:g}, F1 {2:g}"
252+ .format (eval_pre_ts , eval_rec_ts , eval_F1_ts ))
253253
254254 # Predict by topK
255255 logger .info ("Predict by topK:" )
256256 for top_num in range (args .topK ):
257- logger .info ("Top{0}: Precision {1:g}, Recall {2:g}, F {3:g}"
258- .format (top_num + 1 , eval_pre_tk [top_num ], eval_rec_tk [top_num ], eval_F_tk [top_num ]))
257+ logger .info ("Top{0}: Precision {1:g}, Recall {2:g}, F1 {3:g}"
258+ .format (top_num + 1 , eval_pre_tk [top_num ], eval_rec_tk [top_num ], eval_F1_tk [top_num ]))
259259 best_saver .handle (eval_prc , sess , current_step )
260260 if current_step % args .checkpoint_steps == 0 :
261261 checkpoint_prefix = os .path .join (checkpoint_dir , "model" )
0 commit comments