【发布时间】:2019-09-19 09:23:51
【问题描述】:
我有一个 keras 模型,我正在尝试使用大数据集块进行训练
for chunk in pd.read_csv(input_file, chunksize=chunk_size, usecols = FEATURE_COLUMNS, low_memory = False):
(X_train, y_train, X_val, y_val,full_pipeline) = dataPrep.get_data(data=chunk, mean=mean, variance=variance)
print("Clicks in Training Set => {} , in CV => {}".format(np.sum(y_train == 1),np.sum(y_val == 1)))
print("SMS in Training Set => {} , in CV => {}".format(np.sum(y_train == 0),np.sum(y_val == 0)))
preprocessor = full_pipeline
model.fit(X_train, y_train, batch_size=batch_size, validation_data=(X_val, y_val), epochs=epochs, verbose=1)
score = model.evaluate(X_val, y_val, verbose=0)
print('Validation Metrics :', score)
在整个数据集上训练后,有什么方法可以测量模型的 f1 分数吗?而不是每个块的性能
【问题讨论】:
标签: python pandas validation keras deep-learning