0
我想以最高的准确度保存模型,我需要在每个步骤中采取一批验证数据进行验证,以便在每一步训练之后,训练数据集将因时代而重复使用,但是如果train_batch_size
等于validation_batch_size
,验证数据集也将被重用?因为验证数据集远小于训练数据集。我应该怎么做?我的意思是重用验证集没有任何问题?或者我分别设置不同的尺寸。validation_batch_size等于训练CNN中的train_batch_size?
MAX_EPOCH = 10
for epoch in range(MAX_EPOCH):
# training
train_step = int(80000/TRAIN_BATCH_SIZE)
train_loss, train_acc = 0, 0
for step in range(epoch * train_step, (epoch + 1) * train_step):
x_train, y_train = sess.run([x_train_batch, y_train_batch])
train_summary, _, err, ac = sess.run([merged, train_op, loss, acc],
feed_dict={x: x_train, y_: y_train,
mode: learn.ModeKeys.TRAIN,
global_step: step})
train_loss += err
train_acc += ac
if (step + 1) % 100 == 0:
train_writer.add_summary(train_summary, step)
print("Epoch %d,train loss= %.2f,train accuracy=%.2f%%" % (
epoch, (train_loss/train_step), (train_acc/train_step * 100.0)))
# validation
val_step = int(20000/VAL_BATCH_SIZE)
val_loss, val_acc = 0, 0
for step in range(epoch * val_step, (epoch + 1) * val_step):
x_val, y_val = sess.run([x_val_batch, y_val_batch])
val_summary, err, ac = sess.run([merged, loss, acc],
feed_dict={x: x_val, y_: y_val, mode: learn.ModeKeys.EVAL,
global_step: step})
val_loss += err
val_acc += ac
if (step + 1) % 100 == 0:
valid_writer.add_summary(val_summary, step)
print(
"Epoch %d,validation loss= %.2f,validation accuracy=%.2f%%" % (
epoch, (val_loss/val_step), (val_acc/val_step * 100.0)))
我刚修改我的问题,根据你的意思添加一些代码,上面的代码是正确的?每个时代我都有20000个验证集样本迭代完成验证,最后保存模型的最精确时代?在另一个问题中,我展示了我所有的代码,也许你可以看看它[https://stackoverflow.com/questions/45953858/how-to-adjust-parameters-when-the-cnn-model-is-over -配件)。 – Gary