【发布时间】:2016-04-19 00:12:43
【问题描述】:
嘿,我是 tensorflow 的新手,即使经过很多努力也无法添加 L1 正则化项到误差项
x = tf.placeholder("float", [None, n_input])
# Weights and biases to hidden layer
ae_Wh1 = tf.Variable(tf.random_uniform((n_input, n_hidden1), -1.0 / math.sqrt(n_input), 1.0 / math.sqrt(n_input)))
ae_bh1 = tf.Variable(tf.zeros([n_hidden1]))
ae_h1 = tf.nn.tanh(tf.matmul(x,ae_Wh1) + ae_bh1)
ae_Wh2 = tf.Variable(tf.random_uniform((n_hidden1, n_hidden2), -1.0 / math.sqrt(n_hidden1), 1.0 / math.sqrt(n_hidden1)))
ae_bh2 = tf.Variable(tf.zeros([n_hidden2]))
ae_h2 = tf.nn.tanh(tf.matmul(ae_h1,ae_Wh2) + ae_bh2)
ae_Wh3 = tf.transpose(ae_Wh2)
ae_bh3 = tf.Variable(tf.zeros([n_hidden1]))
ae_h1_O = tf.nn.tanh(tf.matmul(ae_h2,ae_Wh3) + ae_bh3)
ae_Wh4 = tf.transpose(ae_Wh1)
ae_bh4 = tf.Variable(tf.zeros([n_input]))
ae_y_pred = tf.nn.tanh(tf.matmul(ae_h1_O,ae_Wh4) + ae_bh4)
ae_y_actual = tf.placeholder("float", [None,n_input])
meansq = tf.reduce_mean(tf.square(ae_y_actual - ae_y_pred))
train_step = tf.train.GradientDescentOptimizer(0.05).minimize(meansq)
在此之后,我使用
运行上面的图表init = tf.initialize_all_variables()
sess = tf.Session()
sess.run(init)
n_rounds = 100
batch_size = min(500, n_samp)
for i in range(100):
sample = np.random.randint(n_samp, size=batch_size)
batch_xs = input_data[sample][:]
batch_ys = output_data_ae[sample][:]
sess.run(train_step, feed_dict={x: batch_xs, ae_y_actual:batch_ys})
以上是4层自动编码器的代码, “meansq”是我的平方损失函数。如何为网络中的权重矩阵(张量)添加 L1 重保证?
【问题讨论】:
-
L1 可以用 sum 和 abs 运算符实现,这两种运算符都存在于 tensorflow 中(包括它们的梯度)
-
0.001*tf.reduce_sum(tf.abs(parameters))为您提供参数向量的 L1 范数(在这种情况下,从技术上讲,可能是更高等级的张量),因此以此来惩罚您的学习 -
非常感谢 +yaroslav。所以对于我来说,它应该像 (?) meanq = tf.reduce_mean(tf.square(ae_y_actual - ae_y_pred)) + 0.001*tf.reduce_sum(tf.abs(ae_Wh1)) + 0.001*tf.reduce_sum(tf. abs(ae_Wh1)) 我说的对吗?
-
嗨@Abhishek 我想知道你的 l_1 正则化器的实现是否有效,以及它是否可以在 tensorFlow 中导出。这是正确的?谢谢
标签: python neural-network tensorflow deep-learning