【发布时间】:2019-10-28 07:04:26
【问题描述】:
我一直在阅读 Jonathan Hui 博客 (https://jhui.github.io/2018/02/11/Keras-tutorial/) 上发布的 Keras 教程,正如他所说,该教程直接来自 Keras 官方文档。
有些部分代码我不太理解,这些与嵌入层的尺寸有关。
我们来看两个例子:
第一个例子:
from keras.models import Sequential
from keras.layers import Dense, Dropout
from keras.layers import Embedding
from keras.layers import LSTM
import numpy as np
max_features = 10
x_train = np.random.random((1000, max_features))
y_train = np.random.randint(2, size=(1000, 1))
x_test = np.random.random((100, max_features))
y_test = np.random.randint(2, size=(100, 1))
model = Sequential()
model.add(Embedding(max_features, output_dim=256))
model.add(LSTM(128))
model.add(Dropout(0.5))
model.add(Dense(1, activation='sigmoid'))
model.compile(loss='binary_crossentropy',
optimizer='rmsprop',
metrics=['accuracy'])
model.fit(x_train, y_train, batch_size=16, epochs=10)
score = model.evaluate(x_test, y_test, batch_size=16)
这里的 X 是一个包含随机数的矩阵 (1000 x 10),每个随机数都可能是唯一的。然而,嵌入层接收作为输入维度 max_features 的参数(即本例中的 10)。但是当我们将输入维度传递给嵌入层时,我们不是在计算我们正在编码的变量的唯一可能值吗?换句话说,根据关于其值源自的空间维度的假设(例如词汇表)对变量进行虚拟化会产生的列数不是吗?
第二个例子:
import keras
import numpy as np
from keras.layers import Input, Embedding, LSTM, Dense
from keras.models import Model
# The first input
main_input = Input(shape=(100,), dtype='int32', name='main_input')
# This embedding layer will encode the input sequence
# into a sequence of dense 512-dimensional vectors.
x = Embedding(output_dim=512, input_dim=10000, input_length=100)(main_input)
# A LSTM will transform the vector sequence into a single vector,
# containing information about the entire sequence
lstm_out = LSTM(32)(x)
auxiliary_output = Dense(1, activation='sigmoid', name='aux_output')(lstm_out)
# Second input
auxiliary_input = Input(shape=(5,), name='aux_input')
x = keras.layers.concatenate([lstm_out, auxiliary_input])
# We stack a deep densely-connected network on top
x = Dense(64, activation='relu')(x)
x = Dense(64, activation='relu')(x)
x = Dense(64, activation='relu')(x)
# And finally we add the main logistic regression layer
main_output = Dense(1, activation='sigmoid', name='main_output')(x)
model = Model(inputs=[main_input, auxiliary_input], outputs=[main_output, auxiliary_output])
model.compile(optimizer='rmsprop', loss='binary_crossentropy',
loss_weights=[1., 0.2])
headline_data = np.random.random((1000, 100))
additional_data = np.random.random((1000, 5))
labels = np.random.random((1000, 1))
model.fit([headline_data, additional_data], [labels, labels],
epochs=50, batch_size=32)
这里的输入维度设置为 10,000,但我们再次处理的输入 (headline_data) 可能具有 1000 x 100 = 100,000 个唯一值。我们如何使用 10,000 维的输入空间来表示这 100,000 个值?
【问题讨论】:
标签: python machine-learning keras neural-network nlp