[英]Negative dimension error while using keras Convolutional1D Layer
I'm trying to create a char cnn using Keras. 我正在尝试使用Keras创建一个char cnn。 That type of cnn requires you to use
Convolutional1D
layer. 这种类型的cnn要求你使用
Convolutional1D
层。 But all the ways I try to add them to my model, it gives me errors at creation stage. 但是我试图将它们添加到我的模型的所有方法,它在创建阶段给我错误。 Here is my code:
这是我的代码:
def char_cnn(n_vocab, max_len, n_classes):
conv_layers = [[256, 7, 3],
[256, 7, 3],
[256, 3, None],
[256, 3, None],
[256, 3, None],
[256, 3, 3]]
fully_layers = [1024, 1024]
th = 1e-6
embedding_size = 128
inputs = Input(shape=(max_len,), name='sent_input', dtype='int64')
# Embedding layer
x = Embedding(n_vocab, embedding_size, input_length=max_len)(inputs)
# Convolution layers
for cl in conv_layers:
x = Convolution1D(cl[0], cl[1])(x)
x = ThresholdedReLU(th)(x)
if not cl[2] is None:
x = MaxPooling1D(cl[2])(x)
x = Flatten()(x)
#Fully connected layers
for fl in fully_layers:
x = Dense(fl)(x)
x = ThresholdedReLU(th)(x)
x = Dropout(0.5)(x)
predictions = Dense(n_classes, activation='softmax')(x)
model = Model(input=inputs, output=predictions)
model.compile(optimizer='adam', loss='categorical_crossentropy')
return model
And here is the error I receive when I try to call char_cnn
function 这是我尝试调用
char_cnn
函数时收到的错误
InvalidArgumentError Traceback (most recent call last)
/usr/local/lib/python3.6/dist-packages/tensorflow/python/framework/common_shapes.py in _call_cpp_shape_fn_impl(op, input_tensors_needed, input_tensors_as_shapes_needed, require_shape_fn)
685 graph_def_version, node_def_str, input_shapes, input_tensors,
--> 686 input_tensors_as_shapes, status)
687 except errors.InvalidArgumentError as err:
/usr/local/lib/python3.6/dist-packages/tensorflow/python/framework/errors_impl.py in __exit__(self, type_arg, value_arg, traceback_arg)
515 compat.as_text(c_api.TF_Message(self.status.status)),
--> 516 c_api.TF_GetCode(self.status.status))
517 # Delete the underlying status object from memory otherwise it stays alive
InvalidArgumentError: Negative dimension size caused by subtracting 3 from 1 for 'conv1d_26/convolution/Conv2D' (op: 'Conv2D') with input shapes: [?,1,1,256], [1,3,256,256].
How to fix it? 怎么解决?
Your downsampling is too aggressive and the key argument here is max_len
: when it's too small, the sequence becomes too short to perform either a convolution or a max-pooling. 你的下采样过于激进,这里的关键参数是
max_len
:当它太小时,序列变得太短而无法执行卷积或最大池化。 You set pool_size=3
, hence it shrinks the sequence by a factor of 3
after each pooling (see the example below). 你设置
pool_size=3
,因此它在每次汇集后将序列缩小了3
倍(参见下面的例子)。 I suggest you try pool_size=2
. 我建议你尝试
pool_size=2
。
The minimal max_len
that this network can handle is max_len=123
. 最小
max_len
该网络可以处理被max_len=123
。 In this case x
shape is transformed in the following way (according to conv_layers
): 在这种情况下,
x
形状按以下方式转换(根据conv_layers
):
(?, 123, 128)
(?, 39, 256)
(?, 11, 256)
(?, 9, 256)
(?, 7, 256)
(?, 5, 256)
Setting a smaller value, like max_len=120
causes x.shape=(?, 4, 256)
before the last layer and this can't be performed. 设置较小的值,如
max_len=120
会导致最后一层之前的x.shape=(?, 4, 256)
,并且无法执行此操作。
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.