Home > Mobile >  TFIDF vector into LSTM model
TFIDF vector into LSTM model

Time:06-30

I am trying to feed my TFIDF vector into an LSTM model. TfidfVectorizer(ngram_range=(1,2), use_idf=True, analyzer='word', max_features = 5000)

Here's the Vector shapes train_vector.shape = (22895, 5000) test_vector.shape = (5724, 5000)

I am defining a model like below:

model = models.Sequential()

model.add(layers.LSTM(64, input_shape=(5000, 1), activation='relu'))
model.add(layers.Dropout(0.2))

model.add(layers.Dense(32, activation='relu'))
model.add(layers.Dropout(0.2))
model.add(layers.Dense(1, activation='sigmoid'))

Other Parameters

model.compile(loss='sparse_categorical_crossentropy', optimizer='adam', metrics=['accuracy'])
model.fit(train_vector, y_train, validation_data=(test_vector, y_test), epochs=10, batch_size=1024)

Tensorflow is being used here.

I am getting this error

ValueError: Input 0 of layer sequential_2 is incompatible with the layer: expected ndim=3, found ndim=2. Full shape received: (None, 5000)

I am trying to reshape the arrays, but the errors are still showing. I know LSTM needs 3D array. So how can I shape my arrays in a way that can be fed into LSTM???

CodePudding user response:

To add a new dimension to your training and test data, you can try:

train_vector = train_vector[..., None] # or tf.newaxis instead of None
test_vector = test_vector[..., None] # or tf.newaxis instead of None

or

train_vector = tf.expand_dims(train_vector, axis=-1)
test_vector = tf.expand_dims(test_vector, axis=-1)

Also, note that if you have one node in your output layer and you are using a sigmoid activation function, you usually combine it with the binary_crossentropy loss function instead of sparse_categorical_crossentropy, which is usually used for more than 2 classes.

  • Related