简体   繁体   中英

Tf.keras model.predict() returns class probabilities that are higher than 1?

I am trying to call model.predict() in tf.keras on a CNN to predict the class for a single image. For some reason, the class probabilities are coming back higher than 1 which is nonsensical. I am unsure why this is occurring. Below is how I train my CNN:

class_names = ['Angry','Disgust','Fear','Happy','Sad','Surprise','Neutral']
model = models.Sequential()
model.add(layers.Conv2D(64, (3, 3), activation='relu', input_shape=(48, 48, 1), kernel_regularizer=tf.keras.regularizers.l1(0.01)))
model.add(layers.Conv2D(128, (3, 3), padding='same', activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(tf.keras.layers.Dropout(0.5))
model.add(layers.Conv2D(64, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(tf.keras.layers.Dropout(0.5))
model.add(layers.Conv2D(64, (3, 3), activation='relu'))


model.summary()

model.add(layers.Flatten())
model.add(layers.Dense(64, activation='relu'))
model.add(layers.Dense(7))


#model.summary()
model.compile(optimizer='adam',loss=tf.keras.losses.CategoricalCrossentropy(from_logits=True),metrics=['accuracy'])
lr_reducer = tf.keras.callbacks.ReduceLROnPlateau(monitor='val_loss', factor=0.9, patience=3) #monitors the validation loss for signs of a plateau and then alter the learning rate by the specified factor if a plateau is detected

early_stopper = tf.keras.callbacks.EarlyStopping(monitor='val_accuracy', min_delta=0, patience=6, mode='auto')  #This will monitor and stop the model training if it is not further converging

checkpointer = tf.keras.callbacks.ModelCheckpoint('C:\\Users\\rtlum\\Documents\\DataSci_Projects\\PythonTensorFlowProjects\\Datasets\\FER2013_Model_Weights\\Model\\weights.hd5', monitor='val_loss', verbose=1, save_best_only=True) #This allows checkpoints to be saved each epoch just in case the model stops training

epochs = 100
batch_size = 64
learning_rate = 0.001

model.fit(
          train_data,
          train_labels,
          epochs = epochs,
          batch_size = batch_size,
          validation_split = 0.2,
          shuffle = True,
          callbacks=[lr_reducer, checkpointer, early_stopper]
          )

Below is how I call model.predict() and pass in a single image to predict:

    model = tf.keras.models.load_model('Model\\weights.hd5')
    img = Image.open(test_image).convert('L')
    img = img.resize([48, 48])
    image_data = np.asarray(img, dtype=np.uint8)
    #image_data = np.resize(img,3072)
    image_data = image_data / 255
    image_data_test = image_data.reshape((1, 48, 48, 1))
    class_names = ['Angry','Disgust','Fear','Happy','Sad','Surprise','Neutral']
    x = model.predict(image_data_test)
    app.logger.info(x)
    image_pred = np.argmax(x)
    y = round(x[0][np.argmax(x)], 2)
    confidence = y * 100
    print(class_names[image_pred], confidence)

And finally, below is the class probabilities I receive from model.predict():

>>> x = model.predict(image_data_test)
>>> x
array([[ 1.0593076 , -3.5140653 ,  0.7505076 ,  2.1341033 ,  0.02394461,
        -0.08749148,  0.6640976 ]], dtype=float32)

Your last layer model.add(layers.Dense(7)) is using linear activation function. To get probability of 7 classes, you should use softmax activation.

Change your last layer to

model.add(layers.Dense(7 , activation='softmax'))

add an Activation Layer to convert your output value into value of [0,1]

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM