![](/img/trans.png)
[英]How can I write confusion_matrix and classification_report to txt
[英]How to do a classification_report and a confusion_matrix from the predict_generator output using a segmentation images dataset ? ( Keras Tensorflow)
我為 ImageDataGenerator 創建了一個分類器,它使用加載我的數據集的flow_from_directory
創建,然后我執行模型和預測的訓練。
我的問題是如何從classifier.predict_generator 的輸出中獲取指標(即acc、recall、FPR 等)?
如果我沒有記錯的話,使用confusion_matrix 和classification_report 方法會有很大幫助。 圖像(.tif 文件)位於
/data/test/image ---> RGB images
/data/test/label ---> Binary mask images
/data/train/image ---> RGB images
/data/train/label ---> Binary mask images
圖像如下: RGB 圖像遮罩圖像。 predict_generator 方法返回這樣的圖像:預測圖像
我已經嘗試過類似以下的代碼來生成混淆矩陣,但無法正常工作:
predicted_classes_indices = np.argmax(results,axis=1)
labels = (image_generator.class_indices)
labels = dict((v, k) for k, v in labels.items())
predictions = [labels[k] for k in predicted_classes_indices]
cm = confusion_matrix(labels, predicted_classes_indices)
所有代碼:
from redeUnet import get_unet
import matplotlib.pyplot as plt
import numpy as np
import os
from tensorflow.python.keras.preprocessing.image import ImageDataGenerator
from tensorflow.python.keras.models import model_from_json
from tensorflow.python.keras.callbacks import EarlyStopping
from tensorflow.python.keras.callbacks import ModelCheckpoint
PATH_TRAIN = "..\\data\\train\\"
btSize = 4
alt = 256 # image row
larg = 256 # image col
image_folder = 'image'
mask_folder = 'label'
image_color_mode = 'rgb'
mask_color_mode = 'grayscale'
clMode = 'None' #'binary'
epocas = 5
qtdPatience = 60
'''
Data augmentation
'''
data_gen_args = dict(featurewise_center=False,
samplewise_center=False, # set each sample mean to 0
featurewise_std_normalization=False, # divide inputs by std of the dataset
samplewise_std_normalization=False, # divide each input by its std
zca_whitening=False, # apply ZCA whitening
rotation_range=40, # randomly rotate images in the range (degrees, 0 to 180)
zoom_range = 0.2, # Randomly zoom image
width_shift_range=0.2, # randomly shift images horizontally (fraction of total width)
height_shift_range=0.2, # randomly shift images vertically (fraction of total height)
horizontal_flip=True, # randomly flip images
vertical_flip=False,
rescale=1./255,
validation_split = 0.2) # randomly flip images
train_image_datagen = ImageDataGenerator(**data_gen_args)
train_mask_datagen = ImageDataGenerator(**data_gen_args)
'''
DATASET prepare and load (20% Validation)
'''
# Load RGB images TRAINING
image_generator = train_image_datagen.flow_from_directory(PATH_TRAIN,
classes = [image_folder],
class_mode = None,
color_mode = image_color_mode,
target_size = (larg, alt),
batch_size = btSize,
save_to_dir = None,
shuffle = False,
subset = 'training',
seed = 1)
# Load BINARY (Mask) images TRAINING
mask_generator = train_mask_datagen.flow_from_directory(PATH_TRAIN,
classes = [mask_folder],
class_mode = None,
color_mode = mask_color_mode,
target_size = (larg, alt),
batch_size = btSize,
save_to_dir = None,
shuffle = False,
subset = 'training',
seed = 1)
train_generator = zip(image_generator, mask_generator)
#-------------------------------------------------
# VALIDATION images RGB
valid_image_generator = train_image_datagen.flow_from_directory(PATH_TRAIN,
classes = [image_folder],
class_mode = None,
color_mode = image_color_mode,
target_size = (larg, alt),
batch_size = btSize,
save_to_dir = None,
shuffle = False,
subset = 'validation',
seed = 1)
# VALIDATION images BINARY (Mask)
valid_mask_generator = train_mask_datagen.flow_from_directory(PATH_TRAIN,
classes = [mask_folder],
class_mode = None,
color_mode = mask_color_mode,
target_size = (larg, alt),
batch_size = btSize,
save_to_dir = None,
shuffle = False,
subset = 'validation',
seed = 1)
valid_generator = zip(valid_image_generator, valid_mask_generator)
#-------------------------------------------------
'''
RUN TRAINING
'''
# Get UNET
classificador = get_unet(larg, alt, 3)
classificador.compile(optimizer = Adam(lr = 1e-4), loss = 'binary_crossentropy', metrics = ['accuracy']) #metrics = ['accuracy', minhaMetrica])
# Salvando o Modelo e Pesos (Best Model, StopEarly)
es = EarlyStopping(monitor = 'val_loss', mode = 'min', verbose = 1, patience = qtdPatience)
mc = ModelCheckpoint('best_polyp_unet_model.h5', monitor = 'val_loss', verbose = 1, save_best_only = True)
history = classificador.fit_generator(train_generator,
steps_per_epoch = image_generator.n // btSize,
validation_data = valid_generator,
validation_steps = valid_image_generator.n // btSize,
epochs = epocas, callbacks=[es, mc])
resultados = classificador.predict_generator(valid_generator,
steps = valid_image_generator.n,
verbose = 1)
#-------------------------------------------------
#HOW TO GET THE METRICS?
predicted_classes_indices = np.argmax(resultados,axis=1)
labels = (image_generator.class_indices)
labels = dict((v, k) for k, v in labels.items())
predictions = [labels[k] for k in predicted_classes_indices]
cm = confusion_matrix(ground_truth, predicted_classes)
#-------------------------------------------------
我在這一行收到一條錯誤消息: predictions = [labels[k] for k in predicted_classes_indices]
錯誤:不可散列類型:'numpy.ndarray'
當我通過運行以下命令檢查預測的輸出變量(“resultados”)時: resultados.shape
。顯示:
(480, 256, 256, 1)
U-net 預測生成了 480 張圖像。
但是,例如,我如何轉換此信息以與“confusion_matrix”或“classification_report”相匹配? 我認為這更困難,因為這是一個分割問題。
任何建議將不勝感激。
您需要展平您的預測和基本事實 (y):
import numpy as np
predictions_flat = predictions.flatten()
y_flat = y.flatten()
然后您可以在展平的矩陣上運行分類報告和混淆矩陣
from sklearn.metrics import classification_report
from sklearn.metrics import confusion_matrix
print('Train report', classification_report(y_flat, predictions_flat))
print('Train conf matrix', confusion_matrix(y_flat, predictions_flat))
聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.