How to output per-class accuracy in Keras?

前端 未结 3 750
有刺的猬
有刺的猬 2020-12-25 13:12

Caffe can not only print overall accuracy, but also per-class accuracy.

In Keras log, there\'s only overall accuracy. It\'s hard for me to calculate the separate cla

相关标签:
3条回答
  • 2020-12-25 13:48

    Precision & recall are more useful measures for multi-class classification (see definitions). Following the Keras MNIST CNN example (10-class classification), you can get the per-class measures using classification_report from sklearn.metrics:

    from sklearn.metrics import classification_report
    import numpy as np
    
    Y_test = np.argmax(y_test, axis=1) # Convert one-hot to index
    y_pred = model.predict_classes(x_test)
    print(classification_report(Y_test, y_pred))
    

    Here is the result:

             precision    recall  f1-score   support
    
          0       0.99      1.00      1.00       980
          1       0.99      0.99      0.99      1135
          2       1.00      0.99      0.99      1032
          3       0.99      0.99      0.99      1010
          4       0.98      1.00      0.99       982
          5       0.99      0.99      0.99       892
          6       1.00      0.99      0.99       958
          7       0.97      1.00      0.99      1028
          8       0.99      0.99      0.99       974
          9       0.99      0.98      0.99      1009
    
    avg / total   0.99      0.99      0.99     10000
    
    0 讨论(0)
  • 2020-12-25 13:58

    For train per-class accuracy: implement below on training dataset - after (and/or before) training on the dataset.


    For raw per-class validation accuracy:

    def per_class_accuracy(y_preds,y_true,class_labels):
        return [np.mean([
            (y_true[pred_idx] == np.round(y_pred)) for pred_idx, y_pred in enumerate(y_preds) 
          if y_true[pred_idx] == int(class_label)
                        ]) for class_label in class_labels]
    
    def update_val_history():
        [val_history[class_label].append(np.mean( np.asarray(temp_history).T[class_idx] )
                                 ) for class_idx, class_label in enumerate(class_labels)]
    

    Example:

    class_labels = ['0','1','2','3']
    val_history = {class_label:[] for class_label in class_labels}
    
    y_true   = np.asarray([0,0,0,0, 1,1,1,1, 2,2,2,2, 3,3,3,3])
    y_preds1 = np.asarray([0,3,3,3, 1,1,0,0, 2,2,2,0, 3,3,3,3])
    y_preds2 = np.asarray([0,0,3,3, 0,1,0,0, 2,2,2,2, 0,0,0,0])
    
    y_preds1 = model.predict(x1)
    temp_hist.append(per_class_accuracy(y_preds1,y_true,class_labels))
    update_val_history()
    y_preds2 = model.predict(x2)
    temp_hist.append(per_class_accuracy(y_preds2,y_true,class_labels))
    update_val_history()
    
    print(val_history)
    

    >>{
    '0': [0.25, 0.50],
    '1': [0.50, 0.25],
    '2': [0.75, 1.00],
    '3': [1.00, 0.00]
    }

    0 讨论(0)
  • 2020-12-25 14:10

    You are probably looking to use a callback, which you can easily add to the model.fit() call.

    For example, you can define your own class using the keras.callbacks.Callback interface. I recommend using the on_epoch_end() function since it will format nicely inside of your training summary if you decide to print with that verbosity setting. Please note that this particular code block is set to use 3 classes, but you can of course change it to your desired number.

    # your class labels
    classes = ["class_1","class_2", "class_3"]
    
    class AccuracyCallback(tf.keras.callbacks.Callback):
    
        def __init__(self, test_data):
            self.test_data = test_data
    
        def on_epoch_end(self, epoch, logs=None):
            x_data, y_data = self.test_data
    
            correct = 0
            incorrect = 0
    
            x_result = self.model.predict(x_data, verbose=0)
    
            x_numpy = []
    
            for i in classes:
                self.class_history.append([])
    
            class_correct = [0] * len(classes)
            class_incorrect = [0] * len(classes)
    
            for i in range(len(x_data)):
                x = x_data[i]
                y = y_data[i]
    
                res = x_result[i]
    
                actual_label = np.argmax(y)
                pred_label = np.argmax(res)
    
                if(pred_label == actual_label):
                    x_numpy.append(["cor:", str(y), str(res), str(pred_label)])     
                    class_correct[actual_label] += 1   
                    correct += 1
                else:
                    x_numpy.append(["inc:", str(y), str(res), str(pred_label)])
                    class_incorrect[actual_label] += 1
                    incorrect += 1
    
            print("\tCorrect: %d" %(correct))
            print("\tIncorrect: %d" %(incorrect))
    
            for i in range(len(classes)):
                tot = float(class_correct[i] + class_incorrect[i])
                class_acc = -1
                if (tot > 0):
                    class_acc = float(class_correct[i]) / tot
    
                print("\t%s: %.3f" %(classes[i],class_acc)) 
    
            acc = float(correct) / float(correct + incorrect)  
    
            print("\tCurrent Network Accuracy: %.3f" %(acc))
    

    Then, you are going to want to configure your new callback to your model fit. Assuming your validation data (val_data) is some tuple pair, you can use the following:

    accuracy_callback = AccuracyCallback(val_data)
    
    # you can use the history if desired
    history = model.fit( x=_, y=_, verbose=1, 
               epochs=_, shuffle=_, validation_data = val_data,
               callbacks=[accuracy_callback], batch_size=_
             )
    

    Please note that the _ indicates values likely to change based on your configuration

    0 讨论(0)
提交回复
热议问题