1
我有一個二進制分類問題,我試圖在張量流中解決。我正在使用一個簡單的多層感知器。我試圖理解我得到的混淆矩陣。示例輸出是:瞭解二元分類的張量流混淆矩陣
Epoch: 0001 cost=882.103631592
Epoch: 0002 cost=496.739675903
Epoch: 0003 cost=403.711282349
Epoch: 0004 cost=389.798379517
Epoch: 0005 cost=324.857388306
Optimization Finished!
Accuracy: 0.889306
CM=
[[797 260]
[ 0 1071]]
標籤是AWAKE和NOT_AWAKE。它看起來像從一個單一的熱門編碼,我有[1,0] AWAKE和[0,1] NOT_AWAKE(我只是將數組保存到一個文件,並目測檢查)。
如何解釋混淆矩陣? 相信這樣的輸出:
CM=
[[797 260]
[ 0 1071]]
可被解釋爲:
Pred: 0 | Pred: 1
Actual 0: 797 | 260
Actual 1: 0 | 1071
不[1,0](一個熱爲AWAKE編碼)成爲第1行中的混淆矩陣? 大部分運行mlp的代碼如下。
# Parameters
learning_rate = 0.00001
training_epochs = 4
display_step = 1
keep_prob_training = 0.75
# Network Parameters
n_hidden_1 = 2048 # 1st layer number of neurons
n_hidden_2 = 2048 # 2nd layer number of neurons
n_input = 9 # channels
n_classes = 2 # total classes
print("Some hyper params: training_epochs = %s,learning_rate = %f,keep_prob_training = %s, n_hidden_1 = %s,n_hidden_2 = %s" % (training_epochs, learning_rate, keep_prob_training, n_hidden_1, n_hidden_2))
print ("Misc shape info: X_train.shape = %s, X_test.shape = %s, y_train.shape = %s, y_test.shape = %s" % (np.shape(X_train), np.shape(X_test), np.shape(y_train), np.shape(y_test)))
# tf Graph input
X = tf.placeholder("float", [None, n_input])
Y = tf.placeholder("float", [None, n_classes])
keep_prob = tf.placeholder(tf.float32)
# placeholder for confusion matrix
y_ = tf.placeholder(tf.float32, shape = [None, 2])
# Store layers weight & bias
weights = {
'h1': tf.Variable(tf.random_normal([n_input, n_hidden_1])),
'h2': tf.Variable(tf.random_normal([n_hidden_1, n_hidden_2])),
'out': tf.Variable(tf.random_normal([n_hidden_2, n_classes]))
}
biases = {
'b1': tf.Variable(tf.random_normal([n_hidden_1])),
'b2': tf.Variable(tf.random_normal([n_hidden_2])),
'out': tf.Variable(tf.random_normal([n_classes]))
}
# Create model
def multilayer_perceptron(x):
# Hidden fully connected layer
layer_1 = tf.add(tf.matmul(x, weights['h1']), biases['b1'])
layer_1 = tf.nn.relu(layer_1)
# apply DropOut to hidden layer
drop_out_1 = tf.nn.dropout(layer_1, keep_prob)
# Hidden fully connected layer
layer_2 = tf.add(tf.matmul(drop_out_1, weights['h2']), biases['b2'])
layer_2 = tf.nn.relu(layer_2)
drop_out_2 = tf.nn.dropout(layer_2, keep_prob)
# Output fully connected layer with a neuron for each class
out_layer = tf.matmul(drop_out_2, weights['out']) + biases['out']
return out_layer
# Construct model
logits = multilayer_perceptron(X)
# obtain cm after training
confusion_matrix_tf = tf.confusion_matrix(tf.argmax(logits, 1), tf.argmax(y_, 1))
# Define loss and optimizer
loss_op = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits = logits, labels = Y))
optimizer = tf.train.AdamOptimizer(learning_rate)
train_op = optimizer.minimize(loss_op)
# Initializing the variables
init = tf.global_variables_initializer()
with tf.Session() as sess:
sess.run(init)
# Training cycle
for epoch in range(training_epochs):
avg_cost = 0.
# Run optimization op (backprop) and cost op (to get loss value)
_, c = sess.run([train_op, loss_op], feed_dict = {X: X_train, Y: y_train, keep_prob : keep_prob_training})
# Compute average loss
# Display logs per epoch step
if epoch % display_step == 0:
print("Epoch:", '%04d' % (epoch + 1) , "cost={:.9f}".format(c))
print("Optimization Finished!")
# Test model
pred = tf.nn.softmax(logits) # Apply softmax to logits
correct_prediction = tf.equal(tf.argmax(pred, 1), tf.argmax(Y, 1))
# Calculate accuracy
accuracy = tf.reduce_mean(tf.cast(correct_prediction, "float"))
print("Accuracy:", accuracy.eval({X: X_test, Y: y_test, keep_prob : 1.0}))
cm = confusion_matrix_tf.eval(feed_dict = {X: X_train, y_: y_train, keep_prob: 1.0})
print("CM=\n", cm)
這是我如何編碼我的標籤:
label_encoder = LabelEncoder()
integer_encoded = label_encoder.fit_transform(df_combined['Label'])
# binary encode
onehot_encoder = OneHotEncoder(sparse = False)
integer_encoded = integer_encoded.reshape(len(integer_encoded), 1)
all_y = onehot_encoder.fit_transform(integer_encoded)
我在上面更新的代碼中顯示''''integer_encoded'''。它看起來像AWAKE被賦予0,然後是[1,0]作爲一個熱門編碼。 NOT_AWAKE給出1,然後給出[0,1]。所以我現在明白了混淆矩陣。 「實際0」和「Pred:0」是AWAKE。謝謝! –