爲了學習的需要,我想檢查tf.Metrics.mean_squared_error的準確性。令我驚訝的是,他們有很大的不同。我正在尋求解釋。這裏是我的體驗簡介,然後是我的示例代碼:tf.metrics.mean_squared_error的準確性
1)通過tf.Metrics.mean_squared_error評估整個訓練數據的訓練過的玩具模型;和
2)在步驟1之後立即再次評估,首先收集給定相同整個訓練數據的所有「Xs」(或圖像)的預測,然後用所有的基本事實(或標籤)的訓練數據和預測。
我有兩個未經驗證的解釋:(1)浮點精度損失累積和(2)tf.Metrics.mean_square_error在其實現中應用看似移動的平均值,導致不準確。
任何相關的想法,非常感謝!謝謝!
import tensorflow as tf
from numpy import genfromtxt
tf.logging.set_verbosity(tf.logging.INFO)
# (hyper)parameters
batch_size = 200
num_epochs = 1000
steps = 1000
# prepare data
with tf.Session() as sess:
training_x = sess.run(tf.random_normal([2048, 16], mean=-1, stddev=4, dtype=tf.float64))
training_y = norm = sess.run(tf.random_normal([2048, 1], mean=-1, stddev=4, dtype=tf.float64))
# input function
_input_fn = lambda _input_path: genfromtxt(_input_path, delimiter=',')
input_training = tf.contrib.learn.io.numpy_input_fn({"input": training_x}, training_y,
batch_size=batch_size, num_epochs=num_epochs)
input_evaluate_train_data = tf.contrib.learn.io.numpy_input_fn({"input": training_x}, training_y)
# remember to give the same column name as used in _input_fn
features = [tf.contrib.layers.real_valued_column('input', dimension=16)]
regressor = tf.contrib.learn.DNNRegressor(feature_columns=features,
hidden_units=[32, 8],
dropout=0.1,
model_dir="testDNNR/result",
optimizer=tf.train.AdamOptimizer(learning_rate=0.008),
activation_fn=tf.nn.elu)
# training
regressor.fit(input_fn=input_training, steps=steps)
# testing with training data
eval_metric_ops = {
"mse": lambda targets, predictions: tf.metrics.mean_squared_error(tf.cast(targets, tf.float64), predictions)
}
ev = regressor.evaluate(input_fn=input_evaluate_train_data, steps=1, metrics=eval_metric_ops)
pred = regressor.predict(input_fn=input_evaluate_train_data, as_iterable=False)
# using my MSE
mse = ((training_y - pred) ** 2).mean()
print ("evaluation result given training data using my MSE: " + str(mse))
print ("evaluation result given training data using the library built-in MSE: " + str(ev))