Bad results with tensor flow DNNClassifier and cross_val_score

I am using python 3.5, tensorflow 0.10 and its DNNClassifier. If I perform one phase of training and testing, as shown below, the test result is acceptable: accuracy = 0.9333

import tensorflow as tf
from tensorflow.contrib import learn
from sklearn.cross_validation import cross_val_score, ShuffleSplit, train_test_split
from sklearn.metrics import accuracy_score
import numpy as np
from sklearn.metrics import accuracy_score
from sklearn import datasets, cross_validation

iris = datasets.load_iris()

feature_columns = learn.infer_real_valued_columns_from_input(iris.data)

x_train, x_test, y_train, y_test = train_test_split(iris.data, iris.target, test_size=0.20, random_state = 20)

model = learn.DNNClassifier(hidden_units=[5], 
                             n_classes=3, 
                             feature_columns=feature_columns, 
                            )

model.fit(x_train, y_train, steps=1000)
predicted = model.predict(x_test)

print('Accuracy on test set: %f' % accuracy_score(y_test, predicted))

If I use sklearn cross_val_score, then the final results are much worse, approximately 0.33 accuracy:

model = learn.DNNClassifier(hidden_units=[5], 
                             n_classes=3, 
                             feature_columns=feature_columns, 
                            )

scores = cross_val_score(estimator=model, 
                         X=iris.data, 
                         y=iris.target, 
                         scoring = 'accuracy',
                         cv=5,
                         fit_params={'steps': 1000},
#                          verbose=100
                        )

print(scores)
print(np.mean(scores))

Estimates and their average value:

[ 0.          0.33333333  1.          0.33333333  0.        ]
0.333333333333

What is wrong with my code in evaluating cross validation?

+4
source share

Source: https://habr.com/ru/post/1652055/


All Articles