本文整理汇总了Python中sklearn.neural_network.MLPClassifier.predict方法的典型用法代码示例。如果您正苦于以下问题:Python MLPClassifier.predict方法的具体用法?Python MLPClassifier.predict怎么用?Python MLPClassifier.predict使用的例子?那么恭喜您, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在类sklearn.neural_network.MLPClassifier
的用法示例。
在下文中一共展示了MLPClassifier.predict方法的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Python代码示例。
示例1: neuralNetwork
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def neuralNetwork():
import pydotplus
a,b,c,d,e,f = traing_test_data_set();
for feature_number in range(1, 6):
print("Feature Number : " + str(feature_number));
train_data, train_label = a[feature_number - 1], b[feature_number - 1];
test_data, test_label = c[feature_number - 1], d[feature_number - 1];
validation_data,validation_label = e[feature_number-1],f[feature_number-1];
from sklearn.neural_network import MLPClassifier
clf = MLPClassifier(solver='lbfgs', alpha=.003, hidden_layer_sizes=(10,), random_state=1, activation='relu')
clf.fit(train_data, train_label)
tot = len(test_label);
cnt = 0;
prediction = clf.predict(test_data);
for i in range(0, len(test_data)):
if clf.predict([test_data[i]])[0] != test_label[i]:
# print(str(i)+str(clf.predict([test_data[i]]))+" "+str(test_label[i]));
cnt += 1;
from sklearn.metrics import accuracy_score
from sklearn.metrics import precision_score
from sklearn.metrics import f1_score
print("Complete for Feature :" + str(feature_number));
print("Train Score : " + str(clf.score(train_data, train_label)));
print("Total test set size : " + str(len(test_label)));
print("Correct prediction : " + str(tot - cnt));
print("Incorrect Prediction : " + str(cnt));
print("Accuracy : " + str(accuracy_score(test_label, prediction) * 100.0))
print("Precision : " + str(precision_score(test_label, prediction, average='weighted') * 100.0))
print("F1 Score : " + str(f1_score(test_label, prediction, average='weighted') * 100.0))
print("Error Rate : " + str(cnt / tot * 100.0));
print("---------------------------------------\n");
示例2: mlp_cv_architecture
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def mlp_cv_architecture(X,Y):
kfold = KFold(X.shape[0], n_folds = 10)
architectures = ( (500,2), (400,2), (400,100,2), (400,200,2), (400,100,50,2), (400,200,50,2) )
res_dict = {}
for architecture in architectures:
mlp = MLPClassifier( algorithm = 'sgd',
learning_rate = 'adaptive',
hidden_layer_sizes = architecture,
random_state = 1)
train_times = []
train_accuracy = []
test_accuracy = []
for train, test in kfold:
t_tr = time.time()
mlp.fit( X[train], Y[train] )
train_times.append( time.time() - t_tr )
acc_train = np.sum( np.equal( mlp.predict( X[train]), Y[train] ) ) / float(X[train].shape[0])
acc_test = np.sum( np.equal( mlp.predict( X[test]), Y[test] ) ) / float(X[test].shape[0])
train_accuracy.append( acc_train )
test_accuracy.append( acc_test )
res_dict[str(architecture)] = (np.mean(train_accuracy), np.std(train_accuracy),
np.mean(test_accuracy), np.std(test_accuracy),
np.mean(train_times), np.std(train_times))
with open('./../results/res_nncv_architecture.pkl', 'w') as f:
pickle.dump(res_dict,f)
示例3: create
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def create(self):
csvPath = self.sourceCsvFile
dataset = np.loadtxt( csvPath, dtype='int', delimiter=",", skiprows=1,converters={ \
4: convertCell, \
5: convertCell, \
6: convertCell, \
7: convertCell, \
8: convertCell, \
9: convertCell, \
10: convertCell, \
11: convertCell, \
12: convertCell, \
13: convertCell, \
14: convertCell, \
15: convertCell \
} )
non_cat_data = dataset[:, [0,1,2] ]
cat_data = dataset[:, [4,5,6,7,8,9,10,11,12,13,14,15] ]
output_data = dataset[:, 3]
enc = preprocessing.OneHotEncoder()
enc.fit(cat_data)
cat_out = enc.transform(cat_data).toarray()
merge_data = np.concatenate((non_cat_data,cat_data),axis=1)
d(merge_data[0])
clf = MLPClassifier(algorithm='l-bfgs', alpha=1e-5, hidden_layer_sizes=(5, 2), random_state=1)
#clf = tree.DecisionTreeClassifier()
clf = clf.fit(merge_data, output_data)
s = pickle.dumps(clf)
dtFileName = "%s\\save.pkl"%self.outDir
dtFile = open( dtFileName, 'w' )
print dtFileName
dtFile.write( s );
dtFile.close()
choicesFile = open( "%s\\choices.pkl"%self.outDir, 'w' )
s = pickle.dumps(choiceArr)
choicesFile.write( s );
choicesFile.close()
sample_inputs = []
for i in range( 100 ):
sample_inputs.append( merge_data[i*500] )
file = open( "%s\\sampleInputs.pkl"%self.outDir, 'w' )
file.write( pickle.dumps(sample_inputs) )
file.close()
file = open( "%s\\def.txt"%self.outDir, 'w' )
file.write( "input file: %s\n"%self.sourceCsvFile )
file.close()
print dataset[722]
print merge_data[722]
print output_data[722]
print clf.predict( sample_inputs )
示例4: BCISignal
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
class BCISignal():
def __init__(self, fs, bands, ch_names, states_labels, indexes):
self.states_labels = states_labels
self.bands = bands
self.prefilter = FilterSequence([ButterFilter((0.5, 45), fs, len(ch_names))])
self.csp_pools = [SpatialDecompositionPool(ch_names, fs, bands, 'csp', indexes) for _label in states_labels]
self.csp_transformer = None
self.var_detector = InstantaneousVarianceFilter(len(bands)*len(indexes)*len(states_labels), n_taps=fs//2)
self.classifier = MLPClassifier(hidden_layer_sizes=(), early_stopping=True, verbose=True)
#self.classifier = RandomForestClassifier(max_depth=3, min_samples_leaf=100)
def fit(self, X, y=None):
X = self.prefilter.apply(X)
for csp_pool, label in zip(self.csp_pools, self.states_labels):
csp_pool.fit(X, y == label)
self.csp_transformer = FilterStack([pool.get_filter_stack() for pool in self.csp_pools])
X = self.csp_transformer.apply(X)
X = self.var_detector.apply(X)
self.classifier.fit(X, y)
print('Fit accuracy {}'.format(sum(self.classifier.predict(X) == y)/len(y)))
def apply(self, chunk: np.ndarray):
chunk = self.prefilter.apply(chunk)
chunk = self.csp_transformer.apply(chunk)
chunk = self.var_detector.apply(chunk)
predicted_labels = self.classifier.predict(chunk)
return predicted_labels
示例5: neuralNetworkIterationLogistic
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def neuralNetworkIterationLogistic():
import pydotplus
a,b,c,d,e,f = traing_test_data_set();
for feature_number in range(1, 6):
iteration_output = "Iteration,Training Error,Validation Error\n";
print("Feature Number : " + str(feature_number));
train_data, train_label = a[feature_number - 1], b[feature_number - 1];
test_data, test_label = c[feature_number - 1], d[feature_number - 1];
validation_data,validation_label = e[feature_number-1],f[feature_number-1];
from sklearn.neural_network import MLPClassifier
clf = MLPClassifier(alpha=1, hidden_layer_sizes=(15,), random_state=1, activation='logistic',
warm_start=True,max_iter=1);
for iteration in range(1,350):
clf.fit(train_data, train_label)
tot = len(validation_data);
cnt = 0;
prediction = clf.predict(validation_data);
for i in range(0, len(validation_data)):
if clf.predict([validation_data[i]])[0] != validation_label[i]:
# print(str(i)+str(clf.predict([test_data[i]]))+" "+str(test_label[i]));
cnt += 1;
from sklearn.metrics import accuracy_score
from sklearn.metrics import precision_score
from sklearn.metrics import f1_score
iteration_output+=str(str(iteration) +","+ str(100-clf.score(train_data, train_label)*100.0)+","+str(100-accuracy_score(validation_label, prediction) * 100.0));
iteration_output+="\n";
print(str(str(iteration) +","+ str(100-clf.score(train_data, train_label)*100.0)+","+str(100-accuracy_score(validation_label, prediction) * 100.0)))
file_name = "Feature No "+str(feature_number)+" Iteration data"+".csv";
print(file_name);
datafile = open(file_name,"w",encoding="utf-8");
datafile.write(iteration_output);
datafile.close();
示例6: test2
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def test2():
X = [[0., 0.], [1., 1.]]
y = [0, 1]
clf = MLPClassifier(solver='lbfgs', alpha=1e-5, hidden_layer_sizes=(3), random_state=1, activation='relu')
clf.fit(X,y)
test_sample = [[2., 2.], [-1., -2.]]
print clf.predict(test_sample)
print clf.predict_proba(test_sample)
output_mlp(clf)
示例7: neuralNetwork
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def neuralNetwork():
import pydotplus
a,b,c,d,e,f = traing_test_data_set();
for feature_number in range(1, 2):
print("Feature Number : " + str(feature_number));
train_data, train_label = a[feature_number - 1], b[feature_number - 1];
test_data, test_label = c[feature_number - 1], d[feature_number - 1];
validation_data,validation_label = e[feature_number-1],f[feature_number-1];
from sklearn.preprocessing import StandardScaler
scaler = StandardScaler();
scaler.fit(train_data);
train_data = scaler.transform(train_data);
test_data = scaler.transform(test_data);
validation_data = scaler.transform(validation_data);
from sklearn.neural_network import MLPClassifier
clf = MLPClassifier(alpha=1, hidden_layer_sizes=(100,), random_state=1, activation='logistic', max_iter=1000);
clf.fit(train_data, train_label)
tot = len(test_label);
cnt = 0;
prediction = clf.predict(test_data);
for i in range(0, len(test_data)):
if prediction[i] != test_label[i]:
print(str(i)+str(prediction[i])+" "+str(test_label[i]));
cnt += 1;
from sklearn.metrics import accuracy_score
from sklearn.metrics import precision_score
from sklearn.metrics import f1_score
print("Complete for Feature :" + str(feature_number));
print("Train data set size : " + str(len(train_data)));
print("Train Score : " + str(clf.score(train_data, train_label)));
print("Total test set size : " + str(len(test_label)));
print("Correct prediction : " + str(tot - cnt));
print("Incorrect Prediction : " + str(cnt));
print("Accuracy : " + str(accuracy_score(test_label, prediction) * 100.0))
print("Precision : " + str(precision_score(test_label, prediction, average='weighted') * 100.0))
print("F1 Score : " + str(f1_score(test_label, prediction, average='weighted') * 100.0))
print("Error Rate : " + str(cnt / tot * 100.0));
print("---------------------------------------\n");
tot = len(validation_label);
cnt = 0;
prediction = clf.predict(validation_data);
for i in range(0, len(validation_label)):
if prediction[i] != validation_label[i]:
print(str(i)+str(prediction[i])+" "+str(validation_label[i]));
cnt += 1;
print("Total validation set size : " + str(len(validation_label)));
print("Correct prediction : " + str(tot - cnt));
print("Incorrect Prediction : " + str(cnt));
print("Accuracy : " + str(accuracy_score(validation_label, prediction) * 100.0))
print("Precision : " + str(precision_score(validation_label, prediction, average='weighted') * 100.0))
print("F1 Score : " + str(f1_score(validation_label, prediction, average='weighted') * 100.0))
print("Error Rate : " + str(cnt / tot * 100.0));
print("---------------------------------------\n");
示例8: test_bool_and
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def test_bool_and(self):
x = ((0, 0), (1, 1), (1, 0), (0, 1))
y = ( 0, 1, 0, 0)
mlp = MLPClassifier(hidden_layer_sizes=(), activation='logistic', max_iter=2, alpha=1e-4,
algorithm='l-bfgs', verbose=False, tol=1e-4, random_state=1,
learning_rate_init=.1)
mlp.fit(x, y)
assert mlp.predict(((0, 0))) == 0
assert mlp.predict(((0, 1))) == 0
assert mlp.predict(((1, 0))) == 0
assert mlp.predict(((1, 1))) == 1
示例9: main
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def main():
enc = OneHotEncoder(n_values=[7,7,7,7,7,7])
burgers = pandas.read_hdf('../../../machine/data.h5', 'df')
X = burgers.drop(['output'], axis=1)
y = burgers['output']
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.5)
clf = MLPClassifier(solver='adam', activation='relu',
hidden_layer_sizes=64,
verbose=False,
max_iter=10000,
tol=1e-9,
random_state=1)
classes = numpy.unique(y)
i = 0
while True:
burgers = X_train[y_train == 1]
notburgers = X_train[y_train == 0]
# Pull 32 samples from training data,
# where half the samples come from each class
sample = burgers.sample(16).join(y_train)
sample = sample.append(notburgers.sample(16).join(y_train))
sample_X_train = sample.drop(['output'], axis=1)
sample_y_train = sample['output']
sample_X_train_categoricals = sample_X_train[column_names]
tX_sample_train_categoricals = enc.fit_transform(sample_X_train_categoricals)
clf.partial_fit(tX_sample_train_categoricals, sample_y_train.as_matrix().astype(int), classes=classes)
if (i % 5) == 0:
print(i)
X_test_categoricals = X_test[column_names]
tX_test_categoricals = enc.fit_transform(X_test_categoricals)
prediction = clf.predict(tX_test_categoricals)
print_eval(y_test, prediction)
print(classification_report(y_test, prediction))
i += 1
X_train_categoricals = X_train[column_names]
tX_train_categoricals = enc.fit_transform(X_train_categoricals)
probs = clf.predict_proba(tX_train_categoricals)
# Store the probabilities
X_train_copy = X_train.copy()
X_train_copy['prob_notburger'] = probs[:,0]
X_train_copy['prob_burger'] = probs[:,1]
X_train_categoricals = X_train_copy[column_names]
tX_train_categoricals = enc.fit_transform(X_train_categoricals)
prediction = clf.predict(tX_train_categoricals)
pickle.dump(clf, open("clf.pkl.tmp", "wb"))
os.rename("clf.pkl.tmp", "clf.pkl")
示例10: test_sparse_matrices
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def test_sparse_matrices():
# Test that sparse and dense input matrices output the same results."""
X = X_digits_binary[:50]
y = y_digits_binary[:50]
X_sparse = csr_matrix(X)
mlp = MLPClassifier(random_state=1, hidden_layer_sizes=15)
mlp.fit(X, y)
pred1 = mlp.decision_function(X)
mlp.fit(X_sparse, y)
pred2 = mlp.decision_function(X_sparse)
assert_almost_equal(pred1, pred2)
pred1 = mlp.predict(X)
pred2 = mlp.predict(X_sparse)
assert_array_equal(pred1, pred2)
示例11: test_sparse_matrices
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def test_sparse_matrices():
# Test that sparse and dense input matrices output the same results.
X = X_digits_binary[:50]
y = y_digits_binary[:50]
X_sparse = csr_matrix(X)
mlp = MLPClassifier(algorithm='l-bfgs', hidden_layer_sizes=15,
random_state=1)
mlp.fit(X, y)
pred1 = mlp.predict(X)
mlp.fit(X_sparse, y)
pred2 = mlp.predict(X_sparse)
assert_almost_equal(pred1, pred2)
pred1 = mlp.predict(X)
pred2 = mlp.predict(X_sparse)
assert_array_equal(pred1, pred2)
示例12: main
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def main():
enc = OneHotEncoder(n_values=[7,7,7,7,7,7])
conn = sqlite3.connect('server.db')
cursor = conn.cursor()
all_ = pandas.read_sql_query('SELECT layers.burger, labels.output, layers.layer0, layers.layer1, layers.layer2, layers.layer3, layers.layer4, layers.layer5 FROM layers,labels WHERE layers.burger = labels.burger', conn, index_col='burger')
X = all_.drop(['output'], axis=1)
y = all_['output']
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.5)
clf = MLPClassifier(solver='adam', activation='relu',
verbose=False,
max_iter=10000,
tol=1e-9,
random_state=1)
X_train_categoricals = X_train[column_names]
tX_train_categoricals = enc.fit_transform(X_train_categoricals)
clf.fit(tX_train_categoricals, y_train.as_matrix().astype(int))
X_test_categoricals = X_test[column_names]
tX_test_categoricals = enc.fit_transform(X_test_categoricals)
prediction = clf.predict(tX_test_categoricals)
print(classification_report(y_test, prediction))
print_eval(y_test, prediction)
示例13: init_Q
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def init_Q():
# make some dummy training set
board = init_board()
board_vec = board2vec(board)
X = np.array([board_vec])
y = [(BOARD_SIZE-1)**2]
board_vec = np.invert(board_vec)
X = np.append(X,np.array([board_vec]),axis=0)
y.append(0)
edges = get_potential_moves(board) # all the edges, since the board is empty
for edge in edges:
i = edge2ind(edge)
board_vec[i] = False
X = np.append(X,np.array([board_vec]),axis=0)
y.append(check_surrounding_squares(board,edge,0))
board_vec[i] = True
Q = MLPClassifier(warm_start=True,
hidden_layer_sizes=(BOARD_SIZE,10*BOARD_SIZE,BOARD_SIZE),
tol = 1e-10,
)
# Q = DecisionTreeRegressor()
# shf = range(len(y))
# for j in xrange(100):
# random.shuffle(shf)
# Xshf = [X[i] for i in shf]
# yshf = [y[i] for i in shf]
triedy = range((BOARD_SIZE-1)**2+1)
Q.partial_fit(np.repeat(X,100,axis=0),np.repeat(y,100,axis=0),classes=triedy)
print(Q.predict(X))
return(Q)
示例14: neuralNetworkIteration
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
def neuralNetworkIteration():
import pydotplus
a,b,c,d,e,f = traing_test_data_set();
alphalist = [.00001,.00003,.0001,.0003,.001,.003,.01,.03,1,10]
for feature_number in range(1, 2):
print("Feature Number : " + str(feature_number));
train_data, train_label = a[feature_number - 1], b[feature_number - 1];
test_data, test_label = c[feature_number - 1], d[feature_number - 1];
validation_data,validation_label = e[feature_number-1],f[feature_number-1];
for new_alpha in alphalist:
iteration_output = "Iteration,Training Error,Validation Error\n";
from sklearn.neural_network import MLPClassifier
clf = MLPClassifier(alpha=new_alpha, hidden_layer_sizes=(200,), random_state=1, activation='logistic',
warm_start=True,max_iter=1);
for iteration in range(1,500):
clf.fit(train_data, train_label)
prediction = clf.predict(validation_data);
from sklearn.metrics import accuracy_score
iteration_output+=str(str(iteration) +","+ str(100-clf.score(train_data, train_label)*100.0)+","+str(100-accuracy_score(validation_label, prediction) * 100.0));
iteration_output+="\n";
print(str(str(iteration) +","+ str(100-clf.score(train_data, train_label)*100.0)+","+str(100-accuracy_score(validation_label, prediction) * 100.0)))
file_name = "For All Feature. Alpha = "+str(new_alpha)+" "+" Iteration data"+".csv";
print(file_name);
datafile = open(file_name,"w",encoding="utf-8");
datafile.write(iteration_output);
datafile.close();
示例15: NeuralLearner
# 需要导入模块: from sklearn.neural_network import MLPClassifier [as 别名]
# 或者: from sklearn.neural_network.MLPClassifier import predict [as 别名]
class NeuralLearner(Learner.Learner):
def __init__(self, FeatureMask):
super(NeuralLearner, self).__init__(FeatureMask)
self.expected = FeatureMask.LabelsForAllPoints
#self.model = MLPClassifier(algorithm='sgd', hidden_layer_sizes=(64,32))
self.model = MLPClassifier(algorithm = 'sgd',
learning_rate = 'constant',
momentum = .9,
nesterovs_momentum = True,
learning_rate_init = 0.2)
def FitAndPredict(self, mask):
return self.Predict(self.Fit(mask))
def SetupInputActivations(self, FeatureMask):
arr = np.hstack([FeatureMask.ForceStd.reshape(-1,1),
FeatureMask.ForceMinMax.reshape(-1,1),
FeatureMask.CannyFilter.reshape(-1,1)])
expected = FeatureMask.LabelsForAllPoints
return arr, expected
def Fit(self, mask):
arr, expected = self.SetupInputActivations(mask)
self.model.fit(arr, expected)
def Predict(self, mask):
arr, expected = self.SetupInputActivations(mask)
return self.model.predict(arr).reshape(-1,1)