1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135
| import pandas as pd import numpy as np import pickle import random import matplotlib.pyplot as plt from sklearn.metrics import mean_squared_error from sklearn.preprocessing import StandardScaler from sklearn.neural_network import MLPRegressor from sklearn.model_selection import cross_val_score from sklearn.model_selection import learning_curve from sklearn.model_selection import train_test_split from hyperopt import fmin, tpe, hp, STATUS_OK, Trials from sklearn import preprocessing from sklearn.decomposition import PCA
train_df = pd.read_csv('历史数据.csv') predict_df = pd.read_csv('需预测数据.csv')
date_columns = ['order_date', 'year', 'month', 'day'] for col in date_columns: train_df[col] = pd.to_datetime(train_df['order_date']) predict_df[col] = pd.to_datetime(predict_df['order_date'])
new_order = ['year', 'month', 'day', 'sales_region_code', 'item_code', 'first_cate_code', 'second_cate_code', 'ord_qty'] train_df = train_df.drop(['item_price', 'order_date', 'sales_chan_name'], axis=1) predict_df = predict_df.reindex(columns=new_order) train_df = train_df.reindex(columns=new_order)
df = pd.concat([train_df, predict_df], axis=0, ignore_index=True) df = pd.get_dummies(df, columns=['sales_region_code', 'item_code', 'first_cate_code', 'second_cate_code'], prefix='cat')
train_df = df.head(n=len(train_df)) predict_df = df[-len(predict_df):] predict_df = predict_df.drop(['ord_qty'], axis=1)
min_max_scaler = preprocessing.MinMaxScaler() df0 = min_max_scaler.fit_transform(train_df) train_df = pd.DataFrame(df0, columns=train_df.columns)
X = df.drop(columns=['ord_qty']) y = df['ord_qty'] pca = PCA(n_components=0.9) X = pca.fit_transform(X)
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)
def objective(params): mlp_reg = MLPRegressor(hidden_layer_sizes=(params['hidden_layer_size'], ), activation=params['activation'], solver=params['solver'], alpha=params['alpha'], learning_rate=params['learning_rate'], learning_rate_init=params['learning_rate_init'], power_t=params['power_t'], max_iter=params['max_iter'], shuffle=True, random_state=42, tol=0.0001, verbose=False, warm_start=False, momentum=params['momentum'], nesterovs_momentum=True, early_stopping=False, validation_fraction=0.1, beta_1=params['beta_1'], beta_2=params['beta_2'], epsilon=params['epsilon'], n_iter_no_change=10, max_fun=15000) mlp_reg.fit(X_train, y_train) y_pred = mlp_reg.predict(X_test) rmse = mean_squared_error(y_test, y_pred) ** 0.5 return {'loss': rmse, 'status': STATUS_OK}
space = { 'hidden_layer_size': hp.choice('hidden_layer_size', [64, 128, 256, 512]), 'activation': hp.choice('activation', ['identity', 'logistic', 'tanh', 'relu']), 'solver': hp.choice('solver', ['adam', 'sgd', 'lbfgs']), 'alpha': hp.uniform('alpha', 0.0001, 0.1), 'learning_rate': hp.choice('learning_rate', ['constant', 'invscaling', 'adaptive']), 'learning_rate_init': hp.uniform('learning_rate_init', 0.0001, 0.1), 'power_t': hp.uniform('power_t', 0.1, 0.9), 'max_iter': hp.choice('max_iter', range(100, 1000)), 'momentum': hp.uniform('momentum', 0.1, 0.9), 'beta_1': hp.uniform('beta_1', 0.1, 0.9), 'beta_2': hp.uniform('beta_2', 0.1, 0.999), 'epsilon': hp.uniform('epsilon', 1e-8, 1e-6), }
trials = Trials() best = fmin(fn=objective, space=space, algo=tpe.suggest, max_evals=100, trials=trials, rstate=random.seed(42))
print('Best hyperparameters: ', best)
best_model = MLPRegressor(hidden_layer_sizes=(best['hidden_layer_size'], ), activation=best['activation'], solver=best['solver'], alpha=best['alpha'], learning_rate=best['learning_rate'], learning_rate_init=best['learning_rate_init'], power_t=best['power_t'], max_iter=best['max_iter'], shuffle=True, random_state=42, tol=0.0001, verbose=False, warm_start=False, momentum=best['momentum'], nesterovs_momentum=True, early_stopping=False, validation_fraction=0.1, beta_1=best['beta_1'], beta_2=best['beta_2'], epsilon=best['epsilon'], n_iter_no_change=10, max_fun=15000) best_model.fit(X, y)
with open('mlp_reg.pickle', 'wb') as f: pickle.dump(best_model, f)
scores = cross_val_score(best_model, X, y, cv=5, scoring='neg_mean_squared_error') rmse_scores = np.sqrt(-scores) print('交叉验证结果:', rmse_scores) print('平均RMSE:', rmse_scores.mean()) print('RMSE标准差:', rmse_scores.std())
train_sizes, train_scores, test_scores = learning_curve(best_model, X, y, cv=5, scoring='neg_mean_squared_error', train_sizes=np.linspace(0.1, 1.0, 10)) train_rmse_scores = np.sqrt(-train_scores) test_rmse_scores = np.sqrt(-test_scores)
train_rmse_mean = np.mean(train_rmse_scores, axis=1) train_rmse_std = np.std(train_rmse_scores, axis=1) test_rmse_mean = np.mean(test_rmse_scores, axis=1) test_rmse_std = np.std(test_rmse_scores, axis=1)
plt.plot(train_sizes, train_rmse_mean, 'o-', color='r', label='训练集RMSE') plt.plot(train_sizes, test_rmse_mean, 'o-', color='g', label='验证集RMSE') plt.fill_between(train_sizes, train_rmse_mean-train_rmse_std, train_rmse_mean+train_rmse_std, alpha=0.1, color='r') plt.fill_between(train_sizes, test_rmse_mean-test_rmse_std, test_rmse_mean+test_rmse_std, alpha=0.1, color='g') plt.xlabel('训练集样本数量') plt.ylabel('RMSE') plt.legend(loc='best') plt.show()
dff = pd.DataFrame({"ord_qty": best_model.predict(predict_df)}) dff.to_csv("预测结果.csv", index=False)
|