标签:agg gpu lambda 默认参数 grid gre _id sea ESS
最近在做kaggle比赛,xgboost调参是个大问题。耗时,耗力啊。一个参数调半个小时啊。
看得懂吧,每个参数逐步的,调整取值范围。
建议:
每次调一个参数。
每次一个参数,输入3个数,例如:默认参数是 1, 候选范围你可以选择 【0.1,1,10】,一定要差一个数量级,这样可以圈定范围。然后通过调整粒度,使参数越调约精巧。
param = {‘subsample‘:[0.000001,0.00001,0.0001,0.0005]}
# ‘learning_rate‘:[0.001,0.005,0.007,0.01,0.02,0.03]
# ‘colsample_bytree‘:[0.6,0.7,0.8,0.9,1,1.1,1.2],
# ‘n_estimators‘:[400,500,600,700,1000],
# ‘min_child_weight‘:[1,2,3,4,5,6],
# ‘max_depth‘:[3,4,5,6,7,8,9,10],
# ‘gamma‘: [0.1, 0.2, 0.3, 0.4, 0.5, 0.6],
# ‘reg_alpha‘: [0.05, 0.1, 1, 2, 3],
# ‘reg_lambda‘: [0.05, 0.1, 1, 2, 3]
XGBR=XGBRegressor(gpu_id=0,
single_precision_histogram=True,
n_estimators=500,
min_child_weight=1,
tree_method=‘gpu_hist‘,
eval_metric=‘mae‘,
objective=‘reg:linear‘,
booster=‘gblinear‘,
silent=1,
nthread=-1,
learning_rate=0.02,
gamma=0,
subsample=0.8,
colsample_bytree=0.8,
max_depth=5,
reg_alpha=0,
reg_lambda=1,
verbose_eval=5)
model = GridSearchCV(XGBR, param_grid=param,cv=5,scoring=‘neg_mean_absolute_error‘)
model.fit(X_train, Y_train)
print("Best score: %0.3f" % model.best_score_)
print("Best parameters set:")
best_parameters = model.best_estimator_.get_params()
for param_name in sorted(best_parameters.keys()):
print("\t%s: %r" % (param_name, best_parameters[param_name]))
标签:agg gpu lambda 默认参数 grid gre _id sea ESS
原文地址:https://www.cnblogs.com/duoba/p/12623804.html