Guest User

Untitled

a guest
Nov 18th, 2017
128
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.03 KB | None | 0 0
  1. params = {'task': 'train',
  2. 'boosting_type': 'gbdt',
  3. 'objective': 'multiclass',
  4. 'num_class':3,
  5. 'metric': 'multi_logloss',
  6. 'learning_rate': 0.002296,
  7. 'max_depth': 7,
  8. 'num_leaves': 17,
  9. 'feature_fraction': 0.4,
  10. 'bagging_fraction': 0.6,
  11. 'bagging_freq': 17}
  12.  
  13. lgb_cv = lgbm.cv(params, d_train, num_boost_round=10000, nfold=3, shuffle=True, stratified=True, verbose_eval=20, early_stopping_rounds=100)
  14.  
  15. nround = lgb_cv['multi_logloss-mean'].index(np.min(lgb_cv['multi_logloss-mean']))
  16. print(nround)
  17.  
  18. model = lgbm.train(params, d_train, num_boost_round=nround)
  19.  
  20. preds = model.predict(test)
  21. print(preds)
  22.  
  23. [[ 7.93856847e-06 9.99989550e-01 2.51164967e-06]
  24. [ 7.26332978e-01 1.65316511e-05 2.73650491e-01]
  25. [ 7.28564308e-01 8.36756769e-06 2.71427325e-01]
  26. ...,
  27. [ 7.26892634e-01 1.26915179e-05 2.73094674e-01]
  28. [ 5.93217601e-01 2.07172044e-04 4.06575227e-01]
  29. [ 5.91722491e-05 9.99883828e-01 5.69994435e-05]]
  30.  
  31. predictions = []
  32.  
  33. for x in preds:
  34. predictions.append(np.argmax(x))
Add Comment
Please, Sign In to add comment