Advertisement
Guest User

Untitled

a guest
Aug 21st, 2019
69
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.16 KB | None | 0 0
  1. >>> from numpy import allclose
  2. >>> from pyspark.ml.linalg import Vectors
  3. >>> from pyspark.ml.regression import RandomForestRegressor
  4. >>> df = spark.createDataFrame([
  5. ... (1.0, Vectors.dense(1.0)),
  6. ... (0.0, Vectors.sparse(1, [], []))], ["label", "features"])
  7. >>> rf = RandomForestRegressor(numTrees=2, maxDepth=2, seed=42)
  8. >>> model = rf.fit(df)
  9. >>> model.featureImportances
  10. SparseVector(1, {0: 1.0})
  11. >>> allclose(model.treeWeights, [1.0, 1.0])
  12. True
  13. >>> test0 = spark.createDataFrame([(Vectors.dense(-1.0),)], ["features"])
  14. >>> model.transform(test0).head().prediction
  15. 0.0
  16. >>> model.numFeatures
  17. 1
  18. >>> model.trees
  19. [DecisionTreeRegressionModel (uid=...) of depth..., DecisionTreeRegressionModel...]
  20. >>> model.getNumTrees
  21. 2
  22. >>> test1 = spark.createDataFrame([(Vectors.sparse(1, [0], [1.0]),)], ["features"])
  23. >>> model.transform(test1).head().prediction
  24. 0.5
  25. >>> rfr_path = temp_path + "/rfr"
  26. >>> rf.save(rfr_path)
  27. >>> rf2 = RandomForestRegressor.load(rfr_path)
  28. >>> rf2.getNumTrees()
  29. 2
  30. >>> model_path = temp_path + "/rfr_model"
  31. >>> model.save(model_path)
  32. >>> model2 = RandomForestRegressionModel.load(model_path)
  33. >>> model.featureImportances == model2.featureImportances
  34. True
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement