Advertisement
Guest User

Untitled

a guest
Apr 24th, 2014
44
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.51 KB | None | 0 0
  1. from sklearn import svm
  2. from sklearn import datasets
  3.  
  4. # train set
  5. iris = datasets.load_iris()
  6. X = iris.data[0::2, :2]
  7. Y = iris.target[0::2]
  8.  
  9. clf = svm.SVC(probability=True)
  10. clf.fit(X, Y)
  11.  
  12. # test set
  13. Z = iris.data[1::2, :2]
  14.  
  15. Y_predict = clf.predict(Z)
  16. Y_actual = iris.target[1::2]
  17. Y_probas = clf.predict_proba(Z) # probabilities of each classification
  18.  
  19. >>> import numpy as np
  20. >>> from sklearn.metrics import roc_auc_score
  21. >>> y_true = np.array([0, 0, 1, 1])
  22. >>> y_scores = np.array([0.1, 0.4, 0.35, 0.8])
  23. >>> roc_auc_score(y_true, y_scores)
  24. 0.75
  25.  
  26. >>> import numpy as np
  27. >>> X = np.random.randn(10, 4)
  28. >>> y = np.random.randint(0, 2, 10)
  29. >>> from sklearn.svm import LinearSVC
  30. >>> svm = LinearSVC().fit(X, y)
  31. >>> svm.decision_function(X)
  32. array([-0.92744332, 0.78697484, -0.71569751, -0.19938963, -0.15521737,
  33. 0.45962204, 0.1326111 , 0.44614422, 0.95731802, 0.8980536 ])
  34.  
  35. >>> from sklearn.linear_model import LogisticRegression
  36. >>> lr = LogisticRegression().fit(X, y)
  37. >>> lr.predict_proba(X)
  38. array([[ 0.73987796, 0.26012204],
  39. [ 0.26009545, 0.73990455],
  40. [ 0.63918314, 0.36081686],
  41. [ 0.62055698, 0.37944302],
  42. [ 0.54361598, 0.45638402],
  43. [ 0.38383357, 0.61616643],
  44. [ 0.50740302, 0.49259698],
  45. [ 0.39236783, 0.60763217],
  46. [ 0.32553896, 0.67446104],
  47. [ 0.20791651, 0.79208349]])
  48. >>> lr.predict_proba(X)[:, 1]
  49. array([ 0.26012204, 0.73990455, 0.36081686, 0.37944302, 0.45638402,
  50. 0.61616643, 0.49259698, 0.60763217, 0.67446104, 0.79208349])
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement