Sindhu315

BulletDataSample

Feb 1st, 2016
229
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 2.27 KB | None | 0 0
  1. The scikit-learn flow chart.
  2.  
  3. Many of the following tutorials and exercises will be driven by the iPython (Jupyter) Notebook, which is an interactive environment for executing Python. These iPython notebooks can optionally be viewed online or downloaded and interacted with locally on your own computer.
  4.  
  5. ▪ iPython Notebook Overview from Stanford
  6.  
  7. Also note that the tutorials below are from a number of online sources. All Notebooks have been attributed to the authors; if, for some reason, you find that someone has not been properly credited for their work, please let me know and the situation will be rectified ASAP. In particular, I would like to tip my hat to Jake VanderPlas, Randal Olson, Donne Martin, Kevin Markham, and Colin Raffel for their fantastic freely-available resources.
  8.  
  9. Our first tutorials for getting our feet wet with scikit-learn follow. I suggest doing all of these in order before moving to the following steps.
  10.  
  11. A general introduction to scikit-learn, Python's most-used general purpose machine learning library, covering the k-nearest neighbors algorithm:
  12.  
  13. ▪ An Introduction to scikit-learn by Jake VanderPlas
  14.  
  15. A more in-depth and expanded introduction, including a starter project with a well-known dataset from start to finish:
  16.  
  17. ▪ Example Machine Learning Notebook by Randal Olson
  18.  
  19. A focus on strategies for evaluating different models in scikit-learn, covering train/test dataset splits:
  20.  
  21. ▪ Model Evaluation by Kevin Markham
  22.  
  23. Step 5: Machine Learning Topics with Python
  24.  
  25. With a foundation having been laid in scikit-learn, we can move on to some more in-depth explorations of the various common, and useful, algorithms. We start with k-means clustering, one of the most well-known machine learning algorithms. It is a simple and often effective method for solving unsupervised learning problems:
  26.  
  27. ▪ k-means Clustering by Jake VanderPlas
  28.  
  29. Next, we move back toward classification, and take a look at one of the most historically popular classification methods:
  30.  
  31. ▪ Decision Trees via The Grimm Scientist
  32.  
  33. From classification, we look at continuous numeric prediction:
  34.  
  35. ▪ Linear Regression by Jake VanderPlas
  36.  
  37. We can then leverage regression for classification problems, via logistic regression:
  38.  
  39. ▪ Logistic Regression by Kevin Markham
Add Comment
Please, Sign In to add comment