Advertisement
Guest User

Untitled

a guest
Mar 22nd, 2019
103
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 0.76 KB | None | 0 0
  1. # Import libraries
  2. import os
  3. import argparse
  4.  
  5. # Import modules
  6. from Modules.WikipediaCrawler import WikipediaCrawler
  7.  
  8. # Define arguments
  9. parser = argparse.ArgumentParser()
  10. parser.add_argument('-c', '--category', dest='category')
  11. parser.add_argument('-d', '--depth', dest='depth')
  12. values = parser.parse_args()
  13.  
  14. # Define variables based on arguments
  15. if (values.depth and values.category):
  16. category = 'Category:' + str(values.category)
  17. search_depth = values.depth
  18. else:
  19. print('Error: Please enter both category and depth')
  20. exit(1)
  21.  
  22. # Create data directory
  23. if not os.path.exists('data'):
  24. os.makedirs('data')
  25.  
  26. # Collect data
  27. crawler = WikipediaCrawler('data/wikiData.db')
  28. crawler.collect_data(category, int(search_depth))
  29. print("The database has been generated")
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement