Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- import sys
- from urllib import request
- import requests as rq
- from bs4 import BeautifulSoup as bs
- import pandas as pd
- if len(sys.argv) < 3:
- print("Usage: python3 scrape.py <date> <xslx>")
- exit(1)
- desired_date = sys.argv[1]
- outpath = sys.argv[2]
- url = 'https://discoveratlanta.com/events/all/'
- ## Your code here
- result = rq.get(url)
- soup = bs(result.content, 'html.parser')
- events = soup.find_all('div', class_='all-listings')
- event_list = []
- print(events)
- for event in events:
- event_date = event.find('article', class_='data-eventdates').text
- if event_date == desired_date:
- event_title = event.find('h4', class_='listing-title').text
- event_list.append(event_title)
- df = pd.DataFrame(event_list)
- df.to_excel(outpath)
- df.to_csv('result.csv', index=False)
- #df.to_excel('result.xlsx', index=False)
Advertisement
Add Comment
Please, Sign In to add comment