Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- def soething(self):
- self.agency_url_list = []
- for i, page_url in enumerate(pages_urls):
- yield scrapy.Request(
- page_url, callback=self.parse_agency_urls, dont_filter=True
- )
- print self.agency_url_list
- def parse_agency_urls(self, response):
- urls = response.xpath("//a[contains(@class, 'catalog__cell-content')]/@href").re('.*')
- urls = filter(lambda x: x != '', urls)
- for url in urls:
- self.agency_url_list.append(url)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement