What is this Scrapy error: ReactorNotRestartable?
What is this Scrapy error: ReactorNotRestartable?
我不明白为什么我的蜘蛛不会 运行。我单独测试了css选择器,所以我认为不是解析方法
回溯消息:
ReactorNotRestartable:
class espn_spider(scrapy.Spider):
name = "fsu2021_spider"
def start_requests(self):
urls = "https://www.espn.com/college-football/team/_/id/52"
for url in urls:
yield scrapy.Request(url = url, callback = self.parse_front)
def parse(self, response):
schedule_link = response.css('div.global-nav-container li > a::attr(href)')
process = CrawlerProcess()
process.crawl(espn_spider)
process.start()
urls = "https://www.espn.com/college-football/team/_/id/52"
for url in urls:
你正在遍历“urls”的字符,将其更改为列表:
urls = ["https://www.espn.com/college-football/team/_/id/52"]
...
...
另外你没有“parse_front”功能,如果你只是没有将它添加到代码片段中那么忽略它,如果它是一个错误则将其更改为:
yield scrapy.Request(url=url, callback=self.parse)
我不明白为什么我的蜘蛛不会 运行。我单独测试了css选择器,所以我认为不是解析方法
回溯消息: ReactorNotRestartable:
class espn_spider(scrapy.Spider):
name = "fsu2021_spider"
def start_requests(self):
urls = "https://www.espn.com/college-football/team/_/id/52"
for url in urls:
yield scrapy.Request(url = url, callback = self.parse_front)
def parse(self, response):
schedule_link = response.css('div.global-nav-container li > a::attr(href)')
process = CrawlerProcess()
process.crawl(espn_spider)
process.start()
urls = "https://www.espn.com/college-football/team/_/id/52" for url in urls:
你正在遍历“urls”的字符,将其更改为列表:
urls = ["https://www.espn.com/college-football/team/_/id/52"]
...
...
另外你没有“parse_front”功能,如果你只是没有将它添加到代码片段中那么忽略它,如果它是一个错误则将其更改为:
yield scrapy.Request(url=url, callback=self.parse)