Crawler ( scrapy.crawler ) is the main entry point to the Scrapy API. It provides access to all the main components of Scrapy, and it is used to intercept extension functions in Scrapy.
Component
A scraper ( scrapy.core.scraper ) is responsible for analyzing responses and extracting information from them. It starts from Engine and is used to launch your spiders.
scrapy.spiders is a module that contains the basic Spider implementation (which you use to write your spiders) along with some common spiders available out of the box (for example, CrawlSpider for scanning rule sets, SitemapSpider for scanning based on a sitemap or XMLFeedSpider for crawling XML feeds).
Additional information is available on the official pages of the documentation:
http://doc.scrapy.org/en/latest/topics/spiders.html?highlight=crawlspider#spiders http://doc.scrapy.org/en/latest/topics/api.html?highlight=scrapy.crawler # module-scrapy.crawler
source share