i am creating generic spider (scrapy spider) for multiple websites. below is my project directory structure.
myproject <Directory>
--- __init__.py
--- common.py
--- scrapy.cfg
--- myproject <Directory>
---__init__.py
---items.py
---pipelines.py
---settings.py
---spiders <Directory>
---__init__.py
---spider.py (generic spider)
---stackoverflow_com.py (spider per website)
---anotherwebsite1_com.py (spider per website)
---anotherwebsite2_com.py (spider per website)
common.py
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
#
'''
common file
'''
from scrapy.spider import BaseSpider
from scrapy.selector import HtmlXPathSelector
from scrapy.crawler import CrawlerProcess
from scrapy.stats import stats
from scrapy.http import Request
from WSS.items import WssItem
import MySQLdb, time
import urllib2, sys
#Database connection
def open_database_connection():
connection = MySQLdb.connect(user=db_user, passwd=db_password, db=database, host=db_host, port=db_port, charset="utf8", use_unicode=True)
cursor = connection.cursor()
return connection, cursor
def close_database_connection(cursor, connection):
cursor.close()
connection.close()
return
class Domain_:
def __init__(self, spider_name, allowed_domains, start_urls, extract_topics_xpath, extract_viewed_xpath):
self.spider_name = spider_name
self.extract_topics_xpath = extract_topics_xpath
self.extract_viewed_xpath = extract_viewed_xpath
self.allowed_domains = allowed_domains
self.start_urls = start_urls
spider.py
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
#
from common import *
class DomainSpider(BaseSpider):
name = "generic_spider"
def __init__(self, current_domain):
self.allowed_domains = current_domain.allowed_domains
self.start_urls = current_domain.start_urls
self.current_domain = current_domain
def parse(self, response):
hxs = HtmlXPathSelector(response)
for topics in list(set(hxs.select(self.current_domain.extract_topics_xpath).extract())):
yield Request(topics, dont_filter=True, callback=self.extract_all_topics_data)
def extract_all_topics_data(self, response):
hxs = HtmlXPathSelector(response)
item = WssItem()
print "Processing "+response.url
connection, cursor = open_database_connection()
for viewed in hxs.select(self.current_domain.extract_viewed_xpath).extract():
item['TopicURL'] = response.url
item['Topic_viewed'] = viewed
yield item
close_database_connection(cursor, connection)
return
stackoverflow_com.py
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
#
from common import *
current_domain = Domain_(
spider_name = 'stackoverflow_com',
allowed_domains = ["stackoverflow.com"],
start_urls = ["http://stackoverflow.com/"],
extract_topics_xpath = '//div[contains(@class,\"bottomOrder\")]/a/@href',
extract_viewed_xpath = '//div[contains(@class,\"views\")]/text()'
)
import WSS.spiders.spider as spider
StackOverflowSpider = spider.DomainSpider(current_domain)
from the above scripts, i don't want to touch spider.py (assuming that all websites having same structure so i can use spider.py for all spiders)
i just want to create new spiders per website same as stackoverflow_com.py and i want to call spider.py for crawling process.
can you please advice me is there anything wrong on my code?. it showing below error message
output1: if i run "scrapy crawl stackoverflow_com" it shows below error message
C:\myproject>scrapy crawl stackoverflow_com
2013-08-05 09:41:45+0400 [scrapy] INFO: Scrapy 0.16.4 started (bot: WSS)
2013-08-05 09:41:45+0400 [scrapy] DEBUG: Enabled extensions: LogStats, TelnetConsole, CloseSpider, WebService, CoreStats, SpiderState
2013-08-05 09:41:45+0400 [scrapy] DEBUG: Enabled downloader middlewares: HttpAuthMiddleware, DownloadTimeoutMiddleware, UserAgentMiddleware, RetryMiddleware, DefaultHeadersMiddleware, RedirectMiddleware, CookiesMiddleware, HttpCompressionMiddleware, ChunkedTransferMiddleware, DownloaderStats
2013-08-05 09:41:45+0400 [scrapy] DEBUG: Enabled spider middlewares: HttpErrorMiddleware, OffsiteMiddleware, RefererMiddleware, UrlLengthMiddleware, DepthMiddleware
2013-08-05 09:41:45+0400 [scrapy] DEBUG: Enabled item pipelines: WssPipeline
Traceback (most recent call last):
File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "C:\Python27\lib\runpy.py", line 72, in _run_code
exec code in run_globals
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 156, in <module>
execute()
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 131, in execute
_run_print_help(parser, _run_command, cmd, args, opts)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 76, in _run_print_help
func(*a, **kw)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 138, in _run_command
cmd.run(args, opts)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\commands\crawl.py", line 43, in run
spider = self.crawler.spiders.create(spname, **opts.spargs)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\spidermanager.py", line 43, in create
raise KeyError("Spider not found: %s" % spider_name)
KeyError: 'Spider not found: stackoverflow_com'
output2: if i run "scrapy crawl generic_spider" it shows below error message
C:\myproject>scrapy crawl generic_spider
2013-08-05 12:25:15+0400 [scrapy] INFO: Scrapy 0.16.4 started (bot: WSS)
2013-08-05 12:25:15+0400 [scrapy] DEBUG: Enabled extensions: LogStats, TelnetConsole, CloseSpider, WebService, CoreStats, SpiderState
2013-08-05 12:25:16+0400 [scrapy] DEBUG: Enabled downloader middlewares: HttpAuthMiddleware, DownloadTimeoutMiddleware, UserAgentMiddleware, RetryMiddleware, DefaultHeadersMiddleware, RedirectMiddleware, CookiesMiddleware, HttpCompressionMiddleware, ChunkedTransferMiddleware, DownloaderStats
2013-08-05 12:25:16+0400 [scrapy] DEBUG: Enabled spider middlewares: HttpErrorMiddleware, OffsiteMiddleware, RefererMiddleware, UrlLengthMiddleware, DepthMiddleware
2013-08-05 12:25:16+0400 [scrapy] DEBUG: Enabled item pipelines: WssPipeline
Traceback (most recent call last):
File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "C:\Python27\lib\runpy.py", line 72, in _run_code
exec code in run_globals
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 156, in <module>
execute()
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 131, in execute
_run_print_help(parser, _run_command, cmd, args, opts)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 76, in _run_print_help
func(*a, **kw)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\cmdline.py", line 138, in _run_command
cmd.run(args, opts)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\commands\crawl.py", line 43, in run
spider = self.crawler.spiders.create(spname, **opts.spargs)
File "C:\Python27\lib\site-packages\scrapy-0.16.4-py2.7.egg\scrapy\spidermanager.py", line 44, in create
return spcls(**spider_kwargs)
TypeError: __init__() takes exactly 2 arguments (1 given)
Thanking you in advance :)