Uploading images to scrapy

I am trying to upload an image via scrapy. Here are my different files:

items.py

class DmozItem(Item): title = Field() image_urls = Field() images = Field() 

settings.py

 BOT_NAME = 'tutorial' SPIDER_MODULES = ['tutorial.spiders'] NEWSPIDER_MODULE = 'tutorial.spiders' ITEM_PIPELINES = ['scrapy.contrib.pipeline.images.ImagesPipeline'] IMAGES= '/home/mayank/Desktop/sc/tutorial/tutorial' 

spider

 class DmozSpider(BaseSpider): name = "wikipedia" allowed_domains = ["wikipedia.org"] start_urls = [ "http://en.wikipedia.org/wiki/Pune" ] def parse(self, response): hxs = HtmlXPathSelector(response) items = [] images=hxs.select('//a[@class="image"]') for image in images: item = DmozItem() link=image.select('@href').extract()[0] link = 'http://en.wikipedia.com'+link item['image_urls']=link items.append(item) 

Despite all these settings, my pipeline does not activate. Please, help. I am new to this structure.

+7
source share
1 answer

Settings.py first: IMAGES → IMAGES_STORE

Secondly, spider : you must return item so that ImagesPipeline can load these images.

 item = DmozItem() image_urls = hxs.select('//img/@src').extract() item['image_urls'] = ["http:" + x for x in image_urls] return item 
+10
source

All Articles