问题
I'm new with python and scrapy. I'm tring to follow the Scrapy tutorial but I don't understand the logic of the storage step.
scrapy crawl spidername -o items.json -t json
scrapy crawl spidername --set FEED_URI=output.csv --set FEED_FORMAT=csv
I dont understand the signification of :
- -o
- -t
- --set
Thank you for your help
回答1:
You can view a list of available commands by typing scrapy crawl -h from within your project directory.
scrapy crawl spidername -o items.json -t json
-ospecifies the output filename for dumped items (items.json)-tspecifies the format for dumping items (json)
scrapy crawl spidername --set FEED_URI=output.csv --set FEED_FORMAT=csv
--setis used to set/override a settingFEED_URIis used to set the storage backend for the item dumping. In this instance it is set to "output.csv" which is using the local filesystem ie a simple output file.(for current example - output.csv)FEED_FORMATis used to set the serialization format for the (output) feed ie (for current example csv)
References (Scrapy documentation):
- Available tool commands (for the command line)
- Feed exports
来源:https://stackoverflow.com/questions/14073442/scrapy-storing-the-data