FetchAnnualReportFromCNINFO

所属分类:硬件设计
开发工具:Python
文件大小:0KB
下载次数:0
上传日期:2018-03-14 07:40:14
上 传 者sh-1993
说明:  爬虫旨在通过[www.cninfo...](http:www.cninfo.com.cn)、,
(The crawler is designed to fetch annual reports of listed companies on main board of Shanghai and Shenzhen Stock Exchanges via www.cninfo.com.cn)

文件列表:
FAR/ (0, 2018-03-14)
FAR/items.py (282, 2018-03-14)
FAR/middlewares.py (3589, 2018-03-14)
FAR/pipelines.py (283, 2018-03-14)
FAR/settings.py (3457, 2018-03-14)
FAR/spiders/ (0, 2018-03-14)
FAR/spiders/__init__.py (161, 2018-03-14)
FAR/spiders/crawler.py (3780, 2018-03-14)
LICENSE (35345, 2018-03-14)

# FetchAnnualReportFromCNINFO The crawler is designed to fetch annual reports of listed companies on main board of Shanghai and Shenzhen Stock Exchanges via www.cninfo.com.cn Use this crawler: Download all files 1. Install scrapy and scrapy-splash. 2. Install other pacakge needed in crawler.py 3. Install splash on a computer, record it's IP. 4. Edit the settings.py file find the line "SPLASH_URL = 'http://IPofyoursplashserver:8050'" change the IPofyoursplashserver to the IP of your splash server in step 3. 5. Edit crawler.py fill tickers = ['600001','000777',] with the stock tickers you need. (remember to delete the example: 600001 and 000777) 6. Edit crawler.py change ((not file.find('2007') == -1) or (not file.find('2015') == -1) ) to the year you want. for example, if you want three years 2002,2005 and 2008, it should be ((not file.find('2002') == -1) or (not file.find('2005') == -1) ) or (not file.find('2008') == -1) ) 5. Locate the FAR dirctory via commandline. 6. Run scrapy crawl fetchthereports

近期下载者

相关文件


收藏者