pyscraper Code
Brought to you by:
lonsty
File | Date | Author | Commit |
---|---|---|---|
scraper | 2020-07-22 | Allen Shaw | [89dab8] 🎨 improve structure and format |
screenshots | 2020-07-21 | Allen Shaw | [bca1b5] 📝 update README.md |
tests | 2020-07-22 | Allen Shaw | [89dab8] 🎨 improve structure and format |
.gitignore | 2019-09-08 | lonsty | [8f3f90] 1. redownload from failed; 2. test for better |
LICENSE | 2019-09-07 | Allen Shaw | [00ba90] Initial commit |
Pipfile | 2020-07-22 | Allen Shaw | [89dab8] 🎨 improve structure and format |
Pipfile.lock | 2020-07-22 | Allen Shaw | [89dab8] 🎨 improve structure and format |
README.md | 2020-07-22 | Allen Shaw | [89dab8] 🎨 improve structure and format |
_config.yml | 2019-10-31 | Allen Shaw | [02908c] Set theme jekyll-theme-minimal |
crawler.py | 2020-07-22 | Allen Shaw | [c07279] 🎨 improve code structure |
requirements.txt | 2020-07-22 | Allen Shaw | [89dab8] 🎨 improve structure and format |
极速爬取下载站酷(https://www.zcool.com.cn/)设计师/用户
上传的全部 图片/照片/插画
。
🎉 🎉 🎉 此下载工具已发布到 PyPI
pip install -U zcooldl
zcooldl -u <username>
PS: scraper
本来是规划用来存放各式各样的爬虫程序的。站酷仅仅是当初构想中的一个,因为太懒而没有新增其他爬虫。
想不到 zcool.py 竟然从原来的几十行代码,逐步增加到现在的 500+ 行 😂 😂 😂。
--thumbnail
下载缩略图(宽最大 1280px,约 500KB)python3.6
及以上$ git clone https://github.com/lonsty/scraper.git
$ cd scraper
$ pip install -r requirements.txt
下载用户名为 username
的所有图片到路径 path
下:
$ python crawler.py -u <username> -d <path>
运行截图
爬取结果
$ python crawler.py -u <username> -t <topic1>,<topic2>,...
$ python crawler.py -u <username1>,<username2>,...
$ python crawler.py -u <username> -d <last-saved-path>
$ python crawler.py --help
Usage: crawler.py [OPTIONS]
ZCool picture crawler. Download ZCool (https://www.zcool.com.cn/)
Designer's or User's pictures, photos and illustrations.
Options:
-u, --usernames TEXT One or more user names, separated by commas.
-i, --ids TEXT One or more user ids, separated by commas.
-t, --topics TEXT Specific topics of this user to download, separated
by commas.
-d, --destination TEXT Destination to save images.
-R, --retries INTEGER Repeat download for failed images. [default: 3]
-r, --redownload TEXT Redownload images from failed records.
-o, --overwrite Override existing files.
--thumbnail Download thumbnails with a maximum width of 1280px.
--max-pages INTEGER Maximum pages to download.
--max-topics INTEGER Maximum topics per page to download.
--max-workers INTEGER Maximum thread workers. [default: 20]
--help Show this message and exit.
新功能:
--thumbnail
下载缩略图(宽最大 1280px,约 500KB)新功能:
BUG 修复:
主要功能:
此项目使用 MIT 开源协议
注意:使用此工具下载的所有作品,版权归原作者所有,请谨慎使用!