所谓爬取其实就是获取链接的内容保存到本地。所以爬之前需要先知道要爬的链接是什么。
要爬取的页面是这个:http://findicons.com/pack/2787/beautiful_flat_icons
里面有很多不错的图标,目标就是把这些文件图片爬下来,保存成本地图片。
用python3怎么做呢?
第一步:获取要爬取的母网页的内容
import urllib.request import re url = "http://findicons.com/pack/2787/beautiful_flat_icons" webPage=urllib.request.urlopen(url) data = webPage.read() data = data.decode('UTF-8')
第二步:对母网页内容处理,提取出里面的图片链接
k = re.split(r's+',data) s = [] sp = [] si = [] for i in k : if (re.match(r'src',i) or re.match(r'href',i)): if (not re.match(r'href="#"',i)): if (re.match(r'.*?png"',i) or re.match(r'.*?ico"',i)): if (re.match(r'src',i)): s.append(i) for it in s : if (re.match(r'.*?png"',it)): sp.append(it)
第三步:获取这些图片链接的内容,并保存成本地图片
cnt = 0 cou = 1 for it in sp: m = re.search(r'src="(.*?)"',it) iturl = m.group(1) print(iturl) if (iturl[0]=='/'): continue; web = urllib.request.urlopen(iturl) itdata = web.read() if (cnt%3==1 and cnt>=4 and cou<=30): f = open('d:/pythoncode/simplecodes/image/'+str(cou)+'.png',"wb") cou = cou+1 f.write(itdata) f.close() print(it) cnt = cnt+1
保存目录可以自行设定。
以下是综合起来的代码:
import urllib.request import re url = "http://findicons.com/pack/2787/beautiful_flat_icons" webPage=urllib.request.urlopen(url) data = webPage.read() data = data.decode('UTF-8') k = re.split(r's+',data) s = [] sp = [] si = [] for i in k : if (re.match(r'src',i) or re.match(r'href',i)): if (not re.match(r'href="#"',i)): if (re.match(r'.*?png"',i) or re.match(r'.*?ico"',i)): if (re.match(r'src',i)): s.append(i) for it in s : if (re.match(r'.*?png"',it)): sp.append(it) cnt = 0 cou = 1 for it in sp: m = re.search(r'src="(.*?)"',it) iturl = m.group(1) print(iturl) if (iturl[0]=='/'): continue; web = urllib.request.urlopen(iturl) itdata = web.read() if (cnt%3==1 and cnt>=4 and cou<=30): f = open('d:/pythoncode/simplecodes/image/'+str(cou)+'.png',"wb") cou = cou+1 f.write(itdata) f.close() print(it) cnt = cnt+1