• Python-批量刷博客园访问量脚本


    今早无聊。。。7点起来突然想写个刷访问量的。。那就动手吧

    仅供测试,不建议刷访问量哦~~

    很简单的思路,第一步提取代理ip,第二步模拟访问。

    0x01 提取HTTP代理IP

    网上很多收费的代理和免费的代理IP

    如:

    无论哪个网站,我们需要的就是爬取上面的ip和端口号,整理到一起。

    具体的网站根据具体的结构爬取 比如上面那个网站,ip和端口在td标签

    这里利用bs4爬取即可。贴上脚本

    ##获取代理ip
    def Get_proxy_ip():
        print("==========批量提取ip刷博客园访问量 By 卿=========")
        print("          Blogs:https://www.cnblogs.com/-qing-/")
        print("          Started!      ")
        global proxy_list
        proxy_list = []
        url = "https://www.kuaidaili.com/free/inha/"
        headers = {
                "Accept":"text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8",
                "Accept-Encoding":"gzip, deflate, sdch, br",
                "Accept-Language":"zh-CN,zh;q=0.8",
                "Cache-Control":"max-age=0",
                "Connection":"keep-alive",
                "Cookie":"channelid=0; sid=1561681200472193; _ga=GA1.2.762166746.1561681203; _gid=GA1.2.971407760.1561681203; _gat=1; Hm_lvt_7ed65b1cc4b810e9fd37959c9bb51b31=1561681203; Hm_lpvt_7ed65b1cc4b810e9fd37959c9bb51b31=1561681203",
                "Host":"www.kuaidaili.com",
                "Upgrade-Insecure-Requests":"1",
                "User-Agent":"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36 SE 2.X MetaSr 1.0",
                "Referrer Policy":"no-referrer-when-downgrade",
                }
        for i in range(1,100):
            url = url = "https://www.kuaidaili.com/free/inha/"+str(i)
            html = requests.get(url = url,headers = headers).content
            soup = BeautifulSoup(html,'html.parser')
            ip_list = '';
            port_list = '';
            protocol_list = '';
            for ip in soup.find_all('td'):
                if "IP" in ip.get('data-title') :
                    ip_list = ip.get_text()##获取ip             
                if "PORT" in ip.get('data-title'):
                    port_list = ip.get_text()##获取port
                if ip_list != '' and port_list != '':
                    proxy = ip_list+":"+port_list
                    ip_list = '';
                    port_list = '';
                    proxy_list.append(proxy)
            iv_main()
            time.sleep(2)
            proxy_list = []

    这样就把 提取的ip和端口放到列表里

    0x02 模拟访问刷博客园文章

    这里就很简单了 ,遍历上面那个代理ip的列表,使用requests模块取访问就是了

    def iv_main():
        proxies = {}
        requests.packages.urllib3.disable_warnings()
        #proxy_ip = random.choice(proxy_list)
        url = 'https://www.cnblogs.com/-qing-/p/11080845.html'
        for proxy_ip in proxy_list:
            headers2 = {
                'accept':'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8',
                'accept-encoding':'gzip, deflate, sdch, br',
                'accept-language':'zh-CN,zh;q=0.8',
                'cache-control':'max-age=0',
                'cookie':'__gads=ID=8c6fd85d91262bb1:T=1561554219:S=ALNI_MZwz0CMKQJK-L19DrX5DPDtYvp63Q; _gat=1; _ga=GA1.2.359634670.1561535095; _gid=GA1.2.1087331661.1561535095',
                'if-modified-since':'Fri, 28 Jun 2019 02:10:23 GMT',
                'referer':'https://www.cnblogs.com/',
                'upgrade-insecure-requests':'1',
                'user-agent':random.choice(user_agent_list),
                }
            proxies['HTTP'] = proxy_ip
            #user_agent = random.choice(user_agent_list)
            try:
                r = requests.get(url,headers=headers2,proxies=proxies,verify=False) #verify是否验证服务器的SSL证书
                print("[*]"+proxy_ip+"访问成功!")
            except:
                print("[-]"+proxy_ip+"访问失败!")

    最好带上随机的ua请求头

    user_agent_list = [
        "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) "
                          "Chrome/45.0.2454.85 Safari/537.36 115Browser/6.0.3",
        "Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_6_8; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50",
        "Mozilla/5.0 (Windows; U; Windows NT 6.1; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50",
        "Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.0; Trident/4.0)",
        "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 6.0)",
        "Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1",
        "Opera/9.80 (Windows NT 6.1; U; en) Presto/2.8.131 Version/11.11",
        "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_0) AppleWebKit/535.11 (KHTML, like Gecko) Chrome/17.0.963.56 Safari/535.11",
        "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Trident/4.0; SE 2.X MetaSr 1.0; SE 2.X MetaSr 1.0; .NET CLR 2.0.50727; SE 2.X MetaSr 1.0)",
        "Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; Trident/5.0",
        "Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1",
    ]

    0x03 优化整合

    这里可以稍微优化下,加入队列线程优化(虽然python这个没啥用)

    最终代码整合:

    # -*- coding:utf-8 -*-
    #By 卿 
    #Blog:https://www.cnblogs.com/-qing-/
    
    import requests
    from bs4 import BeautifulSoup
    import re
    import time
    import random
    import threading
    
    print("==========批量提取ip刷博客园访问量 By 卿=========")
    print("          Blogs:https://www.cnblogs.com/-qing-/")
    print("          Started!      ")
    user_agent_list = [
        "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) "
                          "Chrome/45.0.2454.85 Safari/537.36 115Browser/6.0.3",
        "Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_6_8; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50",
        "Mozilla/5.0 (Windows; U; Windows NT 6.1; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 Safari/534.50",
        "Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.0; Trident/4.0)",
        "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 6.0)",
        "Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1",
        "Opera/9.80 (Windows NT 6.1; U; en) Presto/2.8.131 Version/11.11",
        "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_0) AppleWebKit/535.11 (KHTML, like Gecko) Chrome/17.0.963.56 Safari/535.11",
        "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Trident/4.0; SE 2.X MetaSr 1.0; SE 2.X MetaSr 1.0; .NET CLR 2.0.50727; SE 2.X MetaSr 1.0)",
        "Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; Trident/5.0",
        "Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1",
    ]
    def iv_main():
        proxies = {}
        requests.packages.urllib3.disable_warnings()
        #proxy_ip = random.choice(proxy_list)
        url = 'https://www.cnblogs.com/-qing-/p/11080845.html'
        for proxy_ip in proxy_list:
            headers2 = {
                'accept':'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8',
                'accept-encoding':'gzip, deflate, sdch, br',
                'accept-language':'zh-CN,zh;q=0.8',
                'cache-control':'max-age=0',
                'cookie':'__gads=ID=8c6fd85d91262bb1:T=1561554219:S=ALNI_MZwz0CMKQJK-L19DrX5DPDtYvp63Q; _gat=1; _ga=GA1.2.359634670.1561535095; _gid=GA1.2.1087331661.1561535095',
                'if-modified-since':'Fri, 28 Jun 2019 02:10:23 GMT',
                'referer':'https://www.cnblogs.com/',
                'upgrade-insecure-requests':'1',
                'user-agent':random.choice(user_agent_list),
                }
            proxies['HTTP'] = proxy_ip
            #user_agent = random.choice(user_agent_list)
            try:
                r = requests.get(url,headers=headers2,proxies=proxies,verify=False) #verify是否验证服务器的SSL证书
                print("[*]"+proxy_ip+"访问成功!")
            except:
                print("[-]"+proxy_ip+"访问失败!")
    
    ##获取代理ip
    def Get_proxy_ip():
        global proxy_list
        proxy_list = []
        url = "https://www.kuaidaili.com/free/inha/"
        headers = {
                "Accept":"text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8",
                "Accept-Encoding":"gzip, deflate, sdch, br",
                "Accept-Language":"zh-CN,zh;q=0.8",
                "Cache-Control":"max-age=0",
                "Connection":"keep-alive",
                "Cookie":"channelid=0; sid=1561681200472193; _ga=GA1.2.762166746.1561681203; _gid=GA1.2.971407760.1561681203; _gat=1; Hm_lvt_7ed65b1cc4b810e9fd37959c9bb51b31=1561681203; Hm_lpvt_7ed65b1cc4b810e9fd37959c9bb51b31=1561681203",
                "Host":"www.kuaidaili.com",
                "Upgrade-Insecure-Requests":"1",
                "User-Agent":"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36 SE 2.X MetaSr 1.0",
                "Referrer Policy":"no-referrer-when-downgrade",
                }
        for i in range(1,100):
            url = url = "https://www.kuaidaili.com/free/inha/"+str(i)
            html = requests.get(url = url,headers = headers).content
            soup = BeautifulSoup(html,'html.parser')
            ip_list = '';
            port_list = '';
            protocol_list = '';
            for ip in soup.find_all('td'):
                if "IP" in ip.get('data-title') :
                    ip_list = ip.get_text()##获取ip             
                if "PORT" in ip.get('data-title'):
                    port_list = ip.get_text()##获取port
                if ip_list != '' and port_list != '':
                    proxy = ip_list+":"+port_list
                    ip_list = '';
                    port_list = '';
                    proxy_list.append(proxy)
            iv_main()
            time.sleep(2)
            proxy_list = []
            
    
    th=[]
    th_num=10
    for x in range(th_num):
            t=threading.Thread(target=Get_proxy_ip)
            th.append(t)
    for x in range(th_num):
            th[x].start()
    for x in range(th_num):
            th[x].join()
    
        

  • 相关阅读:
    安卓自带浏览器 position:fixed时 图片模糊问题
    建在ghost的blog
    javascript活动对象
    将json对象通过控制台保存到本地文件
    从用户体验出发的性能指标分析-DOM Ready(转)
    像诗一样的 Javascript 代码(转)
    给数组原型对象添加一个去重的方法,并返回去重后的数组
    如何避免FOUC(转)
    使用JS调用腾讯接口获取天气
    CSS好文推荐
  • 原文地址:https://www.cnblogs.com/-qing-/p/11101414.html
Copyright © 2020-2023  润新知