1. github
import requests from lxml import etree class Login(object): def __init__(self): self.headers = { 'Origin': 'https://github.com', 'Referer': 'https: // github.com /', 'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/57.0.2987.133 Safari/537.36', 'Host': 'github.com' } # 获取登录页面与token self.login_url = 'https://github.com/login' # 登录页面 self.login_post_url = 'https://github.com/session' self.session=requests.session() # 获取token def get_token(self): headers = { 'Host': 'github.com', 'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/72.0.3626.119 Safari/537.36', } res_text = self.session.get(url=self.login_url, headers=headers).text tree = etree.HTML(res_text) token = tree.xpath('//div//input[2]/@value')[0] return token # 模拟登录 def login(self,username, password): token=self.get_token() print(token) post_data = { "login": username, "password":password, "commit": "Sign in", "utf8": "✓", "authenticity_token": token } res = self.session.post(url=self.login_post_url, data=post_data) print(res.status_code) if res.status_code == 200: self.get_email_page() with open("github.html", "wb") as f: f.write(res.content) #获取个人页 def get_email_page(self): print('获取个人页') email_data = self.session.get('https://github.com/tjp40922').text with open('xxxxxx.html', 'w', encoding='utf8') as f: f.write(email_data) if __name__ == '__main__': login = Login() username = input('请输入用户名') password = input('请输入密码') login.login(username, password)
注意点:
1.要获取token,而且要一一对应
2.headers,两次请求的请求头是不一样的,否则会报错,422状态码
2.scrapy模拟登陆人人
# -*- coding: utf-8 -*- import scrapy from scrapy.http import Request,FormRequest class LoginrrSpider(scrapy.Spider): name = 'loginrr' allowed_domains = ['renren.com'] start_urls = ['http://renren.com/'] headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36' } # 这里我们先编写start_requests方法(会首先执行) def start_requests(self): # 首先访问一遍登录页面(在这里我们可以监测有无验证码,若有则需要添加验证码字段,若无则不需要) # 然后回调parse方法 # 设置meta参数,将cookiejar设置为1,表示cookjar开启 return [Request("http://www.renren.com/PLogin.do", meta={"cookiejar": 1}, callback=self.parse)] def parse(self, response): # 需要发送的表单,这次没有验证码 data = { # 这里输入你们自己的账号密码 'email': 'xxxxxx', 'password': 'xxxxxxxx' } print("正在登陆...") # 通过 FormRequest.from_response方法来进行登录 return [FormRequest.from_response(response, # 设置cookie信息 meta={"cookiejar": response.meta["cookiejar"]}, headers=self.headers, formdata=data, # 回调到next方法 callback=self.next, )] def next(self, response): # 打印登录成功后的界面源代码 print(response.body)
3.scrapy模拟登录豆瓣
import scrapy from faker import Factory
#伪造ua的 f = Factory.create() class MailSpider(scrapy.Spider): name = 'douban-mail' allowed_domains = ['accounts.douban.com', 'douban.com'] start_urls = [ 'https://www.douban.com/' ] headers = { 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', 'Accept-Encoding': 'gzip, deflate, br', 'Accept-Language': 'zh-CN,zh;q=0.8,en-US;q=0.5,en;q=0.3', 'Connection': 'keep-alive', 'Host': 'accounts.douban.com', 'User-Agent': f.user_agent() } formdata = { 'form_email': '您的账号', 'form_password': '您的密码', # 'captcha-solution': '', # 'captcha-id': '', 'login': '登录', 'redir': 'https://www.douban.com/', 'source': 'None' } def start_requests(self): return [scrapy.Request(url='https://www.douban.com/accounts/login', headers=self.headers, meta={'cookiejar': 1}, callback=self.parse_login)] def parse_login(self, response): # 如果有验证码要人为处理 if 'captcha_image' in response.body: print 'Copy the link:' link = response.xpath('//img[@class="captcha_image"]/@src').extract()[0] print link captcha_solution = raw_input('captcha-solution:') captcha_id = urlparse.parse_qs(urlparse.urlparse(link).query, True)['id'] self.formdata['captcha-solution'] = captcha_solution self.formdata['captcha-id'] = captcha_id return [scrapy.FormRequest.from_response(response, formdata=self.formdata, headers=self.headers, meta={'cookiejar': response.meta['cookiejar']}, callback=self.after_login )] def after_login(self, response): print response.status self.headers['Host'] = "www.douban.com" return scrapy.Request(url='https://www.douban.com/doumail/', meta={'cookiejar': response.meta['cookiejar']}, headers=self.headers, callback=self.parse_mail) def parse_mail(self, response): print response.status for item in response.xpath('//div[@class="doumail-list"]/ul/li'): mail = DoubanMailItem() mail['sender_time'] = item.xpath('div[2]/div/span[1]/text()').extract()[0] mail['sender_from'] = item.xpath('div[2]/div/span[2]/text()').extract()[0] mail['url'] = item.xpath('div[2]/p/a/@href').extract()[0] mail['title'] = item.xpath('div[2]/p/a/text()').extract()[0] print mail yield mail