失眠网,内容丰富有趣,生活中的好帮手!
失眠网 > python爬取网页数据(模拟用户名密码登录)

python爬取网页数据(模拟用户名密码登录)

时间:2023-05-22 02:30:44

相关推荐

python爬取网页数据(模拟用户名密码登录)

python爬取网页数据(模拟用户登录)

简介:python模拟用户登录,获取网页数据。

# -*- coding: utf-8 -*-import requestsimport jsonimport http.cookiejar as cookielibfrom bs4 import BeautifulSoupfrom lxml import etreeimport reimport xlwtwangyuSession = requests.session()wangyuSession.cookies = cookielib.LWPCookieJar(filename="wangyuCookies")userAgent = "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/80.0.3987.132 Safari/537.36"headers = {"Referer": "http://192.168.14.2:88/leadsec-cvs/cvs/checkItem/itemDetail?ci_ids=4659&alias=/server/Linux",'User-Agent': userAgent,}def Login():wookbook = xlwt.Workbook()sheet = wookbook.add_sheet('sheet2',cell_overwrite_ok=True)row = 0for i in range(1,10000):print("网页登录")postUrl = "http://192.168.14.2:88/leadsec-cvs/signin"postData = {"username": 'username',"password": 'password',"redirect": '/cvs/checkItem/itemDetail?ci_ids={}&alias=/server/Linux/.html?src=lm&ls=n478bfd1a95'.format(i),'pwd-encrypted': 'True',}# 使用session直接发起post请求responseRes = wangyuSession.post(postUrl, data=postData, headers=headers, json=True )print(f"statusCode = {responseRes.status_code}")wangyuSession.cookies.save()mes = responseRes.content.decode(encoding='utf-8')html = etree.HTML(mes)###获取idlinux_id = html.xpath("/html/body/div[1]/form/div[2]/div[1]/ul/li[3]/div/input[@id='benchmark']/@value")###获取json串result = re.findall(r'baselineDataStr\s=\s({(?:.|\n)*})?\r\n\tvar', str(mes))str_result = ''.join(result)json_result = json.loads(str_result)ns = 0for r in json_result['beans']:row += 1ns += 1sheet.write(row, 0, row)sheet.write(row, 1, linux_id)sheet.write(row, 2, ns-1)sheet.write(row, 3, r['info'])sheet.write(row, 4, r['opSign'])sheet.write(row, 5, r['valueItems'])sheet.write(row, 6, json_result['relation'])wookbook.save('wangyu1.xls')if __name__ == '__main__':Login()

如果觉得《python爬取网页数据(模拟用户名密码登录)》对你有帮助,请点赞、收藏,并留下你的观点哦!

本内容不代表本网观点和政治立场,如有侵犯你的权益请联系我们处理。
网友评论
网友评论仅供其表达个人看法,并不表明网站立场。