流程分析
一、导入需要的三方库
import re #正则表表达式文字匹配 import requests #指定url,获取网页数据 import json #转化json格式
import os #创建文件
二、确定目标地址
快手的目标地址需要在https://www.kuaishou.com网站截取请求,需要在控制台network截取,找到真正的地址https://www.kuaishou.com/graphql
url = "https://www.kuaishou.com/graphql" #目标网站url
三、确定数据位置
通过在network中搜索网页中显示的特定数据来确定data数据包,可以发现数据在data中。
四、伪装请求头、复制data数据
在控制台找到headers,复制里面的信息并手动转换成键值对的形式
在负载中找到以下信息,并封装到data里用于post请求
data = { #封装data数据 'operationName': "visionSearchPhoto", 'query': "fragment photoContent on PhotoEntity {\n id\n duration\n caption\n likeCount\n viewCount\n realLikeCount\n coverUrl\n photoUrl\n photoH265Url\n manifest\n manifestH265\n videoResource\n coverUrls {\n url\n __typename\n }\n timestamp\n expTag\n animatedCoverUrl\n distance\n videoRatio\n liked\n stereoType\n profileUserTopPhoto\n __typename\n}\n\nfragment FeedContent on Feed {\n type\n author {\n id\n name\n headerUrl\n following\n headerUrls {\n url\n __typename\n }\n __typename\n }\n photo {\n ...photoContent\n __typename\n }\n canAddComment\n llsid\n status\n currentPcursor\n __typename\n}\n\nquery visionSearchPhoto($keyword: String, $pcursor: String, $searchSessionId: String, $page: String, $webPageArea: String) {\n visionSearchPhoto(keyword: $keyword, pcursor: $pcursor, searchSessionId: $searchSessionId, page: $page, webPageArea: $webPageArea) {\n result\n llsid\n webPageArea\n Feeds {\n ...FeedContent\n __typename\n }\n searchSessionId\n pcursor\n aladdinBanner {\n imgurl\n link\n __typename\n }\n __typename\n }\n}\n", 'variables': {'keyword': keyword,'pcursor': str(page),'page':"search",'searchSessionId':"MTRfNTU2NDg2NDBfMTY1MTE0Njk1MTk5OV_mhaLmkYdfNjA0NQ"} } data = json.dumps(data) #将字典数据data转换成需要的json headers = { #伪装请求头,模拟浏览器访问 'accept': "*/*", 'Accept-Encoding' : "gzip,deflate,br", 'Accept-Language': "zh-CN,zh;q=0.9,en;q=0.8,en-GB;q=0.7,en-US;q=0.6", 'Connection': "keep-alive", 'Content-Length': "1279", 'content-type': "application/json", 'Cookie': "clientid=3; did=web_3a9d9e97c96c8d33eddd663ef9362703; client_key=65890b29; kpf=PC_WEB; kpn=KUAISHOU_VISION; userId=55648640; kuaishou.server.web_st=ChZrdWFpc2hvdS5zZXJ2ZXIud2ViLnN0EqABVSIs0T2ccYTa7qgMDgpadriErWNJyU1r87iKEYuIWEo_oJSNGq4aTBB5sq0bA7iLeCMOoX0grrEbBkpPmehOVWEs_tC-cDytf6dxSLnrE9-tRQaVcHziopazhh5rroA2XZmHEjHAe6z9-AHD0ZTxV9nJPHeI0-k0wfn9DHvDcj8ZUgQexbGwXDeH2wBV_WSuzutsd3d5oLNmN-90a33TXhoS6uws2LN-siMyPVYdMaXTUH7FIiBd27-7kiEEP6BHgx67BTXhdNPtfA88SZZL4Q_uQ09AuCgFMAE; kuaishou.server.web_ph=b0d04539ad2d77e13d866a70f9e56371c4a6", 'Host' : "www.kuaishou.com", 'Origin': "https://www.kuaishou.com", 'Referer': "https://www.kuaishou.com/search/video?searchKey=%E6%85%A2%E6%91%87", 'sec-ch-ua': '" Not A;Brand";v="99", "Chromium";v="100", "Microsoft Edge";v="100"', 'sec-ch-ua-mobile' : "?0", 'sec-ch-ua-platform': '" "Windows" "', 'Sec-Fetch-Dest': "empty", 'sec-fetch-mode':"cors", 'sec-fetch-site': "same-origin", "User-Agent": "Mozilla / 5.0(Linux;Android6.0;Nexus5 Build / MRA58N) AppleWebKit / 537.36(KHTML, likeGecko) Chrome / 99.0.4844.51Mobile Safari / 537.36" }
五、发出请求
向网站发起请求,需要用到第三方requests库,进行访问。
response = requests.post(url=url,headers=headers,data=data) #向目标网站发出请求
六、数据解析
通过re库和xpath进行数据解析,解析到需要的数据封装到列表中。
json_data = response.json() #转换json数据 #print(json_data) Feed_list = json_data['data']['visionSearchPhoto']['Feeds'] #在标签中找出每一项的列表 #print(Feed_list) for Feeds in Feed_list: #循环每一项 title = Feeds['photo']['caption'] #在网页中找出标题 new_title = re.sub(r'[\/:*?"<>\n]','_',title) #去除标题中的特殊字符 photoUrl = Feeds['photo']['photoUrl'] #找到视频的URL地址 mp4_data = requests.get(photoUrl).content #转换视频二进制形式
七、保存数据
def mkdir(path): #创建文件夹 # 引入模块 import os #引入第三方库 # 去除首位空格 path = path.strip() # 去除尾部 \ 符号 path = path.rstrip("\\") # 判断路径是否存在 # 存在 True # 不存在 False isExists = os.path.exists(path) # 判断结果 if not isExists: # 如果不存在则创建目录 # 创建目录操作函数 os.makedirs(path) print(path + ' 创建成功') return True else: # 如果目录存在则不创建,并提示目录已存在 print(path + ' 目录已存在') return False
with open(keyword+'/'+ new_title +'.mp4',mode="wb") as f: #设定保存数据路径 f.write(mp4_data) #保存数据 print("保存成功!",new_title)
完整源代码
# -*- coding = utf-8 -*- # @Time : 2022/4/28 10:48 # @Author :王敬博 # @File : 快手小姐姐.py # @Software: PyCharm import re #正则表表达式文字匹配 import requests #指定url,获取网页数据 import json #转化json格式 def mkdir(path): #创建文件夹 # 引入模块 import os #引入第三方库 # 去除首位空格 path = path.strip() # 去除尾部 \ 符号 path = path.rstrip("\\") # 判断路径是否存在 # 存在 True # 不存在 False isExists = os.path.exists(path) # 判断结果 if not isExists: # 如果不存在则创建目录 # 创建目录操作函数 os.makedirs(path) print(path + ' 创建成功') return True else: # 如果目录存在则不创建,并提示目录已存在 print(path + ' 目录已存在') return False url = "https://www.kuaishou.com/graphql" #目标网站url keyword = input("请输入想要搜索的关键词") #搜索关键字 mkpath = "d:\\pywork\\demo02\\"+keyword #创建的文件路径 mkdir(mkpath) #调用函数,创建文件夹 for page in range(1,5): #爬取4页 print("==================第%d页================="%page) #打印爬取的页数 data = { #封装data数据 'operationName': "visionSearchPhoto", 'query': "fragment photoContent on PhotoEntity {\n id\n duration\n caption\n likeCount\n viewCount\n realLikeCount\n coverUrl\n photoUrl\n photoH265Url\n manifest\n manifestH265\n videoResource\n coverUrls {\n url\n __typename\n }\n timestamp\n expTag\n animatedCoverUrl\n distance\n videoRatio\n liked\n stereoType\n profileUserTopPhoto\n __typename\n}\n\nfragment FeedContent on Feed {\n type\n author {\n id\n name\n headerUrl\n following\n headerUrls {\n url\n __typename\n }\n __typename\n }\n photo {\n ...photoContent\n __typename\n }\n canAddComment\n llsid\n status\n currentPcursor\n __typename\n}\n\nquery visionSearchPhoto($keyword: String, $pcursor: String, $searchSessionId: String, $page: String, $webPageArea: String) {\n visionSearchPhoto(keyword: $keyword, pcursor: $pcursor, searchSessionId: $searchSessionId, page: $page, webPageArea: $webPageArea) {\n result\n llsid\n webPageArea\n Feeds {\n ...FeedContent\n __typename\n }\n searchSessionId\n pcursor\n aladdinBanner {\n imgurl\n link\n __typename\n }\n __typename\n }\n}\n", 'variables': {'keyword': keyword,'pcursor': str(page),'page':"search",'searchSessionId':"MTRfNTU2NDg2NDBfMTY1MTE0Njk1MTk5OV_mhaLmkYdfNjA0NQ"} } data = json.dumps(data) #将字典数据data转换成需要的json headers = { #伪装请求头,模拟浏览器访问 'accept': "*/*", 'Accept-Encoding' : "gzip,deflate,br", 'Accept-Language': "zh-CN,zh;q=0.9,en;q=0.8,en-GB;q=0.7,en-US;q=0.6", 'Connection': "keep-alive", 'Content-Length': "1279", 'content-type': "application/json", 'Cookie': "clientid=3; did=web_3a9d9e97c96c8d33eddd663ef9362703; client_key=65890b29; kpf=PC_WEB; kpn=KUAISHOU_VISION; userId=55648640; kuaishou.server.web_st=ChZrdWFpc2hvdS5zZXJ2ZXIud2ViLnN0EqABVSIs0T2ccYTa7qgMDgpadriErWNJyU1r87iKEYuIWEo_oJSNGq4aTBB5sq0bA7iLeCMOoX0grrEbBkpPmehOVWEs_tC-cDytf6dxSLnrE9-tRQaVcHziopazhh5rroA2XZmHEjHAe6z9-AHD0ZTxV9nJPHeI0-k0wfn9DHvDcj8ZUgQexbGwXDeH2wBV_WSuzutsd3d5oLNmN-90a33TXhoS6uws2LN-siMyPVYdMaXTUH7FIiBd27-7kiEEP6BHgx67BTXhdNPtfA88SZZL4Q_uQ09AuCgFMAE; kuaishou.server.web_ph=b0d04539ad2d77e13d866a70f9e56371c4a6", 'Host' : "www.kuaishou.com", 'Origin': "https://www.kuaishou.com", 'Referer': "https://www.kuaishou.com/search/video?searchKey=%E6%85%A2%E6%91%87", 'sec-ch-ua': '" Not A;Brand";v="99", "Chromium";v="100", "Microsoft Edge";v="100"', 'sec-ch-ua-mobile' : "?0", 'sec-ch-ua-platform': '" "Windows" "', 'Sec-Fetch-Dest': "empty", 'sec-fetch-mode':"cors", 'sec-fetch-site': "same-origin", "User-Agent": "Mozilla / 5.0(Linux;Android6.0;Nexus5 Build / MRA58N) AppleWebKit / 537.36(KHTML, likeGecko) Chrome / 99.0.4844.51Mobile Safari / 537.36" } response = requests.post(url=url,headers=headers,data=data) #向目标网站发出请求 json_data = response.json() #转换json数据 #print(json_data) Feed_list = json_data['data']['visionSearchPhoto']['Feeds'] #在标签中找出每一项的列表 #print(Feed_list) for Feeds in Feed_list: #循环每一项 title = Feeds['photo']['caption'] #在网页中找出标题 new_title = re.sub(r'[\/:*?"<>\n]','_',title) #去除标题中的特殊字符 photoUrl = Feeds['photo']['photoUrl'] #找到视频的URL地址 mp4_data = requests.get(photoUrl).content #转换视频二进制形式 with open(keyword+'/'+ new_title +'.mp4',mode="wb") as f: #设定保存数据路径 f.write(mp4_data) #保存数据 print("保存成功!",new_title)
效果展示
总结
这次爬取的流程以及源代码是从bili的课程中学到的,大概用了3个小时左右,个人感觉第一次爬取动态网页还是找到数据是比较难的。通过这次实践我终于体会到了程序员的快乐(谁不喜欢看小姐姐呢),嘿嘿嘿!
如果有问题可以私聊我,如果对你有帮助的话,记得个赞哦!
版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。