commit
e1cb4683bd
@ -0,0 +1,84 @@
|
||||
import re
|
||||
from os import remove
|
||||
|
||||
import requests
|
||||
import json
|
||||
from bs4 import BeautifulSoup
|
||||
|
||||
|
||||
# 获取搜索结果
|
||||
def get_search_result_bv(key_word, limit):
|
||||
bv_list = []
|
||||
count = 0
|
||||
page_no = 1
|
||||
while count < limit:
|
||||
search_url = (
|
||||
"https://search.bilibili.com/all?keyword="
|
||||
+ key_word
|
||||
+ "&pages="
|
||||
+ page_no.__str__()
|
||||
)
|
||||
head = {
|
||||
"User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/128.0.0.0 Safari/537.36",
|
||||
"referer": "https://www.bilibili.com/",
|
||||
}
|
||||
response_data = requests.get(url=search_url, headers=head)
|
||||
soup = BeautifulSoup(response_data.text, "lxml")
|
||||
results = soup("div", "bili-video-card__info--right")
|
||||
try:
|
||||
for result in results:
|
||||
curr_href = result.a["href"].strip("/")
|
||||
# 过滤搜索结果中的直播(live.bilibili)
|
||||
info = curr_href.split("/")
|
||||
if info[1] == "video":
|
||||
bv_list.append(info[2])
|
||||
count += 1
|
||||
if count >= limit:
|
||||
break
|
||||
except IndexError:
|
||||
print("Out of Index at [" + __name__ + "]! 搜索结果获取失败?")
|
||||
print("http code: {code}".format(code=response_data.status_code))
|
||||
break
|
||||
page_no += 1
|
||||
return bv_list
|
||||
|
||||
|
||||
def get_cid(bv_list):
|
||||
cid_list = []
|
||||
head = {
|
||||
"User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/128.0.0.0 Safari/537.36",
|
||||
"referer": "https://www.bilibili.com/",
|
||||
}
|
||||
for bv in bv_list:
|
||||
url = "https://api.bilibili.com/x/player/pagelist?bvid=" + bv
|
||||
response_data = requests.get(url=url, headers=head)
|
||||
data = json.loads(response_data.text)
|
||||
cid_list.append(data["data"][0]["cid"])
|
||||
return cid_list
|
||||
def get_comments(cid_list):
|
||||
comments_list = []
|
||||
for cid in cid_list:
|
||||
curr_url="https://comment.bilibili.com/" + cid.__str__() + ".xml"
|
||||
head = {
|
||||
"User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/128.0.0.0 Safari/537.36",
|
||||
"referer": "https://www.bilibili.com/",
|
||||
}
|
||||
response_data = requests.get(url=curr_url, headers=head)
|
||||
response_data.encoding="utf-8"
|
||||
soup = BeautifulSoup(response_data.text, "xml")
|
||||
comments = soup.find_all("d")
|
||||
for comment in comments:
|
||||
comments_list.append(comment.text)
|
||||
return comments_list
|
||||
def comments_filter(rules, comments_list):
|
||||
patterns=[]
|
||||
res=[]
|
||||
for rule in rules:
|
||||
patterns.append(re.compile(rule, re.IGNORECASE))
|
||||
for comment in comments_list:
|
||||
for pattern in patterns:
|
||||
if re.match(pattern, comment):
|
||||
if not res.__contains__(comment):
|
||||
res.append(comment)
|
||||
|
||||
return res
|
Loading…
Reference in new issue