import requests from bs4 import BeautifulSoup import time import csv headers = { "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/97.0.4692.71 Safari/537.36" } def get_info(url): web_data = requests.get(url, headers=headers) web_data.raise_for_status() soup = BeautifulSoup(web_data.text, 'lxml') ranks = soup.select('span.pc_temp_num') titles = soup.select('div.pc_temp_songlist > ul > li > a') times = soup.select('span.pc_temp_tips_r > span') songs_data = [] for rank, title, time in zip(ranks, titles, times): data = { "rank": rank.get_text().strip(), "song": title.get_text().replace("\n", "").replace("\t", "").split('-')[0], "singer": title.get_text().replace("\n", "").replace("\t", "").split('-')[1], "time": time.get_text().strip() } songs_data.append(data) return songs_data if __name__ == '__main__': urls = ["https://www.kugou.com/yy/rank/home/{}-8888.html".format(str(i)) for i in range(1, 24)] with open('qqmusic_top_songs.csv', 'w', newline='', encoding='utf-8') as csvfile: fieldnames = ["rank", "song", "singer", "time"] writer = csv.DictWriter(csvfile, fieldnames=fieldnames) writer.writeheader() for url in urls: songs_data = get_info(url) for song in songs_data: writer.writerow(song) time.sleep(1) print("所有歌曲数据已保存到CSV文件中")