关于爬虫的更近一步的学习
import requests from bs4 import BeautifulSoup import pymysql headers = { "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/101.0.4951.41 Safari/537.36 Edg/101.0.1210.32", # 关闭多余的连接,因为默认是keep-alive "Connection": "close" } page = requests.get(url="http://www.4399.com/", headers=headers) page.encoding = 'gbk' soup = BeautifulSoup(page.text, 'lxml') li_list = soup.select(".tm_list>li") name_list = [] for item in li_list: name_list.append(item.text) conn = pymysql.connect( host='localhost', port=3306, user='root', passwd='root', db='4399', charset='utf8' ) cursor = conn.cursor() sql = "INSERT INTO game(name) VALUES (%s)" # 执行SQL语句 cursor.executemany(sql, name_list) conn.commit() # 关闭光标对象 cursor.close() # 关闭数据库连接 conn.close()
over

浙公网安备 33010602011771号