关于爬虫的更近一步的学习

import requests
from bs4 import BeautifulSoup
import pymysql

headers = {
    "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/101.0.4951.41 Safari/537.36 Edg/101.0.1210.32",
    # 关闭多余的连接,因为默认是keep-alive
    "Connection": "close"
}
page = requests.get(url="http://www.4399.com/", headers=headers)
page.encoding = 'gbk'
soup = BeautifulSoup(page.text, 'lxml')
li_list = soup.select(".tm_list>li")
name_list = []
for item in li_list:
    name_list.append(item.text)

conn = pymysql.connect(
    host='localhost',
    port=3306,
    user='root',
    passwd='root',
    db='4399',
    charset='utf8'
)
cursor = conn.cursor()
sql = "INSERT INTO game(name) VALUES (%s)"
# 执行SQL语句
cursor.executemany(sql, name_list)
conn.commit()
# 关闭光标对象
cursor.close()
# 关闭数据库连接
conn.close()

over

posted @ 2022-06-01 00:37  湖南陈冠希  阅读(40)  评论(0)    收藏  举报