python爬虫基础之requests

基础的爬虫包括发送get,post请求。其中包括请求头请求体和标识等。一些常用的代码和方法如下

import requests
from requests.auth import HTTPProxyAuth

request = requests.post(
    #地址
    url="http://www.baidu.com",
    #请求头
    headers={ "user-agent": "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/57.0.2987.133 Safari/537.36"},
    #cookie
    cookies="",
    #请求参数
    params={"k1":"v1","k2":"v2"},
    #post请求体(data or json 都行)
    data = {"username":"admin", "pwd":"admin"},
    json = {"username":"admin", "pwd":"admin"},
    #代理ip
    proxies = { "http": "http://192.168.0.1:8080", "https": "https://192.168.0.1:8080" },
    #登录代理的账号密码
    auth = HTTPProxyAuth("username","passwd"),
    #上传文件
    file = {"f1":open("C:\Finish.log","rb")},
    #超时
    timeout = 2,
    #允许重定向
    allow_redirects = False
)

#保持cookie(保持登录状态等)
session = requests.session()
session.get(
    url="http://www.baidu.com"
)
print(request.text)

 

posted @ 2022-02-15 16:44  mywink  阅读(141)  评论(0)    收藏  举报