python爬虫基础之requests
基础的爬虫包括发送get,post请求。其中包括请求头请求体和标识等。一些常用的代码和方法如下
import requests from requests.auth import HTTPProxyAuth request = requests.post( #地址 url="http://www.baidu.com", #请求头 headers={ "user-agent": "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/57.0.2987.133 Safari/537.36"}, #cookie cookies="", #请求参数 params={"k1":"v1","k2":"v2"}, #post请求体(data or json 都行) data = {"username":"admin", "pwd":"admin"}, json = {"username":"admin", "pwd":"admin"}, #代理ip proxies = { "http": "http://192.168.0.1:8080", "https": "https://192.168.0.1:8080" }, #登录代理的账号密码 auth = HTTPProxyAuth("username","passwd"), #上传文件 file = {"f1":open("C:\Finish.log","rb")}, #超时 timeout = 2, #允许重定向 allow_redirects = False ) #保持cookie(保持登录状态等) session = requests.session() session.get( url="http://www.baidu.com" ) print(request.text)

浙公网安备 33010602011771号