1. import requests
    2. #这例有个坑,这个登录地址不是直接看到的登录地址,这是人人网的反爬机制
    3. url = "http://www.renren.com/PLogin.do"
    4. # url = "http://www.renren.com/SysHome.do"
    5. # 在此输入账号密码,注意“email”“password”是网页中查到的key值
    6. data = {"email": "18904041689", "password": "123686"}
    7. headers = {
    8. "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/72.0.3626.121 Safari/537.36"
    9. }
    10. session = requests.Session()
    11. session.post(url, data = data, headers = headers)
    12. response = session.get("http://www.renren.com/personal/413066191")
    13. with open('renrenzhuye.html', 'w', encoding="utf-8") as fp:
    14. fp.write(response.text)
    1. # coding=utf-8
    2. import requests
    3. session = requests.session()
    4. # 登录的表单url
    5. post_url = "http://www.renren.com/PLogin.do"
    6. post_data = {"email": "your_email", "password": "your_password"}
    7. headers = {
    8. "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.84 Safari/537.36"
    9. }
    10. # 使用session发送post请求,cookie保存在其中
    11. session.post(post_url, data=post_data, headers=headers)
    12. # 在使用session进行请求登陆之后才能访问的地址
    13. # 这是个人首页url
    14. r = session.get("http://www.renren.com/327550088/profile", headers=headers)
    15. # 保存页面到本地
    16. with open("renren1.html", "w", encoding="utf-8") as f:
    17. f.write(r.content.decode('utf-8'))