码迷,mamicode.com
首页 > 其他好文 > 详细

二、抓取网络上的资源信息

时间:2017-10-28 22:04:44      阅读:127      评论:0      收藏:0      [点我收藏+]

标签:text   .com   lis   .sh   zip   image   ges   方式   tps   

 

一、获取到网络上的网页

from bs4 import  BeautifulSoup
import requests

url = https://www.tripadvisor.cn/Attractions-g60763-Activities-New_York_City_New_York.html
web_data = requests.get(url)
soup = BeautifulSoup(web_data.text,lxml)
print(soup)

技术分享

二、获取想要的数据

技术分享

from bs4 import  BeautifulSoup
import requests

url = https://www.tripadvisor.cn/Attractions-g60763-Activities-New_York_City_New_York.html
web_data = requests.get(url)
soup = BeautifulSoup(web_data.text,lxml)
titles = soup.select(#taplc_attraction_coverpage_attraction_0 > div:nth-of-type(1) > div > div > div.shelf_item_container > div:nth-of-type(1) > div.poi > div > div.item.name > a)
print(titles)

但不是所有的,使用下面的方式获取所有

from bs4 import  BeautifulSoup
import requests

url = https://www.tripadvisor.cn/Attractions-g60763-Activities-New_York_City_New_York.html
web_data = requests.get(url)
soup = BeautifulSoup(web_data.text,lxml)
titles = soup.select(div.item.name)
imgs = soup.select(img[width="200"])
cates = soup.select(div.poi > div > div:nth-of-type(3))
#taplc_attraction_coverpage_attraction_0 > div:nth-child(1) > div > div > div.shelf_item_container > div:nth-child(4) > div.poi > div > div:nth-child(3)
# print(titles,imgs,cates,sep=‘\n-----------\n‘)
#验证下
# for title in titles:
#     print(title.get_text())
# for img in imgs:
#     print(img.get(‘src‘))
# for cate in cates:
#     print(cate.get_text())
for title,img,cate in zip(titles,imgs,cates):
    data={
        title:title.get_text(),
        img:img.get(src),
        cate:list(cate.stripped_strings),
    }
    print(data)

二、伪造登陆

技术分享

 

 三、爬取多个网页

技术分享

技术分享

技术分享

 

 四、应对js-爬取手机端

技术分享

 

技术分享

 

二、抓取网络上的资源信息

标签:text   .com   lis   .sh   zip   image   ges   方式   tps   

原文地址:http://www.cnblogs.com/Michael2397/p/7748049.html

(0)
(0)
   
举报
评论 一句话评论(0
登录后才能评论!
© 2014 mamicode.com 版权所有  联系我们:gaon5@hotmail.com
迷上了代码!