Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
48 changes: 48 additions & 0 deletions week1大作业提交/hekailong/第一周大作业.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,48 @@
from bs4 import BeautifulSoup
import requests
import time
url = 'http://bj.58.com/pbdn/1/?PGTID=0d305a36-0000-1e96-9601-c083f8999a80&ClickID=1'
wb_data = requests.get(url)
soup = BeautifulSoup(wb_data.text,'lxml')


titles = soup.select(' tr > td.t > a.t')
prices = soup.select(' tr > td > b')
addresses = soup.select(' tr > td.t > span.fl > a')
sellerurlfirst = soup.select(' tr > td.t > a.t')


for title,price,address,sellerurl in zip(titles,prices,addresses,sellerurlfirst):
info = {
'title':title.get_text(),
'price':price.get_text()+'元',
'address':address.get_text(),
'sellerurl':sellerurl.get('href')


}


#获取卖家信息
seller_url= info['sellerurl']
seller_data = requests.get(seller_url)
sellersoup = BeautifulSoup(seller_data.text,'lxml')
issuedates = sellersoup.select('#index_show > ul.mtit_con_left.fl > li.time')
pageviews = sellersoup.select('#index_show > ul.mtit_con_left.fl > li.count')
catagries = sellersoup.select('#header > div.breadCrumb.f12 > span:nth-of-type(3) > a')
seller_types = u'商家' if sellersoup.select('#divOwner > ul > li > em') else u'个人'
#print(seller_types)

for issuedate,pageview,catagry,seller_type in zip(issuedates,pageviews,catagries,seller_types):
infonext = {
'issuedate':issuedate.get_text(),
'pageview':pageview.get_text(),
'catagry':catagry.get_text(),
'seller_type':seller_types


}
info.update(infonext)
info.pop('sellerurl')
print(info)
time.sleep(2)
3 changes: 2 additions & 1 deletion 花名册
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
请在下面加上你的姓名和一句话感言,或者联系方式什么都行,注意一人一行哦:
请在下面加上你的姓名和一句话感言,或者联系方式什么都行,注意一人一行哦:

林茜茜 Hello,world!
何凯隆 加油!