title_node = soup.find('dd',class_='lemmaWgt-lemmaTitle-title').find('h1')
title_data = title_node.get_text()
title_data = title_node.get_text()
2018-05-18
最新回答 / 慕粉1224386790
?URL 需要先確定。必須是能夠正常打開(kāi)的網(wǎng)址(URL),比如 www.hao123.com? 和 www.baidu.com?之類(lèi)的都可以。
print('獲取P')
pnode = soup.find_all('p', class_ = 'title')
for link in pnode:
print(link.name, link.get_text)
pnode = soup.find_all('p', class_ = 'title')
for link in pnode:
print(link.name, link.get_text)
2018-05-11
print(link.name, link['href'], link.get_text() )
2018-05-11
print('獲取Lacie的鏈接')
link_node = soup.find_all('a', href = 'http://example.come/Lacie')
for link in link_node:
print(link.name, link['href'], link.get_text() )
print('正則匹配')
link_node = soup.find_all( 'a', href = re.compile(r"ill") )
for link in link_node:
link_node = soup.find_all('a', href = 'http://example.come/Lacie')
for link in link_node:
print(link.name, link['href'], link.get_text() )
print('正則匹配')
link_node = soup.find_all( 'a', href = re.compile(r"ill") )
for link in link_node:
2018-05-11
# coding = gbk
from bs4 import BeautifulSoup
import re
'''
'''
soup = BeautifulSoup(html_doc, 'html.parser')
print('獲取所有的鏈接:')
links = soup.find_all('a')
for link in links:
print( link.name, link['href'], link.get_text() )
from bs4 import BeautifulSoup
import re
'''
'''
soup = BeautifulSoup(html_doc, 'html.parser')
print('獲取所有的鏈接:')
links = soup.find_all('a')
for link in links:
print( link.name, link['href'], link.get_text() )
2018-05-11
最贊回答 / 小圖兔
在Python3中我們用urllib.requet 代替urllib2看你的代碼應(yīng)該是Python3,其中包含了request方法,所以結(jié)果第一行返回結(jié)果一樣的話(huà),程序會(huì)認(rèn)為你下面的request是你上面第一行的返回,所以出錯(cuò)。