怎么去除爬下来的网站中的一些转义字符串
from bs4 import BeautifulSoup
import re
import urllib.request, urllib.error
def first(baseurl):
return getData(baseurl)
findLink = re.compile(r'<a href="https://q.cnblogs.com/q/130384/(.?)" target="_blank">', re.S)
findTttle = re.compile(r'<!-- headline start -->(.?)<!-- headline end -->', re.S)
findContant1 = re.compile(r'<p>(.)</p>', re.S)
findContant2 = re.compile(r'justify;">(.)</p>', re.S)
dr=re.compile(r'<[^>]+>', re.S)
def getData(baseurl):
datalist = []
url = baseurl
html = askUrl(url)
soup = BeautifulSoup(html, "html.parser")
for presscolumn in soup.find_all('div', class_="presscolumn"):
item = str(presscolumn)
link = re.findall(findLink, item)
datalist.append(link[0])
return datalist
def askUrl(url):
head = {
"User-Agent": "Mozilla/5.0(Windows NT 10.0; Win64; x64)AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.198 Safari/537.36"
}
request = urllib.request.Request(url, headers=head)
html = ""
try:
response = urllib.request.urlopen(request)
html = response.read().decode("utf-8")
except urllib.error.URLError as e:
if hasattr(e, "code"):
print(e.code)
if hasattr(e, "reason"):
print(e.reason)
return html
def getData2(baseurl2):
datalist2 = []
url = baseurl2
html = askUrl(url)
soup = BeautifulSoup(html, "html.parser")
for overall in soup.find_all('div', class_="overall"):
item = str(overall)
title = re.findall(findTttle, item)[0]
datalist2.append(title)
contant =re.findall(findContant1, item)
if len(contant) <= 0:
contant = re.findall(findContant2, item)
contant = dr.sub('', str(contant))
datalist2.append(contant)
return datalist2
def main():
baseurl = input("请输入您要的网站:")
data = []
baseurl2 = first(baseurl)
for n in range(0, len(baseurl2)):
print(getData2(baseurl2[n]))
data.append(getData2(baseurl2[n]))
if name == "main":
main()
兄弟萌,这些\和后面的东西怎么去掉啊
回答
爬的网站url 发一下,都啥年代了还用这种古老的模块
以上是 怎么去除爬下来的网站中的一些转义字符串 的全部内容, 来源链接: utcz.com/a/69088.html