参考:http://blog.csdn.net/ajinglingmofashi/article/details/53611168
>>> import urllib.request
>>> shenda = urllib.request.urlopen(url)>>> shenda = "http://zsb.szu.edu.cn/"
>>> page = urllib.request.urlopen(shenda)
>>> from bs4 import BeautifulSoup
>>> soup = BeautifulSoup(page,"lxml")
>>> print(soup.prettify())
----------------------------------------------
>>> import re
>>> import requests
>>> r = requests.get(shenda)
>>> html = r.text
>>> j = 1
>>> title = re.findall(r'\raw_title\"\:\".*?"',html)
>>> price = re.findall(r'"view_price\"\:\"[\d\.]*\"',html)
---------------------------------------------
--------------------------------------------
---------------------------
获取网站的编码信息:参考:https://bbs.csdn.net/topics/390774501
--------------------------
参考:http://blog.csdn.net/f156207495/article/details/78074240
BeautifulSoup中的.text方法和get_text()方法的区别
-----------------------------------
参考:http://blog.csdn.net/my_precious/article/details/52948362
用Python的Lxml库抓取网页内容
-----------------------------------
参考:https://www.cnblogs.com/gopythoner/p/6390381.html
BeautifulSoup4----利用find_all和get方法来获取信息
------------------------------------------------------------------------------------
------------------------
--------------------------------------------------
-------------------------------------------------