清华大佬耗费三个月吐血整理的几百G的资源,免费分享!....>>>
import requests
import lxml
from lxml import etree
def get_content(url):
try:
headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.2; WOW64; rv:37.0) Gecko/20100101 Firefox/37.0'}
r = requests.get(url, headers = headers)
return '<html>' + r.content.decode('utf-8') + '</html>'
except Exception as exc:
#raise exc
return None
class User(object):
pass
def parse_content(content):
if content is None: return []
try:
xml = etree.fromstring(content)
divs= xml.findall('li/div')
page_info = []
for div in divs:
u = User()
tweets = div.findall('p')
if len(tweets) < 2:
continue
user_and_text = list(tweets[0].itertext())
u.name,u.txt = user_and_text if len(user_and_text) == 2 else [None]*2
data_comments = list(tweets[1].itertext())
u.data,u.commtents,u.client = data_comments if len(data_comments) ==3 else [None]*3
u.link = tweets[1].xpath('a')[0].get('href')
page_info.append(u)
return page_info
except Exception as exc:
raise exc
def test():
url = "http://www.oschina.net/fetch_tweets?p=40"
content = get_content(url)
infos = parse_content(content)
for u in infos:
print(u.name)
print(u.txt)
print(u.client.split('\n')[1].strip() if u.client else u.client)
#print(u.name, u.txt)
if __name__ == '__main__':
test()