1.先给出爬取效果图,存成一个csv文件
具体python代码
# -*- coding: utf-8 -*-
"""
Spyder Editor
This is a temporary script file.
"""
from urllib.parse import urlencode
from selenium import webdriver
from selenium.webdriver.common.by import By
from selenium.webdriver.common.action_chains import ActionChains
from selenium.webdriver.support import expected_conditions as EC
from lxml import etree
import time,random
import csv
from selenium.webdriver.support.wait import WebDriverWait
class spiderall(object):
#设置一个构造器
def __init__(self,url):
self.url=url
#调用spider函数
self.spider(url)
def spider(self,url):
#获取浏览器
browser=webdriver.Chrome()
#打开地址
browser.get(url)
#窗口最大化
# browser.maximize.window()
#打开一个csv文件
fo=open("湖人数据文件.csv",'a',encoding="utf-8")
out=csv.writer(fo)
out.writerow(["对阵球队","詹姆斯","布拉德利","波谱","霍华德","戴维斯","格林","卡鲁索","首发得分","替补得分","对阵球队","对方主力得分","对方替补得分","胜负"])
fo.flush()
fo.close()#关闭输入流
for i in range(1,80):
try:
#先找到数据中心
data_center=browser.find_element_by_xpath('//table/tbody/tr[@class="left"]['+str(i)+']/td[5]/a')
#开赛时间
sdate=browser<