from selenium import webdriver
import time
import json
import re
import random
from selenium.webdriver.common.by import By
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.support import expected_conditions as EC
from selenium.webdriver.support.wait import WebDriverWait
from selenium.common.exceptions import TimeoutException, NoSuchElementException
from pyquery import PyQuery as pq
import os,os.path
import urllib.request
#配置模拟器
chrome_driver = r'C:\Users\Chysial\AppData\Local\Google\Chrome\Application\chromedriver.exe'
browser = webdriver.Chrome(chrome_driver)
wait = WebDriverWait(browser,10)
browser.get('https://www.baidu.com/')
#cookie模拟登陆,get一般运行一次就行,我因为调试后边把它注释了
def get_cookies(browser):
#cookies没有时限方法,一种方法直接在里面把expiry改成int,另一种就是输出时候改写然后写入文件不然就需要修改了
time.sleep(50)
#这段时间里在运行selenium的网址里自己扫码登陆,返回的cookie会有记录的,但是在运行的时候一定要刷新不然不显示
with open('C:\\Users\\Chysial\\Desktop\\cookies.txt','w') as cookief:
cookief.write(json.dumps(browser.get_cookies()))
browser.close()
def open_chrome(browser):
browser.delete_all_cookies()
with open('C:\\Users\\Chysial\\Desktop\\cookies.txt') as cookief:
cookieslist = json.load(cookief)
for cookies in cookieslist:
browser.add_cookie(cookies)
#从主页面baidu一步一步爬到图片里去
def search():
try:
input = wait.until(
EC.presence_of_element_located((By.CSS_SELECTOR ,'#kw'))
)
submit = wait.until(
EC.presence_of_element_located((By.CSS_SELECTOR ,'#su'))
)
input.send_keys("百度图片")
submit.click()
wait.until(
EC.presence_of_element_located((By.CSS_SELECTOR, '#page > a.n'))
)
html = browser.page_source
doc = pq(html)
items = doc('#1 .t a')
links = []
for each in items.items():
links.append(each.attr('href'))
links[0]
browser.get(links[0])
input = wait.until(
EC.presence_of_element_located((By.CSS_SELECTOR ,'#kw'))
)
input.send_keys("坂井泉水")
input.send_keys(Keys.ENTER)
wait.until(
EC.presence_of_element_located((By.CSS_SELECTOR, '.imgpage .imglist .imgitem'))
)
scorll_num(3)
wait.until(
EC.presence_of_element_located((By.CSS_SELECTOR, '.imgpage .imglist .imgitem'))
)
result = get_products()
return result
except TimeoutException:
browser.refresh()
#获取姐姐图片的地址
def get_products():
html = browser.page_source
doc = pq(html)
items = doc('.imgpage .imglist .imgitem').items()
i = 0
product = {}
for each in items:
i += 1
product[ 'picture'+str(i)] = each.attr('data-objurl')
return product
def scorll_num(num = 10):
for i in range(num):
target = browser.find_element_by_id("pageMoreWrap")
browser.execute_script("arguments[0].scrollIntoView();", target) #拖动到可见的元素去
wait.until(
EC.presence_of_element_located((By.CSS_SELECTOR, '#pageMoreWrap'))
)
#保存
def save_class(name,url):
req = urllib.request.Request(url)
response = urllib.request.urlopen(req)
img = response.read()
if '.jpg' in url:
with open(name+'.jpg','wb') as f:
f.write(img)
elif '.jpeg' in url:
with open(name+'.jpeg','wb') as f:
f.write(img)
elif '.png' in url:
with open(name+'.png','wb') as f:
f.write(img)
#下载
def download_picture(folder='坂井泉水'):
os.chdir("C:\\Users\\Chysial\\Desktop")
os.mkdir(folder)
os.chdir(folder)
each = search()
for i in each:
try:
save_class(i,each[i])
except:
continue
def main():
# 运行一次就行了,我这里就不重复了get_cookies(browser)
open_chrome(browser)
browser.refresh()
download_picture(folder='我的泉水姐姐')
if __name__ == '__main__':
main()