登陆之后的信息会保存在cookie中,
所以需要保存cookie
import scrapy,random
from pachong_scrapy.items import PachongScrapyItem
from scrapy.http import Request
from scrapy import FormRequest
class DdSpider(scrapy.Spider):
name = 'dd'
allowed_domains = ['iqianyue.com']
start_urls = ['http://edu.iqianyue.com/index_user_login']
uapools = [
"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.140 Safari/537.36 Edge/17.17134",
]
thisua = random.choice(uapools)
header = {"user-agent": thisua}
def start_requests(self):
return [Request("http://edu.iqianyue.com/index_user_login",meta={"cookiejar":1}, callback=self.parse)]
def parse(self, response):
login_data = {
"number":"weijc",
"passwd":"weijc7789",
}
print("loging....")
return [FormRequest.from_response(response,
meta={"cookiejar":response.meta["cookiejar"]},
headers = self.header,
formdata= login_data,
callback = self.next,
)]
def next(self, response):
yield Request("http://edu.iqianyue.com/index_user_index",
meta = {"cookiejar":True},
callback = self.next2)
def next2(self, response):
print(response.xpath("/html/head/title/text()").extract())
A = response.body 是该页面所有的数据