Scrape Center爬虫平台
lingyuncelia
这个作者很懒,什么都没留下…
展开
-
Scrape Center爬虫平台之spa10案例--JJEncode混淆
去除最后括号,复制,粘贴,变成下图改写代码:function anonymous() { const players = [{ name: '凯文-杜兰特', image: 'durant.png', birthday: '1988-09-29', height: '208cm', weight: '108.9KG' }, { name: '勒布朗-詹姆斯', imag.原创 2021-08-08 21:45:40 · 427 阅读 · 0 评论 -
Scrape Center爬虫平台之spa13案例--Obfuscator混淆
https://spa13.scrape.center/将main.js用https://tool.lu/js/index.html在线工具解密const _0x4afa = ['1993-03-11', '79.4KG', '1984-05-29', 'stringify', '128.8KG', '1991-06-29', '198cm', 'davis.png', '208cm', '卡尔-安东尼-唐斯', '188cm', '196cm', 'antetokounmpo.png', '83.9K原创 2021-08-01 21:27:48 · 382 阅读 · 0 评论 -
Scrape Center爬虫平台之spa12案例--JSFuck混淆
JSFuck注意最后一个右括号下的横线,找寻前面相对应的左括号由第一行开始往下找,但愿你没有眼花,看到左括号下有横线了。将括号内的东东copy,放在console内运行,得到JSFuck混淆后的结果...原创 2021-08-01 11:45:41 · 356 阅读 · 0 评论 -
Scrape Center爬虫平台之spa3+spa4案例
import requestsdef getHTMLText(url): try: r=requests.get(url,timeout=60) r.raise_for_status() r.encoding='utf-8' return r.json() except: print('url:',url)for j in range(10): url=f"https://spa3.scrape.ce原创 2021-07-21 20:16:10 · 460 阅读 · 0 评论 -
Scrape Center爬虫平台之spa1案例
import requestsimport osdef getHTMLText(url): try: r=requests.get(url,timeout=30) r.raise_for_status() r.encoding='utf-8' return r.json() except: passdef parseHTML(html,i): id=html['results'][i][原创 2021-07-21 20:12:30 · 570 阅读 · 0 评论 -
Scrape Center爬虫平台之ssr3案例
如果是IE浏览器的话,无须输入账号+密码爬虫的话,要设置好URL协议://用户名:密码@服务域名或IP:端口号/接口地址?查询参数以下是正确姿势:import requestsimport timefrom lxml import etreeurl="https://admin:admin@ssr3.scrape.center/"r=requests.get(url)r.encoding='utf-8'r=r.textprint(r) #Internal Server Error原创 2021-07-21 20:07:47 · 560 阅读 · 0 评论 -
Scrape Center爬虫平台之ssr1+ssr2案例
import requestsimport timefrom lxml import etreefor i in range(1,11): url=f"https://ssr1.scrape.center/page/{i}" r=requests.get(url) r.encoding='utf-8' r=r.text selector=etree.HTML(r) for j in range(1,11): x1=f'//*[@id="i原创 2021-07-03 12:41:56 · 962 阅读 · 0 评论 -
Scrape Center爬虫平台之ssr4案例
#异步爬取详情页import timefrom requests.exceptions import Timeoutt1=time.time()import requestsfrom lxml import etree#异步爬取详情页import asyncioimport aiohttptemplate = 'https://ssr4.scrape.center/detail/{page}'async def get(session, queue): while True:.原创 2021-07-03 15:43:22 · 725 阅读 · 0 评论 -
Scrape Center爬虫平台之spa2案例
参考:知乎LLI ,ibra146会修电脑的程序猿scrapy学习之爬虫练习平台2B站https://www.bilibili.com/video/BV1Mf4y1s7ds?p=42主要就是破解这个token值思路分析:1:当下时间戳time.time()取整,得t,假设t为16255727362:["/api/movie", 0, “1625572736”]----》/api/movie,0,1625572736将 /api/movie,0,1625572736 用SHA1加原创 2021-07-06 22:34:10 · 1552 阅读 · 0 评论 -
Scrape Center爬虫平台之spa5案例
import requestsimport timet1=time.time()import asyncioimport aiohttpasync def get(session, queue): while True: try: page = queue.get_nowait() except asyncio.QueueEmpty: return url = f"https://spa5.sc原创 2021-07-08 22:40:04 · 474 阅读 · 0 评论 -
Scrape Center爬虫平台之spa6案例
先看明白Scrape Center爬虫平台之spa2案例import requestsimport timeimport hashlibimport base64def getHTMLText(url): try: r=requests.get(url,timeout=60) r.raise_for_status() r.encoding='utf-8' return r.json() except:原创 2021-07-09 22:39:37 · 823 阅读 · 0 评论 -
Scrape Center爬虫平台之spa7案例
import requestsdef getHTMLText(url): try: r=requests.get(url,timeout=30) r.raise_for_status() r.encoding='utf-8' return r.text[17:2000] except: passurl="https://spa7.scrape.center/js/main.js"html=getHTMLT.原创 2021-07-15 20:53:55 · 854 阅读 · 3 评论 -
Scrape Center爬虫平台之spa8案例
#效果太差,爬那么一点点数据就耗时一分钟,还是搞js逆向吧。import timet1=time.time()from selenium import webdriverfrom selenium.webdriver.support.ui import WebDriverWaitfrom selenium.webdriver.support import expected_conditions as EC, select, wait from selenium.webdriver.common.原创 2021-07-17 16:09:18 · 1294 阅读 · 0 评论 -
Scrape Center爬虫平台之spa9案例
import requestsimport redef getHTMLText(url): try: r=requests.get(url,timeout=30) r.raise_for_status() r.encoding='utf-8' return r.text except: passurl="https://spa9.scrape.center/"html=getHTMLTex原创 2021-07-18 16:59:51 · 620 阅读 · 1 评论 -
Scrape Center爬虫平台之spa10+spa11+spa12+spa13案例
import requestsimport redef getHTMLText(url): try: r=requests.get(url,timeout=30) r.raise_for_status() r.encoding='utf-8' return r.text except: passurl="https://spa9.scrape.center/"html=getHTMLTex原创 2021-07-18 16:58:29 · 540 阅读 · 2 评论