在网络怕爬虫过程中遇到,编码错误
'gbk' codec can't encode character '\ue5d1' in position 0: illegal multibyte sequence
response = requests.get(url)
html_uid = response.text
# name=html_uid["result"]['name']
# path = "E:\\python_py\\景区详情\\省份\\吉林省\\"#文件路径"
file_path =cities_folder+province+city+'_'+str(i)+"_"+str(j)+'.json'
file = open(file_path, 'w')
file.writelines(html_uid)
file.close()
修改部分
file = open(file_path, 'w')增加encoding='utf-8'即 file =open(file_path,'w',encoding='utf-8')

本文介绍了一种常见的网络爬虫编码错误:gbk编码无法处理特定字符的问题,并提供了解决方案,通过设置文件写入编码为utf-8来避免非法多字节序列错误。
10万+

被折叠的 条评论
为什么被折叠?



