如何解决请求 socialblade.com 使用超级错误获取 403
我的蜘蛛网址:socialblade.com
我使用超级解决特殊标题
我把chrome中request header涉及的字段全部复制到代码中,bug不行。 return 403 or socket is time out
我的代码
url = "https://socialblade.com/youtube/channel/UCqmld-BIYME2i_ooRTo1EOg"
path = url.replace("https://socialblade.com","")
headers = {
":authority": "socialblade.com",":method": "GET",":path": path,":scheme": "https","accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.9","accept-encoding": "gzip,deflate,br","accept-language": "zh-CN,zh;q=0.9","sec-ch-ua": 'Not;A Brand";v="99","Google Chrome";v="91","Chromium";v="91"',"sec-ch-ua-mobile": "?0","sec-fetch-dest": "document","sec-fetch-mode": "navigate","sec-fetch-site": "none","sec-fetch-user": "?1","upgrade-insecure-requests": "1","User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML,like Gecko) Chrome/91.0.4472.124 Safari/537.36",}
def get():
sessions = requests.session()
sessions.mount("https://socialblade.com",HTTP20Adapter())
response = sessions.get(url,headers=headers,timeout=30)
status_code = response.status_code
print(f"path:{path} 请求状态:{status_code}")
if status_code != 200:
return
text=response.text
print(text)
soup=BeautifulSoup(text,'lxml')
div_ele=soup.find_all('div',id='socialblade-user-content')
print(div_ele)
版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。