Spaces:
Build error
Build error
File size: 5,271 Bytes
3b2ecd2 494cdcf 973d2bc ca11d6d baf3cbb 7835185 e37298f 973d2bc e37298f efebe44 aee7417 efebe44 aee7417 efebe44 aee7417 a8ab2da cbb85f0 a8ab2da a1e6f2e 973d2bc 65230b0 19f5db9 992ce0a 247ee08 34b052b 247ee08 65230b0 247ee08 77269d1 247ee08 02fa17e 3895bea 247ee08 65230b0 247ee08 65230b0 3895bea 247ee08 efebe44 c7f05f3 efebe44 c7f05f3 4efdbc0 efebe44 992ce0a c7f05f3 efebe44 4196efc 34b052b 992ce0a 0d18c41 a1e6f2e 0d18c41 1bf1c75 7835185 1bf1c75 0e44ac3 7835185 0e44ac3 1e513f0 7835185 efebe44 0d18c41 3895bea efebe44 8996078 efebe44 1ca82d8 7274bf1 efebe44 7274bf1 8996078 efebe44 8996078 efebe44 8996078 efebe44 8996078 efebe44 8996078 1497d6f a1e6f2e cbb85f0 a1e6f2e d02bd7e 51c48b8 d02bd7e cbb85f0 992ce0a 8996078 a8ab2da 4efdbc0 a1e6f2e 8996078 e37298f 973d2bc ca9d760 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 |
from seleniumwire import webdriver
from selenium.webdriver.chrome.options import Options
from fastapi import FastAPI, Request
import uvicorn
import time
import json
from urllib.parse import unquote, urlparse, quote_plus
app = FastAPI()
# 解析cookie字符串为字典
def convert_cookies_to_dict(cookies):
cookie_items = cookies.split("; ")
parsed_cookies = {item.split("=", 1)[0].strip(): item.split("=", 1)[1].strip() if "=" in item else "" for item in cookie_items}
return parsed_cookies#
# 获取域名字符串的根域
def get_root_domain(url):
parsed_url = urlparse(url)
domain = parsed_url.netloc
parts = domain.split('.')
if len(parts) > 1:
return '.'.join(parts[-2:])
else:
return domain
# 尝试对字符串做 json 解析,如果失败则返回原字符串
def try_json_decode(headers):
try:
return json.loads(str(headers))
except Exception as e:
return headers
@app.get("/")
def main():
return {"code": 200,"msg":"Success"}
@app.get("/chrome")
def chrome(url:str=None,wait:int=5,header:str=None,cookie:str=None,cookie_domain:str=None):
# 开启捕获HAR数据功能,允许使用 driver.har 进行检索
seleniumwire_options = {
'enable_har': True
}
# 必须有目标url
if type(url) == str:
target_url = unquote(url)
target_domain = get_root_domain(target_url)
else:
return {"code": 500,"msg":"No target URL"}
# 等待时间必须在 0 到 30 之间
if wait in range(0, 31):
wait_time = wait
else:
return {"code": 500,"msg":"The waiting time must be between 0 and 30"}
header_array = {}
# header可以覆写,但必须传入json
try:
if type(header) == str:
header_array.update(json.loads(unquote(header)))
except Exception as e:
return {"code": 500,"msg":"The header field is not JSON"}
# 如果输入了cookie
if type(cookie) == str:
header_array.update({"cookie":unquote(cookie)})
# 初始化浏览器
options = Options()
# 设置为无头模式
options.add_argument('--headless')
# 实例化
driver = webdriver.Chrome(options=options,seleniumwire_options=seleniumwire_options)
# 需要打开网址页面,才能用 driver.add_cookie 进行cookie追加
driver.get(target_url)
# 清除本次打开网址页面,可能存储在本地的cookie、sessionStorage、localStorage,并删除因此次访问所产生的 network 和 har 记录
driver.delete_all_cookies()
driver.execute_script("window.sessionStorage.clear();")
driver.execute_script("window.localStorage.clear();")
del driver.requests
# 对浏览器追加我们传递进来的cookie
if 'cookie' in header_array:
cookie_array = convert_cookies_to_dict(header_array['cookie'])
del header_array['cookie']
if type(cookie_domain) == str:
domain = cookie_domain
else:
domain = f'.{target_domain}'
for key, value in cookie_array.items():
try:
driver.add_cookie({"name": key, "value": quote_plus(value), "domain": domain, "path": "/"})
except Exception as e:
print("Error Cookie:")
print({"name": key, "value": quote_plus(value), "domain": domain, "path": "/"})
# 把下次访问中的请求头修改成我们需要的样式(没有修改的项目则保持原样)
driver.header_overrides = header_array
# 再次访问网址
driver.get(target_url)
# 输出此时访问的网页源码
# print(driver.page_source)
# 等待多少秒,来预估网页完全的加载完成(执行完内部的所有js,因为部分js可能涉及到请求后的动态处理,或者延时跳转)
if wait_time > 0:
time.sleep(wait_time)
# 获取完全加载完成时,页面的URL
current_url = driver.current_url
# 获取完全加载完成时,页面的源代码
page_source = driver.page_source
# 获取完全加载完成时,页面的cookie
cookies = driver.get_cookies()
# 完全加载完成时,页面是否有发生过 301 302 跳转过
is_jump = (target_url != current_url)
network = []
# 遍历输出过程中的 network(使用非 har 文件的摘要方式输出)
for request in driver.requests:
if request.response:
network.append({
"method":request.method,
"status":request.response.status_code ,
"url":request.url,
"responseheaders":{k: try_json_decode(v) for k, v in request.response.headers.items()},
"requestheaders":{k: try_json_decode(v) for k, v in request.headers.items()},
})
# driver.har 将调用 har 记录,输出最为完整的 network 数据流
# print(driver.har)
data = {
"url": current_url,
"page_source": page_source,
"end_cookies": cookies,
"is_jump": is_jump,
"network": network
}
driver.quit()
return {"code": 200,"data":data}
if __name__ == '__main__':
uvicorn.run(app='app:app', host="0.0.0.0", port=7860) |