Add hoster engine with header support
This commit is contained in:
2
.env
2
.env
@@ -20,7 +20,7 @@ JELLYFIN_MOVIES_DIR=/jellyfin/Filme
|
|||||||
JELLYFIN_SERIES_DIR=/jellyfin/Serien
|
JELLYFIN_SERIES_DIR=/jellyfin/Serien
|
||||||
|
|
||||||
# Engines
|
# Engines
|
||||||
ENGINE_DEFAULT=auto # auto|ytdlp|direct
|
ENGINE_DEFAULT=auto # auto|ytdlp|direct|hoster
|
||||||
YTDLP_FORMAT=bestvideo+bestaudio/best
|
YTDLP_FORMAT=bestvideo+bestaudio/best
|
||||||
|
|
||||||
# Proxy pool (ONLY used for downloads; not for SFTP / webgui internal calls)
|
# Proxy pool (ONLY used for downloads; not for SFTP / webgui internal calls)
|
||||||
|
|||||||
@@ -3,6 +3,7 @@
|
|||||||
Dieses Projekt lädt Links über:
|
Dieses Projekt lädt Links über:
|
||||||
- **yt-dlp** (YouTube & unterstützte Video-Plattformen)
|
- **yt-dlp** (YouTube & unterstützte Video-Plattformen)
|
||||||
- **aria2c** (direkte HTTP/HTTPS-Links, z.B. .mkv/.mp4)
|
- **aria2c** (direkte HTTP/HTTPS-Links, z.B. .mkv/.mp4)
|
||||||
|
- **hoster** (aria2c mit optionalen HTTP-Headern, z.B. Cookies)
|
||||||
|
|
||||||
Danach:
|
Danach:
|
||||||
- erzeugt es eine **MD5** und speichert sie als Sidecar
|
- erzeugt es eine **MD5** und speichert sie als Sidecar
|
||||||
@@ -22,3 +23,7 @@ Danach:
|
|||||||
## Proxies
|
## Proxies
|
||||||
- Proxies werden **nur** an yt-dlp/aria2 übergeben (pro Job), beeinflussen also nicht SFTP/Jellyfin.
|
- Proxies werden **nur** an yt-dlp/aria2 übergeben (pro Job), beeinflussen also nicht SFTP/Jellyfin.
|
||||||
- `PROXY_LIST` enthält eine Zeile pro Proxy: `socks5://IP:PORT`, `http://IP:PORT`, ...
|
- `PROXY_LIST` enthält eine Zeile pro Proxy: `socks5://IP:PORT`, `http://IP:PORT`, ...
|
||||||
|
|
||||||
|
## Hoster-Engine
|
||||||
|
- Engine `hoster` nutzt **aria2c** und akzeptiert zusätzliche HTTP-Header (z.B. `Cookie:` oder `User-Agent:`) im Formular.
|
||||||
|
- Ein Header pro Zeile. Leere Zeilen/Kommentare werden ignoriert.
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
+26-2
|
|
||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
from __future__ import annotations
|
from __future__ import annotations
|
||||||
|
|
||||||
@@ -54,9 +53,11 @@ SERIES_RE = re.compile(r"(?:^|[^a-z0-9])S(\d{1,2})E(\d{1,2})(?:[^a-z0-9]|$)", re
|
|||||||
app = FastAPI()
|
app = FastAPI()
|
||||||
app.mount("/static", StaticFiles(directory="static"), name="static")
|
app.mount("/static", StaticFiles(directory="static"), name="static")
|
||||||
|
|
||||||
|
|
||||||
def _auth_enabled() -> bool:
|
def _auth_enabled() -> bool:
|
||||||
return bool(BASIC_AUTH_USER and BASIC_AUTH_PASS)
|
return bool(BASIC_AUTH_USER and BASIC_AUTH_PASS)
|
||||||
|
|
||||||
|
|
||||||
def _check_basic_auth(req: Request) -> bool:
|
def _check_basic_auth(req: Request) -> bool:
|
||||||
if not _auth_enabled():
|
if not _auth_enabled():
|
||||||
return True
|
return True
|
||||||
@@ -69,17 +70,39 @@ def _check_basic_auth(req: Request) -> bool:
|
|||||||
except Exception:
|
except Exception:
|
||||||
return False
|
return False
|
||||||
if ":" not in raw:
|
if ":" not in raw:
|
||||||
@@ -82,88 +88,106 @@ class Job:
|
return False
|
||||||
|
user, pw = raw.split(":", 1)
|
||||||
|
return user == BASIC_AUTH_USER and pw == BASIC_AUTH_PASS
|
||||||
|
|
||||||
|
|
||||||
|
def _auth_challenge() -> HTMLResponse:
|
||||||
|
return HTMLResponse("Authentication required", status_code=401, headers={"WWW-Authenticate": 'Basic realm="media-webgui"'})
|
||||||
|
|
||||||
|
|
||||||
|
@app.middleware("http")
|
||||||
|
async def basic_auth_middleware(request: Request, call_next):
|
||||||
|
if not _check_basic_auth(request):
|
||||||
|
return _auth_challenge()
|
||||||
|
return await call_next(request)
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class Job:
|
||||||
|
id: str
|
||||||
|
url: str
|
||||||
engine: str
|
engine: str
|
||||||
library: str
|
library: str
|
||||||
proxy: str
|
proxy: str
|
||||||
|
headers: List[str]
|
||||||
status: str
|
status: str
|
||||||
message: str
|
message: str
|
||||||
|
|
||||||
|
|
||||||
jobs: Dict[str, Job] = {}
|
jobs: Dict[str, Job] = {}
|
||||||
lock = threading.Lock()
|
lock = threading.Lock()
|
||||||
_rr_idx = 0
|
_rr_idx = 0
|
||||||
|
|
||||||
|
|
||||||
def parse_proxy_list(raw: str) -> List[str]:
|
def parse_proxy_list(raw: str) -> List[str]:
|
||||||
out = []
|
out = []
|
||||||
for line in (raw or "").splitlines():
|
for line in (raw or "").splitlines():
|
||||||
@@ -95,6 +118,7 @@ def parse_proxy_list(raw: str) -> List[str]:
|
|||||||
dedup.append(x)
|
dedup.append(x)
|
||||||
return dedup
|
return dedup
|
||||||
|
|
||||||
|
|
||||||
def pick_proxy(forced_proxy: str = "") -> str:
|
def pick_proxy(forced_proxy: str = "") -> str:
|
||||||
global _rr_idx
|
global _rr_idx
|
||||||
if forced_proxy:
|
if forced_proxy:
|
||||||
@@ -107,6 +131,7 @@ def pick_proxy(forced_proxy: str = "") -> str:
|
|||||||
_rr_idx += 1
|
_rr_idx += 1
|
||||||
return p
|
return p
|
||||||
|
|
||||||
|
|
||||||
def format_proxy_lines(raw: str, scheme: str) -> str:
|
def format_proxy_lines(raw: str, scheme: str) -> str:
|
||||||
scheme = scheme.strip().lower()
|
scheme = scheme.strip().lower()
|
||||||
if scheme not in {"socks5", "socks4", "http", "https"}:
|
if scheme not in {"socks5", "socks4", "http", "https"}:
|
||||||
@@ -125,16 +150,20 @@ def format_proxy_lines(raw: str, scheme: str) -> str:
|
|||||||
if not host or not port.isdigit():
|
if not host or not port.isdigit():
|
||||||
continue
|
continue
|
||||||
out.append(f"{scheme}://{host}:{port}")
|
out.append(f"{scheme}://{host}:{port}")
|
||||||
seen=set(); ded=[]
|
seen = set()
|
||||||
|
ded = []
|
||||||
for x in out:
|
for x in out:
|
||||||
if x not in seen:
|
if x not in seen:
|
||||||
seen.add(x); ded.append(x)
|
seen.add(x)
|
||||||
|
ded.append(x)
|
||||||
return "\n".join(ded)
|
return "\n".join(ded)
|
||||||
|
|
||||||
|
|
||||||
def fetch_proxy_source(url: str) -> str:
|
def fetch_proxy_source(url: str) -> str:
|
||||||
with urlopen(url, timeout=20) as resp:
|
with urlopen(url, timeout=20) as resp:
|
||||||
return resp.read().decode("utf-8", "replace")
|
return resp.read().decode("utf-8", "replace")
|
||||||
|
|
||||||
|
|
||||||
def load_proxy_sources() -> List[str]:
|
def load_proxy_sources() -> List[str]:
|
||||||
chunks = []
|
chunks = []
|
||||||
for scheme, url in PROXY_SOURCES.items():
|
for scheme, url in PROXY_SOURCES.items():
|
||||||
@@ -149,8 +178,27 @@ def load_proxy_sources() -> List[str]:
|
|||||||
combined = "\n".join(chunks)
|
combined = "\n".join(chunks)
|
||||||
return parse_proxy_list(combined)
|
return parse_proxy_list(combined)
|
||||||
|
|
||||||
|
|
||||||
PROXIES = parse_proxy_list("\n".join([PROXY_LIST_RAW, "\n".join(load_proxy_sources())]))
|
PROXIES = parse_proxy_list("\n".join([PROXY_LIST_RAW, "\n".join(load_proxy_sources())]))
|
||||||
|
|
||||||
|
|
||||||
|
def parse_header_lines(raw: str) -> List[str]:
|
||||||
|
headers = []
|
||||||
|
for line in (raw or "").splitlines():
|
||||||
|
s = line.strip()
|
||||||
|
if not s or s.startswith("#"):
|
||||||
|
continue
|
||||||
|
if ":" not in s:
|
||||||
|
raise ValueError(f"Invalid header line: {s}")
|
||||||
|
name, value = s.split(":", 1)
|
||||||
|
name = name.strip()
|
||||||
|
value = value.strip()
|
||||||
|
if not name or not value:
|
||||||
|
raise ValueError(f"Invalid header line: {s}")
|
||||||
|
headers.append(f"{name}: {value}")
|
||||||
|
return headers
|
||||||
|
|
||||||
|
|
||||||
def pick_engine(url: str, forced: str) -> str:
|
def pick_engine(url: str, forced: str) -> str:
|
||||||
forced = (forced or "").strip().lower()
|
forced = (forced or "").strip().lower()
|
||||||
if forced and forced != "auto":
|
if forced and forced != "auto":
|
||||||
@@ -162,17 +210,370 @@ def pick_engine(url: str, forced: str) -> str:
|
|||||||
return "direct"
|
return "direct"
|
||||||
return "direct"
|
return "direct"
|
||||||
|
|
||||||
|
|
||||||
def run_ytdlp(url: str, out_dir: str, fmt: str, proxy: str):
|
def run_ytdlp(url: str, out_dir: str, fmt: str, proxy: str):
|
||||||
cmd = ["yt-dlp", "-f", fmt, "-o", f"{out_dir}/%(title)s.%(ext)s", url]
|
cmd = ["yt-dlp", "-f", fmt, "-o", f"{out_dir}/%(title)s.%(ext)s", url]
|
||||||
if proxy:
|
if proxy:
|
||||||
cmd += ["--proxy", proxy]
|
cmd += ["--proxy", proxy]
|
||||||
subprocess.check_call(cmd)
|
subprocess.check_call(cmd)
|
||||||
|
|
||||||
def run_aria2(url: str, out_dir: str, proxy: str):
|
|
||||||
|
def run_aria2(url: str, out_dir: str, proxy: str, headers: List[str] | None = None):
|
||||||
cmd = ["aria2c", "--dir", out_dir, "--allow-overwrite=true", "--auto-file-renaming=false", url]
|
cmd = ["aria2c", "--dir", out_dir, "--allow-overwrite=true", "--auto-file-renaming=false", url]
|
||||||
if proxy:
|
if proxy:
|
||||||
cmd += ["--all-proxy", proxy]
|
cmd += ["--all-proxy", proxy]
|
||||||
|
for header in headers or []:
|
||||||
|
cmd += ["--header", header]
|
||||||
subprocess.check_call(cmd)
|
subprocess.check_call(cmd)
|
||||||
|
|
||||||
|
|
||||||
def md5_file(path: str) -> str:
|
def md5_file(path: str) -> str:
|
||||||
h = hashlib.md5()
|
h = hashlib.md5()
|
||||||
|
with open(path, "rb") as f:
|
||||||
|
for chunk in iter(lambda: f.read(1024 * 1024), b""):
|
||||||
|
h.update(chunk)
|
||||||
|
return h.hexdigest()
|
||||||
|
|
||||||
|
|
||||||
|
def write_md5_sidecar(local_file: str, md5_hex: str) -> str:
|
||||||
|
os.makedirs(MD5_DIR, exist_ok=True)
|
||||||
|
base = os.path.basename(local_file)
|
||||||
|
md5p = os.path.join(MD5_DIR, base + ".md5")
|
||||||
|
with open(md5p, "w", encoding="utf-8") as f:
|
||||||
|
f.write(f"{md5_hex} {base}\n")
|
||||||
|
return md5p
|
||||||
|
|
||||||
|
|
||||||
|
def ssh_connect() -> paramiko.SSHClient:
|
||||||
|
if not JELLYFIN_USER:
|
||||||
|
raise RuntimeError("JELLYFIN_USER missing")
|
||||||
|
ssh = paramiko.SSHClient()
|
||||||
|
ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy())
|
||||||
|
ssh.connect(
|
||||||
|
hostname=JELLYFIN_HOST,
|
||||||
|
port=JELLYFIN_PORT,
|
||||||
|
username=JELLYFIN_USER,
|
||||||
|
key_filename=JELLYFIN_SSH_KEY,
|
||||||
|
timeout=30,
|
||||||
|
)
|
||||||
|
return ssh
|
||||||
|
|
||||||
|
|
||||||
|
def sftp_mkdirs(sftp: paramiko.SFTPClient, remote_dir: str):
|
||||||
|
parts = [p for p in remote_dir.split("/") if p]
|
||||||
|
cur = ""
|
||||||
|
for p in parts:
|
||||||
|
cur += "/" + p
|
||||||
|
try:
|
||||||
|
sftp.stat(cur)
|
||||||
|
except IOError:
|
||||||
|
sftp.mkdir(cur)
|
||||||
|
|
||||||
|
|
||||||
|
def sftp_upload(ssh: paramiko.SSHClient, local_path: str, remote_path: str):
|
||||||
|
sftp = ssh.open_sftp()
|
||||||
|
try:
|
||||||
|
sftp_mkdirs(sftp, os.path.dirname(remote_path))
|
||||||
|
sftp.put(local_path, remote_path)
|
||||||
|
except Exception as e:
|
||||||
|
raise RuntimeError(f"SFTP upload failed: local={local_path} remote={remote_path} error={e}")
|
||||||
|
finally:
|
||||||
|
sftp.close()
|
||||||
|
|
||||||
|
|
||||||
|
def remote_md5sum(ssh: paramiko.SSHClient, remote_path: str) -> str:
|
||||||
|
cmd = f"md5sum {shlex.quote(remote_path)}"
|
||||||
|
_, stdout, stderr = ssh.exec_command(cmd, timeout=120)
|
||||||
|
out = stdout.read().decode("utf-8", "replace").strip()
|
||||||
|
err = stderr.read().decode("utf-8", "replace").strip()
|
||||||
|
if err and not out:
|
||||||
|
raise RuntimeError(f"Remote md5sum failed: {err}")
|
||||||
|
if not out:
|
||||||
|
raise RuntimeError("Remote md5sum returned empty output")
|
||||||
|
return out.split()[0]
|
||||||
|
|
||||||
|
|
||||||
|
def choose_target_dir(library: str, filename: str) -> str:
|
||||||
|
library = (library or "auto").lower()
|
||||||
|
if library == "series":
|
||||||
|
return JELLYFIN_SERIES_DIR
|
||||||
|
if library == "movies":
|
||||||
|
return JELLYFIN_MOVIES_DIR
|
||||||
|
if SERIES_RE.search(filename):
|
||||||
|
return JELLYFIN_SERIES_DIR
|
||||||
|
return JELLYFIN_MOVIES_DIR
|
||||||
|
|
||||||
|
|
||||||
|
def list_output_files(before: set) -> List[str]:
|
||||||
|
now = set()
|
||||||
|
for root, _, files in os.walk(OUTPUT_DIR):
|
||||||
|
for fn in files:
|
||||||
|
now.add(os.path.join(root, fn))
|
||||||
|
new = [p for p in sorted(now) if p not in before]
|
||||||
|
final = []
|
||||||
|
for p in new:
|
||||||
|
low = p.lower()
|
||||||
|
if low.endswith((".part", ".tmp", ".crdownload")):
|
||||||
|
continue
|
||||||
|
final.append(p)
|
||||||
|
return final
|
||||||
|
|
||||||
|
|
||||||
|
def worker(jobid: str):
|
||||||
|
try:
|
||||||
|
with lock:
|
||||||
|
job = jobs[jobid]
|
||||||
|
|
||||||
|
os.makedirs(OUTPUT_DIR, exist_ok=True)
|
||||||
|
|
||||||
|
before = set()
|
||||||
|
for root, _, files in os.walk(OUTPUT_DIR):
|
||||||
|
for fn in files:
|
||||||
|
before.add(os.path.join(root, fn))
|
||||||
|
|
||||||
|
engine = pick_engine(job.url, job.engine)
|
||||||
|
proxy = job.proxy
|
||||||
|
headers = job.headers
|
||||||
|
|
||||||
|
with lock:
|
||||||
|
header_note = f"Headers={len(headers)}" if headers else "Headers=none"
|
||||||
|
job.status = "downloading"
|
||||||
|
job.message = f"Engine={engine} Proxy={'none' if not proxy else proxy} {header_note}"
|
||||||
|
|
||||||
|
if engine == "ytdlp":
|
||||||
|
run_ytdlp(job.url, OUTPUT_DIR, YTDLP_FORMAT, proxy)
|
||||||
|
elif engine == "hoster":
|
||||||
|
run_aria2(job.url, OUTPUT_DIR, proxy, headers=headers)
|
||||||
|
else:
|
||||||
|
run_aria2(job.url, OUTPUT_DIR, proxy)
|
||||||
|
|
||||||
|
new_files = list_output_files(before)
|
||||||
|
if not new_files:
|
||||||
|
raise RuntimeError("No output file detected in /output")
|
||||||
|
|
||||||
|
ssh = ssh_connect()
|
||||||
|
try:
|
||||||
|
for f in new_files:
|
||||||
|
if not os.path.isfile(f):
|
||||||
|
continue
|
||||||
|
|
||||||
|
md5_hex = md5_file(f)
|
||||||
|
md5_path = write_md5_sidecar(f, md5_hex)
|
||||||
|
|
||||||
|
target_dir = choose_target_dir(job.library, os.path.basename(f))
|
||||||
|
remote_file = f"{target_dir}/{os.path.basename(f)}"
|
||||||
|
remote_md5f = remote_file + ".md5"
|
||||||
|
|
||||||
|
with lock:
|
||||||
|
job.status = "upload"
|
||||||
|
job.message = f"Uploading: {os.path.basename(f)} -> {remote_file}"
|
||||||
|
|
||||||
|
sftp_upload(ssh, f, remote_file)
|
||||||
|
sftp_upload(ssh, md5_path, remote_md5f)
|
||||||
|
|
||||||
|
remote_md5 = remote_md5sum(ssh, remote_file)
|
||||||
|
if remote_md5.lower() != md5_hex.lower():
|
||||||
|
raise RuntimeError(f"MD5 mismatch: local={md5_hex} remote={remote_md5}")
|
||||||
|
|
||||||
|
try:
|
||||||
|
os.remove(f)
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
try:
|
||||||
|
os.remove(md5_path)
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
|
finally:
|
||||||
|
ssh.close()
|
||||||
|
|
||||||
|
with lock:
|
||||||
|
job.status = "finished"
|
||||||
|
job.message = f"OK ({len(new_files)} file(s))"
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
with lock:
|
||||||
|
jobs[jobid].status = "failed"
|
||||||
|
jobs[jobid].message = str(e)
|
||||||
|
|
||||||
|
|
||||||
|
def render_nav(active: str) -> str:
|
||||||
|
def link(label: str, href: str, key: str) -> str:
|
||||||
|
style = "font-weight:700;" if active == key else ""
|
||||||
|
return f"<a href='{href}' style='margin-right:14px; {style}'>{label}</a>"
|
||||||
|
|
||||||
|
return "<div style='margin: 8px 0 14px 0;'>" + link("Downloads", "/", "downloads") + link("Proxies", "/proxies", "proxies") + "</div>"
|
||||||
|
|
||||||
|
|
||||||
|
def render_downloads(error: str = "") -> str:
|
||||||
|
rows = ""
|
||||||
|
with lock:
|
||||||
|
job_list = list(jobs.values())[::-1]
|
||||||
|
|
||||||
|
for j in job_list:
|
||||||
|
rows += (
|
||||||
|
f"<tr>"
|
||||||
|
f"<td><code>{j.id}</code></td>"
|
||||||
|
f"<td style='max-width:560px; word-break:break-all;'>{j.url}</td>"
|
||||||
|
f"<td>{j.engine}</td>"
|
||||||
|
f"<td>{j.library}</td>"
|
||||||
|
f"<td>{'none' if not j.proxy else j.proxy}</td>"
|
||||||
|
f"<td><b>{j.status}</b><br/><small>{j.message}</small></td>"
|
||||||
|
f"</tr>"
|
||||||
|
)
|
||||||
|
|
||||||
|
err_html = f"<p class='error'>{error}</p>" if error else ""
|
||||||
|
proxy_note = f"{len(PROXIES)} configured, mode={PROXY_MODE}" if PROXIES else "none configured"
|
||||||
|
return f"""
|
||||||
|
<html><head>
|
||||||
|
<link rel="stylesheet" href="/static/style.css">
|
||||||
|
<meta charset="utf-8">
|
||||||
|
<title>Media WebGUI</title>
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<h1>Media WebGUI</h1>
|
||||||
|
{render_nav("downloads")}
|
||||||
|
{err_html}
|
||||||
|
|
||||||
|
<form method="post" action="/submit">
|
||||||
|
<div class="row"><label>Link</label><br/>
|
||||||
|
<input name="url" placeholder="https://..." required />
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="row"><label>Engine</label><br/>
|
||||||
|
<select name="engine">
|
||||||
|
<option value="auto">auto</option>
|
||||||
|
<option value="ytdlp">ytdlp</option>
|
||||||
|
<option value="direct">direct (aria2)</option>
|
||||||
|
<option value="hoster">hoster (aria2 + headers)</option>
|
||||||
|
</select>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="row"><label>Library</label><br/>
|
||||||
|
<select name="library">
|
||||||
|
<option value="auto">auto</option>
|
||||||
|
<option value="movies">movies</option>
|
||||||
|
<option value="series">series</option>
|
||||||
|
</select>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="row"><label>Proxy (optional)</label><br/>
|
||||||
|
<input name="proxy" placeholder="leer = auto ({proxy_note})" />
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="row"><label>Hoster-Header (optional)</label><br/>
|
||||||
|
<textarea name="headers" rows="4" placeholder="Cookie: ...\nUser-Agent: ..." style="width:100%; max-width:860px; padding:10px; border:1px solid #ccc; border-radius:8px;"></textarea>
|
||||||
|
<div class="hint">Nur für Engine <code>hoster</code>. Ein Header pro Zeile.</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<button type="submit">Start</button>
|
||||||
|
</form>
|
||||||
|
|
||||||
|
<p class="hint">
|
||||||
|
Output: <code>{OUTPUT_DIR}</code> | MD5: <code>{MD5_DIR}</code> | Proxies: <code>{proxy_note}</code>
|
||||||
|
</p>
|
||||||
|
|
||||||
|
<table>
|
||||||
|
<thead><tr><th>JobID</th><th>URL</th><th>Engine</th><th>Library</th><th>Proxy</th><th>Status</th></tr></thead>
|
||||||
|
<tbody>
|
||||||
|
{rows if rows else "<tr><td colspan='6'><em>No jobs yet.</em></td></tr>"}
|
||||||
|
</tbody>
|
||||||
|
</table>
|
||||||
|
</body></html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/", response_class=HTMLResponse)
|
||||||
|
def index():
|
||||||
|
return HTMLResponse(render_downloads())
|
||||||
|
|
||||||
|
|
||||||
|
@app.post("/submit")
|
||||||
|
def submit(
|
||||||
|
url: str = Form(...),
|
||||||
|
engine: str = Form("auto"),
|
||||||
|
library: str = Form("auto"),
|
||||||
|
proxy: str = Form(""),
|
||||||
|
headers: str = Form(""),
|
||||||
|
):
|
||||||
|
url = url.strip()
|
||||||
|
if not URL_RE.match(url):
|
||||||
|
return HTMLResponse(render_downloads("Only http(s) URLs supported"), status_code=400)
|
||||||
|
|
||||||
|
engine = (engine or ENGINE_DEFAULT).strip().lower()
|
||||||
|
library = (library or "auto").strip().lower()
|
||||||
|
|
||||||
|
try:
|
||||||
|
header_lines = parse_header_lines(headers)
|
||||||
|
except ValueError as exc:
|
||||||
|
return HTMLResponse(render_downloads(str(exc)), status_code=400)
|
||||||
|
|
||||||
|
chosen_proxy = pick_proxy(proxy.strip())
|
||||||
|
jobid = str(int(time.time() * 1000))
|
||||||
|
|
||||||
|
with lock:
|
||||||
|
jobs[jobid] = Job(
|
||||||
|
id=jobid,
|
||||||
|
url=url,
|
||||||
|
engine=engine,
|
||||||
|
library=library,
|
||||||
|
proxy=chosen_proxy,
|
||||||
|
headers=header_lines,
|
||||||
|
status="queued",
|
||||||
|
message="queued",
|
||||||
|
)
|
||||||
|
|
||||||
|
t = threading.Thread(target=worker, args=(jobid,), daemon=True)
|
||||||
|
t.start()
|
||||||
|
return RedirectResponse(url="/", status_code=303)
|
||||||
|
|
||||||
|
|
||||||
|
def render_proxies_page(error: str = "", s5: str = "", s4: str = "", hp: str = "", out_text: str = "") -> str:
|
||||||
|
err_html = f"<p class='error'>{error}</p>" if error else ""
|
||||||
|
return f"""
|
||||||
|
<html><head>
|
||||||
|
<link rel="stylesheet" href="/static/style.css">
|
||||||
|
<meta charset="utf-8">
|
||||||
|
<title>Proxies</title>
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<h1>Media WebGUI</h1>
|
||||||
|
{render_nav("proxies")}
|
||||||
|
{err_html}
|
||||||
|
|
||||||
|
<form method="post" action="/proxies">
|
||||||
|
<div class="row"><label>SOCKS5 (IP:PORT je Zeile)</label><br/>
|
||||||
|
<textarea name="s5" rows="6" style="width:100%; max-width:860px; padding:10px; border:1px solid #ccc; border-radius:8px;">{s5}</textarea>
|
||||||
|
</div>
|
||||||
|
<div class="row"><label>SOCKS4 (IP:PORT je Zeile)</label><br/>
|
||||||
|
<textarea name="s4" rows="6" style="width:100%; max-width:860px; padding:10px; border:1px solid #ccc; border-radius:8px;">{s4}</textarea>
|
||||||
|
</div>
|
||||||
|
<div class="row"><label>HTTP (IP:PORT je Zeile)</label><br/>
|
||||||
|
<textarea name="hp" rows="6" style="width:100%; max-width:860px; padding:10px; border:1px solid #ccc; border-radius:8px;">{hp}</textarea>
|
||||||
|
</div>
|
||||||
|
<button type="submit">In Import-Format umwandeln</button>
|
||||||
|
</form>
|
||||||
|
|
||||||
|
<h2 style="margin-top:18px;">Import-Liste (zum Kopieren)</h2>
|
||||||
|
<p class="hint">Keine Prüfung/Validierung. In .env in <code>PROXY_LIST</code> einfügen (eine Zeile pro Proxy).</p>
|
||||||
|
<textarea id="out" rows="12" readonly style="width:100%; max-width:860px; padding:10px; border:1px solid #ccc; border-radius:8px;">{out_text}</textarea><br/>
|
||||||
|
<button type="button" onclick="navigator.clipboard.writeText(document.getElementById('out').value)">Kopieren</button>
|
||||||
|
</body></html>
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/proxies", response_class=HTMLResponse)
|
||||||
|
def proxies_get():
|
||||||
|
return HTMLResponse(render_proxies_page())
|
||||||
|
|
||||||
|
|
||||||
|
@app.post("/proxies", response_class=HTMLResponse)
|
||||||
|
def proxies_post(s5: str = Form(""), s4: str = Form(""), hp: str = Form("")):
|
||||||
|
try:
|
||||||
|
o1 = format_proxy_lines(s5, "socks5")
|
||||||
|
o2 = format_proxy_lines(s4, "socks4")
|
||||||
|
o3 = format_proxy_lines(hp, "http")
|
||||||
|
combined = "\n".join([x for x in [o1, o2, o3] if x.strip()])
|
||||||
|
return HTMLResponse(render_proxies_page(s5=s5, s4=s4, hp=hp, out_text=combined))
|
||||||
|
except Exception as e:
|
||||||
|
return HTMLResponse(render_proxies_page(error=str(e), s5=s5, s4=s4, hp=hp, out_text=""), status_code=400)
|
||||||
|
|||||||
Reference in New Issue
Block a user