Compare commits
10 Commits
codex/add-
...
codex/conf
| Author | SHA1 | Date | |
|---|---|---|---|
| c3aac479fe | |||
| 1350b50199 | |||
| 6b06134edf | |||
| be4785b04a | |||
| db39f2b55e | |||
| a0e7ed91c7 | |||
| 7443a0e0ca | |||
| 3cf7581797 | |||
| e9ccb51f13 | |||
| a549ba66ba |
151
jd-webgui/app.py
151
jd-webgui/app.py
@@ -18,7 +18,7 @@ from typing import Any, Dict, List, Optional, Tuple
|
||||
from myjdapi import Myjdapi
|
||||
import paramiko
|
||||
from fastapi import FastAPI, Form, Request
|
||||
from fastapi.responses import HTMLResponse, RedirectResponse
|
||||
from fastapi.responses import HTMLResponse, PlainTextResponse, RedirectResponse
|
||||
from fastapi.staticfiles import StaticFiles
|
||||
|
||||
# ============================================================
|
||||
@@ -57,9 +57,12 @@ POLL_SECONDS = float(os.environ.get("POLL_SECONDS", "5"))
|
||||
# JDownloader writes here inside container
|
||||
JD_OUTPUT_PATH = "/output"
|
||||
PROXY_EXPORT_PATH = os.environ.get("PROXY_EXPORT_PATH", "/output/jd-proxies.jdproxies")
|
||||
LOG_BUFFER_LIMIT = int(os.environ.get("LOG_BUFFER_LIMIT", "500"))
|
||||
|
||||
URL_RE = re.compile(r"^https?://", re.I)
|
||||
|
||||
NO_PROXY_OPENER = urllib.request.build_opener(urllib.request.ProxyHandler({}))
|
||||
|
||||
VIDEO_EXTS = {
|
||||
".mkv", ".mp4", ".m4v", ".avi", ".mov", ".wmv", ".flv", ".webm",
|
||||
".ts", ".m2ts", ".mts", ".mpg", ".mpeg", ".vob", ".ogv",
|
||||
@@ -123,6 +126,21 @@ class Job:
|
||||
|
||||
jobs: Dict[str, Job] = {}
|
||||
lock = threading.Lock()
|
||||
log_lock = threading.Lock()
|
||||
connection_logs: List[str] = []
|
||||
|
||||
def log_connection(message: str) -> None:
|
||||
timestamp = time.strftime("%Y-%m-%d %H:%M:%S")
|
||||
line = f"[{timestamp}] {message}"
|
||||
with log_lock:
|
||||
connection_logs.append(line)
|
||||
if len(connection_logs) > LOG_BUFFER_LIMIT:
|
||||
excess = len(connection_logs) - LOG_BUFFER_LIMIT
|
||||
del connection_logs[:excess]
|
||||
|
||||
def get_connection_logs() -> str:
|
||||
with log_lock:
|
||||
return "\n".join(connection_logs)
|
||||
|
||||
# ============================================================
|
||||
# Core helpers
|
||||
@@ -149,6 +167,7 @@ def ensure_env():
|
||||
|
||||
def get_device():
|
||||
jd = Myjdapi()
|
||||
log_connection(f"MyJDownloader connect as {MYJD_EMAIL or 'unknown'}")
|
||||
jd.connect(MYJD_EMAIL, MYJD_PASSWORD)
|
||||
|
||||
wanted = (MYJD_DEVICE or "").strip()
|
||||
@@ -246,6 +265,7 @@ def ffprobe_ok(path: str) -> bool:
|
||||
def ssh_connect() -> paramiko.SSHClient:
|
||||
ssh = paramiko.SSHClient()
|
||||
ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy())
|
||||
log_connection(f"SSH connect {JELLYFIN_USER}@{JELLYFIN_HOST}:{JELLYFIN_PORT}")
|
||||
ssh.connect(
|
||||
hostname=JELLYFIN_HOST,
|
||||
port=JELLYFIN_PORT,
|
||||
@@ -268,6 +288,7 @@ def sftp_mkdirs(sftp: paramiko.SFTPClient, remote_dir: str):
|
||||
def sftp_upload(ssh: paramiko.SSHClient, local_path: str, remote_path: str):
|
||||
sftp = ssh.open_sftp()
|
||||
try:
|
||||
log_connection(f"SFTP upload {local_path} -> {remote_path}")
|
||||
sftp_mkdirs(sftp, os.path.dirname(remote_path))
|
||||
sftp.put(local_path, remote_path)
|
||||
finally:
|
||||
@@ -276,6 +297,7 @@ def sftp_upload(ssh: paramiko.SSHClient, local_path: str, remote_path: str):
|
||||
def remote_md5sum(ssh: paramiko.SSHClient, remote_path: str) -> str:
|
||||
quoted = shlex.quote(remote_path)
|
||||
cmd = f"md5sum {quoted}"
|
||||
log_connection(f"SSH exec {cmd}")
|
||||
stdin, stdout, stderr = ssh.exec_command(cmd, timeout=120)
|
||||
out = stdout.read().decode("utf-8", "replace").strip()
|
||||
err = stderr.read().decode("utf-8", "replace").strip()
|
||||
@@ -290,7 +312,8 @@ def remote_md5sum(ssh: paramiko.SSHClient, remote_path: str) -> str:
|
||||
# ============================================================
|
||||
def _http_get_json(url: str, headers: Optional[Dict[str, str]] = None) -> Any:
|
||||
req = urllib.request.Request(url, headers=headers or {})
|
||||
with urllib.request.urlopen(req, timeout=20) as r:
|
||||
log_connection(f"HTTP GET {url} (no-proxy)")
|
||||
with NO_PROXY_OPENER.open(req, timeout=20) as r:
|
||||
return json.loads(r.read().decode("utf-8", "replace"))
|
||||
|
||||
def tmdb_search_movie(query: str) -> Optional[Dict[str, Any]]:
|
||||
@@ -363,18 +386,58 @@ def format_proxy_lines(raw: str, scheme: str) -> str:
|
||||
|
||||
def fetch_proxy_list(url: str) -> str:
|
||||
req = urllib.request.Request(url)
|
||||
with urllib.request.urlopen(req, timeout=20) as resp:
|
||||
log_connection(f"HTTP GET {url} (no-proxy)")
|
||||
with NO_PROXY_OPENER.open(req, timeout=20) as resp:
|
||||
return resp.read().decode("utf-8", "replace")
|
||||
|
||||
def build_jdproxies_payload(text: str) -> Dict[str, Any]:
|
||||
if not text.strip():
|
||||
raise ValueError("Keine Proxy-Einträge zum Speichern.")
|
||||
blacklist_filter = {
|
||||
"entries": [
|
||||
"# Dies ist ein Kommentar",
|
||||
"// Dies ist auch ein Kommentar",
|
||||
"# Für jdownloader.org auskommentieren",
|
||||
"# jdownloader.org",
|
||||
"# unten für alle Accounts mit der ID 'test *' @ jdownloader.org auskommentieren",
|
||||
"#test@jdownloader.org",
|
||||
"# Kommentar unten für ein Konto mit der ID 'test' @ jdownloader.org",
|
||||
"#test$@jdownloader.org",
|
||||
"# Sie können Muster für Konto-ID und Host verwenden, z. B. accountPattern @ hostPattern",
|
||||
"",
|
||||
"my.jdownloader.org",
|
||||
"",
|
||||
"api.jdownloader.org",
|
||||
"",
|
||||
"*.jdownloader.org",
|
||||
"",
|
||||
"*.your-server.de",
|
||||
],
|
||||
"type": "BLACKLIST",
|
||||
}
|
||||
entries: List[Dict[str, Any]] = []
|
||||
type_map = {
|
||||
"socks5": "SOCKS5",
|
||||
"socks4": "SOCKS4",
|
||||
"http": "HTTP",
|
||||
}
|
||||
entries.append({
|
||||
"filter": None,
|
||||
"proxy": {
|
||||
"address": None,
|
||||
"password": None,
|
||||
"port": 80,
|
||||
"type": "NONE",
|
||||
"username": None,
|
||||
"connectMethodPrefered": False,
|
||||
"preferNativeImplementation": False,
|
||||
"resolveHostName": False,
|
||||
},
|
||||
"enabled": True,
|
||||
"pac": False,
|
||||
"rangeRequestsSupported": True,
|
||||
"reconnectSupported": True,
|
||||
})
|
||||
for line in text.splitlines():
|
||||
s = line.strip()
|
||||
if not s:
|
||||
@@ -386,7 +449,7 @@ def build_jdproxies_payload(text: str) -> Dict[str, Any]:
|
||||
if not proxy_type:
|
||||
continue
|
||||
entries.append({
|
||||
"filter": None,
|
||||
"filter": blacklist_filter,
|
||||
"proxy": {
|
||||
"address": parsed.hostname,
|
||||
"password": None,
|
||||
@@ -412,6 +475,8 @@ def save_proxy_export(text: str) -> str:
|
||||
export_dir = os.path.dirname(export_path)
|
||||
if export_dir:
|
||||
os.makedirs(export_dir, exist_ok=True)
|
||||
if os.path.exists(export_path):
|
||||
os.remove(export_path)
|
||||
with open(export_path, "w", encoding="utf-8") as handle:
|
||||
handle.write(json.dumps(payload, indent=2))
|
||||
handle.write("\n")
|
||||
@@ -488,7 +553,8 @@ def jellyfin_refresh_library():
|
||||
try:
|
||||
url = JELLYFIN_API_BASE + path
|
||||
req = urllib.request.Request(url, headers=headers, method="POST")
|
||||
with urllib.request.urlopen(req, timeout=20) as r:
|
||||
log_connection(f"HTTP POST {url} (no-proxy)")
|
||||
with NO_PROXY_OPENER.open(req, timeout=20) as r:
|
||||
_ = r.read()
|
||||
return
|
||||
except Exception:
|
||||
@@ -782,7 +848,19 @@ def worker(jobid: str):
|
||||
def favicon():
|
||||
return HTMLResponse(status_code=204)
|
||||
|
||||
def render_page(error: str = "") -> str:
|
||||
@app.get("/jobs", response_class=HTMLResponse)
|
||||
def jobs_get():
|
||||
return HTMLResponse(render_job_rows())
|
||||
|
||||
@app.get("/logs", response_class=HTMLResponse)
|
||||
def logs_get():
|
||||
return HTMLResponse(render_logs_page())
|
||||
|
||||
@app.get("/logs/data", response_class=PlainTextResponse)
|
||||
def logs_data():
|
||||
return PlainTextResponse(get_connection_logs())
|
||||
|
||||
def render_job_rows() -> str:
|
||||
rows = ""
|
||||
with lock:
|
||||
job_list = list(jobs.values())[::-1]
|
||||
@@ -812,6 +890,13 @@ def render_page(error: str = "") -> str:
|
||||
f"</tr>"
|
||||
)
|
||||
|
||||
if not rows:
|
||||
rows = "<tr><td colspan='5'><em>No jobs yet.</em></td></tr>"
|
||||
return rows
|
||||
|
||||
def render_page(error: str = "") -> str:
|
||||
rows = render_job_rows()
|
||||
|
||||
err_html = f"<p class='error'>{error}</p>" if error else ""
|
||||
auth_note = "aktiv" if _auth_enabled() else "aus"
|
||||
return f"""
|
||||
@@ -821,10 +906,18 @@ def render_page(error: str = "") -> str:
|
||||
<meta charset="utf-8">
|
||||
<title>JD → Jellyfin</title>
|
||||
<script>
|
||||
setInterval(() => {{
|
||||
async function refreshJobs() {{
|
||||
if (document.hidden) return;
|
||||
window.location.reload();
|
||||
}}, 5000);
|
||||
try {{
|
||||
const resp = await fetch('/jobs');
|
||||
if (!resp.ok) return;
|
||||
const html = await resp.text();
|
||||
const tbody = document.getElementById('jobs-body');
|
||||
if (tbody) tbody.innerHTML = html;
|
||||
}} catch (e) {{
|
||||
}}
|
||||
}}
|
||||
setInterval(refreshJobs, 5000);
|
||||
</script>
|
||||
</head>
|
||||
<body>
|
||||
@@ -862,8 +955,8 @@ def render_page(error: str = "") -> str:
|
||||
<thead>
|
||||
<tr><th>JobID</th><th>URL</th><th>Paket</th><th>Ziel</th><th>Status</th></tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
{rows if rows else "<tr><td colspan='5'><em>No jobs yet.</em></td></tr>"}
|
||||
<tbody id="jobs-body">
|
||||
{rows}
|
||||
</tbody>
|
||||
</table>
|
||||
</body>
|
||||
@@ -878,9 +971,45 @@ def render_nav(active: str) -> str:
|
||||
"<div style='margin: 8px 0 14px 0;'>"
|
||||
+ link("Downloads", "/", "downloads")
|
||||
+ link("Proxies", "/proxies", "proxies")
|
||||
+ link("Logs", "/logs", "logs")
|
||||
+ "</div>"
|
||||
)
|
||||
|
||||
def render_logs_page() -> str:
|
||||
return f"""
|
||||
<html>
|
||||
<head>
|
||||
<link rel="stylesheet" href="/static/style.css">
|
||||
<meta charset="utf-8">
|
||||
<title>JD → Jellyfin (Logs)</title>
|
||||
<script>
|
||||
async function refreshLogs() {{
|
||||
if (document.hidden) return;
|
||||
try {{
|
||||
const resp = await fetch('/logs/data');
|
||||
if (!resp.ok) return;
|
||||
const text = await resp.text();
|
||||
const area = document.getElementById('log-body');
|
||||
if (area) {{
|
||||
area.value = text;
|
||||
area.scrollTop = area.scrollHeight;
|
||||
}}
|
||||
}} catch (e) {{
|
||||
}}
|
||||
}}
|
||||
setInterval(refreshLogs, 2000);
|
||||
window.addEventListener('load', refreshLogs);
|
||||
</script>
|
||||
</head>
|
||||
<body>
|
||||
<h1>JD → Jellyfin</h1>
|
||||
{render_nav("logs")}
|
||||
<p class="hint">Verbindungs-Debugger (Echtzeit). Letzte {LOG_BUFFER_LIMIT} Einträge.</p>
|
||||
<textarea id="log-body" class="log-area" rows="20" readonly></textarea>
|
||||
</body>
|
||||
</html>
|
||||
"""
|
||||
|
||||
def render_proxies_page(
|
||||
error: str = "",
|
||||
message: str = "",
|
||||
|
||||
@@ -19,3 +19,4 @@ code { font-family: ui-monospace, SFMono-Regular, Menlo, Consolas, monospace; fo
|
||||
.progress-row { display:flex; align-items:center; gap:8px; margin-top:6px; }
|
||||
.progress-text { font-size:12px; color:#333; min-width:48px; }
|
||||
.inline-form { margin-top:6px; }
|
||||
.log-area { width:100%; max-width: 920px; padding:10px; border:1px solid #ccc; border-radius:8px; background:#fff; }
|
||||
|
||||
Reference in New Issue
Block a user