Compare commits
6 Commits
codex/uber
...
main
| Author | SHA1 | Date | |
|---|---|---|---|
| 7a7f9979cd | |||
| 6b17e8bc06 | |||
| 54ae313563 | |||
| b34d4062a4 | |||
| 79230d62a2 | |||
| 71b6645157 |
@@ -26,4 +26,4 @@ services:
|
|||||||
volumes:
|
volumes:
|
||||||
- ./data/jd-output:/output:rw
|
- ./data/jd-output:/output:rw
|
||||||
- ./data/md5:/md5:rw
|
- ./data/md5:/md5:rw
|
||||||
- /root/.ssh/id_ed25519:/ssh/id_ed25519:ro
|
- ${SSH_KEY_PATH:-/root/.ssh/id_ed25519}:/ssh/id_ed25519:ro
|
||||||
|
|||||||
@@ -2,19 +2,17 @@ FROM python:3.12-slim
|
|||||||
|
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
|
|
||||||
RUN apt-get update \
|
RUN apt-get update && apt-get install -y --no-install-recommends ffmpeg && rm -rf /var/lib/apt/lists/*
|
||||||
&& apt-get install -y --no-install-recommends ffmpeg \
|
|
||||||
&& rm -rf /var/lib/apt/lists/*
|
|
||||||
|
|
||||||
RUN pip install --no-cache-dir \
|
COPY requirements.txt .
|
||||||
fastapi \
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
uvicorn \
|
|
||||||
myjdapi \
|
|
||||||
paramiko \
|
|
||||||
python-multipart
|
|
||||||
|
|
||||||
COPY app.py /app/app.py
|
RUN useradd -m -u 1000 appuser && chown appuser:appuser /app
|
||||||
COPY static /app/static
|
|
||||||
|
USER appuser
|
||||||
|
|
||||||
|
COPY --chown=appuser:appuser app.py .
|
||||||
|
COPY --chown=appuser:appuser static ./static
|
||||||
|
|
||||||
EXPOSE 8080
|
EXPOSE 8080
|
||||||
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "8080"]
|
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "8080"]
|
||||||
|
|||||||
@@ -3,11 +3,14 @@ from __future__ import annotations
|
|||||||
|
|
||||||
import base64
|
import base64
|
||||||
import hashlib
|
import hashlib
|
||||||
|
import hmac
|
||||||
import html as html_mod
|
import html as html_mod
|
||||||
|
import ipaddress
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import shlex
|
import shlex
|
||||||
|
import socket
|
||||||
import subprocess
|
import subprocess
|
||||||
import threading
|
import threading
|
||||||
import time
|
import time
|
||||||
@@ -106,7 +109,7 @@ def _check_basic_auth(req: Request) -> bool:
|
|||||||
if ":" not in raw:
|
if ":" not in raw:
|
||||||
return False
|
return False
|
||||||
user, pw = raw.split(":", 1)
|
user, pw = raw.split(":", 1)
|
||||||
return user == BASIC_AUTH_USER and pw == BASIC_AUTH_PASS
|
return hmac.compare_digest(user, BASIC_AUTH_USER) and hmac.compare_digest(pw, BASIC_AUTH_PASS)
|
||||||
|
|
||||||
def _auth_challenge() -> HTMLResponse:
|
def _auth_challenge() -> HTMLResponse:
|
||||||
return HTMLResponse(
|
return HTMLResponse(
|
||||||
@@ -237,8 +240,26 @@ def is_demo_link(name: str) -> bool:
|
|||||||
lower = name.lower().replace("-", "_").replace(".", " ")
|
lower = name.lower().replace("-", "_").replace(".", " ")
|
||||||
return any(pat in lower for pat in DEMO_PATTERNS)
|
return any(pat in lower for pat in DEMO_PATTERNS)
|
||||||
|
|
||||||
|
def _is_ssrf_target(url: str) -> bool:
|
||||||
|
"""Return True if the URL resolves to a private/loopback address (SSRF protection)."""
|
||||||
|
try:
|
||||||
|
host = urllib.parse.urlparse(url).hostname or ""
|
||||||
|
try:
|
||||||
|
addr = ipaddress.ip_address(host)
|
||||||
|
except ValueError:
|
||||||
|
try:
|
||||||
|
host = socket.gethostbyname(host)
|
||||||
|
addr = ipaddress.ip_address(host)
|
||||||
|
except Exception:
|
||||||
|
return False
|
||||||
|
return addr.is_private or addr.is_loopback or addr.is_link_local or addr.is_reserved
|
||||||
|
except Exception:
|
||||||
|
return False
|
||||||
|
|
||||||
def check_url_reachable(url: str) -> Optional[str]:
|
def check_url_reachable(url: str) -> Optional[str]:
|
||||||
"""Try a HEAD request to verify the URL is reachable. Returns error string or None."""
|
"""Try a HEAD request to verify the URL is reachable. Returns error string or None."""
|
||||||
|
if _is_ssrf_target(url):
|
||||||
|
return "URL zeigt auf eine interne/private Adresse (nicht erlaubt)"
|
||||||
try:
|
try:
|
||||||
req = urllib.request.Request(url, method="HEAD")
|
req = urllib.request.Request(url, method="HEAD")
|
||||||
req.add_header("User-Agent", "Mozilla/5.0")
|
req.add_header("User-Agent", "Mozilla/5.0")
|
||||||
@@ -435,13 +456,22 @@ def format_proxy_lines(raw: str, scheme: str) -> str:
|
|||||||
|
|
||||||
return "\n".join(dedup)
|
return "\n".join(dedup)
|
||||||
|
|
||||||
|
_PROXY_FETCH_LIMIT = 2 * 1024 * 1024 # 2 MB cap to prevent memory exhaustion
|
||||||
|
_proxy_cache: Dict[str, Tuple[float, str]] = {}
|
||||||
|
_PROXY_CACHE_TTL = 300.0 # 5 minutes
|
||||||
|
|
||||||
def fetch_proxy_list(url: str) -> str:
|
def fetch_proxy_list(url: str) -> str:
|
||||||
|
now = time.time()
|
||||||
|
cached_ts, cached_text = _proxy_cache.get(url, (0.0, ""))
|
||||||
|
if cached_text and now - cached_ts < _PROXY_CACHE_TTL:
|
||||||
|
return cached_text
|
||||||
req = urllib.request.Request(url)
|
req = urllib.request.Request(url)
|
||||||
log_connection(f"HTTP GET {url} (no-proxy)")
|
log_connection(f"HTTP GET {url} (no-proxy)")
|
||||||
with NO_PROXY_OPENER.open(req, timeout=20) as resp:
|
with NO_PROXY_OPENER.open(req, timeout=20) as resp:
|
||||||
text = resp.read().decode("utf-8", "replace")
|
text = resp.read(_PROXY_FETCH_LIMIT).decode("utf-8", "replace")
|
||||||
if "\n" not in text and re.search(r"\s", text):
|
if "\n" not in text and re.search(r"\s", text):
|
||||||
return re.sub(r"\s+", "\n", text.strip())
|
text = re.sub(r"\s+", "\n", text.strip())
|
||||||
|
_proxy_cache[url] = (now, text)
|
||||||
return text
|
return text
|
||||||
|
|
||||||
def build_jdproxies_payload(text: str) -> Dict[str, Any]:
|
def build_jdproxies_payload(text: str) -> Dict[str, Any]:
|
||||||
@@ -1181,7 +1211,10 @@ def submit(url: str = Form(...), package_name: str = Form(""), library: str = Fo
|
|||||||
log_connection(f"URL-Check fehlgeschlagen: {url} -> {url_err}")
|
log_connection(f"URL-Check fehlgeschlagen: {url} -> {url_err}")
|
||||||
return HTMLResponse(render_page(f"Link nicht erreichbar: {url_err}"), status_code=400)
|
return HTMLResponse(render_page(f"Link nicht erreichbar: {url_err}"), status_code=400)
|
||||||
|
|
||||||
dev = get_device()
|
try:
|
||||||
|
dev = get_device()
|
||||||
|
except Exception as e:
|
||||||
|
return HTMLResponse(render_page(f"JDownloader nicht erreichbar: {e}"), status_code=503)
|
||||||
resp = dev.linkgrabber.add_links([{
|
resp = dev.linkgrabber.add_links([{
|
||||||
"links": url,
|
"links": url,
|
||||||
"autostart": True,
|
"autostart": True,
|
||||||
|
|||||||
5
jd-webgui/requirements.txt
Normal file
5
jd-webgui/requirements.txt
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
fastapi
|
||||||
|
uvicorn
|
||||||
|
myjdapi
|
||||||
|
paramiko
|
||||||
|
python-multipart
|
||||||
Reference in New Issue
Block a user