chore: initial backend + travel-proxy dockerized setup for NAS deployment
This commit is contained in:
17
.env.example
Normal file
17
.env.example
Normal file
@@ -0,0 +1,17 @@
|
||||
# timezone
|
||||
TZ=Asia/Seoul
|
||||
|
||||
COMPOSE_PROJECT_NAME=webpage
|
||||
|
||||
# backend lotto collector sources
|
||||
LOTTO_ALL_URL=https://smok95.github.io/lotto/results/all.json
|
||||
LOTTO_LATEST_URL=https://smok95.github.io/lotto/results/latest.json
|
||||
|
||||
# travel-proxy
|
||||
TRAVEL_ROOT=/data/travel
|
||||
TRAVEL_THUMB_ROOT=/data/thumbs
|
||||
TRAVEL_MEDIA_BASE=/media/travel
|
||||
TRAVEL_CACHE_TTL=300
|
||||
|
||||
# CORS (travel-proxy)
|
||||
CORS_ALLOW_ORIGINS=*
|
||||
65
.gitignore
vendored
Normal file
65
.gitignore
vendored
Normal file
@@ -0,0 +1,65 @@
|
||||
################################
|
||||
# 기본
|
||||
################################
|
||||
.DS_Store
|
||||
Thumbs.db
|
||||
*.swp
|
||||
*.log
|
||||
|
||||
################################
|
||||
# Python
|
||||
################################
|
||||
__pycache__/
|
||||
*.py[cod]
|
||||
*.pyo
|
||||
*.pyd
|
||||
*.egg-info/
|
||||
.eggs/
|
||||
.env
|
||||
.env.*
|
||||
!.env.example
|
||||
.venv/
|
||||
venv/
|
||||
|
||||
################################
|
||||
# FastAPI / uvicorn
|
||||
################################
|
||||
*.sqlite
|
||||
*.db
|
||||
*.db-journal
|
||||
|
||||
################################
|
||||
# Docker
|
||||
################################
|
||||
*.tar
|
||||
docker-compose.override.yml
|
||||
|
||||
################################
|
||||
# Node / Frontend (소스만 관리)
|
||||
################################
|
||||
node_modules/
|
||||
dist/
|
||||
build/
|
||||
.vite/
|
||||
.cache/
|
||||
|
||||
################################
|
||||
# NAS / Synology
|
||||
################################
|
||||
@eaDir/
|
||||
*.@__thumb
|
||||
._*
|
||||
|
||||
################################
|
||||
# Runtime data (운영에서만 생성)
|
||||
################################
|
||||
data/
|
||||
travel-thumbs/
|
||||
thumbs/
|
||||
uploads/
|
||||
|
||||
################################
|
||||
# Scheduler / temp
|
||||
################################
|
||||
tmp/
|
||||
temp/
|
||||
18
backend/Dockerfile
Normal file
18
backend/Dockerfile
Normal file
@@ -0,0 +1,18 @@
|
||||
FROM python:3.12-slim
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||
ca-certificates curl \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
COPY app/requirements.txt /app/requirements.txt
|
||||
RUN pip install --no-cache-dir -r /app/requirements.txt
|
||||
|
||||
COPY app /app/app
|
||||
|
||||
ENV PYTHONUNBUFFERED=1
|
||||
|
||||
EXPOSE 8000
|
||||
|
||||
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "8000"]
|
||||
55
backend/app/collector.py
Normal file
55
backend/app/collector.py
Normal file
@@ -0,0 +1,55 @@
|
||||
import requests
|
||||
from typing import Dict, Any
|
||||
|
||||
from .db import get_draw, upsert_draw
|
||||
|
||||
def _normalize_item(item: dict) -> dict:
|
||||
# smok95 all.json / latest.json 구조
|
||||
# - draw_no: int
|
||||
# - numbers: [n1..n6]
|
||||
# - bonus_no: int
|
||||
# - date: "YYYY-MM-DD ..."
|
||||
numbers = item["numbers"]
|
||||
return {
|
||||
"drw_no": int(item["draw_no"]),
|
||||
"drw_date": (item.get("date") or "")[:10],
|
||||
"n1": int(numbers[0]),
|
||||
"n2": int(numbers[1]),
|
||||
"n3": int(numbers[2]),
|
||||
"n4": int(numbers[3]),
|
||||
"n5": int(numbers[4]),
|
||||
"n6": int(numbers[5]),
|
||||
"bonus": int(item["bonus_no"]),
|
||||
}
|
||||
|
||||
def sync_all_from_json(all_url: str) -> Dict[str, Any]:
|
||||
r = requests.get(all_url, timeout=60)
|
||||
r.raise_for_status()
|
||||
data = r.json() # list[dict]
|
||||
|
||||
inserted = 0
|
||||
skipped = 0
|
||||
|
||||
for item in data:
|
||||
row = _normalize_item(item)
|
||||
|
||||
if get_draw(row["drw_no"]):
|
||||
skipped += 1
|
||||
continue
|
||||
|
||||
upsert_draw(row)
|
||||
inserted += 1
|
||||
|
||||
return {"mode": "all_json", "url": all_url, "inserted": inserted, "skipped": skipped, "total": len(data)}
|
||||
|
||||
def sync_latest(latest_url: str) -> Dict[str, Any]:
|
||||
r = requests.get(latest_url, timeout=30)
|
||||
r.raise_for_status()
|
||||
item = r.json()
|
||||
|
||||
row = _normalize_item(item)
|
||||
before = get_draw(row["drw_no"])
|
||||
upsert_draw(row)
|
||||
|
||||
return {"mode": "latest_json", "url": latest_url, "was_new": (before is None), "drawNo": row["drw_no"]}
|
||||
|
||||
239
backend/app/db.py
Normal file
239
backend/app/db.py
Normal file
@@ -0,0 +1,239 @@
|
||||
# backend/app/db.py
|
||||
import os
|
||||
import sqlite3
|
||||
import json
|
||||
import hashlib
|
||||
from typing import Any, Dict, Optional, List
|
||||
|
||||
DB_PATH = "/app/data/lotto.db"
|
||||
|
||||
def _conn() -> sqlite3.Connection:
|
||||
os.makedirs(os.path.dirname(DB_PATH), exist_ok=True)
|
||||
conn = sqlite3.connect(DB_PATH)
|
||||
conn.row_factory = sqlite3.Row
|
||||
return conn
|
||||
|
||||
def _ensure_column(conn: sqlite3.Connection, table: str, col: str, ddl: str) -> None:
|
||||
cols = {r["name"] for r in conn.execute(f"PRAGMA table_info({table})").fetchall()}
|
||||
if col not in cols:
|
||||
conn.execute(ddl)
|
||||
|
||||
def init_db() -> None:
|
||||
with _conn() as conn:
|
||||
conn.execute(
|
||||
"""
|
||||
CREATE TABLE IF NOT EXISTS draws (
|
||||
drw_no INTEGER PRIMARY KEY,
|
||||
drw_date TEXT NOT NULL,
|
||||
n1 INTEGER NOT NULL,
|
||||
n2 INTEGER NOT NULL,
|
||||
n3 INTEGER NOT NULL,
|
||||
n4 INTEGER NOT NULL,
|
||||
n5 INTEGER NOT NULL,
|
||||
n6 INTEGER NOT NULL,
|
||||
bonus INTEGER NOT NULL,
|
||||
updated_at TEXT NOT NULL DEFAULT (datetime('now'))
|
||||
);
|
||||
"""
|
||||
)
|
||||
conn.execute("CREATE INDEX IF NOT EXISTS idx_draws_date ON draws(drw_date);")
|
||||
|
||||
conn.execute(
|
||||
"""
|
||||
CREATE TABLE IF NOT EXISTS recommendations (
|
||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
created_at TEXT NOT NULL DEFAULT (datetime('now')),
|
||||
based_on_draw INTEGER,
|
||||
numbers TEXT NOT NULL,
|
||||
params TEXT NOT NULL
|
||||
);
|
||||
"""
|
||||
)
|
||||
conn.execute("CREATE INDEX IF NOT EXISTS idx_reco_created ON recommendations(created_at DESC);")
|
||||
|
||||
# ✅ 확장 컬럼들(기존 DB에도 자동 추가)
|
||||
_ensure_column(conn, "recommendations", "numbers_sorted",
|
||||
"ALTER TABLE recommendations ADD COLUMN numbers_sorted TEXT;")
|
||||
_ensure_column(conn, "recommendations", "dedup_hash",
|
||||
"ALTER TABLE recommendations ADD COLUMN dedup_hash TEXT;")
|
||||
_ensure_column(conn, "recommendations", "favorite",
|
||||
"ALTER TABLE recommendations ADD COLUMN favorite INTEGER NOT NULL DEFAULT 0;")
|
||||
_ensure_column(conn, "recommendations", "note",
|
||||
"ALTER TABLE recommendations ADD COLUMN note TEXT NOT NULL DEFAULT '';")
|
||||
_ensure_column(conn, "recommendations", "tags",
|
||||
"ALTER TABLE recommendations ADD COLUMN tags TEXT NOT NULL DEFAULT '[]';")
|
||||
|
||||
# ✅ UNIQUE 인덱스(중복 저장 방지)
|
||||
conn.execute("CREATE UNIQUE INDEX IF NOT EXISTS uq_reco_dedup ON recommendations(dedup_hash);")
|
||||
|
||||
def upsert_draw(row: Dict[str, Any]) -> None:
|
||||
with _conn() as conn:
|
||||
conn.execute(
|
||||
"""
|
||||
INSERT INTO draws (drw_no, drw_date, n1, n2, n3, n4, n5, n6, bonus)
|
||||
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
ON CONFLICT(drw_no) DO UPDATE SET
|
||||
drw_date=excluded.drw_date,
|
||||
n1=excluded.n1, n2=excluded.n2, n3=excluded.n3,
|
||||
n4=excluded.n4, n5=excluded.n5, n6=excluded.n6,
|
||||
bonus=excluded.bonus,
|
||||
updated_at=datetime('now')
|
||||
""",
|
||||
(
|
||||
int(row["drw_no"]),
|
||||
str(row["drw_date"]),
|
||||
int(row["n1"]), int(row["n2"]), int(row["n3"]),
|
||||
int(row["n4"]), int(row["n5"]), int(row["n6"]),
|
||||
int(row["bonus"]),
|
||||
),
|
||||
)
|
||||
|
||||
def get_latest_draw() -> Optional[Dict[str, Any]]:
|
||||
with _conn() as conn:
|
||||
r = conn.execute("SELECT * FROM draws ORDER BY drw_no DESC LIMIT 1").fetchone()
|
||||
return dict(r) if r else None
|
||||
|
||||
def get_draw(drw_no: int) -> Optional[Dict[str, Any]]:
|
||||
with _conn() as conn:
|
||||
r = conn.execute("SELECT * FROM draws WHERE drw_no = ?", (drw_no,)).fetchone()
|
||||
return dict(r) if r else None
|
||||
|
||||
def count_draws() -> int:
|
||||
with _conn() as conn:
|
||||
r = conn.execute("SELECT COUNT(*) AS c FROM draws").fetchone()
|
||||
return int(r["c"])
|
||||
|
||||
def get_all_draw_numbers():
|
||||
with _conn() as conn:
|
||||
rows = conn.execute(
|
||||
"SELECT drw_no, n1, n2, n3, n4, n5, n6 FROM draws ORDER BY drw_no ASC"
|
||||
).fetchall()
|
||||
return [(int(r["drw_no"]), [int(r["n1"]), int(r["n2"]), int(r["n3"]), int(r["n4"]), int(r["n5"]), int(r["n6"])]) for r in rows]
|
||||
|
||||
# ---------- ✅ recommendation helpers ----------
|
||||
|
||||
def _canonical_params(params: dict) -> str:
|
||||
return json.dumps(params, sort_keys=True, separators=(",", ":"))
|
||||
|
||||
def _numbers_sorted_str(numbers: List[int]) -> str:
|
||||
return ",".join(str(x) for x in sorted(numbers))
|
||||
|
||||
def _dedup_hash(based_on_draw: Optional[int], numbers: List[int], params: dict) -> str:
|
||||
s = f"{based_on_draw or ''}|{_numbers_sorted_str(numbers)}|{_canonical_params(params)}"
|
||||
return hashlib.sha1(s.encode("utf-8")).hexdigest()
|
||||
|
||||
def save_recommendation_dedup(based_on_draw: Optional[int], numbers: List[int], params: dict) -> Dict[str, Any]:
|
||||
"""
|
||||
✅ 동일 추천(번호+params+based_on_draw)이면 중복 저장 없이 기존 id 반환
|
||||
"""
|
||||
ns = _numbers_sorted_str(numbers)
|
||||
h = _dedup_hash(based_on_draw, numbers, params)
|
||||
|
||||
with _conn() as conn:
|
||||
# 이미 있으면 반환
|
||||
r = conn.execute("SELECT id FROM recommendations WHERE dedup_hash = ?", (h,)).fetchone()
|
||||
if r:
|
||||
return {"id": int(r["id"]), "saved": False, "deduped": True}
|
||||
|
||||
cur = conn.execute(
|
||||
"""
|
||||
INSERT INTO recommendations (based_on_draw, numbers, params, numbers_sorted, dedup_hash)
|
||||
VALUES (?, ?, ?, ?, ?)
|
||||
""",
|
||||
(based_on_draw, json.dumps(numbers), json.dumps(params), ns, h),
|
||||
)
|
||||
return {"id": int(cur.lastrowid), "saved": True, "deduped": False}
|
||||
|
||||
def list_recommendations_ex(
|
||||
limit: int = 30,
|
||||
offset: int = 0,
|
||||
favorite: Optional[bool] = None,
|
||||
tag: Optional[str] = None,
|
||||
q: Optional[str] = None,
|
||||
sort: str = "id_desc", # id_desc|created_desc|favorite_desc
|
||||
) -> List[Dict[str, Any]]:
|
||||
import json
|
||||
|
||||
where = []
|
||||
args: list[Any] = []
|
||||
|
||||
if favorite is not None:
|
||||
where.append("favorite = ?")
|
||||
args.append(1 if favorite else 0)
|
||||
|
||||
if q:
|
||||
where.append("note LIKE ?")
|
||||
args.append(f"%{q}%")
|
||||
|
||||
# tags는 JSON 문자열이므로 단순 LIKE로 처리(가볍게 시작)
|
||||
if tag:
|
||||
where.append("tags LIKE ?")
|
||||
args.append(f"%{tag}%")
|
||||
|
||||
where_sql = ("WHERE " + " AND ".join(where)) if where else ""
|
||||
|
||||
if sort == "created_desc":
|
||||
order = "created_at DESC"
|
||||
elif sort == "favorite_desc":
|
||||
# favorite(1)이 먼저, 그 다음 최신
|
||||
order = "favorite DESC, id DESC"
|
||||
else:
|
||||
order = "id DESC"
|
||||
|
||||
sql = f"""
|
||||
SELECT id, created_at, based_on_draw, numbers, params, favorite, note, tags
|
||||
FROM recommendations
|
||||
{where_sql}
|
||||
ORDER BY {order}
|
||||
LIMIT ? OFFSET ?
|
||||
"""
|
||||
args.extend([int(limit), int(offset)])
|
||||
|
||||
with _conn() as conn:
|
||||
rows = conn.execute(sql, args).fetchall()
|
||||
|
||||
out = []
|
||||
for r in rows:
|
||||
out.append({
|
||||
"id": int(r["id"]),
|
||||
"created_at": r["created_at"],
|
||||
"based_on_draw": r["based_on_draw"],
|
||||
"numbers": json.loads(r["numbers"]),
|
||||
"params": json.loads(r["params"]),
|
||||
"favorite": bool(r["favorite"]) if r["favorite"] is not None else False,
|
||||
"note": r["note"],
|
||||
"tags": json.loads(r["tags"]) if r["tags"] else [],
|
||||
})
|
||||
return out
|
||||
|
||||
def update_recommendation(rec_id: int, favorite: Optional[bool] = None, note: Optional[str] = None, tags: Optional[List[str]] = None) -> bool:
|
||||
fields = []
|
||||
args: list[Any] = []
|
||||
|
||||
if favorite is not None:
|
||||
fields.append("favorite = ?")
|
||||
args.append(1 if favorite else 0)
|
||||
if note is not None:
|
||||
fields.append("note = ?")
|
||||
args.append(note)
|
||||
if tags is not None:
|
||||
fields.append("tags = ?")
|
||||
args.append(json.dumps(tags))
|
||||
|
||||
if not fields:
|
||||
return False
|
||||
|
||||
args.append(rec_id)
|
||||
|
||||
with _conn() as conn:
|
||||
cur = conn.execute(
|
||||
f"UPDATE recommendations SET {', '.join(fields)} WHERE id = ?",
|
||||
args,
|
||||
)
|
||||
return cur.rowcount > 0
|
||||
|
||||
def delete_recommendation(rec_id: int) -> bool:
|
||||
with _conn() as conn:
|
||||
cur = conn.execute("DELETE FROM recommendations WHERE id = ?", (rec_id,))
|
||||
return cur.rowcount > 0
|
||||
|
||||
344
backend/app/main.py
Normal file
344
backend/app/main.py
Normal file
@@ -0,0 +1,344 @@
|
||||
import os
|
||||
from typing import Optional, List, Dict, Any, Tuple
|
||||
from fastapi import FastAPI, HTTPException
|
||||
from pydantic import BaseModel
|
||||
from apscheduler.schedulers.background import BackgroundScheduler
|
||||
|
||||
from .db import (
|
||||
init_db, get_draw, get_latest_draw, get_all_draw_numbers,
|
||||
save_recommendation_dedup, list_recommendations_ex, delete_recommendation,
|
||||
update_recommendation,
|
||||
)
|
||||
from .recommender import recommend_numbers
|
||||
from .collector import sync_latest
|
||||
|
||||
app = FastAPI()
|
||||
scheduler = BackgroundScheduler(timezone=os.getenv("TZ", "Asia/Seoul"))
|
||||
|
||||
ALL_URL = os.getenv("LOTTO_ALL_URL", "https://smok95.github.io/lotto/results/all.json")
|
||||
LATEST_URL = os.getenv("LOTTO_LATEST_URL", "https://smok95.github.io/lotto/results/latest.json")
|
||||
|
||||
def calc_metrics(numbers: List[int]) -> Dict[str, Any]:
|
||||
nums = sorted(numbers)
|
||||
s = sum(nums)
|
||||
odd = sum(1 for x in nums if x % 2 == 1)
|
||||
even = len(nums) - odd
|
||||
mn, mx = nums[0], nums[-1]
|
||||
rng = mx - mn
|
||||
|
||||
# 1-10, 11-20, 21-30, 31-40, 41-45
|
||||
buckets = {
|
||||
"1-10": 0,
|
||||
"11-20": 0,
|
||||
"21-30": 0,
|
||||
"31-40": 0,
|
||||
"41-45": 0,
|
||||
}
|
||||
for x in nums:
|
||||
if 1 <= x <= 10:
|
||||
buckets["1-10"] += 1
|
||||
elif 11 <= x <= 20:
|
||||
buckets["11-20"] += 1
|
||||
elif 21 <= x <= 30:
|
||||
buckets["21-30"] += 1
|
||||
elif 31 <= x <= 40:
|
||||
buckets["31-40"] += 1
|
||||
else:
|
||||
buckets["41-45"] += 1
|
||||
|
||||
return {
|
||||
"sum": s,
|
||||
"odd": odd,
|
||||
"even": even,
|
||||
"min": mn,
|
||||
"max": mx,
|
||||
"range": rng,
|
||||
"buckets": buckets,
|
||||
}
|
||||
|
||||
def calc_recent_overlap(numbers: List[int], draws: List[Tuple[int, List[int]]], last_k: int) -> Dict[str, Any]:
|
||||
"""
|
||||
draws: [(drw_no, [n1..n6]), ...] 오름차순
|
||||
last_k: 최근 k회 기준 중복
|
||||
"""
|
||||
if last_k <= 0:
|
||||
return {"last_k": 0, "repeats": 0, "repeated_numbers": []}
|
||||
|
||||
recent = draws[-last_k:] if len(draws) >= last_k else draws
|
||||
recent_set = set()
|
||||
for _, nums in recent:
|
||||
recent_set.update(nums)
|
||||
|
||||
repeated = sorted(set(numbers) & recent_set)
|
||||
return {
|
||||
"last_k": len(recent),
|
||||
"repeats": len(repeated),
|
||||
"repeated_numbers": repeated,
|
||||
}
|
||||
|
||||
@app.on_event("startup")
|
||||
def on_startup():
|
||||
init_db()
|
||||
scheduler.add_job(lambda: sync_latest(LATEST_URL), "cron", hour="9,21", minute=10)
|
||||
scheduler.start()
|
||||
|
||||
@app.get("/health")
|
||||
def health():
|
||||
return {"ok": True}
|
||||
|
||||
@app.get("/api/lotto/latest")
|
||||
def api_latest():
|
||||
row = get_latest_draw()
|
||||
if not row:
|
||||
raise HTTPException(status_code=404, detail="No data yet")
|
||||
return {
|
||||
"drawNo": row["drw_no"],
|
||||
"date": row["drw_date"],
|
||||
"numbers": [row["n1"], row["n2"], row["n3"], row["n4"], row["n5"], row["n6"]],
|
||||
"bonus": row["bonus"],
|
||||
}
|
||||
|
||||
@app.get("/api/lotto/{drw_no:int}")
|
||||
def api_draw(drw_no: int):
|
||||
row = get_draw(drw_no)
|
||||
if not row:
|
||||
raise HTTPException(status_code=404, detail="Not found")
|
||||
return {
|
||||
"drwNo": row["drw_no"],
|
||||
"date": row["drw_date"],
|
||||
"numbers": [row["n1"], row["n2"], row["n3"], row["n4"], row["n5"], row["n6"]],
|
||||
"bonus": row["bonus"],
|
||||
}
|
||||
|
||||
@app.post("/api/admin/sync_latest")
|
||||
def admin_sync_latest():
|
||||
return sync_latest(LATEST_URL)
|
||||
|
||||
# ---------- ✅ recommend (dedup save) ----------
|
||||
@app.get("/api/lotto/recommend")
|
||||
def api_recommend(
|
||||
recent_window: int = 200,
|
||||
recent_weight: float = 2.0,
|
||||
avoid_recent_k: int = 5,
|
||||
|
||||
# ---- optional constraints (Lotto Lab) ----
|
||||
sum_min: Optional[int] = None,
|
||||
sum_max: Optional[int] = None,
|
||||
odd_min: Optional[int] = None,
|
||||
odd_max: Optional[int] = None,
|
||||
range_min: Optional[int] = None,
|
||||
range_max: Optional[int] = None,
|
||||
max_overlap_latest: Optional[int] = None, # 최근 avoid_recent_k 회차와 중복 허용 개수
|
||||
max_try: int = 200, # 조건 맞는 조합 찾기 재시도
|
||||
):
|
||||
draws = get_all_draw_numbers()
|
||||
if not draws:
|
||||
raise HTTPException(status_code=404, detail="No data yet")
|
||||
|
||||
latest = get_latest_draw()
|
||||
|
||||
params = {
|
||||
"recent_window": recent_window,
|
||||
"recent_weight": float(recent_weight),
|
||||
"avoid_recent_k": avoid_recent_k,
|
||||
|
||||
"sum_min": sum_min,
|
||||
"sum_max": sum_max,
|
||||
"odd_min": odd_min,
|
||||
"odd_max": odd_max,
|
||||
"range_min": range_min,
|
||||
"range_max": range_max,
|
||||
"max_overlap_latest": max_overlap_latest,
|
||||
"max_try": int(max_try),
|
||||
}
|
||||
|
||||
def _accept(nums: List[int]) -> bool:
|
||||
m = calc_metrics(nums)
|
||||
if sum_min is not None and m["sum"] < sum_min:
|
||||
return False
|
||||
if sum_max is not None and m["sum"] > sum_max:
|
||||
return False
|
||||
if odd_min is not None and m["odd"] < odd_min:
|
||||
return False
|
||||
if odd_max is not None and m["odd"] > odd_max:
|
||||
return False
|
||||
if range_min is not None and m["range"] < range_min:
|
||||
return False
|
||||
if range_max is not None and m["range"] > range_max:
|
||||
return False
|
||||
|
||||
if max_overlap_latest is not None:
|
||||
ov = calc_recent_overlap(nums, draws, last_k=avoid_recent_k)
|
||||
if ov["repeats"] > max_overlap_latest:
|
||||
return False
|
||||
return True
|
||||
|
||||
chosen = None
|
||||
explain = None
|
||||
|
||||
tries = 0
|
||||
while tries < max_try:
|
||||
tries += 1
|
||||
result = recommend_numbers(
|
||||
draws,
|
||||
recent_window=recent_window,
|
||||
recent_weight=recent_weight,
|
||||
avoid_recent_k=avoid_recent_k,
|
||||
)
|
||||
nums = result["numbers"]
|
||||
if _accept(nums):
|
||||
chosen = nums
|
||||
explain = result["explain"]
|
||||
break
|
||||
|
||||
if chosen is None:
|
||||
raise HTTPException(
|
||||
status_code=400,
|
||||
detail=f"Constraints too strict. No valid set found in max_try={max_try}. "
|
||||
f"Try relaxing sum/odd/range/overlap constraints.",
|
||||
)
|
||||
|
||||
# ✅ dedup save
|
||||
saved = save_recommendation_dedup(
|
||||
latest["drw_no"] if latest else None,
|
||||
chosen,
|
||||
params,
|
||||
)
|
||||
|
||||
metrics = calc_metrics(chosen)
|
||||
overlap = calc_recent_overlap(chosen, draws, last_k=avoid_recent_k)
|
||||
|
||||
return {
|
||||
"id": saved["id"],
|
||||
"saved": saved["saved"],
|
||||
"deduped": saved["deduped"],
|
||||
"based_on_latest_draw": latest["drw_no"] if latest else None,
|
||||
"numbers": chosen,
|
||||
"explain": explain,
|
||||
"params": params,
|
||||
"metrics": metrics,
|
||||
"recent_overlap": overlap,
|
||||
"tries": tries,
|
||||
}
|
||||
|
||||
# ---------- ✅ history list (filter/paging) ----------
|
||||
@app.get("/api/history")
|
||||
def api_history(
|
||||
limit: int = 30,
|
||||
offset: int = 0,
|
||||
favorite: Optional[bool] = None,
|
||||
tag: Optional[str] = None,
|
||||
q: Optional[str] = None,
|
||||
sort: str = "id_desc",
|
||||
):
|
||||
items = list_recommendations_ex(
|
||||
limit=limit,
|
||||
offset=offset,
|
||||
favorite=favorite,
|
||||
tag=tag,
|
||||
q=q,
|
||||
sort=sort,
|
||||
)
|
||||
|
||||
draws = get_all_draw_numbers()
|
||||
|
||||
out = []
|
||||
for it in items:
|
||||
nums = it["numbers"]
|
||||
out.append({
|
||||
**it,
|
||||
"metrics": calc_metrics(nums),
|
||||
"recent_overlap": calc_recent_overlap(
|
||||
nums, draws, last_k=int(it["params"].get("avoid_recent_k", 0) or 0)
|
||||
),
|
||||
})
|
||||
|
||||
return {
|
||||
"items": out,
|
||||
"limit": limit,
|
||||
"offset": offset,
|
||||
"filters": {"favorite": favorite, "tag": tag, "q": q, "sort": sort},
|
||||
}
|
||||
|
||||
@app.delete("/api/history/{rec_id:int}")
|
||||
def api_history_delete(rec_id: int):
|
||||
ok = delete_recommendation(rec_id)
|
||||
if not ok:
|
||||
raise HTTPException(status_code=404, detail="Not found")
|
||||
return {"deleted": True, "id": rec_id}
|
||||
|
||||
# ---------- ✅ history update (favorite/note/tags) ----------
|
||||
class HistoryUpdate(BaseModel):
|
||||
favorite: Optional[bool] = None
|
||||
note: Optional[str] = None
|
||||
tags: Optional[List[str]] = None
|
||||
|
||||
@app.patch("/api/history/{rec_id:int}")
|
||||
def api_history_patch(rec_id: int, body: HistoryUpdate):
|
||||
ok = update_recommendation(rec_id, favorite=body.favorite, note=body.note, tags=body.tags)
|
||||
if not ok:
|
||||
raise HTTPException(status_code=404, detail="Not found or no changes")
|
||||
return {"updated": True, "id": rec_id}
|
||||
|
||||
# ---------- ✅ batch recommend ----------
|
||||
def _batch_unique(draws, count: int, recent_window: int, recent_weight: float, avoid_recent_k: int, max_try: int = 200):
|
||||
items = []
|
||||
seen = set()
|
||||
|
||||
tries = 0
|
||||
while len(items) < count and tries < max_try:
|
||||
tries += 1
|
||||
r = recommend_numbers(draws, recent_window=recent_window, recent_weight=recent_weight, avoid_recent_k=avoid_recent_k)
|
||||
key = tuple(sorted(r["numbers"]))
|
||||
if key in seen:
|
||||
continue
|
||||
seen.add(key)
|
||||
items.append(r)
|
||||
|
||||
return items
|
||||
|
||||
@app.get("/api/lotto/recommend/batch")
|
||||
def api_recommend_batch(
|
||||
count: int = 5,
|
||||
recent_window: int = 200,
|
||||
recent_weight: float = 2.0,
|
||||
avoid_recent_k: int = 5,
|
||||
):
|
||||
count = max(1, min(count, 20))
|
||||
draws = get_all_draw_numbers()
|
||||
if not draws:
|
||||
raise HTTPException(status_code=404, detail="No data yet")
|
||||
|
||||
latest = get_latest_draw()
|
||||
params = {
|
||||
"recent_window": recent_window,
|
||||
"recent_weight": float(recent_weight),
|
||||
"avoid_recent_k": avoid_recent_k,
|
||||
"count": count,
|
||||
}
|
||||
|
||||
items = _batch_unique(draws, count, recent_window, float(recent_weight), avoid_recent_k)
|
||||
|
||||
return {
|
||||
"based_on_latest_draw": latest["drw_no"] if latest else None,
|
||||
"count": count,
|
||||
"items": [{"numbers": it["numbers"], "explain": it["explain"]} for it in items],
|
||||
"params": params,
|
||||
}
|
||||
|
||||
class BatchSave(BaseModel):
|
||||
items: List[List[int]]
|
||||
params: dict
|
||||
|
||||
@app.post("/api/lotto/recommend/batch")
|
||||
def api_recommend_batch_save(body: BatchSave):
|
||||
latest = get_latest_draw()
|
||||
based = latest["drw_no"] if latest else None
|
||||
|
||||
created, deduped = [], []
|
||||
for nums in body.items:
|
||||
saved = save_recommendation_dedup(based, nums, body.params)
|
||||
(created if saved["saved"] else deduped).append(saved["id"])
|
||||
|
||||
return {"saved": True, "created_ids": created, "deduped_ids": deduped}
|
||||
|
||||
68
backend/app/recommender.py
Normal file
68
backend/app/recommender.py
Normal file
@@ -0,0 +1,68 @@
|
||||
import random
|
||||
from collections import Counter
|
||||
from typing import Dict, Any, List, Tuple
|
||||
|
||||
def recommend_numbers(
|
||||
draws: List[Tuple[int, List[int]]],
|
||||
*,
|
||||
recent_window: int = 200,
|
||||
recent_weight: float = 2.0,
|
||||
avoid_recent_k: int = 5,
|
||||
seed: int | None = None,
|
||||
) -> Dict[str, Any]:
|
||||
"""
|
||||
가벼운 통계 기반 추천:
|
||||
- 전체 빈도 + 최근(recent_window) 빈도에 가중치를 더한 가중 샘플링
|
||||
- 최근 avoid_recent_k 회차에 나온 번호는 확률을 낮춤(완전 제외는 아님)
|
||||
"""
|
||||
if seed is not None:
|
||||
random.seed(seed)
|
||||
|
||||
# 전체 빈도
|
||||
all_nums = [n for _, nums in draws for n in nums]
|
||||
freq_all = Counter(all_nums)
|
||||
|
||||
# 최근 빈도
|
||||
recent = draws[-recent_window:] if len(draws) >= recent_window else draws
|
||||
recent_nums = [n for _, nums in recent for n in nums]
|
||||
freq_recent = Counter(recent_nums)
|
||||
|
||||
# 최근 k회차 번호(패널티)
|
||||
last_k = draws[-avoid_recent_k:] if len(draws) >= avoid_recent_k else draws
|
||||
last_k_nums = set(n for _, nums in last_k for n in nums)
|
||||
|
||||
# 가중치 구성
|
||||
weights = {}
|
||||
for n in range(1, 46):
|
||||
w = freq_all[n] + recent_weight * freq_recent[n]
|
||||
if n in last_k_nums:
|
||||
w *= 0.6 # 최근에 너무 방금 나온 건 살짝 덜 뽑히게
|
||||
weights[n] = max(w, 0.1)
|
||||
|
||||
# 중복 없이 6개 뽑기(가중 샘플링)
|
||||
chosen = []
|
||||
pool = list(range(1, 46))
|
||||
for _ in range(6):
|
||||
total = sum(weights[n] for n in pool)
|
||||
r = random.random() * total
|
||||
acc = 0.0
|
||||
for n in pool:
|
||||
acc += weights[n]
|
||||
if acc >= r:
|
||||
chosen.append(n)
|
||||
pool.remove(n)
|
||||
break
|
||||
|
||||
chosen_sorted = sorted(chosen)
|
||||
|
||||
explain = {
|
||||
"recent_window": recent_window,
|
||||
"recent_weight": recent_weight,
|
||||
"avoid_recent_k": avoid_recent_k,
|
||||
"top_all": [n for n, _ in freq_all.most_common(10)],
|
||||
"top_recent": [n for n, _ in freq_recent.most_common(10)],
|
||||
"last_k_draws": [d for d, _ in last_k],
|
||||
}
|
||||
|
||||
return {"numbers": chosen_sorted, "explain": explain}
|
||||
|
||||
5
backend/app/requirements.txt
Normal file
5
backend/app/requirements.txt
Normal file
@@ -0,0 +1,5 @@
|
||||
fastapi==0.115.6
|
||||
uvicorn[standard]==0.30.6
|
||||
requests==2.32.3
|
||||
beautifulsoup4==4.12.3
|
||||
APScheduler==3.10.4
|
||||
47
docker-compose.yml
Normal file
47
docker-compose.yml
Normal file
@@ -0,0 +1,47 @@
|
||||
version: "3.8"
|
||||
|
||||
services:
|
||||
backend:
|
||||
build: ./backend
|
||||
container_name: lotto-backend
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "18000:8000"
|
||||
environment:
|
||||
- TZ=${TZ:-Asia/Seoul}
|
||||
- LOTTO_ALL_URL=${LOTTO_ALL_URL:-https://smok95.github.io/lotto/results/all.json}
|
||||
- LOTTO_LATEST_URL=${LOTTO_LATEST_URL:-https://smok95.github.io/lotto/results/latest.json}
|
||||
volumes:
|
||||
- /volume1/docker/webpage/data:/app/data
|
||||
|
||||
travel-proxy:
|
||||
build: ./travel-proxy
|
||||
container_name: travel-proxy
|
||||
restart: unless-stopped
|
||||
user: "1026:100"
|
||||
ports:
|
||||
- "19000:8000" # 내부 확인용
|
||||
environment:
|
||||
- TZ=${TZ:-Asia/Seoul}
|
||||
- TRAVEL_ROOT=${TRAVEL_ROOT:-/data/travel}
|
||||
- TRAVEL_THUMB_ROOT=${TRAVEL_THUMB_ROOT:-/data/thumbs}
|
||||
- TRAVEL_MEDIA_BASE=${TRAVEL_MEDIA_BASE:-/media/travel}
|
||||
- TRAVEL_CACHE_TTL=${TRAVEL_CACHE_TTL:-300}
|
||||
- CORS_ALLOW_ORIGINS=${CORS_ALLOW_ORIGINS:-*}
|
||||
volumes:
|
||||
- /volume1/web/images/webPage/travel:/data/travel:ro
|
||||
- /volume1/docker/webpage/travel-thumbs:/data/thumbs:rw
|
||||
|
||||
frontend:
|
||||
image: nginx:alpine
|
||||
container_name: lotto-frontend
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "8080:80"
|
||||
volumes:
|
||||
- /volume1/docker/webpage/frontend:/usr/share/nginx/html:ro
|
||||
- /volume1/docker/webpage/nginx/default.conf:/etc/nginx/conf.d/default.conf:ro
|
||||
- /volume1/web/images/webPage/travel:/data/travel:ro
|
||||
- /volume1/docker/webpage/travel-thumbs:/data/thumbs:ro
|
||||
extra_hosts:
|
||||
- "host.docker.internal:host-gateway"
|
||||
79
nginx/default.conf
Normal file
79
nginx/default.conf
Normal file
@@ -0,0 +1,79 @@
|
||||
server {
|
||||
listen 80;
|
||||
server_name _;
|
||||
|
||||
root /usr/share/nginx/html;
|
||||
index index.html;
|
||||
|
||||
# index.html은 캐시 금지 (배포 반영 핵심)
|
||||
location = /index.html {
|
||||
add_header Cache-Control "no-store, no-cache, must-revalidate, proxy-revalidate, max-age=0";
|
||||
try_files $uri =404;
|
||||
}
|
||||
|
||||
# 정적 리소스는 장기 캐시 (Vite 해시 파일)
|
||||
location /assets/ {
|
||||
add_header Cache-Control "public, max-age=31536000, immutable";
|
||||
try_files $uri =404;
|
||||
}
|
||||
|
||||
# travel thumbnails (generated by travel-proxy, stored in /data/thumbs)
|
||||
location ^~ /media/travel/.thumb/ {
|
||||
alias /data/thumbs/;
|
||||
|
||||
expires 30d;
|
||||
add_header Cache-Control "public, max-age=2592000, immutable" always;
|
||||
|
||||
autoindex off;
|
||||
}
|
||||
|
||||
# travel media (images) - nginx가 직접 파일 서빙
|
||||
location ^~ /media/travel/ {
|
||||
alias /data/travel/; # ✅ /media/travel/... -> /data/travel/...
|
||||
|
||||
expires 7d;
|
||||
add_header Cache-Control "public, max-age=604800" always;
|
||||
|
||||
# 옵션: 폴더리스팅 막기
|
||||
autoindex off;
|
||||
}
|
||||
|
||||
# 기타 정적 파일(예: vite.svg 등) 기본 캐시(원하면 조정)
|
||||
location ~* \.(?:ico|png|jpg|jpeg|gif|svg|webp|css|js)$ {
|
||||
add_header Cache-Control "public, max-age=604800";
|
||||
try_files $uri =404;
|
||||
}
|
||||
|
||||
# travel API
|
||||
location /api/travel/ {
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Forwarded-Proto $scheme;
|
||||
proxy_pass http://travel-proxy:8000/api/travel/;
|
||||
}
|
||||
|
||||
# API 프록시 (여기가 포인트: /api/ 중복 제거)
|
||||
location /api/ {
|
||||
proxy_http_version 1.1;
|
||||
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Forwarded-Proto $scheme;
|
||||
|
||||
proxy_pass http://backend:8000;
|
||||
}
|
||||
|
||||
# SPA 라우팅 (마지막에 두는 게 안전)
|
||||
location / {
|
||||
try_files $uri $uri/ /index.html;
|
||||
}
|
||||
|
||||
# gzip (옵션)
|
||||
gzip on;
|
||||
gzip_types text/plain text/css application/json application/javascript application/xml+rss;
|
||||
gzip_min_length 1024;
|
||||
}
|
||||
|
||||
21
script/deploy-nas.sh
Normal file
21
script/deploy-nas.sh
Normal file
@@ -0,0 +1,21 @@
|
||||
#!/usr/bin/env bash
|
||||
set -euo pipefail
|
||||
|
||||
ROOT="/volume1/docker/webpage"
|
||||
|
||||
cd "$ROOT"
|
||||
|
||||
echo "[1/5] git fetch + pull"
|
||||
git fetch --all --prune
|
||||
git pull --ff-only
|
||||
|
||||
echo "[2/5] docker compose build"
|
||||
docker compose build --pull
|
||||
|
||||
echo "[3/5] docker compose up"
|
||||
docker compose up -d
|
||||
|
||||
echo "[4/5] status"
|
||||
docker compose ps
|
||||
|
||||
echo "[5/5] done"
|
||||
15
script/healthcheck.sh
Normal file
15
script/healthcheck.sh
Normal file
@@ -0,0 +1,15 @@
|
||||
#!/usr/bin/env bash
|
||||
set -euo pipefail
|
||||
|
||||
BASE="http://127.0.0.1"
|
||||
|
||||
echo "backend health:"
|
||||
curl -fsS "${BASE}:18000/health" | sed 's/^/ /'
|
||||
|
||||
echo "backend latest:"
|
||||
curl -fsS "${BASE}:18000/api/lotto/latest" | head -c 200; echo
|
||||
|
||||
echo "travel regions:"
|
||||
curl -fsS "${BASE}:19000/api/travel/regions" | head -c 200; echo
|
||||
|
||||
echo "OK"
|
||||
22
travel-proxy/Dockerfile
Normal file
22
travel-proxy/Dockerfile
Normal file
@@ -0,0 +1,22 @@
|
||||
FROM python:3.12-slim
|
||||
|
||||
# Pillow용 시스템 라이브러리
|
||||
RUN apt-get update && apt-get install -y \
|
||||
libjpeg-dev \
|
||||
zlib1g-dev \
|
||||
libpng-dev \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
COPY requirements.txt .
|
||||
RUN pip install --no-cache-dir -r requirements.txt
|
||||
|
||||
COPY app /app/app
|
||||
|
||||
ENV TRAVEL_ROOT=/data/travel
|
||||
EXPOSE 8000
|
||||
|
||||
ENV PYTHONUNBUFFERED=1
|
||||
|
||||
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "8000"]
|
||||
220
travel-proxy/app/main.py
Normal file
220
travel-proxy/app/main.py
Normal file
@@ -0,0 +1,220 @@
|
||||
import os
|
||||
import json
|
||||
import time
|
||||
from pathlib import Path
|
||||
from typing import Dict, Any, List
|
||||
|
||||
from fastapi import FastAPI, HTTPException, Query
|
||||
from fastapi.staticfiles import StaticFiles
|
||||
from fastapi.responses import FileResponse
|
||||
from PIL import Image
|
||||
|
||||
app = FastAPI()
|
||||
|
||||
# -----------------------------
|
||||
# Env / Paths
|
||||
# -----------------------------
|
||||
ROOT = Path(os.getenv("TRAVEL_ROOT", "/data/travel")).resolve()
|
||||
MEDIA_BASE = os.getenv("TRAVEL_MEDIA_BASE", "/media/travel")
|
||||
|
||||
META_DIR = ROOT / "_meta"
|
||||
REGION_MAP_PATH = META_DIR / "region_map.json"
|
||||
REGIONS_GEOJSON_PATH = META_DIR / "regions.geojson"
|
||||
|
||||
THUMB_ROOT = Path(os.getenv("TRAVEL_THUMB_ROOT", "/data/thumbs")).resolve()
|
||||
THUMB_SIZE = (480, 480)
|
||||
|
||||
THUMB_ROOT.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
# 썸네일 정적 서빙
|
||||
app.mount(
|
||||
f"{MEDIA_BASE}/.thumb",
|
||||
StaticFiles(directory=THUMB_ROOT),
|
||||
name="travel-thumbs",
|
||||
)
|
||||
|
||||
IMAGE_EXT = {".jpg", ".jpeg", ".png", ".webp"}
|
||||
|
||||
# -----------------------------
|
||||
# Cache
|
||||
# -----------------------------
|
||||
CACHE: Dict[str, Dict[str, Any]] = {}
|
||||
CACHE_TTL = int(os.getenv("TRAVEL_CACHE_TTL", "300"))
|
||||
META_MTIME_CACHE: Dict[str, float] = {}
|
||||
|
||||
# -----------------------------
|
||||
# Helpers
|
||||
# -----------------------------
|
||||
def _file_mtime(p: Path) -> float:
|
||||
try:
|
||||
return p.stat().st_mtime
|
||||
except FileNotFoundError:
|
||||
return 0.0
|
||||
|
||||
|
||||
def _read_json(path: Path) -> Any:
|
||||
if not path.exists():
|
||||
raise HTTPException(500, f"Missing required file: {path}")
|
||||
with open(path, "r", encoding="utf-8") as f:
|
||||
return json.load(f)
|
||||
|
||||
|
||||
def load_region_map() -> dict:
|
||||
return _read_json(REGION_MAP_PATH)
|
||||
|
||||
|
||||
def load_regions_geojson() -> dict:
|
||||
return _read_json(REGIONS_GEOJSON_PATH)
|
||||
|
||||
|
||||
def _meta_changed_invalidate_cache():
|
||||
cur = _file_mtime(REGION_MAP_PATH) + _file_mtime(REGIONS_GEOJSON_PATH)
|
||||
if META_MTIME_CACHE.get("meta") != cur:
|
||||
CACHE.clear()
|
||||
META_MTIME_CACHE["meta"] = cur
|
||||
|
||||
|
||||
def _get_albums_for_region(region: str, region_map: dict) -> List[str]:
|
||||
if region not in region_map:
|
||||
raise HTTPException(400, "Unknown region")
|
||||
|
||||
v = region_map[region]
|
||||
if isinstance(v, list):
|
||||
return v
|
||||
if isinstance(v, dict) and isinstance(v.get("albums"), list):
|
||||
return v["albums"]
|
||||
|
||||
raise HTTPException(500, "Invalid region_map format")
|
||||
|
||||
|
||||
def _thumb_path_for(src: Path, album: str) -> Path:
|
||||
"""
|
||||
썸네일 저장 위치 결정.
|
||||
- TRAVEL_THUMB_ROOT가 있으면: THUMB_ROOT/album/<filename>
|
||||
- 없으면: 원본 폴더 album/.thumb/<filename>
|
||||
"""
|
||||
if THUMB_ROOT:
|
||||
base = THUMB_ROOT / album
|
||||
base.mkdir(parents=True, exist_ok=True)
|
||||
return base / src.name
|
||||
|
||||
thumb_dir = src.parent / ".thumb"
|
||||
thumb_dir.mkdir(exist_ok=True)
|
||||
return thumb_dir / src.name
|
||||
|
||||
def ensure_thumb(src: Path, album: str) -> Path:
|
||||
out = _thumb_path_for(src, album) # THUMB_ROOT/album/<filename> or album/.thumb/<filename>
|
||||
|
||||
if out.exists():
|
||||
return out
|
||||
|
||||
out.parent.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
# ✅ 확장자 유지: IMG_3281.tmp.JPG (끝이 .JPG로 끝나게)
|
||||
tmp = out.with_name(out.stem + ".tmp" + out.suffix)
|
||||
|
||||
try:
|
||||
with Image.open(src) as im:
|
||||
im.thumbnail(THUMB_SIZE)
|
||||
|
||||
# ✅ 확장자 기준으로 포맷 명시 (대문자 .JPG도 대응)
|
||||
ext = out.suffix.lower()
|
||||
if ext in (".jpg", ".jpeg"):
|
||||
fmt = "JPEG"
|
||||
elif ext == ".png":
|
||||
fmt = "PNG"
|
||||
elif ext == ".webp":
|
||||
fmt = "WEBP"
|
||||
else:
|
||||
# 혹시 모를 경우: Pillow가 읽은 포맷을 사용
|
||||
fmt = (im.format or "").upper() or "JPEG"
|
||||
|
||||
im.save(tmp, format=fmt, quality=85, optimize=True)
|
||||
|
||||
# ✅ 생성 완료 후 교체
|
||||
tmp.replace(out)
|
||||
return out
|
||||
|
||||
finally:
|
||||
# 실패 시 tmp 찌꺼기 정리
|
||||
try:
|
||||
if tmp.exists():
|
||||
tmp.unlink()
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
def scan_album(album: str) -> List[Dict[str, Any]]:
|
||||
album_dir = ROOT / album
|
||||
if not album_dir.exists():
|
||||
return []
|
||||
|
||||
items = []
|
||||
for p in album_dir.iterdir():
|
||||
if p.is_file() and p.suffix.lower() in IMAGE_EXT:
|
||||
# ✅ 썸네일 생성 보장
|
||||
ensure_thumb(p, album)
|
||||
|
||||
items.append({
|
||||
"album": album,
|
||||
"file": p.name,
|
||||
"url": f"{MEDIA_BASE}/{album}/{p.name}",
|
||||
"thumb": f"{MEDIA_BASE}/.thumb/{album}/{p.name}",
|
||||
})
|
||||
return items
|
||||
|
||||
# -----------------------------
|
||||
# Routes
|
||||
# -----------------------------
|
||||
@app.get("/api/travel/regions")
|
||||
def regions():
|
||||
_meta_changed_invalidate_cache()
|
||||
return load_regions_geojson()
|
||||
|
||||
|
||||
@app.get("/api/travel/photos")
|
||||
def photos(
|
||||
region: str = Query(...),
|
||||
limit: int = Query(500, le=5000),
|
||||
):
|
||||
_meta_changed_invalidate_cache()
|
||||
|
||||
now = time.time()
|
||||
cached = CACHE.get(region)
|
||||
if cached and now - cached["ts"] < CACHE_TTL:
|
||||
return cached["data"]
|
||||
|
||||
region_map = load_region_map()
|
||||
albums = _get_albums_for_region(region, region_map)
|
||||
|
||||
all_items = []
|
||||
matched = []
|
||||
|
||||
for album in albums:
|
||||
items = scan_album(album)
|
||||
matched.append({"album": album, "count": len(items)})
|
||||
all_items.extend(items)
|
||||
|
||||
all_items.sort(key=lambda x: (x["album"], x["file"]))
|
||||
|
||||
data = {
|
||||
"region": region,
|
||||
"matched_albums": matched,
|
||||
"items": all_items[:limit],
|
||||
"total": len(all_items),
|
||||
"cached_at": int(now),
|
||||
"cache_ttl": CACHE_TTL,
|
||||
}
|
||||
|
||||
CACHE[region] = {"ts": now, "data": data}
|
||||
return data
|
||||
|
||||
|
||||
@app.get("/media/travel/.thumb/{album}/{filename}")
|
||||
def get_thumb(album: str, filename: str):
|
||||
src = (ROOT / album / filename).resolve()
|
||||
if not p.exists() or not p.is_file():
|
||||
raise HTTPException(404, "Thumbnail not found")
|
||||
|
||||
# src로부터 thumb 생성/확인 (원본 확장자 유지)
|
||||
p = ensure_thumb(src, album)
|
||||
return FileResponse(str(p))
|
||||
4
travel-proxy/requirements.txt
Normal file
4
travel-proxy/requirements.txt
Normal file
@@ -0,0 +1,4 @@
|
||||
fastapi
|
||||
uvicorn
|
||||
pillow
|
||||
python-multipart
|
||||
Reference in New Issue
Block a user