refactor: rename stock-lab → stock (graduation)
- git mv stock-lab/ → stock/ - docker-compose.yml: 서비스 키 + container_name + build.context + frontend.depends_on + agent-office STOCK_LAB_URL → STOCK_URL - agent-office/app: config.py, service_proxy.py, agents/stock.py, tests/ STOCK_LAB_URL → STOCK_URL - nginx/default.conf: proxy_pass http://stock-lab → http://stock (3 lines) - CLAUDE.md / README.md / STATUS.md / scripts/ 문구 갱신 - stock/ 내부 자기 참조 갱신 lab 네이밍 정책 (feedback_lab_naming.md) graduation. API URL / Python import / DB 파일명 변경 없음.
This commit is contained in:
0
stock/app/screener/nodes/__init__.py
Normal file
0
stock/app/screener/nodes/__init__.py
Normal file
36
stock/app/screener/nodes/ai_news.py
Normal file
36
stock/app/screener/nodes/ai_news.py
Normal file
@@ -0,0 +1,36 @@
|
||||
"""AI 뉴스 호재/악재 점수 노드.
|
||||
|
||||
ScreenContext.news_sentiment (DataFrame: ticker, score_raw, news_count) 를
|
||||
min_news_count 로 필터한 뒤 percentile_rank 로 0~100 변환.
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode, percentile_rank
|
||||
|
||||
|
||||
class AiNewsSentiment(ScoreNode):
|
||||
name = "ai_news"
|
||||
label = "AI 뉴스 호재/악재"
|
||||
default_params = {"min_news_count": 1}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"min_news_count": {
|
||||
"type": "integer", "minimum": 0, "default": 1,
|
||||
"description": "최소 분석 뉴스 수. 미만이면 점수 미산출.",
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
df = getattr(ctx, "news_sentiment", None)
|
||||
if df is None or df.empty:
|
||||
return pd.Series(dtype=float)
|
||||
min_news = int(params.get("min_news_count", 1))
|
||||
df = df[df["news_count"] >= min_news]
|
||||
if df.empty:
|
||||
return pd.Series(dtype=float)
|
||||
return percentile_rank(df.set_index("ticker")["score_raw"])
|
||||
40
stock/app/screener/nodes/base.py
Normal file
40
stock/app/screener/nodes/base.py
Normal file
@@ -0,0 +1,40 @@
|
||||
"""Node base classes + helpers."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
from abc import ABC, abstractmethod
|
||||
from typing import Any, ClassVar
|
||||
|
||||
import pandas as pd
|
||||
|
||||
|
||||
class ScoreNode(ABC):
|
||||
name: ClassVar[str]
|
||||
label: ClassVar[str]
|
||||
default_params: ClassVar[dict]
|
||||
param_schema: ClassVar[dict]
|
||||
|
||||
@abstractmethod
|
||||
def compute(self, ctx: "Any", params: dict) -> pd.Series:
|
||||
"""returns Series indexed by ticker, 0..100 float."""
|
||||
|
||||
|
||||
class GateNode(ABC):
|
||||
name: ClassVar[str]
|
||||
label: ClassVar[str]
|
||||
default_params: ClassVar[dict]
|
||||
param_schema: ClassVar[dict]
|
||||
|
||||
@abstractmethod
|
||||
def filter(self, ctx: "Any", params: dict) -> pd.Index:
|
||||
"""returns surviving tickers."""
|
||||
|
||||
|
||||
def percentile_rank(series: pd.Series) -> pd.Series:
|
||||
"""Percentile rank in [0, 100]. All-equal → 50. NaN preserved."""
|
||||
if series.empty:
|
||||
return series.astype(float)
|
||||
if series.dropna().nunique() == 1:
|
||||
return pd.Series(50.0, index=series.index)
|
||||
ranked = series.rank(pct=True, na_option="keep") * 100.0
|
||||
return ranked
|
||||
33
stock/app/screener/nodes/foreign_buy.py
Normal file
33
stock/app/screener/nodes/foreign_buy.py
Normal file
@@ -0,0 +1,33 @@
|
||||
"""외국인 N일 누적 순매수 강도 (시총 대비)."""
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode, percentile_rank
|
||||
|
||||
|
||||
class ForeignBuy(ScoreNode):
|
||||
name = "foreign_buy"
|
||||
label = "외국인 누적 순매수"
|
||||
default_params = {"window_days": 5}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"window_days": {"type": "integer", "minimum": 1, "maximum": 60, "default": 5}
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
window = int(params.get("window_days", 5))
|
||||
flow = ctx.flow
|
||||
if flow.empty:
|
||||
return pd.Series(dtype=float)
|
||||
|
||||
last_dates = (
|
||||
flow.sort_values("date").groupby("ticker").tail(window)
|
||||
)
|
||||
net_sum = last_dates.groupby("ticker")["foreign_net"].sum()
|
||||
|
||||
market_cap = ctx.master["market_cap"].fillna(0).reindex(net_sum.index)
|
||||
raw = (net_sum / market_cap.replace(0, pd.NA)).astype(float)
|
||||
|
||||
return percentile_rank(raw).fillna(50.0)
|
||||
30
stock/app/screener/nodes/high52w.py
Normal file
30
stock/app/screener/nodes/high52w.py
Normal file
@@ -0,0 +1,30 @@
|
||||
"""52주 신고가 근접도 (룰 기반: 70% 미만 0점, 100% 도달 100점, 선형)."""
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode
|
||||
|
||||
|
||||
class High52WProximity(ScoreNode):
|
||||
name = "high52w"
|
||||
label = "52주 신고가 근접도"
|
||||
default_params = {"window_days": 252}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"window_days": {"type": "integer", "minimum": 60, "maximum": 504, "default": 252}
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
window = int(params.get("window_days", 252))
|
||||
prices = ctx.prices
|
||||
if prices.empty:
|
||||
return pd.Series(dtype=float)
|
||||
|
||||
ordered = prices.sort_values("date")
|
||||
last = ordered.groupby("ticker").tail(window)
|
||||
agg = last.groupby("ticker").agg(close=("close", "last"), high=("high", "max"))
|
||||
proximity = (agg["close"] / agg["high"]).clip(upper=1.0)
|
||||
score = ((proximity - 0.7) / 0.3).clip(lower=0.0, upper=1.0) * 100.0
|
||||
return score.fillna(0.0)
|
||||
81
stock/app/screener/nodes/hygiene.py
Normal file
81
stock/app/screener/nodes/hygiene.py
Normal file
@@ -0,0 +1,81 @@
|
||||
"""HygieneGate — pre-filter for screener."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import GateNode
|
||||
|
||||
|
||||
class HygieneGate(GateNode):
|
||||
name = "hygiene"
|
||||
label = "위생 게이트"
|
||||
default_params = {
|
||||
"min_market_cap_won": 50_000_000_000,
|
||||
"min_avg_value_won": 500_000_000,
|
||||
"min_listed_days": 60,
|
||||
"skip_managed": True,
|
||||
"skip_preferred": True,
|
||||
"skip_spac": True,
|
||||
"skip_halted_days": 3,
|
||||
}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"min_market_cap_won": {"type": "integer", "minimum": 0},
|
||||
"min_avg_value_won": {"type": "integer", "minimum": 0},
|
||||
"min_listed_days": {"type": "integer", "minimum": 0},
|
||||
"skip_managed": {"type": "boolean"},
|
||||
"skip_preferred": {"type": "boolean"},
|
||||
"skip_spac": {"type": "boolean"},
|
||||
"skip_halted_days": {"type": "integer", "minimum": 0},
|
||||
},
|
||||
}
|
||||
|
||||
def filter(self, ctx, params: dict) -> pd.Index:
|
||||
master = ctx.master.copy()
|
||||
prices = ctx.prices
|
||||
|
||||
# 시총
|
||||
master = master[master["market_cap"].fillna(0) >= params["min_market_cap_won"]]
|
||||
|
||||
# 우선주·관리·스팩
|
||||
if params.get("skip_preferred", True):
|
||||
master = master[master["is_preferred"] == 0]
|
||||
if params.get("skip_managed", True):
|
||||
master = master[master["is_managed"] == 0]
|
||||
if params.get("skip_spac", True):
|
||||
master = master[master["is_spac"] == 0]
|
||||
|
||||
candidates = master.index
|
||||
|
||||
# 20일 평균 거래대금
|
||||
if not prices.empty:
|
||||
recent20 = (
|
||||
prices[prices["ticker"].isin(candidates)]
|
||||
.sort_values("date")
|
||||
.groupby("ticker")
|
||||
.tail(20)
|
||||
)
|
||||
avg_value = recent20.groupby("ticker")["value"].mean()
|
||||
ok = avg_value[avg_value >= params["min_avg_value_won"]].index
|
||||
candidates = candidates.intersection(ok)
|
||||
|
||||
# 최근 N일 거래정지 (volume==0 N일 이상)
|
||||
halted_days = params.get("skip_halted_days", 3)
|
||||
if halted_days > 0 and not prices.empty:
|
||||
recent = (
|
||||
prices[prices["ticker"].isin(candidates)]
|
||||
.sort_values("date")
|
||||
.groupby("ticker")
|
||||
.tail(halted_days)
|
||||
)
|
||||
zero_count = (
|
||||
recent.assign(z=lambda d: (d["volume"] == 0).astype(int))
|
||||
.groupby("ticker")["z"].sum()
|
||||
)
|
||||
healthy = zero_count[zero_count < halted_days].index
|
||||
candidates = candidates.intersection(healthy)
|
||||
|
||||
# 상장 N일 — MVP에선 listed_date null 허용, null이면 통과
|
||||
return pd.Index(candidates)
|
||||
51
stock/app/screener/nodes/ma_alignment.py
Normal file
51
stock/app/screener/nodes/ma_alignment.py
Normal file
@@ -0,0 +1,51 @@
|
||||
"""이평선 정배열 점수 — 5개 조건 충족 개수 / 5 × 100."""
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode
|
||||
|
||||
|
||||
class MaAlignment(ScoreNode):
|
||||
name = "ma_alignment"
|
||||
label = "이평선 정배열"
|
||||
default_params = {"ma_periods": [50, 150, 200]}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"ma_periods": {"type": "array", "items": {"type": "integer"}}
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
ma_periods = params.get("ma_periods", self.default_params["ma_periods"])
|
||||
if len(ma_periods) != 3:
|
||||
raise ValueError("ma_periods must have 3 entries (short, medium, long)")
|
||||
ma_s, ma_m, ma_l = (int(x) for x in ma_periods)
|
||||
|
||||
prices = ctx.prices
|
||||
if prices.empty:
|
||||
return pd.Series(dtype=float)
|
||||
|
||||
ordered = prices.sort_values("date")
|
||||
min_history = max(252, ma_l)
|
||||
|
||||
def _score(s: pd.Series) -> float:
|
||||
closes = s.astype(float).reset_index(drop=True)
|
||||
if len(closes) < min_history:
|
||||
return float("nan")
|
||||
close = closes.iloc[-1]
|
||||
ma_short = closes.rolling(ma_s).mean().iloc[-1]
|
||||
ma_medium = closes.rolling(ma_m).mean().iloc[-1]
|
||||
ma_long = closes.rolling(ma_l).mean().iloc[-1]
|
||||
low52 = closes.iloc[-252:].min()
|
||||
conds = [
|
||||
close > ma_short,
|
||||
ma_short > ma_medium,
|
||||
ma_medium > ma_long,
|
||||
close > ma_long,
|
||||
close >= low52 * 1.25,
|
||||
]
|
||||
return sum(conds) / 5 * 100.0
|
||||
|
||||
raw = ordered.groupby("ticker", group_keys=False)["close"].apply(_score)
|
||||
return raw.fillna(0.0)
|
||||
34
stock/app/screener/nodes/momentum.py
Normal file
34
stock/app/screener/nodes/momentum.py
Normal file
@@ -0,0 +1,34 @@
|
||||
"""20일 모멘텀."""
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode, percentile_rank
|
||||
|
||||
|
||||
class Momentum20(ScoreNode):
|
||||
name = "momentum"
|
||||
label = "20일 모멘텀"
|
||||
default_params = {"window_days": 20}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"window_days": {"type": "integer", "minimum": 5, "maximum": 120, "default": 20}
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
window = int(params.get("window_days", 20))
|
||||
prices = ctx.prices
|
||||
if prices.empty:
|
||||
return pd.Series(dtype=float)
|
||||
|
||||
ordered = prices.sort_values("date")
|
||||
last = ordered.groupby("ticker").tail(window + 1)
|
||||
|
||||
def _ret(s):
|
||||
if len(s) < window + 1:
|
||||
return float("nan")
|
||||
return s.iloc[-1] / s.iloc[0] - 1
|
||||
|
||||
raw = last.groupby("ticker")["close"].apply(_ret)
|
||||
return percentile_rank(raw).fillna(50.0)
|
||||
48
stock/app/screener/nodes/rs_rating.py
Normal file
48
stock/app/screener/nodes/rs_rating.py
Normal file
@@ -0,0 +1,48 @@
|
||||
"""RS Rating — IBD 가중 (3m=2,6m=1,9m=1,12m=1)."""
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode, percentile_rank
|
||||
|
||||
|
||||
_PERIOD_TO_DAYS = {"3m": 63, "6m": 126, "9m": 189, "12m": 252}
|
||||
|
||||
|
||||
class RsRating(ScoreNode):
|
||||
name = "rs_rating"
|
||||
label = "RS Rating (시장 대비 상대강도)"
|
||||
default_params = {"weights": {"3m": 2, "6m": 1, "9m": 1, "12m": 1}}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"weights": {"type": "object"}
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
weights: dict = params.get("weights", self.default_params["weights"])
|
||||
prices = ctx.prices
|
||||
kospi = ctx.kospi
|
||||
if prices.empty or kospi.empty:
|
||||
return pd.Series(dtype=float)
|
||||
|
||||
ordered = prices.sort_values("date")
|
||||
|
||||
def _excess_for_ticker(g: pd.DataFrame) -> float:
|
||||
closes = g.set_index("date")["close"]
|
||||
total = 0.0
|
||||
wsum = 0.0
|
||||
for period, w in weights.items():
|
||||
k = _PERIOD_TO_DAYS.get(period, 0)
|
||||
if len(closes) <= k or len(kospi) <= k:
|
||||
continue
|
||||
r_stock = closes.iloc[-1] / closes.iloc[-(k + 1)] - 1
|
||||
r_market = kospi.iloc[-1] / kospi.iloc[-(k + 1)] - 1
|
||||
total += w * (r_stock - r_market)
|
||||
wsum += w
|
||||
return total / wsum if wsum else float("nan")
|
||||
|
||||
raw = ordered.groupby("ticker", group_keys=False).apply(
|
||||
_excess_for_ticker, include_groups=False
|
||||
)
|
||||
return percentile_rank(raw).fillna(50.0)
|
||||
40
stock/app/screener/nodes/vcp_lite.py
Normal file
40
stock/app/screener/nodes/vcp_lite.py
Normal file
@@ -0,0 +1,40 @@
|
||||
"""VCP-lite — 단기/장기 일중 변동성 비율 기반 수축률."""
|
||||
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode, percentile_rank
|
||||
|
||||
|
||||
class VcpLite(ScoreNode):
|
||||
name = "vcp_lite"
|
||||
label = "VCP-lite (변동성 수축)"
|
||||
default_params = {"short_window": 40, "long_window": 252}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"short_window": {"type": "integer", "minimum": 10, "maximum": 120, "default": 40},
|
||||
"long_window": {"type": "integer", "minimum": 60, "maximum": 504, "default": 252},
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
short_w = int(params.get("short_window", 40))
|
||||
long_w = int(params.get("long_window", 252))
|
||||
prices = ctx.prices
|
||||
if prices.empty:
|
||||
return pd.Series(dtype=float)
|
||||
|
||||
ordered = prices.sort_values("date").copy()
|
||||
ordered["range_pct"] = (ordered["high"] - ordered["low"]) / ordered["close"]
|
||||
|
||||
def _ratio(s: pd.Series) -> float:
|
||||
if len(s) < long_w:
|
||||
return float("nan")
|
||||
short_vol = s.tail(short_w).mean()
|
||||
long_vol = s.tail(long_w).mean()
|
||||
if long_vol == 0 or pd.isna(long_vol):
|
||||
return float("nan")
|
||||
return 1 - (short_vol / long_vol)
|
||||
|
||||
raw = ordered.groupby("ticker", group_keys=False)["range_pct"].apply(_ratio)
|
||||
return percentile_rank(raw).fillna(50.0)
|
||||
40
stock/app/screener/nodes/volume_surge.py
Normal file
40
stock/app/screener/nodes/volume_surge.py
Normal file
@@ -0,0 +1,40 @@
|
||||
"""거래량 급증 — log1p(recent/baseline)."""
|
||||
|
||||
import numpy as np
|
||||
import pandas as pd
|
||||
|
||||
from .base import ScoreNode, percentile_rank
|
||||
|
||||
|
||||
class VolumeSurge(ScoreNode):
|
||||
name = "volume_surge"
|
||||
label = "거래량 급증"
|
||||
default_params = {"baseline_days": 20, "eval_days": 3}
|
||||
param_schema = {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"baseline_days": {"type": "integer", "minimum": 5, "maximum": 60, "default": 20},
|
||||
"eval_days": {"type": "integer", "minimum": 1, "maximum": 10, "default": 3},
|
||||
},
|
||||
}
|
||||
|
||||
def compute(self, ctx, params: dict) -> pd.Series:
|
||||
baseline = int(params.get("baseline_days", 20))
|
||||
eval_d = int(params.get("eval_days", 3))
|
||||
prices = ctx.prices
|
||||
if prices.empty:
|
||||
return pd.Series(dtype=float)
|
||||
|
||||
ordered = prices.sort_values("date")
|
||||
last_recent = ordered.groupby("ticker").tail(eval_d).groupby("ticker")["volume"].mean()
|
||||
last_baseline = (
|
||||
ordered.groupby("ticker")
|
||||
.tail(baseline + eval_d)
|
||||
.groupby("ticker")
|
||||
.head(baseline)
|
||||
.groupby("ticker")["volume"]
|
||||
.mean()
|
||||
)
|
||||
ratio = last_recent / last_baseline.replace(0, pd.NA)
|
||||
raw = np.log1p(ratio.astype(float))
|
||||
return percentile_rank(raw).fillna(50.0)
|
||||
Reference in New Issue
Block a user