diff options
| author | TheSiahxyz <164138827+TheSiahxyz@users.noreply.github.com> | 2026-04-01 15:56:35 +0900 |
|---|---|---|
| committer | TheSiahxyz <164138827+TheSiahxyz@users.noreply.github.com> | 2026-04-01 15:56:35 +0900 |
| commit | 33b14aaa2344b0fd95d1629627c3d135b24ae102 (patch) | |
| tree | 90b214758bc3b076baa7711226a1a1be6268e72e /shared/src | |
| parent | 9360f1a800aa29b40399a2f3bfbfcf215a04e279 (diff) | |
feat: initial trading platform implementation
Binance spot crypto trading platform with microservices architecture:
- shared: Pydantic models, Redis Streams broker, asyncpg DB layer
- data-collector: Binance WebSocket/REST market data collection
- strategy-engine: Plugin-based strategy execution (RSI, Grid)
- order-executor: Order execution with risk management
- portfolio-manager: Position tracking and PnL calculation
- backtester: Historical strategy testing with simulator
- cli: Click-based CLI for all operations
- Docker Compose orchestration with Redis and PostgreSQL
- 24 test files covering all modules
Diffstat (limited to 'shared/src')
| -rw-r--r-- | shared/src/shared/__init__.py | 1 | ||||
| -rw-r--r-- | shared/src/shared/broker.py | 43 | ||||
| -rw-r--r-- | shared/src/shared/config.py | 16 | ||||
| -rw-r--r-- | shared/src/shared/db.py | 184 | ||||
| -rw-r--r-- | shared/src/shared/events.py | 75 | ||||
| -rw-r--r-- | shared/src/shared/models.py | 72 |
6 files changed, 391 insertions, 0 deletions
diff --git a/shared/src/shared/__init__.py b/shared/src/shared/__init__.py new file mode 100644 index 0000000..d2ee024 --- /dev/null +++ b/shared/src/shared/__init__.py @@ -0,0 +1 @@ +"""Shared library for the trading platform.""" diff --git a/shared/src/shared/broker.py b/shared/src/shared/broker.py new file mode 100644 index 0000000..9a50441 --- /dev/null +++ b/shared/src/shared/broker.py @@ -0,0 +1,43 @@ +"""Redis Streams broker for the trading platform.""" +import json +from typing import Any + +import redis.asyncio + + +class RedisBroker: + """Async Redis Streams broker for publishing and reading events.""" + + def __init__(self, redis_url: str) -> None: + self._redis = redis.asyncio.from_url(redis_url) + + async def publish(self, stream: str, data: dict[str, Any]) -> None: + """Publish a message to a Redis stream.""" + payload = json.dumps(data) + await self._redis.xadd(stream, {"payload": payload}) + + async def read( + self, + stream: str, + last_id: str = "$", + count: int = 10, + block: int = 0, + ) -> list[dict[str, Any]]: + """Read messages from a Redis stream.""" + results = await self._redis.xread( + {stream: last_id}, count=count, block=block + ) + messages = [] + if results: + for _stream, entries in results: + for _msg_id, fields in entries: + payload = fields.get(b"payload") or fields.get("payload") + if payload: + if isinstance(payload, bytes): + payload = payload.decode() + messages.append(json.loads(payload)) + return messages + + async def close(self) -> None: + """Close the Redis connection.""" + await self._redis.aclose() diff --git a/shared/src/shared/config.py b/shared/src/shared/config.py new file mode 100644 index 0000000..1304c5e --- /dev/null +++ b/shared/src/shared/config.py @@ -0,0 +1,16 @@ +"""Shared configuration settings for the trading platform.""" +from pydantic_settings import BaseSettings + + +class Settings(BaseSettings): + binance_api_key: str + binance_api_secret: str + redis_url: str = "redis://localhost:6379" + database_url: str = "postgresql://trading:trading@localhost:5432/trading" + log_level: str = "INFO" + risk_max_position_size: float = 0.1 + risk_stop_loss_pct: float = 5.0 + risk_daily_loss_limit_pct: float = 10.0 + dry_run: bool = True + + model_config = {"env_file": ".env", "env_file_encoding": "utf-8"} diff --git a/shared/src/shared/db.py b/shared/src/shared/db.py new file mode 100644 index 0000000..6bddd7c --- /dev/null +++ b/shared/src/shared/db.py @@ -0,0 +1,184 @@ +"""Database layer using asyncpg for the trading platform.""" +from datetime import datetime, timezone +from typing import Optional + +import asyncpg + +from shared.models import Candle, Signal, Order, OrderStatus + + +_INIT_SQL = """ +CREATE TABLE IF NOT EXISTS candles ( + symbol TEXT NOT NULL, + timeframe TEXT NOT NULL, + open_time TIMESTAMPTZ NOT NULL, + open NUMERIC NOT NULL, + high NUMERIC NOT NULL, + low NUMERIC NOT NULL, + close NUMERIC NOT NULL, + volume NUMERIC NOT NULL, + PRIMARY KEY (symbol, timeframe, open_time) +); + +CREATE TABLE IF NOT EXISTS signals ( + id TEXT PRIMARY KEY, + strategy TEXT NOT NULL, + symbol TEXT NOT NULL, + side TEXT NOT NULL, + price NUMERIC NOT NULL, + quantity NUMERIC NOT NULL, + reason TEXT, + created_at TIMESTAMPTZ NOT NULL +); + +CREATE TABLE IF NOT EXISTS orders ( + id TEXT PRIMARY KEY, + signal_id TEXT REFERENCES signals(id), + symbol TEXT NOT NULL, + side TEXT NOT NULL, + type TEXT NOT NULL, + price NUMERIC NOT NULL, + quantity NUMERIC NOT NULL, + status TEXT NOT NULL DEFAULT 'PENDING', + created_at TIMESTAMPTZ NOT NULL, + filled_at TIMESTAMPTZ +); + +CREATE TABLE IF NOT EXISTS trades ( + id TEXT PRIMARY KEY, + order_id TEXT REFERENCES orders(id), + symbol TEXT NOT NULL, + side TEXT NOT NULL, + price NUMERIC NOT NULL, + quantity NUMERIC NOT NULL, + fee NUMERIC NOT NULL DEFAULT 0, + traded_at TIMESTAMPTZ NOT NULL +); + +CREATE TABLE IF NOT EXISTS positions ( + symbol TEXT PRIMARY KEY, + quantity NUMERIC NOT NULL, + avg_entry_price NUMERIC NOT NULL, + current_price NUMERIC NOT NULL, + updated_at TIMESTAMPTZ NOT NULL +); + +CREATE TABLE IF NOT EXISTS portfolio_snapshots ( + id SERIAL PRIMARY KEY, + total_value NUMERIC NOT NULL, + realized_pnl NUMERIC NOT NULL, + unrealized_pnl NUMERIC NOT NULL, + snapshot_at TIMESTAMPTZ NOT NULL +); +""" + + +class Database: + """Async database access layer backed by asyncpg connection pool.""" + + def __init__(self, database_url: str) -> None: + self._database_url = database_url + self._pool: Optional[asyncpg.Pool] = None + + async def connect(self) -> None: + """Create the asyncpg connection pool.""" + self._pool = await asyncpg.create_pool(self._database_url) + + async def close(self) -> None: + """Close the asyncpg connection pool.""" + if self._pool: + await self._pool.close() + self._pool = None + + async def init_tables(self) -> None: + """Create all tables if they do not exist.""" + async with self._pool as conn: + await conn.execute(_INIT_SQL) + + async def insert_candle(self, candle: Candle) -> None: + """Insert a candle row, ignoring duplicates.""" + sql = """ + INSERT INTO candles (symbol, timeframe, open_time, open, high, low, close, volume) + VALUES ($1, $2, $3, $4, $5, $6, $7, $8) + ON CONFLICT DO NOTHING + """ + async with self._pool as conn: + await conn.execute( + sql, + candle.symbol, + candle.timeframe, + candle.open_time, + candle.open, + candle.high, + candle.low, + candle.close, + candle.volume, + ) + + async def insert_signal(self, signal: Signal) -> None: + """Insert a signal row.""" + sql = """ + INSERT INTO signals (id, strategy, symbol, side, price, quantity, reason, created_at) + VALUES ($1, $2, $3, $4, $5, $6, $7, $8) + """ + async with self._pool as conn: + await conn.execute( + sql, + signal.id, + signal.strategy, + signal.symbol, + signal.side.value, + signal.price, + signal.quantity, + signal.reason, + signal.created_at, + ) + + async def insert_order(self, order: Order) -> None: + """Insert an order row.""" + sql = """ + INSERT INTO orders (id, signal_id, symbol, side, type, price, quantity, status, created_at, filled_at) + VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) + """ + async with self._pool as conn: + await conn.execute( + sql, + order.id, + order.signal_id, + order.symbol, + order.side.value, + order.type.value, + order.price, + order.quantity, + order.status.value, + order.created_at, + order.filled_at, + ) + + async def update_order_status( + self, + order_id: str, + status: OrderStatus, + filled_at: Optional[datetime] = None, + ) -> None: + """Update the status (and optionally filled_at) of an order.""" + sql = """ + UPDATE orders SET status = $2, filled_at = $3 WHERE id = $1 + """ + async with self._pool as conn: + await conn.execute(sql, order_id, status.value, filled_at) + + async def get_candles( + self, symbol: str, timeframe: str, limit: int = 500 + ) -> list[dict]: + """Retrieve candles ordered by open_time descending.""" + sql = """ + SELECT symbol, timeframe, open_time, open, high, low, close, volume + FROM candles + WHERE symbol = $1 AND timeframe = $2 + ORDER BY open_time DESC + LIMIT $3 + """ + async with self._pool as conn: + rows = await conn.fetch(sql, symbol, timeframe, limit) + return [dict(row) for row in rows] diff --git a/shared/src/shared/events.py b/shared/src/shared/events.py new file mode 100644 index 0000000..1db2bee --- /dev/null +++ b/shared/src/shared/events.py @@ -0,0 +1,75 @@ +"""Event types and serialization for the trading platform.""" +from enum import Enum +from typing import Any + +from pydantic import BaseModel + +from shared.models import Candle, Signal, Order + + +class EventType(str, Enum): + CANDLE = "CANDLE" + SIGNAL = "SIGNAL" + ORDER = "ORDER" + + +class CandleEvent(BaseModel): + type: EventType = EventType.CANDLE + data: Candle + + def to_dict(self) -> dict: + return { + "type": self.type, + "data": self.data.model_dump(mode="json"), + } + + @classmethod + def from_raw(cls, raw: dict) -> "CandleEvent": + return cls(type=raw["type"], data=Candle(**raw["data"])) + + +class SignalEvent(BaseModel): + type: EventType = EventType.SIGNAL + data: Signal + + def to_dict(self) -> dict: + return { + "type": self.type, + "data": self.data.model_dump(mode="json"), + } + + @classmethod + def from_raw(cls, raw: dict) -> "SignalEvent": + return cls(type=raw["type"], data=Signal(**raw["data"])) + + +class OrderEvent(BaseModel): + type: EventType = EventType.ORDER + data: Order + + def to_dict(self) -> dict: + return { + "type": self.type, + "data": self.data.model_dump(mode="json"), + } + + @classmethod + def from_raw(cls, raw: dict) -> "OrderEvent": + return cls(type=raw["type"], data=Order(**raw["data"])) + + +_EVENT_TYPE_MAP = { + EventType.CANDLE: CandleEvent, + EventType.SIGNAL: SignalEvent, + EventType.ORDER: OrderEvent, +} + + +class Event: + """Dispatcher for deserializing events from raw dicts.""" + + @staticmethod + def from_dict(data: dict) -> Any: + event_type = EventType(data["type"]) + cls = _EVENT_TYPE_MAP[event_type] + return cls.from_raw(data) diff --git a/shared/src/shared/models.py b/shared/src/shared/models.py new file mode 100644 index 0000000..4cb1081 --- /dev/null +++ b/shared/src/shared/models.py @@ -0,0 +1,72 @@ +"""Shared Pydantic models for the trading platform.""" +import uuid +from decimal import Decimal +from datetime import datetime, timezone +from enum import Enum +from typing import Optional + +from pydantic import BaseModel, Field, computed_field + + +class OrderSide(str, Enum): + BUY = "BUY" + SELL = "SELL" + + +class OrderType(str, Enum): + MARKET = "MARKET" + LIMIT = "LIMIT" + + +class OrderStatus(str, Enum): + PENDING = "PENDING" + FILLED = "FILLED" + CANCELLED = "CANCELLED" + FAILED = "FAILED" + + +class Candle(BaseModel): + symbol: str + timeframe: str + open_time: datetime + open: Decimal + high: Decimal + low: Decimal + close: Decimal + volume: Decimal + + +class Signal(BaseModel): + id: str = Field(default_factory=lambda: str(uuid.uuid4())) + strategy: str + symbol: str + side: OrderSide + price: Decimal + quantity: Decimal + reason: str + created_at: datetime = Field(default_factory=lambda: datetime.now(timezone.utc)) + + +class Order(BaseModel): + id: str = Field(default_factory=lambda: str(uuid.uuid4())) + signal_id: str + symbol: str + side: OrderSide + type: OrderType + price: Decimal + quantity: Decimal + status: OrderStatus = OrderStatus.PENDING + created_at: datetime = Field(default_factory=lambda: datetime.now(timezone.utc)) + filled_at: Optional[datetime] = None + + +class Position(BaseModel): + symbol: str + quantity: Decimal + avg_entry_price: Decimal + current_price: Decimal + + @computed_field + @property + def unrealized_pnl(self) -> Decimal: + return self.quantity * (self.current_price - self.avg_entry_price) |
