Skip to content

Commit 1f68d0c

Browse files
committed
Wire AI trading bot to trained-model decision engine
1 parent d9b425c commit 1f68d0c

6 files changed

Lines changed: 275 additions & 327 deletions

File tree

.github/workflows/daily_trading_bot.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -49,8 +49,8 @@ jobs:
4949
RECIPIENT_EMAIL: ${{ secrets.RECIPIENT_EMAIL }}
5050
NVIDIA_API_KEY: ${{ secrets.NVIDIA_API_KEY }}
5151
NVIDIA_API_KEY_ID: ${{ secrets.NVIDIA_API_KEY_ID }}
52-
NVIDIA_REASONING_API_KEY: ${{ secrets.NVIDIA_REASONING_API_KEY }}
53-
NVIDIA_REASONING_API_KEY_ID: ${{ secrets.NVIDIA_REASONING_API_KEY_ID }}
52+
TRAINED_MODEL_INFERENCE_URL: ${{ secrets.TRAINED_MODEL_INFERENCE_URL }}
53+
TRAINED_MODEL_API_KEY: ${{ secrets.TRAINED_MODEL_API_KEY }}
5454
TWELVEDATA_API_KEYS: ${{ secrets.TWELVEDATA_API_KEYS }}
5555
ARM_LIVE_TRADING: ${{ secrets.ARM_LIVE_TRADING }}
5656
LOG_LEVEL: ${{ secrets.LOG_LEVEL }}

config.yaml

Lines changed: 15 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -104,17 +104,25 @@ ai_trading:
104104
max_shorts: 5
105105
# Prevent Core + AI from holding the same symbols (keeps the strategies clearly distinct).
106106
disallow_core_overlap: true
107-
# AI selection uses a compact market snapshot per symbol (no core-model ranking).
107+
# AI selection uses raw prices plus deterministic feature snapshots as model inputs.
108108
price_lookback_days: 30
109+
feature_lookback_days: 80
110+
news_window_days: 7
109111
prompt_candidates_limit: 80
110112
# Force AI weights to deploy most available cash each run.
111113
min_total_weight: 0.90
112-
# Use a separate reasoning model/key for trade selection.
113-
llm_model: "nvidia/llama-3.1-nemotron-ultra-253b-v1"
114-
llm_fallback_models:
115-
- "meta/llama-3.1-70b-instruct"
116-
- "stockmark/stockmark-2-100b-instruct"
117-
api_key_env: "NVIDIA_REASONING_API_KEY"
114+
decision_engine: "trained_model"
115+
trained_model:
116+
backend: "http"
117+
inference_url: ""
118+
inference_url_env: "TRAINED_MODEL_INFERENCE_URL"
119+
api_key_env: "TRAINED_MODEL_API_KEY"
120+
timeout_seconds: 60
121+
base_model: "Qwen/Qwen2.5-7B-Instruct"
122+
adapter_dir: "./models/lora_solid_adapter"
123+
max_new_tokens: 64
124+
temperature: 0.0
125+
cpu_threads: 4
118126

119127

120128

0 commit comments

Comments
 (0)