530 lines
19 KiB
Python
Executable File
530 lines
19 KiB
Python
Executable File
|
|
from datetime import datetime, time as datetime_time
|
|
import schedule
|
|
import time
|
|
import subprocess
|
|
import threading # Import threading module for parallel execution
|
|
import logging # Import logging module
|
|
from logging.handlers import RotatingFileHandler
|
|
|
|
|
|
from dotenv import load_dotenv
|
|
import os
|
|
load_dotenv()
|
|
|
|
useast_ip_address = os.getenv('USEAST_IP_ADDRESS')
|
|
|
|
|
|
# Setup logging
|
|
log_file = 'logs/cron_job.log'
|
|
logger = logging.getLogger('CronJobLogger')
|
|
logger.setLevel(logging.INFO)
|
|
handler = RotatingFileHandler(log_file, maxBytes=5*1024*1024, backupCount=5) # 5MB per file, 5 backup files
|
|
formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')
|
|
handler.setFormatter(formatter)
|
|
logger.addHandler(handler)
|
|
|
|
|
|
|
|
# Set the system's timezone to Berlin at the beginning
|
|
subprocess.run(["timedatectl", "set-timezone", "Europe/Berlin"])
|
|
|
|
|
|
# Function to run commands and log output
|
|
def run_command(command):
|
|
process = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True)
|
|
stdout, stderr = process.communicate()
|
|
|
|
# Log stdout and stderr
|
|
logger.info(f"Command: {' '.join(command)}")
|
|
logger.info("Output:\n" + stdout)
|
|
if stderr:
|
|
logger.error("Error:\n" + stderr)
|
|
|
|
def run_json_job():
|
|
# Run the asynchronous function inside an asyncio loop
|
|
run_command(["python3", "restart_json.py"])
|
|
run_command(["pm2", "restart","fastapi"])
|
|
|
|
def run_pocketbase():
|
|
# Run the asynchronous function inside an asyncio loop
|
|
run_command(["python3", "cron_pocketbase.py"])
|
|
|
|
def run_cron_insider_trading():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_insider_trading.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/insider-trading",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_congress_trading():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_congress_trading.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/congress-trading",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_dividend_list():
|
|
week = datetime.today().weekday()
|
|
current_time = datetime.now().time()
|
|
start_time = datetime_time(15, 30)
|
|
end_time = datetime_time(22, 30)
|
|
|
|
if week <= 4 and start_time <= current_time < end_time:
|
|
run_command(["python3", "cron_dividend_kings.py"])
|
|
run_command(["python3", "cron_dividend_aristocrats.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/stocks-list",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_cron_var():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_var.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/var",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_analyst_estimate():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_analyst_estimate.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/analyst-estimate",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_shareholders():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_shareholders.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/shareholders",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_share_statistics():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_share_statistics.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/share-statistics",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_retail_volume():
|
|
run_command(["python3", "cron_retail_volume.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/retail-volume",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_cron_market_movers():
|
|
run_command(["python3", "cron_market_movers.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/market-movers",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
|
|
def run_cron_market_news():
|
|
run_command(["python3", "cron_market_news.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/market-news",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_cron_heatmap():
|
|
run_command(["python3", "cron_heatmap.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/heatmaps",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_cron_quote():
|
|
week = datetime.today().weekday()
|
|
if week <= 4:
|
|
run_command(["python3", "cron_quote.py"])
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/quote", f"root@{useast_ip_address}:/root/backend/app/json"]
|
|
run_command(command)
|
|
|
|
def run_cron_price_alert():
|
|
week = datetime.today().weekday()
|
|
if week <= 4:
|
|
run_command(["python3", "cron_price_alert.py"])
|
|
|
|
def run_cron_portfolio():
|
|
week = datetime.today().weekday()
|
|
if week <= 4:
|
|
run_command(["python3", "cron_portfolio.py"])
|
|
|
|
def run_cron_options_flow():
|
|
week = datetime.today().weekday()
|
|
current_time = datetime.now().time()
|
|
start_time = datetime_time(15, 30)
|
|
end_time = datetime_time(22, 30)
|
|
|
|
if week <= 4 and start_time <= current_time < end_time:
|
|
run_command(["python3", "cron_options_flow.py"])
|
|
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/options-flow/feed/",
|
|
f"root@{useast_ip_address}:/root/backend/app/json/options-flow/feed/"
|
|
]
|
|
run_command(command)
|
|
|
|
|
|
def run_cron_options_zero_dte():
|
|
week = datetime.today().weekday()
|
|
current_time = datetime.now().time()
|
|
start_time = datetime_time(15, 30)
|
|
end_time = datetime_time(22, 30)
|
|
if week <= 4 and start_time <= current_time < end_time:
|
|
run_command(["python3", "cron_options_zero_dte.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/options-flow/zero-dte/",
|
|
f"root@{useast_ip_address}:/root/backend/app/json/options-flow/zero-dte/"
|
|
]
|
|
run_command(command)
|
|
|
|
|
|
def run_ta_rating():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_ta_rating.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/ta-rating",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_stockdeck():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_stockdeck.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/stockdeck",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_similar_stocks():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_similar_stocks.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/similar-stocks",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_historical_price():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_historical_price.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/historical-price",
|
|
f"root@{useast_ip_address}:/root/backend/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_one_day_price():
|
|
week = datetime.today().weekday()
|
|
if week <= 4:
|
|
run_command(["python3", "cron_one_day_price.py"])
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/one-day-price/", f"root@{useast_ip_address}:/root/backend/app/json/one-day-price/"]
|
|
run_command(command)
|
|
|
|
def run_sec_filings():
|
|
week = datetime.today().weekday()
|
|
if week <= 4:
|
|
run_command(["python3", "cron_sec_filings.py"])
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/sec-filings", f"root@{useast_ip_address}:/root/backend/app/json"]
|
|
run_command(command)
|
|
|
|
def run_executive():
|
|
week = datetime.today().weekday()
|
|
if week <= 4:
|
|
run_command(["python3", "cron_executive.py"])
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/executives", f"root@{useast_ip_address}:/root/backend/app/json"]
|
|
run_command(command)
|
|
|
|
|
|
def run_options_bubble_ticker():
|
|
week = datetime.today().weekday()
|
|
if week <= 4:
|
|
run_command(["python3", "cron_options_bubble.py"])
|
|
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/options-bubble/", f"root@{useast_ip_address}:/root/backend/app/json/options-bubble/"]
|
|
run_command(command)
|
|
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/options-flow/company/", f"root@{useast_ip_address}:/root/backend/app/json/options-flow/company/"]
|
|
run_command(command)
|
|
|
|
def run_analyst_rating():
|
|
run_command(["python3", "cron_analyst_insight.py"])
|
|
run_command(["python3", "cron_analyst_db.py"])
|
|
run_command(["python3", "cron_analyst_ticker.py"])
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/analyst", f"root@{useast_ip_address}:/root/backend/app/json"]
|
|
run_command(command)
|
|
|
|
def run_market_moods():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_bull_bear_say.py"])
|
|
run_command(["python3", "cron_wiim.py"])
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/bull_bear_say", f"root@{useast_ip_address}:/root/backend/app/json"]
|
|
run_command(command)
|
|
command = ["sudo", "rsync", "-avz", "-e", "ssh", "/root/backend/app/json/wiim", f"root@{useast_ip_address}:/root/backend/app/json"]
|
|
run_command(command)
|
|
|
|
|
|
def run_db_schedule_job():
|
|
#update db daily
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["bash", "run_universe.sh"])
|
|
|
|
def run_restart_cache():
|
|
#update db daily
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["pm2", "restart","fastapi"])
|
|
run_command(["pm2", "restart","fastify"])
|
|
#run_command(["python3", "cache_endpoints.py"])
|
|
|
|
def run_dark_pool():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_dark_pool.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/dark-pool",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
|
|
def run_dark_pool_flow():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_dark_pool_flow.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/dark-pool/flow",
|
|
f"root@{useast_ip_address}:/root/backend/app/json/dark-pool"
|
|
]
|
|
run_command(command)
|
|
|
|
|
|
def run_market_maker():
|
|
run_command(["python3", "cron_market_maker.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/market-maker",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_ownership_stats():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_ownership_stats.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/ownership-stats",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_clinical_trial():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_clinical_trial.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/clinical-trial",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_fda_calendar():
|
|
week = datetime.today().weekday()
|
|
if week <= 5:
|
|
run_command(["python3", "cron_fda_calendar.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/fda-calendar",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_borrowed_share():
|
|
run_command(["python3", "cron_borrowed_share.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/borrowed-share",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_implied_volatility():
|
|
run_command(["python3", "cron_implied_volatility.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/implied-volatility",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_options_net_flow():
|
|
run_command(["python3", "cron_options_net_flow.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/options-net-flow",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_government_contract():
|
|
run_command(["python3", "cron_government_contract.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/government-contract",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_dashboard():
|
|
run_command(["python3", "cron_dashboard.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/dashboard",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_reddit_tracker():
|
|
run_command(["python3", "cron_reddit_tracker.py"])
|
|
run_command(["python3", "cron_reddit_statistics.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/reddit-tracker",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
def run_cramer_tracker():
|
|
run_command(["python3", "cron_cramer_tracker.py"])
|
|
command = [
|
|
"sudo", "rsync", "-avz", "-e", "ssh",
|
|
"/root/backend/app/json/cramer-tracker",
|
|
f"root@{useast_ip_address}:/root/backend/app/json"
|
|
]
|
|
run_command(command)
|
|
|
|
# Create functions to run each schedule in a separate thread
|
|
def run_threaded(job_func):
|
|
job_thread = threading.Thread(target=job_func)
|
|
job_thread.start()
|
|
|
|
# Schedule the job to run
|
|
|
|
schedule.every().day.at("01:00").do(run_threaded, run_options_bubble_ticker).tag('options_ticker_job')
|
|
schedule.every().day.at("02:00").do(run_threaded, run_db_schedule_job)
|
|
schedule.every().day.at("03:00").do(run_threaded, run_dark_pool)
|
|
schedule.every().day.at("04:00").do(run_threaded, run_options_net_flow).tag('options_net_flow_job')
|
|
schedule.every().day.at("06:00").do(run_threaded, run_historical_price).tag('historical_job')
|
|
schedule.every().day.at("06:30").do(run_threaded, run_pocketbase).tag('pocketbase_job')
|
|
|
|
schedule.every().day.at("07:00").do(run_threaded, run_ta_rating).tag('ta_rating_job')
|
|
schedule.every().day.at("07:30").do(run_threaded, run_government_contract).tag('government_contract_job')
|
|
|
|
schedule.every().day.at("08:00").do(run_threaded, run_cron_insider_trading).tag('insider_trading_job')
|
|
schedule.every().day.at("09:00").do(run_threaded, run_congress_trading).tag('congress_job')
|
|
schedule.every().day.at("10:00").do(run_threaded, run_shareholders).tag('shareholders_job')
|
|
schedule.every().day.at("10:15").do(run_threaded, run_share_statistics).tag('share_statistics_job')
|
|
schedule.every().day.at("10:30").do(run_threaded, run_sec_filings).tag('sec_filings_job')
|
|
schedule.every().day.at("11:00").do(run_threaded, run_executive).tag('executive_job')
|
|
schedule.every().day.at("03:00").do(run_threaded, run_retail_volume).tag('retail_volume_job')
|
|
schedule.every().day.at("11:45").do(run_threaded, run_clinical_trial).tag('clinical_trial_job')
|
|
schedule.every().day.at("05:00").do(run_threaded, run_implied_volatility).tag('implied_volatility_job')
|
|
|
|
|
|
schedule.every().day.at("13:30").do(run_threaded, run_stockdeck).tag('stockdeck_job')
|
|
schedule.every().day.at("13:40").do(run_threaded, run_analyst_estimate).tag('analyst_estimate_job')
|
|
schedule.every().day.at("13:45").do(run_threaded, run_similar_stocks).tag('similar_stocks_job')
|
|
schedule.every().day.at("14:00").do(run_threaded, run_cron_var).tag('var_job')
|
|
|
|
|
|
schedule.every().day.at("15:45").do(run_threaded, run_restart_cache)
|
|
schedule.every(2).days.at("01:00").do(run_borrowed_share).tag('borrowed_share_job')
|
|
|
|
schedule.every(2).days.at("01:00").do(run_threaded, run_market_maker).tag('markt_maker_job')
|
|
schedule.every().saturday.at("05:00").do(run_threaded, run_ownership_stats).tag('ownership_stats_job')
|
|
|
|
|
|
schedule.every(1).minutes.do(run_threaded, run_cron_portfolio).tag('portfolio_job')
|
|
schedule.every(5).minutes.do(run_threaded, run_cron_market_movers).tag('market_movers_job')
|
|
schedule.every(2).minutes.do(run_threaded, run_dashboard).tag('dashboard_job')
|
|
|
|
schedule.every(30).minutes.do(run_threaded, run_dividend_list).tag('dividend_list_job')
|
|
schedule.every(15).minutes.do(run_threaded, run_cron_market_news).tag('market_news_job')
|
|
schedule.every(10).minutes.do(run_threaded, run_one_day_price).tag('one_day_price_job')
|
|
schedule.every(15).minutes.do(run_threaded, run_cron_heatmap).tag('heatmap_job')
|
|
|
|
schedule.every(10).minutes.do(run_threaded, run_reddit_tracker).tag('reddit_tracker_job')
|
|
schedule.every(10).minutes.do(run_threaded, run_cramer_tracker).tag('cramer_tracker_job')
|
|
|
|
|
|
schedule.every(1).minutes.do(run_threaded, run_cron_quote).tag('quote_job')
|
|
schedule.every(1).minutes.do(run_threaded, run_cron_price_alert).tag('price_alert_job')
|
|
schedule.every(15).minutes.do(run_threaded, run_market_moods).tag('market_moods_job')
|
|
#schedule.every(10).minutes.do(run_threaded, run_dark_pool_flow).tag('dark_pool_flow_job')
|
|
|
|
schedule.every(2).hours.do(run_threaded, run_fda_calendar).tag('fda_calendar_job')
|
|
schedule.every(3).hours.do(run_threaded, run_json_job).tag('json_job')
|
|
schedule.every(6).hours.do(run_threaded, run_analyst_rating).tag('analyst_job')
|
|
|
|
schedule.every(10).seconds.do(run_threaded, run_cron_options_flow).tag('options_flow_job')
|
|
schedule.every(10).seconds.do(run_threaded, run_cron_options_zero_dte).tag('options_zero_dte_job')
|
|
|
|
|
|
# Run the scheduled jobs indefinitelyp
|
|
while True:
|
|
schedule.run_pending()
|
|
time.sleep(3) |