diff --git a/app/cron_options_stats.py b/app/cron_options_stats.py index d92db8d..fb75931 100644 --- a/app/cron_options_stats.py +++ b/app/cron_options_stats.py @@ -93,9 +93,7 @@ chunk_counter = 0 # To keep track of how many chunks have been processed for chunk in chunks: try: - chunk_str = ",".join(chunk) - print(chunk_str) - + chunk_str = ",".join(chunk) url = "https://api.unusualwhales.com/api/screener/stocks" querystring = {"ticker": chunk_str} diff --git a/app/primary_cron_job.py b/app/primary_cron_job.py index 3af98be..872abd4 100755 --- a/app/primary_cron_job.py +++ b/app/primary_cron_job.py @@ -85,6 +85,12 @@ def run_dark_pool_ticker(): run_command(["python3", "cron_dark_pool_ticker.py"]) +def run_options_stats(): + now = datetime.now(ny_tz) + week = now.weekday() + if week <= 5: + run_command(["python3", "cron_options_stats.py"]) + def run_fda_calendar(): now = datetime.now(ny_tz) week = now.weekday() @@ -348,6 +354,9 @@ schedule.every().day.at("02:00").do(run_threaded, run_db_schedule_job) schedule.every().day.at("05:00").do(run_threaded, run_options_gex).tag('options_gex_job') schedule.every().day.at("05:00").do(run_threaded, run_export_price).tag('export_price_job') +schedule.every().day.at("05:30").do(run_threaded, run_options_stats).tag('options_stats_job') + + schedule.every().day.at("06:00").do(run_threaded, run_historical_price).tag('historical_job') schedule.every().day.at("06:30").do(run_threaded, run_ai_score).tag('ai_score_job') diff --git a/app/restart_json.py b/app/restart_json.py index bbc942f..e3e52b1 100755 --- a/app/restart_json.py +++ b/app/restart_json.py @@ -975,8 +975,6 @@ async def get_stock_screener(con): item['netIncomeGrowthYears'] = None item['grossProfitGrowthYears'] = None - - for item in stock_screener_data: for key, value in item.items(): if isinstance(value, float):