update options stats cron job

This commit is contained in:
MuslemRahimi 2025-01-15 19:13:12 +01:00
parent d3d2429a74
commit 75d83f4c21

View File

@ -1,123 +1,143 @@
import requests import aiohttp
import asyncio
import orjson import orjson
from dotenv import load_dotenv from dotenv import load_dotenv
import os import os
import sqlite3 import sqlite3
import time
load_dotenv() load_dotenv()
api_key = os.getenv('UNUSUAL_WHALES_API_KEY') api_key = os.getenv('UNUSUAL_WHALES_API_KEY')
# Connect to the databases # Database connection and symbol retrieval
con = sqlite3.connect('stocks.db') def get_total_symbols():
etf_con = sqlite3.connect('etf.db') with sqlite3.connect('stocks.db') as con:
cursor = con.cursor() cursor = con.cursor()
cursor.execute("PRAGMA journal_mode = wal") cursor.execute("PRAGMA journal_mode = wal")
cursor.execute("SELECT DISTINCT symbol FROM stocks WHERE symbol NOT LIKE '%.%'") cursor.execute("SELECT DISTINCT symbol FROM stocks WHERE symbol NOT LIKE '%.%'")
stocks_symbols = [row[0] for row in cursor.fetchall()] stocks_symbols = [row[0] for row in cursor.fetchall()]
etf_cursor = etf_con.cursor() with sqlite3.connect('etf.db') as etf_con:
etf_cursor.execute("PRAGMA journal_mode = wal") etf_cursor = etf_con.cursor()
etf_cursor.execute("SELECT DISTINCT symbol FROM etfs") etf_cursor.execute("PRAGMA journal_mode = wal")
etf_symbols = [row[0] for row in etf_cursor.fetchall()] etf_cursor.execute("SELECT DISTINCT symbol FROM etfs")
etf_symbols = [row[0] for row in etf_cursor.fetchall()]
con.close() return stocks_symbols + etf_symbols
etf_con.close()
# Combine the lists of stock and ETF symbols
total_symbols = stocks_symbols + etf_symbols
def get_tickers_from_directory():
directory = "json/options-historical-data/companies"
try:
# Ensure the directory exists
if not os.path.exists(directory):
raise FileNotFoundError(f"The directory '{directory}' does not exist.")
# Get all tickers from filenames
return [file.replace(".json", "") for file in os.listdir(directory) if file.endswith(".json")]
except Exception as e:
print(f"An error occurred: {e}")
return []
def save_json(data, symbol): def save_json(data, symbol):
directory = "json/options-stats/companies" directory = "json/options-stats/companies"
os.makedirs(directory, exist_ok=True) # Ensure the directory exists os.makedirs(directory, exist_ok=True)
with open(f"{directory}/{symbol}.json", 'wb') as file: # Use binary mode for orjson with open(f"{directory}/{symbol}.json", 'wb') as file:
file.write(orjson.dumps(data)) file.write(orjson.dumps(data))
def safe_round(value): def safe_round(value):
"""Attempt to convert a value to float and round it. Return the original value if not possible."""
try: try:
return round(float(value), 2) return round(float(value), 2)
except (ValueError, TypeError): except (ValueError, TypeError):
return value return value
def calculate_neutral_premium(data_item): def calculate_neutral_premium(data_item):
"""Calculate the neutral premium for a data item."""
call_premium = float(data_item['call_premium']) call_premium = float(data_item['call_premium'])
put_premium = float(data_item['put_premium']) put_premium = float(data_item['put_premium'])
bearish_premium = float(data_item['bearish_premium']) bearish_premium = float(data_item['bearish_premium'])
bullish_premium = float(data_item['bullish_premium']) bullish_premium = float(data_item['bullish_premium'])
total_premiums = bearish_premium + bullish_premium total_premiums = bearish_premium + bullish_premium
observed_premiums = call_premium + put_premium observed_premiums = call_premium + put_premium
neutral_premium = observed_premiums - total_premiums neutral_premium = observed_premiums - total_premiums
return safe_round(neutral_premium) return safe_round(neutral_premium)
def prepare_data(data): def prepare_data(data):
for item in data: for item in data:
symbol = item['ticker']
bearish_premium = float(item['bearish_premium'])
bullish_premium = float(item['bullish_premium'])
neutral_premium = calculate_neutral_premium(item)
new_item = {
key: safe_round(value)
for key, value in item.items()
if key != 'in_out_flow'
}
new_item['premium_ratio'] = [
safe_round(bearish_premium),
neutral_premium,
safe_round(bullish_premium)
]
try: try:
new_item['open_interest_change'] = new_item['total_open_interest'] - (new_item.get('prev_call_oi',0) + new_item.get('prev_put_oi',0)) symbol = item['ticker']
bearish_premium = float(item['bearish_premium'])
bullish_premium = float(item['bullish_premium'])
neutral_premium = calculate_neutral_premium(item)
new_item = {
key: safe_round(value)
for key, value in item.items()
if key != 'in_out_flow'
}
new_item['premium_ratio'] = [
safe_round(bearish_premium),
neutral_premium,
safe_round(bullish_premium)
]
new_item['open_interest_change'] = (
new_item['total_open_interest'] -
(new_item.get('prev_call_oi', 0) + new_item.get('prev_put_oi', 0))
if 'total_open_interest' in new_item else None
)
if new_item:
save_json(new_item, symbol)
except: except:
new_item['open_interest_change'] = None pass
if len(new_item) > 0:
save_json(new_item, symbol)
def chunk_symbols(symbols, chunk_size=50):
for i in range(0, len(symbols), chunk_size):
yield symbols[i:i + chunk_size]
chunks = chunk_symbols(total_symbols) async def fetch_data(session, chunk):
chunk_counter = 0 # To keep track of how many chunks have been processed chunk_str = ",".join(chunk)
url = "https://api.unusualwhales.com/api/screener/stocks"
params = {"ticker": chunk_str}
headers = {
"Accept": "application/json, text/plain",
"Authorization": api_key
}
for chunk in chunks:
try: try:
chunk_str = ",".join(chunk) async with session.get(url, headers=headers, params=params) as response:
url = "https://api.unusualwhales.com/api/screener/stocks" if response.status == 200:
querystring = {"ticker": chunk_str} json_data = await response.json()
data = json_data.get('data', [])
headers = { prepare_data(data)
"Accept": "application/json, text/plain", print(f"Processed chunk with {len(data)} results.")
"Authorization": api_key else:
} print(f"Error fetching chunk {chunk_str}: {response.status}")
response = requests.get(url, headers=headers, params=querystring)
if response.status_code == 200:
data = response.json()['data']
prepare_data(data)
print(f"Chunk processed. Number of results: {len(data)}")
else:
print(f"Error fetching data for chunk {chunk_str}: {response.status_code}")
# Increment the chunk counter
chunk_counter += 1
# If 50 chunks have been processed, sleep for 60 seconds
if chunk_counter == 50:
print("Processed 50 chunks. Sleeping for 60 seconds...")
time.sleep(60) # Sleep for 60 seconds
chunk_counter = 0 # Reset the chunk counter after sleep
except Exception as e: except Exception as e:
print(f"Error processing chunk {chunk_str}: {e}") print(f"Exception fetching chunk {chunk_str}: {e}")
async def main():
total_symbols = get_tickers_from_directory()
if len(total_symbols) < 3000:
total_symbols = get_total_symbols()
print(f"Number of tickers: {len(total_symbols)}")
chunk_size = 50
chunks = [total_symbols[i:i + chunk_size] for i in range(0, len(total_symbols), chunk_size)]
async with aiohttp.ClientSession() as session:
for i in range(0, len(chunks), 200): # Process 200 chunks at a time
try:
tasks = [fetch_data(session, chunk) for chunk in chunks[i:i + 200]]
await asyncio.gather(*tasks)
print("Processed 200 chunks. Sleeping for 60 seconds...")
await asyncio.sleep(60) # Avoid API rate limits
except:
pass
if __name__ == "__main__":
asyncio.run(main())