refactor code
This commit is contained in:
parent
38aeae5b4b
commit
97072bc844
58
app/main.py
58
app/main.py
@ -3469,61 +3469,7 @@ async def get_reddit_tracker(api_key: str = Security(get_api_key)):
|
|||||||
headers={"Content-Encoding": "gzip"}
|
headers={"Content-Encoding": "gzip"}
|
||||||
)
|
)
|
||||||
|
|
||||||
@app.get("/dividend-kings")
|
|
||||||
async def get_dividend_kings():
|
|
||||||
cache_key = f"dividend-kings"
|
|
||||||
cached_result = redis_client.get(cache_key)
|
|
||||||
if cached_result:
|
|
||||||
return StreamingResponse(
|
|
||||||
io.BytesIO(cached_result),
|
|
||||||
media_type="application/json",
|
|
||||||
headers={"Content-Encoding": "gzip"}
|
|
||||||
)
|
|
||||||
try:
|
|
||||||
with open(f"json/dividends/list/dividend-kings.json", 'rb') as file:
|
|
||||||
res = orjson.loads(file.read())
|
|
||||||
except:
|
|
||||||
res = []
|
|
||||||
|
|
||||||
data = orjson.dumps(res)
|
|
||||||
compressed_data = gzip.compress(data)
|
|
||||||
|
|
||||||
redis_client.set(cache_key, compressed_data)
|
|
||||||
redis_client.expire(cache_key, 60*20)
|
|
||||||
|
|
||||||
return StreamingResponse(
|
|
||||||
io.BytesIO(compressed_data),
|
|
||||||
media_type="application/json",
|
|
||||||
headers={"Content-Encoding": "gzip"}
|
|
||||||
)
|
|
||||||
|
|
||||||
@app.get("/dividend-aristocrats")
|
|
||||||
async def get_dividend_kings():
|
|
||||||
cache_key = f"dividend-aristocrats"
|
|
||||||
cached_result = redis_client.get(cache_key)
|
|
||||||
if cached_result:
|
|
||||||
return StreamingResponse(
|
|
||||||
io.BytesIO(cached_result),
|
|
||||||
media_type="application/json",
|
|
||||||
headers={"Content-Encoding": "gzip"}
|
|
||||||
)
|
|
||||||
try:
|
|
||||||
with open(f"json/dividends/list/dividend-aristocrats.json", 'rb') as file:
|
|
||||||
res = orjson.loads(file.read())
|
|
||||||
except:
|
|
||||||
res = []
|
|
||||||
data = orjson.dumps(res)
|
|
||||||
compressed_data = gzip.compress(data)
|
|
||||||
|
|
||||||
redis_client.set(cache_key, compressed_data)
|
|
||||||
redis_client.expire(cache_key, 60*20)
|
|
||||||
|
|
||||||
return StreamingResponse(
|
|
||||||
io.BytesIO(compressed_data),
|
|
||||||
media_type="application/json",
|
|
||||||
headers={"Content-Encoding": "gzip"}
|
|
||||||
)
|
|
||||||
return res
|
|
||||||
|
|
||||||
@app.post("/historical-market-cap")
|
@app.post("/historical-market-cap")
|
||||||
async def get_historical_market_cap(data:TickerData, api_key: str = Security(get_api_key)):
|
async def get_historical_market_cap(data:TickerData, api_key: str = Security(get_api_key)):
|
||||||
@ -3915,7 +3861,6 @@ async def get_statistics(data: TickerData, api_key: str = Security(get_api_key))
|
|||||||
headers={"Content-Encoding": "gzip"}
|
headers={"Content-Encoding": "gzip"}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@app.post("/list-category")
|
@app.post("/list-category")
|
||||||
async def get_statistics(data: FilterStockList, api_key: str = Security(get_api_key)):
|
async def get_statistics(data: FilterStockList, api_key: str = Security(get_api_key)):
|
||||||
filter_list = data.filterList.lower()
|
filter_list = data.filterList.lower()
|
||||||
@ -3934,6 +3879,8 @@ async def get_statistics(data: FilterStockList, api_key: str = Security(get_api_
|
|||||||
category_type = 'industry'
|
category_type = 'industry'
|
||||||
elif filter_list in ['ca','cn','de','gb','il','in','jp','nyse','nasdaq','amex','dowjones','sp500','nasdaq100']:
|
elif filter_list in ['ca','cn','de','gb','il','in','jp','nyse','nasdaq','amex','dowjones','sp500','nasdaq100']:
|
||||||
category_type = 'stocks-list'
|
category_type = 'stocks-list'
|
||||||
|
elif filter_list in ['dividend-kings','dividend-aristocrats']:
|
||||||
|
category_type = 'dividends'
|
||||||
else:
|
else:
|
||||||
category_type = 'market-cap'
|
category_type = 'market-cap'
|
||||||
try:
|
try:
|
||||||
@ -3941,7 +3888,6 @@ async def get_statistics(data: FilterStockList, api_key: str = Security(get_api_
|
|||||||
res = orjson.loads(file.read())
|
res = orjson.loads(file.read())
|
||||||
except:
|
except:
|
||||||
res = []
|
res = []
|
||||||
|
|
||||||
data = orjson.dumps(res)
|
data = orjson.dumps(res)
|
||||||
compressed_data = gzip.compress(data)
|
compressed_data = gzip.compress(data)
|
||||||
|
|
||||||
|
|||||||
@ -362,7 +362,7 @@ schedule.every().saturday.at("10:00").do(run_threaded, run_price_analysis).tag('
|
|||||||
|
|
||||||
|
|
||||||
schedule.every(30).minutes.do(run_threaded, run_dividend_list).tag('dividend_list_job')
|
schedule.every(30).minutes.do(run_threaded, run_dividend_list).tag('dividend_list_job')
|
||||||
schedule.every(60).minutes.do(run_threaded, run_congress_trading).tag('congress_job')
|
schedule.every(3).hours.do(run_threaded, run_congress_trading).tag('congress_job')
|
||||||
schedule.every(15).minutes.do(run_threaded, run_cron_market_news).tag('market_news_job')
|
schedule.every(15).minutes.do(run_threaded, run_cron_market_news).tag('market_news_job')
|
||||||
schedule.every(30).minutes.do(run_threaded, run_cron_industry).tag('industry_job')
|
schedule.every(30).minutes.do(run_threaded, run_cron_industry).tag('industry_job')
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user