mirror of
https://github.com/ChuckBuilds/LEDMatrix.git
synced 2026-04-11 13:23:00 +00:00
Improve SSL handling in stock managers. Add retry logic with exponential backoff, proper SSL verification, and better error handling for both StockManager and StockNewsManager classes.
This commit is contained in:
@@ -12,6 +12,8 @@ from PIL import Image, ImageDraw, ImageFont
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
import hashlib
|
import hashlib
|
||||||
from .cache_manager import CacheManager
|
from .cache_manager import CacheManager
|
||||||
|
from requests.adapters import HTTPAdapter
|
||||||
|
from urllib3.util.retry import Retry
|
||||||
|
|
||||||
# Configure logging
|
# Configure logging
|
||||||
logging.basicConfig(level=logging.INFO)
|
logging.basicConfig(level=logging.INFO)
|
||||||
@@ -58,6 +60,18 @@ class StockManager:
|
|||||||
self.headers = {
|
self.headers = {
|
||||||
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36'
|
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36'
|
||||||
}
|
}
|
||||||
|
|
||||||
|
# Set up session with retry logic
|
||||||
|
self.session = requests.Session()
|
||||||
|
retry_strategy = Retry(
|
||||||
|
total=3, # number of retries
|
||||||
|
backoff_factor=0.5, # wait 0.5, 1, 2 seconds between retries
|
||||||
|
status_forcelist=[500, 502, 503, 504], # HTTP status codes to retry on
|
||||||
|
)
|
||||||
|
adapter = HTTPAdapter(max_retries=retry_strategy)
|
||||||
|
self.session.mount("https://", adapter)
|
||||||
|
self.session.mount("http://", adapter)
|
||||||
|
|
||||||
# Initialize with first update
|
# Initialize with first update
|
||||||
self.update_stock_data()
|
self.update_stock_data()
|
||||||
|
|
||||||
@@ -142,7 +156,15 @@ class StockManager:
|
|||||||
'range': '1d' # 1 day of data
|
'range': '1d' # 1 day of data
|
||||||
}
|
}
|
||||||
|
|
||||||
response = requests.get(url, headers=self.headers, params=params, timeout=5)
|
# Use session with retry logic
|
||||||
|
response = self.session.get(
|
||||||
|
url,
|
||||||
|
headers=self.headers,
|
||||||
|
params=params,
|
||||||
|
timeout=10, # Increased timeout
|
||||||
|
verify=True # Enable SSL verification
|
||||||
|
)
|
||||||
|
|
||||||
if response.status_code != 200:
|
if response.status_code != 200:
|
||||||
logger.error(f"Failed to fetch data for {symbol}: HTTP {response.status_code}")
|
logger.error(f"Failed to fetch data for {symbol}: HTTP {response.status_code}")
|
||||||
return None
|
return None
|
||||||
@@ -199,6 +221,13 @@ class StockManager:
|
|||||||
|
|
||||||
return stock_data
|
return stock_data
|
||||||
|
|
||||||
|
except requests.exceptions.SSLError as e:
|
||||||
|
logger.error(f"SSL error fetching data for {symbol}: {e}")
|
||||||
|
# Try to use cached data as fallback
|
||||||
|
if cached_data and symbol in cached_data:
|
||||||
|
logger.info(f"Using cached data as fallback for {symbol} after SSL error")
|
||||||
|
return cached_data[symbol]
|
||||||
|
return None
|
||||||
except requests.exceptions.RequestException as e:
|
except requests.exceptions.RequestException as e:
|
||||||
logger.error(f"Network error fetching data for {symbol}: {e}")
|
logger.error(f"Network error fetching data for {symbol}: {e}")
|
||||||
# Try to use cached data as fallback
|
# Try to use cached data as fallback
|
||||||
|
|||||||
@@ -11,6 +11,8 @@ import re
|
|||||||
from src.config_manager import ConfigManager
|
from src.config_manager import ConfigManager
|
||||||
from PIL import Image, ImageDraw
|
from PIL import Image, ImageDraw
|
||||||
from .cache_manager import CacheManager
|
from .cache_manager import CacheManager
|
||||||
|
from requests.adapters import HTTPAdapter
|
||||||
|
from urllib3.util.retry import Retry
|
||||||
|
|
||||||
# Configure logging
|
# Configure logging
|
||||||
logging.basicConfig(level=logging.INFO)
|
logging.basicConfig(level=logging.INFO)
|
||||||
@@ -47,23 +49,48 @@ class StockNewsManager:
|
|||||||
self.headers = {
|
self.headers = {
|
||||||
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36'
|
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36'
|
||||||
}
|
}
|
||||||
|
|
||||||
|
# Set up session with retry logic
|
||||||
|
self.session = requests.Session()
|
||||||
|
retry_strategy = Retry(
|
||||||
|
total=3, # number of retries
|
||||||
|
backoff_factor=0.5, # wait 0.5, 1, 2 seconds between retries
|
||||||
|
status_forcelist=[500, 502, 503, 504], # HTTP status codes to retry on
|
||||||
|
)
|
||||||
|
adapter = HTTPAdapter(max_retries=retry_strategy)
|
||||||
|
self.session.mount("https://", adapter)
|
||||||
|
self.session.mount("http://", adapter)
|
||||||
|
|
||||||
# Initialize with first update
|
# Initialize with first update
|
||||||
self.update_news_data()
|
self.update_news_data()
|
||||||
|
|
||||||
def _fetch_news_for_symbol(self, symbol: str) -> List[Dict[str, Any]]:
|
def _fetch_news(self, symbol: str) -> List[Dict[str, Any]]:
|
||||||
"""Fetch news headlines for a stock symbol."""
|
"""Fetch news data for a stock from Yahoo Finance."""
|
||||||
# Try to get cached data first
|
|
||||||
cached_data = self.cache_manager.get_cached_data('stock_news')
|
|
||||||
if cached_data and symbol in cached_data:
|
|
||||||
logger.info(f"Using cached news data for {symbol}")
|
|
||||||
return cached_data[symbol]
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Using Yahoo Finance API to get news
|
# Use Yahoo Finance query1 API for news data
|
||||||
encoded_symbol = urllib.parse.quote(symbol)
|
url = f"https://query1.finance.yahoo.com/v1/finance/search"
|
||||||
url = f"https://query1.finance.yahoo.com/v1/finance/search?q={encoded_symbol}&lang=en-US®ion=US"esCount=0&newsCount={self.stock_news_config.get('max_headlines_per_symbol', 5)}"
|
params = {
|
||||||
|
'q': symbol,
|
||||||
|
'lang': 'en-US',
|
||||||
|
'region': 'US',
|
||||||
|
'quotesCount': 0,
|
||||||
|
'newsCount': 10,
|
||||||
|
'enableFuzzyQuery': False,
|
||||||
|
'quotesQueryId': 'tss_match_phrase_query',
|
||||||
|
'multiQuoteQueryId': 'multi_quote_single_token_query',
|
||||||
|
'newsQueryId': 'news_cie_vespa',
|
||||||
|
'enableCb': True,
|
||||||
|
}
|
||||||
|
|
||||||
|
# Use session with retry logic
|
||||||
|
response = self.session.get(
|
||||||
|
url,
|
||||||
|
headers=self.headers,
|
||||||
|
params=params,
|
||||||
|
timeout=10, # Increased timeout
|
||||||
|
verify=True # Enable SSL verification
|
||||||
|
)
|
||||||
|
|
||||||
response = requests.get(url, headers=self.headers, timeout=5)
|
|
||||||
if response.status_code != 200:
|
if response.status_code != 200:
|
||||||
logger.error(f"Failed to fetch news for {symbol}: HTTP {response.status_code}")
|
logger.error(f"Failed to fetch news for {symbol}: HTTP {response.status_code}")
|
||||||
return []
|
return []
|
||||||
@@ -71,34 +98,30 @@ class StockNewsManager:
|
|||||||
data = response.json()
|
data = response.json()
|
||||||
news_items = data.get('news', [])
|
news_items = data.get('news', [])
|
||||||
|
|
||||||
# Process and format news items
|
processed_news = []
|
||||||
formatted_news = []
|
|
||||||
for item in news_items:
|
for item in news_items:
|
||||||
formatted_news.append({
|
try:
|
||||||
"title": item.get('title', ''),
|
processed_news.append({
|
||||||
"publisher": item.get('publisher', ''),
|
'title': item.get('title', ''),
|
||||||
"link": item.get('link', ''),
|
'link': item.get('link', ''),
|
||||||
"published": item.get('providerPublishTime', 0)
|
'publisher': item.get('publisher', ''),
|
||||||
})
|
'published': datetime.fromtimestamp(item.get('providerPublishTime', 0)),
|
||||||
|
'summary': item.get('summary', '')
|
||||||
logger.info(f"Fetched {len(formatted_news)} news items for {symbol}")
|
})
|
||||||
|
except (ValueError, TypeError) as e:
|
||||||
# Cache the new data
|
logger.error(f"Error processing news item for {symbol}: {e}")
|
||||||
if cached_data is None:
|
continue
|
||||||
cached_data = {}
|
|
||||||
cached_data[symbol] = formatted_news
|
|
||||||
self.cache_manager.update_cache('stock_news', cached_data)
|
|
||||||
|
|
||||||
return formatted_news
|
logger.debug(f"Fetched {len(processed_news)} news items for {symbol}")
|
||||||
|
return processed_news
|
||||||
|
|
||||||
|
except requests.exceptions.SSLError as e:
|
||||||
|
logger.error(f"SSL error fetching news for {symbol}: {e}")
|
||||||
|
return []
|
||||||
except requests.exceptions.RequestException as e:
|
except requests.exceptions.RequestException as e:
|
||||||
logger.error(f"Network error fetching news for {symbol}: {e}")
|
logger.error(f"Network error fetching news for {symbol}: {e}")
|
||||||
# Try to use cached data as fallback
|
|
||||||
if cached_data and symbol in cached_data:
|
|
||||||
logger.info(f"Using cached news data as fallback for {symbol}")
|
|
||||||
return cached_data[symbol]
|
|
||||||
return []
|
return []
|
||||||
except (ValueError, IndexError, KeyError) as e:
|
except (ValueError, KeyError) as e:
|
||||||
logger.error(f"Error parsing news data for {symbol}: {e}")
|
logger.error(f"Error parsing news data for {symbol}: {e}")
|
||||||
return []
|
return []
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
@@ -136,7 +159,7 @@ class StockNewsManager:
|
|||||||
|
|
||||||
# Add a small delay between requests to avoid rate limiting
|
# Add a small delay between requests to avoid rate limiting
|
||||||
time.sleep(random.uniform(0.1, 0.3))
|
time.sleep(random.uniform(0.1, 0.3))
|
||||||
news_items = self._fetch_news_for_symbol(symbol)
|
news_items = self._fetch_news(symbol)
|
||||||
if news_items:
|
if news_items:
|
||||||
new_data[symbol] = news_items
|
new_data[symbol] = news_items
|
||||||
success = True
|
success = True
|
||||||
|
|||||||
Reference in New Issue
Block a user