mirror of
https://github.com/ChuckBuilds/LEDMatrix.git
synced 2026-04-11 13:23:00 +00:00
found random timezone error from old time import method
This commit is contained in:
@@ -161,7 +161,7 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nfl_scoreboard": {
|
"nfl_scoreboard": {
|
||||||
"enabled": true,
|
"enabled": false,
|
||||||
"show_odds": false,
|
"show_odds": false,
|
||||||
"test_mode": false,
|
"test_mode": false,
|
||||||
"update_interval_seconds": 3600,
|
"update_interval_seconds": 3600,
|
||||||
@@ -179,7 +179,7 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"ncaa_fb_scoreboard": {
|
"ncaa_fb_scoreboard": {
|
||||||
"enabled": true,
|
"enabled": false,
|
||||||
"show_odds": false,
|
"show_odds": false,
|
||||||
"test_mode": false,
|
"test_mode": false,
|
||||||
"update_interval_seconds": 3600,
|
"update_interval_seconds": 3600,
|
||||||
|
|||||||
@@ -69,6 +69,12 @@ class BaseNBAManager:
|
|||||||
self.logger.info(f"Initialized NBA manager with display dimensions: {self.display_width}x{self.display_height}")
|
self.logger.info(f"Initialized NBA manager with display dimensions: {self.display_width}x{self.display_height}")
|
||||||
self.logger.info(f"Logo directory: {self.logo_dir}")
|
self.logger.info(f"Logo directory: {self.logo_dir}")
|
||||||
|
|
||||||
|
def _get_timezone(self):
|
||||||
|
try:
|
||||||
|
return pytz.timezone(self.config_manager.get_timezone())
|
||||||
|
except pytz.UnknownTimeZoneError:
|
||||||
|
return pytz.utc
|
||||||
|
|
||||||
def _should_log(self, message_type: str, cooldown: int = 300) -> bool:
|
def _should_log(self, message_type: str, cooldown: int = 300) -> bool:
|
||||||
"""Check if a message should be logged based on cooldown period."""
|
"""Check if a message should be logged based on cooldown period."""
|
||||||
current_time = time.time()
|
current_time = time.time()
|
||||||
|
|||||||
@@ -162,21 +162,20 @@ class BaseNCAAFBManager: # Renamed class
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
self.logger.error(f"Error fetching odds for game {game.get('id', 'N/A')}: {e}")
|
self.logger.error(f"Error fetching odds for game {game.get('id', 'N/A')}: {e}")
|
||||||
|
|
||||||
@classmethod
|
def _fetch_shared_data(self, past_days: int, future_days: int, date_str: str = None) -> Optional[Dict]:
|
||||||
def _fetch_shared_data(cls, past_days: int, future_days: int, date_str: str = None) -> Optional[Dict]:
|
|
||||||
"""Fetch and cache data for all managers to share."""
|
"""Fetch and cache data for all managers to share."""
|
||||||
current_time = time.time()
|
current_time = time.time()
|
||||||
|
|
||||||
if cls._shared_data and (current_time - cls._last_shared_update) < 300:
|
if BaseNCAAFBManager._shared_data and (current_time - BaseNCAAFBManager._last_shared_update) < 300:
|
||||||
return cls._shared_data
|
return BaseNCAAFBManager._shared_data
|
||||||
|
|
||||||
try:
|
try:
|
||||||
cache_key = date_str if date_str else 'today_ncaafb' # Changed cache key prefix
|
cache_key = date_str if date_str else 'today_ncaafb' # Changed cache key prefix
|
||||||
cached_data = cls.cache_manager.get(cache_key, max_age=300)
|
cached_data = BaseNCAAFBManager.cache_manager.get(cache_key, max_age=300)
|
||||||
if cached_data:
|
if cached_data:
|
||||||
cls.logger.info(f"[NCAAFB] Using cached data for {cache_key}")
|
BaseNCAAFBManager.logger.info(f"[NCAAFB] Using cached data for {cache_key}")
|
||||||
cls._shared_data = cached_data
|
BaseNCAAFBManager._shared_data = cached_data
|
||||||
cls._last_shared_update = current_time
|
BaseNCAAFBManager._last_shared_update = current_time
|
||||||
return cached_data
|
return cached_data
|
||||||
|
|
||||||
url = ESPN_NCAAFB_SCOREBOARD_URL # Use NCAA FB URL
|
url = ESPN_NCAAFB_SCOREBOARD_URL # Use NCAA FB URL
|
||||||
@@ -187,21 +186,21 @@ class BaseNCAAFBManager: # Renamed class
|
|||||||
response = requests.get(url, params=params)
|
response = requests.get(url, params=params)
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
data = response.json()
|
data = response.json()
|
||||||
cls.logger.info(f"[NCAAFB] Successfully fetched data from ESPN API")
|
BaseNCAAFBManager.logger.info(f"[NCAAFB] Successfully fetched data from ESPN API")
|
||||||
|
|
||||||
cls.cache_manager.set(cache_key, data)
|
BaseNCAAFBManager.cache_manager.set(cache_key, data)
|
||||||
cls._shared_data = data
|
BaseNCAAFBManager._shared_data = data
|
||||||
cls._last_shared_update = current_time
|
BaseNCAAFBManager._last_shared_update = current_time
|
||||||
|
|
||||||
if not date_str:
|
if not date_str:
|
||||||
today = datetime.now(cls._get_timezone()).date()
|
today = datetime.now(self._get_timezone()).date()
|
||||||
dates_to_fetch = []
|
dates_to_fetch = []
|
||||||
# Generate dates from past_days ago to future_days ahead
|
# Generate dates from past_days ago to future_days ahead
|
||||||
for i in range(-past_days, future_days + 1):
|
for i in range(-past_days, future_days + 1):
|
||||||
fetch_dt = today + timedelta(days=i)
|
fetch_dt = today + timedelta(days=i)
|
||||||
dates_to_fetch.append(fetch_dt.strftime('%Y%m%d'))
|
dates_to_fetch.append(fetch_dt.strftime('%Y%m%d'))
|
||||||
|
|
||||||
cls.logger.info(f"[NCAAFB] Fetching data for dates: {dates_to_fetch}")
|
BaseNCAAFBManager.logger.info(f"[NCAAFB] Fetching data for dates: {dates_to_fetch}")
|
||||||
|
|
||||||
all_events = []
|
all_events = []
|
||||||
# Fetch data for each date (excluding today if already fetched)
|
# Fetch data for each date (excluding today if already fetched)
|
||||||
@@ -210,9 +209,9 @@ class BaseNCAAFBManager: # Renamed class
|
|||||||
continue
|
continue
|
||||||
|
|
||||||
date_cache_key = f"{fetch_date}_ncaafb" # Changed cache key suffix
|
date_cache_key = f"{fetch_date}_ncaafb" # Changed cache key suffix
|
||||||
cached_date_data = cls.cache_manager.get(date_cache_key, max_age=300)
|
cached_date_data = BaseNCAAFBManager.cache_manager.get(date_cache_key, max_age=300)
|
||||||
if cached_date_data:
|
if cached_date_data:
|
||||||
cls.logger.info(f"[NCAAFB] Using cached data for date {fetch_date}")
|
BaseNCAAFBManager.logger.info(f"[NCAAFB] Using cached data for date {fetch_date}")
|
||||||
if "events" in cached_date_data:
|
if "events" in cached_date_data:
|
||||||
all_events.extend(cached_date_data["events"])
|
all_events.extend(cached_date_data["events"])
|
||||||
continue
|
continue
|
||||||
@@ -223,19 +222,19 @@ class BaseNCAAFBManager: # Renamed class
|
|||||||
date_data = response.json()
|
date_data = response.json()
|
||||||
if date_data and "events" in date_data:
|
if date_data and "events" in date_data:
|
||||||
all_events.extend(date_data["events"])
|
all_events.extend(date_data["events"])
|
||||||
cls.logger.info(f"[NCAAFB] Fetched {len(date_data['events'])} events for date {fetch_date}")
|
BaseNCAAFBManager.logger.info(f"[NCAAFB] Fetched {len(date_data['events'])} events for date {fetch_date}")
|
||||||
cls.cache_manager.set(date_cache_key, date_data)
|
BaseNCAAFBManager.cache_manager.set(date_cache_key, date_data)
|
||||||
|
|
||||||
if all_events:
|
if all_events:
|
||||||
if "events" not in data: data["events"] = [] # Ensure 'events' key exists
|
if "events" not in data: data["events"] = [] # Ensure 'events' key exists
|
||||||
data["events"].extend(all_events)
|
data["events"].extend(all_events)
|
||||||
cls.logger.info(f"[NCAAFB] Combined {len(data['events'])} total events from all dates")
|
BaseNCAAFBManager.logger.info(f"[NCAAFB] Combined {len(data['events'])} total events from all dates")
|
||||||
cls._shared_data = data
|
BaseNCAAFBManager._shared_data = data
|
||||||
cls._last_shared_update = current_time
|
BaseNCAAFBManager._last_shared_update = current_time
|
||||||
|
|
||||||
return data
|
return data
|
||||||
except requests.exceptions.RequestException as e:
|
except requests.exceptions.RequestException as e:
|
||||||
cls.logger.error(f"[NCAAFB] Error fetching data from ESPN: {e}")
|
BaseNCAAFBManager.logger.error(f"[NCAAFB] Error fetching data from ESPN: {e}")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def _fetch_data(self, date_str: str = None) -> Optional[Dict]:
|
def _fetch_data(self, date_str: str = None) -> Optional[Dict]:
|
||||||
|
|||||||
@@ -162,21 +162,20 @@ class BaseNFLManager: # Renamed class
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
self.logger.error(f"Error fetching odds for game {game.get('id', 'N/A')}: {e}")
|
self.logger.error(f"Error fetching odds for game {game.get('id', 'N/A')}: {e}")
|
||||||
|
|
||||||
@classmethod
|
def _fetch_shared_data(self, past_days: int, future_days: int, date_str: str = None) -> Optional[Dict]:
|
||||||
def _fetch_shared_data(cls, past_days: int, future_days: int, date_str: str = None) -> Optional[Dict]:
|
|
||||||
"""Fetch and cache data for all managers to share."""
|
"""Fetch and cache data for all managers to share."""
|
||||||
current_time = time.time()
|
current_time = time.time()
|
||||||
|
|
||||||
if cls._shared_data and (current_time - cls._last_shared_update) < 300:
|
if BaseNFLManager._shared_data and (current_time - BaseNFLManager._last_shared_update) < 300:
|
||||||
return cls._shared_data
|
return BaseNFLManager._shared_data
|
||||||
|
|
||||||
try:
|
try:
|
||||||
cache_key = date_str if date_str else 'today_nfl' # Changed cache key prefix
|
cache_key = date_str if date_str else 'today_nfl' # Changed cache key prefix
|
||||||
cached_data = cls.cache_manager.get(cache_key, max_age=300)
|
cached_data = BaseNFLManager.cache_manager.get(cache_key, max_age=300)
|
||||||
if cached_data:
|
if cached_data:
|
||||||
cls.logger.info(f"[NFL] Using cached data for {cache_key}")
|
BaseNFLManager.logger.info(f"[NFL] Using cached data for {cache_key}")
|
||||||
cls._shared_data = cached_data
|
BaseNFLManager._shared_data = cached_data
|
||||||
cls._last_shared_update = current_time
|
BaseNFLManager._last_shared_update = current_time
|
||||||
return cached_data
|
return cached_data
|
||||||
|
|
||||||
url = ESPN_NFL_SCOREBOARD_URL # Use NFL URL
|
url = ESPN_NFL_SCOREBOARD_URL # Use NFL URL
|
||||||
@@ -187,21 +186,21 @@ class BaseNFLManager: # Renamed class
|
|||||||
response = requests.get(url, params=params)
|
response = requests.get(url, params=params)
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
data = response.json()
|
data = response.json()
|
||||||
cls.logger.info(f"[NFL] Successfully fetched data from ESPN API")
|
BaseNFLManager.logger.info(f"[NFL] Successfully fetched data from ESPN API")
|
||||||
|
|
||||||
cls.cache_manager.set(cache_key, data)
|
BaseNFLManager.cache_manager.set(cache_key, data)
|
||||||
cls._shared_data = data
|
BaseNFLManager._shared_data = data
|
||||||
cls._last_shared_update = current_time
|
BaseNFLManager._last_shared_update = current_time
|
||||||
|
|
||||||
if not date_str:
|
if not date_str:
|
||||||
today = datetime.now(cls._get_timezone()).date()
|
today = datetime.now(self._get_timezone()).date()
|
||||||
dates_to_fetch = []
|
dates_to_fetch = []
|
||||||
# Generate dates from past_days ago to future_days ahead
|
# Generate dates from past_days ago to future_days ahead
|
||||||
for i in range(-past_days, future_days + 1):
|
for i in range(-past_days, future_days + 1):
|
||||||
fetch_dt = today + timedelta(days=i)
|
fetch_dt = today + timedelta(days=i)
|
||||||
dates_to_fetch.append(fetch_dt.strftime('%Y%m%d'))
|
dates_to_fetch.append(fetch_dt.strftime('%Y%m%d'))
|
||||||
|
|
||||||
cls.logger.info(f"[NFL] Fetching data for dates: {dates_to_fetch}")
|
BaseNFLManager.logger.info(f"[NFL] Fetching data for dates: {dates_to_fetch}")
|
||||||
|
|
||||||
all_events = []
|
all_events = []
|
||||||
# Fetch data for each date (excluding today if already fetched)
|
# Fetch data for each date (excluding today if already fetched)
|
||||||
@@ -210,9 +209,9 @@ class BaseNFLManager: # Renamed class
|
|||||||
continue
|
continue
|
||||||
|
|
||||||
date_cache_key = f"{fetch_date}_nfl"
|
date_cache_key = f"{fetch_date}_nfl"
|
||||||
cached_date_data = cls.cache_manager.get(date_cache_key, max_age=300)
|
cached_date_data = BaseNFLManager.cache_manager.get(date_cache_key, max_age=300)
|
||||||
if cached_date_data:
|
if cached_date_data:
|
||||||
cls.logger.info(f"[NFL] Using cached data for date {fetch_date}")
|
BaseNFLManager.logger.info(f"[NFL] Using cached data for date {fetch_date}")
|
||||||
if "events" in cached_date_data:
|
if "events" in cached_date_data:
|
||||||
all_events.extend(cached_date_data["events"])
|
all_events.extend(cached_date_data["events"])
|
||||||
continue
|
continue
|
||||||
@@ -223,19 +222,19 @@ class BaseNFLManager: # Renamed class
|
|||||||
date_data = response.json()
|
date_data = response.json()
|
||||||
if date_data and "events" in date_data:
|
if date_data and "events" in date_data:
|
||||||
all_events.extend(date_data["events"])
|
all_events.extend(date_data["events"])
|
||||||
cls.logger.info(f"[NFL] Fetched {len(date_data['events'])} events for date {fetch_date}")
|
BaseNFLManager.logger.info(f"[NFL] Fetched {len(date_data['events'])} events for date {fetch_date}")
|
||||||
cls.cache_manager.set(date_cache_key, date_data)
|
BaseNFLManager.cache_manager.set(date_cache_key, date_data)
|
||||||
|
|
||||||
if all_events:
|
if all_events:
|
||||||
if "events" not in data: data["events"] = [] # Ensure 'events' key exists
|
if "events" not in data: data["events"] = [] # Ensure 'events' key exists
|
||||||
data["events"].extend(all_events)
|
data["events"].extend(all_events)
|
||||||
cls.logger.info(f"[NFL] Combined {len(data['events'])} total events from all dates")
|
BaseNFLManager.logger.info(f"[NFL] Combined {len(data['events'])} total events from all dates")
|
||||||
cls._shared_data = data
|
BaseNFLManager._shared_data = data
|
||||||
cls._last_shared_update = current_time
|
BaseNFLManager._last_shared_update = current_time
|
||||||
|
|
||||||
return data
|
return data
|
||||||
except requests.exceptions.RequestException as e:
|
except requests.exceptions.RequestException as e:
|
||||||
cls.logger.error(f"[NFL] Error fetching data from ESPN: {e}")
|
BaseNFLManager.logger.error(f"[NFL] Error fetching data from ESPN: {e}")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def _fetch_data(self, date_str: str = None) -> Optional[Dict]:
|
def _fetch_data(self, date_str: str = None) -> Optional[Dict]:
|
||||||
|
|||||||
@@ -140,48 +140,47 @@ class BaseNHLManager:
|
|||||||
except pytz.UnknownTimeZoneError:
|
except pytz.UnknownTimeZoneError:
|
||||||
return pytz.utc
|
return pytz.utc
|
||||||
|
|
||||||
@classmethod
|
def _fetch_shared_data(self, date_str: str = None) -> Optional[Dict]:
|
||||||
def _fetch_shared_data(cls, date_str: str = None) -> Optional[Dict]:
|
|
||||||
"""Fetch and cache data for all managers to share."""
|
"""Fetch and cache data for all managers to share."""
|
||||||
current_time = time.time()
|
current_time = time.time()
|
||||||
|
|
||||||
# If we have recent data, use it
|
# If we have recent data, use it
|
||||||
if cls._shared_data and (current_time - cls._last_shared_update) < 300: # 5 minutes
|
if BaseNHLManager._shared_data and (current_time - BaseNHLManager._last_shared_update) < 300: # 5 minutes
|
||||||
return cls._shared_data
|
return BaseNHLManager._shared_data
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Check cache first
|
# Check cache first
|
||||||
cache_key = date_str if date_str else 'today'
|
cache_key = date_str if date_str else 'today'
|
||||||
cached_data = cls.cache_manager.get(cache_key, max_age=300) # 5 minutes cache
|
cached_data = BaseNHLManager.cache_manager.get(cache_key, max_age=300) # 5 minutes cache
|
||||||
if cached_data:
|
if cached_data:
|
||||||
cls.logger.info(f"[NHL] Using cached data for {cache_key}")
|
BaseNHLManager.logger.info(f"[NHL] Using cached data for {cache_key}")
|
||||||
cls._shared_data = cached_data
|
BaseNHLManager._shared_data = cached_data
|
||||||
cls._last_shared_update = current_time
|
BaseNHLManager._last_shared_update = current_time
|
||||||
return cached_data
|
return cached_data
|
||||||
|
|
||||||
# If not in cache or stale, fetch from API
|
# If not in cache or stale, fetch from API
|
||||||
if not date_str:
|
if not date_str:
|
||||||
# Get today's date in YYYY-MM-DD format
|
# Get today's date in YYYY-MM-DD format
|
||||||
today = datetime.now(cls._get_timezone()).date()
|
today = datetime.now(self._get_timezone()).date()
|
||||||
date_str = today.strftime('%Y-%m-%d')
|
date_str = today.strftime('%Y-%m-%d')
|
||||||
|
|
||||||
url = f"{NHL_API_BASE_URL}{date_str}"
|
url = f"{NHL_API_BASE_URL}{date_str}"
|
||||||
cls.logger.info(f"Fetching data from URL: {url}")
|
BaseNHLManager.logger.info(f"Fetching data from URL: {url}")
|
||||||
|
|
||||||
response = requests.get(url)
|
response = requests.get(url)
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
data = response.json()
|
data = response.json()
|
||||||
cls.logger.info(f"[NHL] Successfully fetched data from NHL API")
|
BaseNHLManager.logger.info(f"[NHL] Successfully fetched data from NHL API")
|
||||||
|
|
||||||
# Cache the response
|
# Cache the response
|
||||||
cls.cache_manager.set(cache_key, data)
|
BaseNHLManager.cache_manager.set(cache_key, data)
|
||||||
cls._shared_data = data
|
BaseNHLManager._shared_data = data
|
||||||
cls._last_shared_update = current_time
|
BaseNHLManager._last_shared_update = current_time
|
||||||
|
|
||||||
# If no date specified, fetch data from multiple days
|
# If no date specified, fetch data from multiple days
|
||||||
if not date_str:
|
if not date_str:
|
||||||
# Get today's date in YYYYMMDD format
|
# Get today's date in YYYYMMDD format
|
||||||
today = datetime.now(cls._get_timezone()).date()
|
today = datetime.now(self._get_timezone()).date()
|
||||||
dates_to_fetch = [
|
dates_to_fetch = [
|
||||||
(today - timedelta(days=2)).strftime('%Y-%m-%d'),
|
(today - timedelta(days=2)).strftime('%Y-%m-%d'),
|
||||||
(today - timedelta(days=1)).strftime('%Y-%m-%d'),
|
(today - timedelta(days=1)).strftime('%Y-%m-%d'),
|
||||||
@@ -193,9 +192,9 @@ class BaseNHLManager:
|
|||||||
for fetch_date in dates_to_fetch:
|
for fetch_date in dates_to_fetch:
|
||||||
if fetch_date != today.strftime('%Y-%m-%d'): # Skip today as we already have it
|
if fetch_date != today.strftime('%Y-%m-%d'): # Skip today as we already have it
|
||||||
# Check cache for this date
|
# Check cache for this date
|
||||||
cached_date_data = cls.cache_manager.get(fetch_date, max_age=300)
|
cached_date_data = BaseNHLManager.cache_manager.get(fetch_date, max_age=300)
|
||||||
if cached_date_data:
|
if cached_date_data:
|
||||||
cls.logger.info(f"[NHL] Using cached data for date {fetch_date}")
|
BaseNHLManager.logger.info(f"[NHL] Using cached data for date {fetch_date}")
|
||||||
if "events" in cached_date_data:
|
if "events" in cached_date_data:
|
||||||
all_events.extend(cached_date_data["events"])
|
all_events.extend(cached_date_data["events"])
|
||||||
continue
|
continue
|
||||||
@@ -206,20 +205,20 @@ class BaseNHLManager:
|
|||||||
date_data = response.json()
|
date_data = response.json()
|
||||||
if date_data and "events" in date_data:
|
if date_data and "events" in date_data:
|
||||||
all_events.extend(date_data["events"])
|
all_events.extend(date_data["events"])
|
||||||
cls.logger.info(f"[NHL] Fetched {len(date_data['events'])} events for date {fetch_date}")
|
BaseNHLManager.logger.info(f"[NHL] Fetched {len(date_data['events'])} events for date {fetch_date}")
|
||||||
# Cache the response
|
# Cache the response
|
||||||
cls.cache_manager.set(fetch_date, date_data)
|
BaseNHLManager.cache_manager.set(fetch_date, date_data)
|
||||||
|
|
||||||
# Combine events from all dates
|
# Combine events from all dates
|
||||||
if all_events:
|
if all_events:
|
||||||
data["events"].extend(all_events)
|
data["events"].extend(all_events)
|
||||||
cls.logger.info(f"[NHL] Combined {len(data['events'])} total events from all dates")
|
BaseNHLManager.logger.info(f"[NHL] Combined {len(data['events'])} total events from all dates")
|
||||||
cls._shared_data = data
|
BaseNHLManager._shared_data = data
|
||||||
cls._last_shared_update = current_time
|
BaseNHLManager._last_shared_update = current_time
|
||||||
|
|
||||||
return data
|
return data
|
||||||
except requests.exceptions.RequestException as e:
|
except requests.exceptions.RequestException as e:
|
||||||
cls.logger.error(f"[NHL] Error fetching data from NHL: {e}")
|
BaseNHLManager.logger.error(f"[NHL] Error fetching data from NHL: {e}")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def _fetch_data(self, date_str: str = None) -> Optional[Dict]:
|
def _fetch_data(self, date_str: str = None) -> Optional[Dict]:
|
||||||
|
|||||||
Reference in New Issue
Block a user