improved logic for checking last and next X games, will check up to one year

This commit is contained in:
Chuck
2025-07-21 17:18:25 -05:00
parent e9cf2829da
commit 8feddb3a82
2 changed files with 172 additions and 136 deletions

View File

@@ -179,85 +179,103 @@ class BaseNCAAFBManager: # Renamed class
BaseNCAAFBManager._last_shared_update = current_time
return cached_data
# For game-based fetching, we need to fetch a wider window to ensure we get enough games
# Use a 30-day window to ensure we capture enough games
# Smart game-based fetching: fetch incrementally until we have enough games
today = datetime.now(self._get_timezone()).date()
dates_to_fetch = []
# Generate dates for a 30-day window (15 days past, 15 days future)
for i in range(-15, 16):
fetch_dt = today + timedelta(days=i)
dates_to_fetch.append(fetch_dt.strftime('%Y%m%d'))
BaseNCAAFBManager.logger.info(f"[NCAAFB] Fetching data for 30-day window to find {fetch_past_games} past and {fetch_future_games} future games")
all_events = []
# Fetch data for each date
for fetch_date in dates_to_fetch:
date_cache_key = f"{fetch_date}_ncaafb"
cached_date_data = BaseNCAAFBManager.cache_manager.get(date_cache_key, max_age=300)
past_events = []
future_events = []
# Start with today and expand outward until we have enough games
days_to_check = 0
max_days_to_check = 365 # Limit to 1 year to prevent infinite loops
while (len(past_events) < fetch_past_games or len(future_events) < fetch_future_games) and days_to_check <= max_days_to_check:
# Check dates in both directions
dates_to_check = []
if cached_date_data:
BaseNCAAFBManager.logger.info(f"[NCAAFB] Using cached data for date {fetch_date}")
if "events" in cached_date_data:
all_events.extend(cached_date_data["events"])
continue
url = ESPN_NCAAFB_SCOREBOARD_URL
params = {'dates': fetch_date}
# Check past dates
if len(past_events) < fetch_past_games:
for i in range(1, days_to_check + 1):
past_date = today - timedelta(days=i)
dates_to_check.append(past_date.strftime('%Y%m%d'))
response = requests.get(url, params=params)
response.raise_for_status()
date_data = response.json()
# Check future dates
if len(future_events) < fetch_future_games:
for i in range(1, days_to_check + 1):
future_date = today + timedelta(days=i)
dates_to_check.append(future_date.strftime('%Y%m%d'))
if date_data and "events" in date_data:
all_events.extend(date_data["events"])
BaseNCAAFBManager.logger.info(f"[NCAAFB] Fetched {len(date_data['events'])} events for date {fetch_date}")
BaseNCAAFBManager.cache_manager.set(date_cache_key, date_data)
# Now filter events based on game counts
if all_events:
# Sort events by date
all_events.sort(key=lambda x: x.get('date', ''))
# Also check today if we haven't already
if days_to_check == 0:
dates_to_check.append(today.strftime('%Y%m%d'))
# Separate past and future events
now = datetime.now(self._get_timezone())
past_events = []
future_events = []
BaseNCAAFBManager.logger.info(f"[NCAAFB] Checking {len(dates_to_check)} dates (day {days_to_check}) to find {fetch_past_games} past and {fetch_future_games} future games")
for event in all_events:
try:
event_time = datetime.fromisoformat(event['date'].replace('Z', '+00:00'))
if event_time.tzinfo is None:
event_time = event_time.replace(tzinfo=pytz.UTC)
event_time = event_time.astimezone(self._get_timezone())
if event_time < now:
past_events.append(event)
else:
future_events.append(event)
except Exception as e:
BaseNCAAFBManager.logger.warning(f"[NCAAFB] Could not parse event date: {e}")
# Fetch data for each date
for fetch_date in dates_to_check:
date_cache_key = f"{fetch_date}_ncaafb"
cached_date_data = BaseNCAAFBManager.cache_manager.get(date_cache_key, max_age=300)
if cached_date_data:
BaseNCAAFBManager.logger.info(f"[NCAAFB] Using cached data for date {fetch_date}")
if "events" in cached_date_data:
all_events.extend(cached_date_data["events"])
continue
# Take the specified number of games
selected_past_events = past_events[-fetch_past_games:] if past_events else []
selected_future_events = future_events[:fetch_future_games] if future_events else []
url = ESPN_NCAAFB_SCOREBOARD_URL
params = {'dates': fetch_date}
response = requests.get(url, params=params)
response.raise_for_status()
date_data = response.json()
if date_data and "events" in date_data:
all_events.extend(date_data["events"])
BaseNCAAFBManager.logger.info(f"[NCAAFB] Fetched {len(date_data['events'])} events for date {fetch_date}")
BaseNCAAFBManager.cache_manager.set(date_cache_key, date_data)
# Combine selected events
selected_events = selected_past_events + selected_future_events
# Process newly fetched events
if all_events:
# Sort events by date
all_events.sort(key=lambda x: x.get('date', ''))
# Separate past and future events
now = datetime.now(self._get_timezone())
past_events = []
future_events = []
for event in all_events:
try:
event_time = datetime.fromisoformat(event['date'].replace('Z', '+00:00'))
if event_time.tzinfo is None:
event_time = event_time.replace(tzinfo=pytz.UTC)
event_time = event_time.astimezone(self._get_timezone())
if event_time < now:
past_events.append(event)
else:
future_events.append(event)
except Exception as e:
BaseNCAAFBManager.logger.warning(f"[NCAAFB] Could not parse event date: {e}")
continue
BaseNCAAFBManager.logger.info(f"[NCAAFB] Selected {len(selected_past_events)} past games and {len(selected_future_events)} future games")
# Create the final data structure
data = {"events": selected_events}
BaseNCAAFBManager._shared_data = data
BaseNCAAFBManager._last_shared_update = current_time
return data
else:
BaseNCAAFBManager.logger.warning("[NCAAFB] No events found in the fetched data")
return None
days_to_check += 1
# Take the specified number of games
selected_past_events = past_events[-fetch_past_games:] if past_events else []
selected_future_events = future_events[:fetch_future_games] if future_events else []
# Combine selected events
selected_events = selected_past_events + selected_future_events
BaseNCAAFBManager.logger.info(f"[NCAAFB] Selected {len(selected_past_events)} past games and {len(selected_future_events)} future games after checking {days_to_check} days")
# Create the final data structure
data = {"events": selected_events}
BaseNCAAFBManager._shared_data = data
BaseNCAAFBManager._last_shared_update = current_time
return data
except requests.exceptions.RequestException as e:
BaseNCAAFBManager.logger.error(f"[NCAAFB] Error fetching data from ESPN: {e}")

View File

@@ -179,85 +179,103 @@ class BaseNFLManager: # Renamed class
BaseNFLManager._last_shared_update = current_time
return cached_data
# For game-based fetching, we need to fetch a wider window to ensure we get enough games
# Use a 30-day window to ensure we capture enough games
# Smart game-based fetching: fetch incrementally until we have enough games
today = datetime.now(self._get_timezone()).date()
dates_to_fetch = []
# Generate dates for a 30-day window (15 days past, 15 days future)
for i in range(-15, 16):
fetch_dt = today + timedelta(days=i)
dates_to_fetch.append(fetch_dt.strftime('%Y%m%d'))
BaseNFLManager.logger.info(f"[NFL] Fetching data for 30-day window to find {fetch_past_games} past and {fetch_future_games} future games")
all_events = []
# Fetch data for each date
for fetch_date in dates_to_fetch:
date_cache_key = f"{fetch_date}_nfl"
cached_date_data = BaseNFLManager.cache_manager.get(date_cache_key, max_age=300)
past_events = []
future_events = []
# Start with today and expand outward until we have enough games
days_to_check = 0
max_days_to_check = 365 # Limit to 1 year to prevent infinite loops
while (len(past_events) < fetch_past_games or len(future_events) < fetch_future_games) and days_to_check <= max_days_to_check:
# Check dates in both directions
dates_to_check = []
if cached_date_data:
BaseNFLManager.logger.info(f"[NFL] Using cached data for date {fetch_date}")
if "events" in cached_date_data:
all_events.extend(cached_date_data["events"])
continue
url = ESPN_NFL_SCOREBOARD_URL
params = {'dates': fetch_date}
# Check past dates
if len(past_events) < fetch_past_games:
for i in range(1, days_to_check + 1):
past_date = today - timedelta(days=i)
dates_to_check.append(past_date.strftime('%Y%m%d'))
response = requests.get(url, params=params)
response.raise_for_status()
date_data = response.json()
# Check future dates
if len(future_events) < fetch_future_games:
for i in range(1, days_to_check + 1):
future_date = today + timedelta(days=i)
dates_to_check.append(future_date.strftime('%Y%m%d'))
if date_data and "events" in date_data:
all_events.extend(date_data["events"])
BaseNFLManager.logger.info(f"[NFL] Fetched {len(date_data['events'])} events for date {fetch_date}")
BaseNFLManager.cache_manager.set(date_cache_key, date_data)
# Now filter events based on game counts
if all_events:
# Sort events by date
all_events.sort(key=lambda x: x.get('date', ''))
# Also check today if we haven't already
if days_to_check == 0:
dates_to_check.append(today.strftime('%Y%m%d'))
# Separate past and future events
now = datetime.now(self._get_timezone())
past_events = []
future_events = []
BaseNFLManager.logger.info(f"[NFL] Checking {len(dates_to_check)} dates (day {days_to_check}) to find {fetch_past_games} past and {fetch_future_games} future games")
for event in all_events:
try:
event_time = datetime.fromisoformat(event['date'].replace('Z', '+00:00'))
if event_time.tzinfo is None:
event_time = event_time.replace(tzinfo=pytz.UTC)
event_time = event_time.astimezone(self._get_timezone())
if event_time < now:
past_events.append(event)
else:
future_events.append(event)
except Exception as e:
BaseNFLManager.logger.warning(f"[NFL] Could not parse event date: {e}")
# Fetch data for each date
for fetch_date in dates_to_check:
date_cache_key = f"{fetch_date}_nfl"
cached_date_data = BaseNFLManager.cache_manager.get(date_cache_key, max_age=300)
if cached_date_data:
BaseNFLManager.logger.info(f"[NFL] Using cached data for date {fetch_date}")
if "events" in cached_date_data:
all_events.extend(cached_date_data["events"])
continue
# Take the specified number of games
selected_past_events = past_events[-fetch_past_games:] if past_events else []
selected_future_events = future_events[:fetch_future_games] if future_events else []
url = ESPN_NFL_SCOREBOARD_URL
params = {'dates': fetch_date}
response = requests.get(url, params=params)
response.raise_for_status()
date_data = response.json()
if date_data and "events" in date_data:
all_events.extend(date_data["events"])
BaseNFLManager.logger.info(f"[NFL] Fetched {len(date_data['events'])} events for date {fetch_date}")
BaseNFLManager.cache_manager.set(date_cache_key, date_data)
# Combine selected events
selected_events = selected_past_events + selected_future_events
# Process newly fetched events
if all_events:
# Sort events by date
all_events.sort(key=lambda x: x.get('date', ''))
# Separate past and future events
now = datetime.now(self._get_timezone())
past_events = []
future_events = []
for event in all_events:
try:
event_time = datetime.fromisoformat(event['date'].replace('Z', '+00:00'))
if event_time.tzinfo is None:
event_time = event_time.replace(tzinfo=pytz.UTC)
event_time = event_time.astimezone(self._get_timezone())
if event_time < now:
past_events.append(event)
else:
future_events.append(event)
except Exception as e:
BaseNFLManager.logger.warning(f"[NFL] Could not parse event date: {e}")
continue
BaseNFLManager.logger.info(f"[NFL] Selected {len(selected_past_events)} past games and {len(selected_future_events)} future games")
# Create the final data structure
data = {"events": selected_events}
BaseNFLManager._shared_data = data
BaseNFLManager._last_shared_update = current_time
return data
else:
BaseNFLManager.logger.warning("[NFL] No events found in the fetched data")
return None
days_to_check += 1
# Take the specified number of games
selected_past_events = past_events[-fetch_past_games:] if past_events else []
selected_future_events = future_events[:fetch_future_games] if future_events else []
# Combine selected events
selected_events = selected_past_events + selected_future_events
BaseNFLManager.logger.info(f"[NFL] Selected {len(selected_past_events)} past games and {len(selected_future_events)} future games after checking {days_to_check} days")
# Create the final data structure
data = {"events": selected_events}
BaseNFLManager._shared_data = data
BaseNFLManager._last_shared_update = current_time
return data
except requests.exceptions.RequestException as e:
BaseNFLManager.logger.error(f"[NFL] Error fetching data from ESPN: {e}")