# foodie_weekly_thread.py import json import os import logging import random import signal import sys import fcntl import time from datetime import datetime, timedelta, timezone import tweepy from openai import OpenAI from foodie_utils import post_tweet, AUTHORS, SUMMARY_MODEL, load_json_file, check_author_rate_limit from foodie_config import X_API_CREDENTIALS, RECENT_POSTS_FILE from dotenv import load_dotenv load_dotenv() LOCK_FILE = "/home/shane/foodie_automator/locks/foodie_weekly_thread.lock" LOG_FILE = "/home/shane/foodie_automator/logs/foodie_weekly_thread.log" LOG_PRUNE_DAYS = 30 MAX_RETRIES = 3 RETRY_BACKOFF = 2 def setup_logging(): """Initialize logging with pruning of old logs.""" try: os.makedirs(os.path.dirname(LOG_FILE), exist_ok=True) if os.path.exists(LOG_FILE): with open(LOG_FILE, 'r') as f: lines = f.readlines() cutoff = datetime.now(timezone.utc) - timedelta(days=LOG_PRUNE_DAYS) pruned_lines = [] malformed_count = 0 for line in lines: if len(line) < 19 or not line[:19].replace('-', '').replace(':', '').replace(' ', '').isdigit(): malformed_count += 1 continue try: timestamp = datetime.strptime(line[:19], '%Y-%m-%d %H:%M:%S').replace(tzinfo=timezone.utc) if timestamp > cutoff: pruned_lines.append(line) except ValueError: malformed_count += 1 continue if malformed_count > 0: logging.info(f"Skipped {malformed_count} malformed log lines during pruning") with open(LOG_FILE, 'w') as f: f.writelines(pruned_lines) logging.basicConfig( filename=LOG_FILE, level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s', datefmt='%Y-%m-%d %H:%M:%S' ) console_handler = logging.StreamHandler() console_handler.setFormatter(logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')) logging.getLogger().addHandler(console_handler) logging.getLogger("tweepy").setLevel(logging.WARNING) logging.info("Logging initialized for foodie_weekly_thread.py") except Exception as e: print(f"Failed to setup logging: {e}") sys.exit(1) def acquire_lock(): """Acquire a lock to prevent concurrent runs.""" os.makedirs(os.path.dirname(LOCK_FILE), exist_ok=True) lock_fd = open(LOCK_FILE, 'w') try: fcntl.flock(lock_fd, fcntl.LOCK_EX | fcntl.LOCK_NB) lock_fd.write(str(os.getpid())) lock_fd.flush() return lock_fd except IOError: logging.info("Another instance of foodie_weekly_thread.py is running") sys.exit(0) def signal_handler(sig, frame): """Handle termination signals gracefully.""" logging.info("Received termination signal, exiting...") sys.exit(0) signal.signal(signal.SIGTERM, signal_handler) signal.signal(signal.SIGINT, signal_handler) # Initialize OpenAI client try: client = OpenAI(api_key=os.getenv("OPENAI_API_KEY")) if not os.getenv("OPENAI_API_KEY"): logging.error("OPENAI_API_KEY is not set in environment variables") raise ValueError("OPENAI_API_KEY is required") except Exception as e: logging.error(f"Failed to initialize OpenAI client: {e}", exc_info=True) sys.exit(1) def validate_twitter_credentials(): """Validate Twitter API credentials for all authors.""" logging.info("Validating Twitter API credentials for all authors") valid_credentials = [] for author in AUTHORS: credentials = X_API_CREDENTIALS.get(author["username"]) if not credentials: logging.error(f"No X credentials found for {author['username']} in X_API_CREDENTIALS") continue for attempt in range(MAX_RETRIES): try: twitter_client = tweepy.Client( consumer_key=credentials["api_key"], consumer_secret=credentials["api_secret"], access_token=credentials["access_token"], access_token_secret=credentials["access_token_secret"] ) user = twitter_client.get_me() logging.info(f"Credentials valid for {author['username']} (handle: {credentials['x_username']})") valid_credentials.append(credentials) break except tweepy.TweepyException as e: logging.error(f"Failed to validate credentials for {author['username']} (attempt {attempt + 1}): {e}") if attempt < MAX_RETRIES - 1: time.sleep(RETRY_BACKOFF * (2 ** attempt)) else: logging.error(f"Credentials invalid for {author['username']} after {MAX_RETRIES} attempts") if not valid_credentials: logging.error("No valid Twitter credentials found for any author") raise ValueError("No valid Twitter credentials found") return valid_credentials def load_recent_posts(): """Load and deduplicate posts from recent_posts.json.""" logging.debug(f"Attempting to load posts from {RECENT_POSTS_FILE}") posts = load_json_file(RECENT_POSTS_FILE) if not posts: logging.warning(f"No valid posts loaded from {RECENT_POSTS_FILE}") return [] # Deduplicate posts unique_posts = {} for post in posts: try: required_fields = ["title", "url", "author_username", "timestamp"] if not all(key in post for key in required_fields): logging.warning(f"Skipping invalid post: missing fields {post}") continue datetime.fromisoformat(post["timestamp"].replace('Z', '+00:00')) key = (post["title"], post["url"], post["author_username"]) if key not in unique_posts: unique_posts[key] = post else: logging.debug(f"Skipping duplicate post: {post['title']}") except (KeyError, ValueError) as e: logging.warning(f"Skipping post due to invalid format: {e}") continue deduped_posts = list(unique_posts.values()) logging.info(f"Loaded {len(deduped_posts)} unique posts from {RECENT_POSTS_FILE}") return deduped_posts def filter_posts_for_week(posts, start_date, end_date): """Filter posts within the given week range.""" filtered_posts = [] for post in posts: try: post_date = datetime.fromisoformat(post["timestamp"]) logging.debug(f"Checking post: title={post['title']}, timestamp={post_date}, in range {start_date} to {end_date}") if start_date <= post_date <= end_date: filtered_posts.append(post) logging.debug(f"Included post: {post['title']}") else: logging.debug(f"Excluded post: {post['title']} (timestamp {post_date} outside range)") except (KeyError, ValueError) as e: logging.warning(f"Skipping post due to invalid format: {e}") continue logging.info(f"Filtered to {len(filtered_posts)} posts for the week") return filtered_posts def generate_intro_tweet(author): """Generate an intro tweet for the weekly thread.""" credentials = next((cred for cred in X_API_CREDENTIALS if cred["username"] == author["username"]), None) if not credentials: logging.error(f"No X credentials found for {author['username']}") return None author_handle = credentials["x_username"] logging.debug(f"Generating intro tweet for {author_handle}") prompt = ( f"Generate a concise tweet (under 280 characters) for {author_handle}. " f"Introduce a thread of their top 10 foodie posts of the week on InsiderFoodie.com. " f"Make it engaging, create curiosity, and include a call to action to visit InsiderFoodie.com, follow {author_handle}, or like the thread. " f"Avoid using the word 'elevate'—use more humanized language like 'level up' or 'bring to life'. " f"Do not include emojis, hashtags, or reward-driven incentives (e.g., giveaways)." ) for attempt in range(MAX_RETRIES): try: response = client.chat.completions.create( model=SUMMARY_MODEL, messages=[ {"role": "system", "content": "You are a social media expert crafting engaging tweets."}, {"role": "user", "content": prompt} ], max_tokens=100, temperature=0.7 ) tweet = response.choices[0].message.content.strip() if len(tweet) > 280: tweet = tweet[:277] + "..." logging.debug(f"Generated intro tweet: {tweet}") return tweet except Exception as e: logging.warning(f"Failed to generate intro tweet for {author['username']} (attempt {attempt + 1}): {e}") if attempt < MAX_RETRIES - 1: time.sleep(RETRY_BACKOFF * (2 ** attempt)) else: logging.error(f"Failed to generate intro tweet after {MAX_RETRIES} attempts") fallback = ( f"This week's top 10 foodie finds by {author_handle}! Check out the best on InsiderFoodie.com. " f"Follow {author_handle} for more and like this thread to stay in the loop! Visit us at https://insiderfoodie.com" ) logging.info(f"Using fallback intro tweet: {fallback}") return fallback def generate_final_cta(author): """Generate a final CTA tweet for the weekly thread using GPT.""" credentials = next((cred for cred in X_API_CREDENTIALS if cred["username"] == author["username"]), None) if not credentials: logging.error(f"No X credentials found for {author['username']}") return None author_handle = credentials["x_username"] logging.debug(f"Generating final CTA tweet for {author_handle}") prompt = ( f"Generate a concise tweet (under 280 characters) for {author_handle}. " f"Conclude a thread of their top 10 foodie posts of the week on InsiderFoodie.com. " f"Make it engaging, value-driven, and urgent, in the style of Neil Patel. " f"Include a call to action to visit InsiderFoodie.com and follow {author_handle}. " f"Mention that the top 10 foodie trends are shared every Monday. " f"Avoid using the word 'elevate'—use humanized language like 'level up' or 'bring to life'. " f"Do not include emojis, hashtags, or reward-driven incentives (e.g., giveaways)." ) for attempt in range(MAX_RETRIES): try: response = client.chat.completions.create( model=SUMMARY_MODEL, messages=[ {"role": "system", "content": "You are a social media expert crafting engaging tweets."}, {"role": "user", "content": prompt} ], max_tokens=100, temperature=0.7 ) tweet = response.choices[0].message.content.strip() if len(tweet) > 280: tweet = tweet[:277] + "..." logging.debug(f"Generated final CTA tweet: {tweet}") return tweet except Exception as e: logging.warning(f"Failed to generate final CTA tweet for {author['username']} (attempt {attempt + 1}): {e}") if attempt < MAX_RETRIES - 1: time.sleep(RETRY_BACKOFF * (2 ** attempt)) else: logging.error(f"Failed to generate final CTA tweet after {MAX_RETRIES} attempts") fallback = ( f"Want more foodie insights like these? Check out insiderfoodie.com and follow {author_handle} " f"for the world’s top 10 foodie trends every Monday. Don’t miss out!" ) logging.info(f"Using fallback final CTA tweet: {fallback}") return fallback def post_weekly_thread(): """Generate and post a weekly thread of top posts for each author on Mondays.""" logging.info("Starting foodie_weekly_thread.py") # Check if today is Monday today = datetime.now(timezone.utc) if today.weekday() != 0: # 0 = Monday logging.info(f"Today is not Monday (weekday: {today.weekday()}), skipping weekly thread") return # Calculate date range: 7 days prior to run date start_date = (today - timedelta(days=7)).replace(hour=0, minute=0, second=0, microsecond=0) end_date = (today - timedelta(days=1)).replace(hour=23, minute=59, second=59, microsecond=999999) logging.info(f"Fetching posts from {start_date} to {end_date}") # Load and filter posts recent_posts = load_json_file(RECENT_POSTS_FILE) logging.info(f"Loaded {len(recent_posts)} posts from {RECENT_POSTS_FILE}") # Deduplicate posts seen = set() deduped_posts = [] for post in recent_posts: key = (post["title"], post["url"], post["author_username"]) if key not in seen: seen.add(key) deduped_posts.append(post) logging.info(f"Filtered to {len(deduped_posts)} unique posts after deduplication") weekly_posts = filter_posts_for_week(deduped_posts, start_date, end_date) if not weekly_posts: logging.warning(f"No posts found within the week range {start_date} to {end_date}, exiting post_weekly_thread") return # Group posts by author posts_by_author = {author["username"]: [] for author in AUTHORS} for post in weekly_posts: username = post["author_username"] if username in posts_by_author: posts_by_author[username].append(post) # Post threads for each author for author in AUTHORS: username = author["username"] author_posts = posts_by_author.get(username, []) if not author_posts: logging.info(f"No posts found for {username}, skipping") continue # Check if the author can post before generating the thread if check_author_rate_limit(author): reset_time = datetime.fromtimestamp( load_json_file('/home/shane/foodie_automator/rate_limit_info.json', default={}) .get(username, {}) .get('tweet_reset', time.time()), tz=timezone.utc ).strftime('%Y-%m-%d %H:%M:%S') logging.info(f"Skipping weekly thread for {username} due to rate limit. Reset at: {reset_time}") continue # Select top 2 posts (to fit within 3-tweet limit: lead + 2 posts) author_posts = sorted(author_posts, key=lambda x: datetime.fromisoformat(x["timestamp"]), reverse=True)[:2] logging.info(f"Selected {len(author_posts)} posts for {username}") # Generate and post thread try: # Post lead tweet intro_tweet = generate_intro_tweet(author) if not intro_tweet: logging.error(f"Failed to generate intro tweet for {username}, skipping") continue lead_response = post_tweet(author, intro_tweet) if not lead_response: logging.error(f"Failed to post lead tweet for {username}, skipping") continue lead_tweet_id = lead_response["id"] logging.info(f"Posted lead tweet for {username}: {intro_tweet}") # Post thread tweets (up to 2) for i, post in enumerate(author_posts, 1): thread_tweet = ( f"{i}. {post['title']} " f"Read more: {post['url']}" ) thread_response = post_tweet(author, thread_tweet, reply_to_id=lead_tweet_id) if thread_response: lead_tweet_id = thread_response["id"] logging.info(f"Posted thread tweet {i} for {username}: {thread_tweet}") else: logging.warning(f"Failed to post thread tweet {i} for {username}") # Post final CTA tweet final_cta = generate_final_cta(author) if final_cta: cta_response = post_tweet(author, final_cta, reply_to_id=lead_tweet_id) if cta_response: logging.info(f"Posted final CTA tweet for {username}: {final_cta}") else: logging.warning(f"Failed to post final CTA tweet for {username}") except Exception as e: logging.error(f"Error posting thread for {username}: {e}", exc_info=True) continue logging.info("Completed foodie_weekly_thread.py") def main(): """Main function to run the script.""" lock_fd = None try: lock_fd = acquire_lock() setup_logging() post_weekly_thread() except Exception as e: logging.error(f"Fatal error in main: {e}", exc_info=True) print(f"Fatal error: {e}") sys.exit(1) finally: if lock_fd: fcntl.flock(lock_fd, fcntl.LOCK_UN) lock_fd.close() os.remove(LOCK_FILE) if os.path.exists(LOCK_FILE) else None if __name__ == "__main__": main()