| import aiohttp |
| import asyncio |
| import json |
| import logging |
| import sqlite3 |
| import pandas as pd |
| from datetime import datetime, timedelta |
| from email.utils import parsedate_to_datetime |
| from typing import List, Dict, Set, Optional, Tuple |
| from pathlib import Path |
|
|
| class QuotaLowException(Exception): |
| pass |
|
|
| class ArticleCrawler: |
| def __init__(self, api_key: str, source_db: str, target_user_count: int = 800, output_db: str = "tweets.db"): |
| self.api_key = api_key |
| self.base_url = "https://api.tweetscout.io/v2" |
| self.source_db = source_db |
| self.target_user_count = target_user_count |
| self.db_path = output_db |
| self.headers = { |
| "ApiKey": api_key, |
| "Accept": "application/json" |
| } |
| self.processed_users_file = "processed_users.json" |
| self.no_tweets_users_file = "no_tweets_users.json" |
| self.processed_users: Set[str] = self._load_processed_users() |
| self.no_tweets_users: Set[str] = self._load_no_tweets_users() |
| self.api_call_count = 0 |
| self.warning_threshold = 5500 |
| self.last_save_time = datetime.now() |
| self.save_interval = timedelta(minutes=5) |
| self._setup_logging() |
| self._init_database() |
|
|
| def _load_processed_users(self) -> Set[str]: |
| if Path(self.processed_users_file).exists(): |
| try: |
| with open(self.processed_users_file, 'r') as f: |
| return set(json.load(f)) |
| except Exception as e: |
| logging.error(f"Error loading processed users: {e}") |
| return set() |
| return set() |
|
|
| def _load_no_tweets_users(self) -> Set[str]: |
| if Path(self.no_tweets_users_file).exists(): |
| try: |
| with open(self.no_tweets_users_file, 'r') as f: |
| return set(json.load(f)) |
| except Exception as e: |
| logging.error(f"Error loading no tweets users: {e}") |
| return set() |
| return set() |
|
|
| def _save_processed_users(self): |
| try: |
| with open(self.processed_users_file, 'w') as f: |
| json.dump(list(self.processed_users), f) |
| except Exception as e: |
| logging.error(f"Error saving processed users: {e}") |
|
|
| def _save_no_tweets_users(self): |
| try: |
| with open(self.no_tweets_users_file, 'w') as f: |
| json.dump(list(self.no_tweets_users), f) |
| except Exception as e: |
| logging.error(f"Error saving no tweets users: {e}") |
|
|
| def _get_next_rotation_time(self): |
| """计算下一次日志切换的时间""" |
| now = datetime.now() |
| if now.hour < 12: |
| |
| next_rotation = now.replace(hour=12, minute=0, second=0, microsecond=0) |
| else: |
| |
| next_rotation = (now + timedelta(days=1)).replace(hour=0, minute=0, second=0, microsecond=0) |
| return next_rotation |
|
|
| def get_progress_stats(self) -> Dict: |
| """获取当前进度统计""" |
| try: |
| if not hasattr(self, 'cursor') or self.cursor is None: |
| return { |
| 'user_count': 0, |
| 'tweet_count': 0, |
| 'start_time': None, |
| 'last_update': None, |
| 'processed_users': len(self.processed_users), |
| 'no_tweets_users': len(self.no_tweets_users) |
| } |
| |
| self.cursor.execute(""" |
| SELECT |
| COUNT(DISTINCT author_id) as user_count, |
| COUNT(*) as tweet_count, |
| MIN(collected_at) as start_time, |
| MAX(collected_at) as last_update |
| FROM tweets |
| """) |
| stats = dict(zip(['user_count', 'tweet_count', 'start_time', 'last_update'], |
| self.cursor.fetchone())) |
| stats['processed_users'] = len(self.processed_users) |
| stats['no_tweets_users'] = len(self.no_tweets_users) |
| return stats |
| except Exception as e: |
| self.logger.error(f"Error getting progress stats: {e}") |
| return { |
| 'user_count': 0, |
| 'tweet_count': 0, |
| 'start_time': None, |
| 'last_update': None, |
| 'processed_users': len(self.processed_users), |
| 'no_tweets_users': len(self.no_tweets_users) |
| } |
|
|
| def _rotate_log_file(self): |
| """切换到新的日志文件""" |
| |
| for handler in self.logger.handlers[:]: |
| if isinstance(handler, logging.FileHandler): |
| handler.close() |
| self.logger.removeHandler(handler) |
|
|
| |
| new_log_file = f"crawler_{datetime.now().strftime('%Y%m%d_%H%M')}.log" |
| file_handler = logging.FileHandler(new_log_file, encoding='utf-8') |
| file_handler.setFormatter(logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')) |
| self.logger.addHandler(file_handler) |
|
|
| |
| self.next_log_rotation = self._get_next_rotation_time() |
|
|
| |
| stats = self.get_progress_stats() |
| self.logger.info(f"Rotated to new log file: {new_log_file}") |
| self.logger.info(f"Progress Stats: {json.dumps(stats, default=str)}") |
| def _setup_logging(self): |
| """设置日志系统""" |
| self.logger = logging.getLogger("ArticleCrawler") |
| self.logger.setLevel(logging.INFO) |
|
|
| |
| console_handler = logging.StreamHandler() |
| console_handler.setFormatter(logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')) |
| self.logger.addHandler(console_handler) |
|
|
| |
| self.next_log_rotation = self._get_next_rotation_time() |
| self._rotate_log_file() |
|
|
| def _init_database(self): |
| self.conn = sqlite3.connect( |
| self.db_path, |
| detect_types=sqlite3.PARSE_DECLTYPES|sqlite3.PARSE_COLNAMES, |
| timeout=30 |
| ) |
| self.conn.execute("PRAGMA journal_mode=WAL") |
| self.conn.execute("PRAGMA synchronous=NORMAL") |
| self.cursor = self.conn.cursor() |
|
|
| self.cursor.executescript(''' |
| CREATE TABLE IF NOT EXISTS tweets ( |
| id INTEGER PRIMARY KEY AUTOINCREMENT, |
| tweet_id TEXT UNIQUE, |
| author_id TEXT, |
| author_name TEXT, |
| full_text TEXT, |
| created_at TIMESTAMP, |
| likes_count INTEGER DEFAULT 0, |
| retweets_count INTEGER DEFAULT 0, |
| replies_count INTEGER DEFAULT 0, |
| views_count INTEGER DEFAULT 0, |
| quote_count INTEGER DEFAULT 0, |
| is_quote_status BOOLEAN DEFAULT 0, |
| conversation_id TEXT, |
| in_reply_to_status_id TEXT, |
| quoted_status TEXT, -- 引用推文信息,存储为JSON |
| retweeted_status TEXT, -- 转发推文信息,存储为JSON |
| entities TEXT, -- 推文中的链接、媒体等实体,存储为JSON |
| author_avatar TEXT, -- 作者头像URL |
| author_description TEXT, -- 作者简介 |
| author_followers_count INT, -- 作者粉丝数 |
| author_friends_count INT, -- 作者关注数 |
| collected_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP |
| ); |
| |
| CREATE INDEX IF NOT EXISTS idx_tweet_author ON tweets(author_id); |
| CREATE INDEX IF NOT EXISTS idx_tweet_id ON tweets(tweet_id); |
| CREATE INDEX IF NOT EXISTS idx_tweet_created ON tweets(created_at); |
| ''') |
| self.conn.commit() |
|
|
| def get_successful_user_count(self) -> int: |
| """获取已成功爬取推文的用户数量""" |
| try: |
| self.cursor.execute( |
| "SELECT COUNT(DISTINCT author_id) FROM tweets" |
| ) |
| return self.cursor.fetchone()[0] |
| except Exception as e: |
| self.logger.error(f"Error getting successful user count: {e}") |
| return len(self.processed_users) |
|
|
| def load_users(self) -> List[Dict]: |
| source_conn = sqlite3.connect(self.source_db) |
| |
| |
| successful_count = self.get_successful_user_count() |
| remaining_target = max(0, self.target_user_count - successful_count) |
| |
| |
| if remaining_target <= 0: |
| self.logger.info(f"Already reached target user count: {self.target_user_count}") |
| source_conn.close() |
| return [] |
|
|
| |
| fetch_limit = remaining_target * 2 |
|
|
| |
| excluded_users = self.processed_users.union(self.no_tweets_users) |
| if excluded_users: |
| placeholders = ','.join(['?' for _ in excluded_users]) |
| query = f""" |
| SELECT user_id, name |
| FROM users |
| WHERE user_id NOT IN ({placeholders}) |
| LIMIT {fetch_limit} |
| """ |
| df = pd.read_sql_query(query, source_conn, params=list(excluded_users)) |
| else: |
| query = f""" |
| SELECT user_id, name |
| FROM users |
| LIMIT {fetch_limit} |
| """ |
| df = pd.read_sql_query(query, source_conn) |
| |
| source_conn.close() |
| users = df.to_dict('records') |
| self.logger.info( |
| f"Loaded {len(users)} new users to process. " |
| f"Current successful users: {successful_count}, Target: {self.target_user_count}" |
| ) |
| return users |
|
|
| async def check_api_quota(self) -> bool: |
| """检查API配额,如果配额不足返回False""" |
| try: |
| async with aiohttp.ClientSession() as session: |
| async with session.get(f"{self.base_url}/quota", headers=self.headers) as response: |
| if response.status == 200: |
| quota_data = await response.json() |
| remaining = quota_data.get('remaining', 0) |
| self.logger.info(f"Current API quota remaining: {remaining}") |
| return remaining >= 100 |
| except Exception as e: |
| self.logger.error(f"Error checking API quota: {e}") |
| return False |
| return True |
|
|
| async def get_user_tweets(self, user: Dict, max_tweets: int = 200) -> List[Dict]: |
| try: |
| url = f"{self.base_url}/user-tweets" |
| all_tweets = [] |
| cursor = "" |
| retries = 3 |
|
|
| while len(all_tweets) < max_tweets: |
| if self.api_call_count >= self.warning_threshold: |
| quota_ok = await self.check_api_quota() |
| if not quota_ok and self.api_call_count % 100 == 0: |
| self.logger.warning(f"API calls: {self.api_call_count}, checking quota...") |
| if not await self.check_api_quota(): |
| raise QuotaLowException("API quota running low") |
|
|
| data = { |
| "user_id": user['user_id'], |
| "cursor": cursor, |
| "limit": 200 |
| } |
|
|
| got_response = False |
| for attempt in range(retries): |
| try: |
| async with aiohttp.ClientSession() as session: |
| async with session.post(url, headers=self.headers, json=data, timeout=30) as response: |
| if response.status == 200: |
| got_response = True |
| self.api_call_count += 1 |
| response_data = await response.json() |
| tweets = response_data.get('tweets', []) |
|
|
| if not tweets: |
| self.logger.info( |
| f"No more tweets available for user {user['name']}. " |
| f"Collected {len(all_tweets)} tweets." |
| ) |
| return all_tweets |
|
|
| |
| for tweet in tweets: |
| if tweet.get('retweeted_status'): |
| tweet['tweet_type'] = 'retweet' |
| elif tweet.get('is_quote_status'): |
| tweet['tweet_type'] = 'quote' |
| else: |
| tweet['tweet_type'] = 'original' |
|
|
| all_tweets.extend(tweets) |
|
|
| |
| type_counts = { |
| 'original': len([t for t in tweets if t['tweet_type'] == 'original']), |
| 'retweet': len([t for t in tweets if t['tweet_type'] == 'retweet']), |
| 'quote': len([t for t in tweets if t['tweet_type'] == 'quote']) |
| } |
|
|
| self.logger.info( |
| f"User {user['name']}: Got {len(tweets)} tweets " |
| f"(Original: {type_counts['original']}, " |
| f"Retweet: {type_counts['retweet']}, " |
| f"Quote: {type_counts['quote']}). " |
| f"Current total: {len(all_tweets)}" |
| ) |
|
|
| cursor = response_data.get('next_cursor') |
| if not cursor or len(all_tweets) >= max_tweets: |
| |
| return all_tweets[:max_tweets] |
|
|
| break |
| |
| elif response.status == 429: |
| wait_time = int(response.headers.get('Retry-After', 60)) |
| self.logger.warning(f"Rate limited, waiting {wait_time} seconds") |
| await asyncio.sleep(wait_time) |
| elif response.status == 403: |
| self.logger.error(f"Access forbidden for user {user['name']}") |
| return [] |
| elif response.status == 404: |
| self.logger.error(f"User {user['name']} not found") |
| return [] |
| else: |
| self.logger.error( |
| f"Failed to get tweets for {user['name']}: " |
| f"Status {response.status}, Response: {await response.text()}" |
| ) |
| if attempt == retries - 1: |
| if got_response: |
| return all_tweets |
| return [] |
| except Exception as e: |
| self.logger.error(f"Request error: {str(e)}") |
| if attempt == retries - 1: |
| if got_response: |
| return all_tweets |
| return [] |
| await asyncio.sleep(5) |
|
|
| await asyncio.sleep(3) |
|
|
| return all_tweets |
|
|
| except QuotaLowException: |
| raise |
| except Exception as e: |
| self.logger.error(f"Error getting tweets for {user['name']}: {str(e)}") |
| return [] |
|
|
| def save_tweet(self, user: Dict, tweet_data: Dict) -> bool: |
| try: |
| created_at = parsedate_to_datetime(tweet_data.get('created_at')) |
|
|
| |
| user_info = tweet_data.get('user', {}) |
|
|
| |
| quoted_status = tweet_data.get('quoted_status') |
| retweeted_status = tweet_data.get('retweeted_status') |
|
|
| self.cursor.execute(''' |
| INSERT OR REPLACE INTO tweets |
| (tweet_id, author_id, author_name, full_text, created_at, |
| likes_count, retweets_count, replies_count, views_count, |
| quote_count, is_quote_status, conversation_id, in_reply_to_status_id, |
| quoted_status, retweeted_status, entities, |
| author_avatar, author_description, author_followers_count, |
| author_friends_count, collected_at) |
| VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, CURRENT_TIMESTAMP) |
| ''', ( |
| tweet_data.get('id_str'), |
| user['user_id'], |
| user['name'], |
| tweet_data.get('full_text', ''), |
| created_at, |
| tweet_data.get('favorite_count', 0), |
| tweet_data.get('retweet_count', 0), |
| tweet_data.get('reply_count', 0), |
| tweet_data.get('view_count', 0), |
| tweet_data.get('quote_count', 0), |
| tweet_data.get('is_quote_status', False), |
| tweet_data.get('conversation_id_str'), |
| tweet_data.get('in_reply_to_status_id_str'), |
| json.dumps(quoted_status) if quoted_status else None, |
| json.dumps(retweeted_status) if retweeted_status else None, |
| json.dumps(tweet_data.get('entities', {})), |
| user_info.get('avatar'), |
| user_info.get('description'), |
| user_info.get('followers_count'), |
| user_info.get('friends_count') |
| )) |
| self.conn.commit() |
| return True |
| except Exception as e: |
| self.logger.error(f"Error saving tweet: {str(e)}") |
| self.conn.rollback() |
| return False |
|
|
| async def process_user(self, user: Dict): |
| try: |
| |
| if datetime.now() >= self.next_log_rotation: |
| self._rotate_log_file() |
| |
| |
| if datetime.now() - self.last_save_time > self.save_interval: |
| self._save_processed_users() |
| self._save_no_tweets_users() |
| self.last_save_time = datetime.now() |
| |
| if user['user_id'] in self.processed_users or user['user_id'] in self.no_tweets_users: |
| self.logger.info(f"Skipping already processed user {user['name']}") |
| return |
|
|
| tweets = await self.get_user_tweets(user, max_tweets=100) |
| |
| if not tweets: |
| self.logger.info(f"No tweets found for user {user['name']}") |
| self.no_tweets_users.add(user['user_id']) |
| self._save_no_tweets_users() |
| return |
|
|
| self.logger.info(f"Processing {len(tweets)} tweets for user {user['name']}") |
|
|
| success = True |
| for tweet in tweets: |
| if not self.save_tweet(user, tweet): |
| success = False |
| break |
|
|
| if success: |
| self.processed_users.add(user['user_id']) |
| self._save_processed_users() |
| self.logger.info(f"Successfully processed user {user['name']}") |
|
|
| except QuotaLowException: |
| raise |
| except Exception as e: |
| self.logger.error(f"Error processing user {user['name']}: {str(e)}") |
|
|
| async def run(self): |
| try: |
| while True: |
| users = self.load_users() |
| if not users: |
| self.logger.info("No more users to process or reached target count") |
| break |
|
|
| for user in users: |
| try: |
| await self.process_user(user) |
| |
| if self.get_successful_user_count() >= self.target_user_count: |
| self.logger.info(f"Reached target user count: {self.target_user_count}") |
| return |
| except QuotaLowException: |
| self.logger.info(f"API quota running low, stopping at user {user['name']}") |
| return |
| |
| await asyncio.sleep(2) |
| |
| except Exception as e: |
| self.logger.error(f"Unexpected error in main loop: {str(e)}") |
| raise |
| finally: |
| try: |
| self._save_processed_users() |
| self._save_no_tweets_users() |
| self.conn.close() |
| except Exception as e: |
| self.logger.error(f"Error during cleanup: {str(e)}") |
|
|
| async def main(): |
| crawler = None |
| try: |
| crawler = ArticleCrawler( |
| api_key="Enter your API KEY", |
| source_db="filtered_users.db", |
| target_user_count=800, |
| output_db="tweetsFor800.db" |
| ) |
| await crawler.run() |
| except KeyboardInterrupt: |
| print("\nReceived shutdown signal, cleaning up...") |
| except Exception as e: |
| print(f"Critical error: {str(e)}") |
| finally: |
| if crawler: |
| crawler.conn.close() |
|
|
| if __name__ == "__main__": |
| asyncio.run(main()) |