crawl_gov_zhejiangi_full.py 13 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352
  1. import os
  2. import random
  3. import re
  4. import subprocess
  5. import time
  6. from pathlib import Path
  7. from urllib.parse import urljoin
  8. from faker import Faker
  9. from selenium import webdriver
  10. from selenium.common.exceptions import StaleElementReferenceException
  11. from selenium.webdriver import FirefoxOptions
  12. from selenium.webdriver.common.by import By
  13. from selenium.webdriver.support import expected_conditions as EC
  14. from selenium.webdriver.support.ui import WebDriverWait
  15. import gov_commodity_zhejiang_city
  16. import gov_commodity_zhejiang_country
  17. import gov_commodity_zhejiang_import_export
  18. from utils import base_country_code, base_mysql
  19. download_dir = base_country_code.download_dir
  20. Path(download_dir).mkdir(parents=True, exist_ok=True)
  21. def configure_stealth_options():
  22. """增强型反检测配置[1,4](@ref)"""
  23. opts = FirefoxOptions()
  24. print("当前下载路径:", Path(download_dir).resolve())
  25. # 文件下载配置
  26. opts.set_preference("browser.download.dir", download_dir)
  27. opts.set_preference("browser.download.folderList", 2)
  28. opts.set_preference("browser.download.manager.showWhenStarting", False)
  29. opts.set_preference("browser.helperApps.neverAsk.saveToDisk",
  30. "application/octet-stream, application/vnd.ms-excel") # 覆盖常见文件类型
  31. opts.set_preference("browser.download.manager.useWindow", False) # 禁用下载管理器窗口
  32. opts.set_preference("browser.download.manager.showAlertOnComplete", False) # 关闭完成提示
  33. # 反检测参数
  34. opts.set_preference("dom.webdriver.enabled", False)
  35. opts.set_preference("useAutomationExtension", False)
  36. opts.add_argument("--disable-blink-features=AutomationControlled")
  37. # 动态指纹
  38. fake = Faker()
  39. opts.set_preference("general.useragent.override", fake.firefox())
  40. opts.set_preference("intl.accept_languages", "zh-CN,zh;q=0.9")
  41. # 视口配置
  42. opts.add_argument("--width=1440")
  43. opts.add_argument("--height=900")
  44. opts.add_argument("--headless")
  45. return opts
  46. def crawl_by_year_tabs(driver, base_url):
  47. """按年份Tab导航爬取数据"""
  48. years = ['2023年', '2024年', '2025年']
  49. WebDriverWait(driver, 30).until(
  50. EC.presence_of_element_located((By.CLASS_NAME, "portlet"))
  51. )
  52. year_tabs = driver.find_elements(By.XPATH, '//ul[@class="nav_sj"]//li//a')
  53. for tab in year_tabs:
  54. year_text = tab.text.strip()
  55. if int(year_text[:4]) <= 2022:
  56. print(f"{year_text} 后的数据无需下载")
  57. continue
  58. year_url = tab.get_attribute("href")
  59. if not year_url.startswith(('http://', 'https://')):
  60. year_url = base_url.split('//')[0] + '//' + base_url.split('/')[2] + year_url
  61. # 新标签页打开年份页面
  62. driver.execute_script("window.open(arguments[0]);", year_url)
  63. driver.switch_to.window(driver.window_handles[-1])
  64. print(f"\n正在处理 {year_text} 年份页面")
  65. process_month_tabs(driver, year_text, base_url)
  66. # 返回主窗口
  67. driver.close()
  68. driver.switch_to.window(driver.window_handles[0])
  69. def process_month_tabs(driver, year, base_url):
  70. """处理月份Tab导航(动态获取真实存在的月份)"""
  71. # ✅ 显式等待容器加载
  72. WebDriverWait(driver, 30).until(
  73. EC.presence_of_element_located((By.CLASS_NAME, "portlet"))
  74. )
  75. target_months = ['一月', '二月', '三月', '四月', '五月', '六月',
  76. '七月', '八月', '九月', '十月', '十一月', '十二月']
  77. processed_months = set() # 已处理月份记录
  78. retry_count = 0
  79. # while retry_count < 3: # 最多重试3次
  80. while True: # 最多重试3次
  81. try:
  82. # 全量获取所有月份Tab
  83. month_items = driver.find_elements(By.XPATH, '//ul[@class="nav_tab"]//li')
  84. if not month_items:
  85. print(f"{year}年没有月份Tab,停止处理")
  86. break
  87. all_found = True
  88. month_text = ''
  89. found = False
  90. for i,item in enumerate(month_items):
  91. a_tag = item.find_element(By.XPATH, './/a')
  92. month_text = a_tag.text.strip()
  93. if month_text in processed_months:
  94. continue
  95. if not month_text in target_months:
  96. continue # 跳过已处理月份
  97. print(f"点击月份Tab:{year}-{month_text}")
  98. a_tag.click()
  99. # 处理详情页逻辑
  100. WebDriverWait(driver, 30).until(
  101. EC.presence_of_element_located((By.CLASS_NAME, "portlet"))
  102. )
  103. detail_link_arr = get_behind_detail_link(driver, base_url)
  104. if not detail_link_arr:
  105. print(f"{year}-{month_text} 未找到详情链接")
  106. for detail_link in detail_link_arr:
  107. print(f"{year}-{month_text} 详情链接:{detail_link}")
  108. driver.get(detail_link)
  109. download_file_from_detail_page(driver)
  110. driver.back()
  111. WebDriverWait(driver, 30).until(
  112. EC.presence_of_element_located((By.CLASS_NAME, "portlet"))
  113. )
  114. processed_months.add(month_text)
  115. found = True
  116. if not found:
  117. print(f"{year}年未找到 {month_text} Tab")
  118. all_found = False
  119. if all_found:
  120. print(f"{year}年所有目标月份处理完成")
  121. break
  122. else:
  123. # 部分月份未找到,重新获取元素
  124. # retry_count += 1
  125. print(f"第 {retry_count} 次重试获取月份Tab...")
  126. time.sleep(2)
  127. except StaleElementReferenceException:
  128. print("页面刷新,重新获取月份Tab列表...")
  129. # retry_count += 1
  130. time.sleep(2)
  131. print(f"{year}年最终处理的月份:{processed_months}")
  132. def get_behind_detail_link(driver, base_url):
  133. """获取点击月份Tab后 conList_ul 下所有 li 的 a 标签完整链接"""
  134. href_arr = []
  135. try:
  136. elements = WebDriverWait(driver, 30).until(
  137. EC.element_to_be_clickable((By.XPATH, '//ul[@class="conList_ul"]/li/a'))
  138. )
  139. elements = elements.find_elements(By.XPATH, '//ul[@class="conList_ul"]/li/a')
  140. for element in elements:
  141. href = element.get_attribute("href")
  142. full_url = urljoin(base_url, href) # 自动处理相对路径
  143. href_arr.append(full_url)
  144. return href_arr
  145. except Exception as e:
  146. print(f"获取详情链接失败: {str(e)}")
  147. return []
  148. def download_file_from_detail_page(driver):
  149. WebDriverWait(driver, 30).until(
  150. EC.presence_of_element_located((By.CLASS_NAME, "portlet"))
  151. )
  152. try:
  153. elements = driver.find_elements(By.XPATH, '//div[@class="easysite-news-content"]//div[@id="easysiteText"]//p//a')
  154. if not elements:
  155. print("详情页未找到目标文件链接")
  156. return
  157. for download_btn in elements:
  158. file_name = download_btn.text.strip()
  159. if not file_name:
  160. continue
  161. file_url = download_btn.get_attribute("href")
  162. if not file_url.lower().endswith(('.xls', '.xlsx')):
  163. print(f"跳过非 Excel 文件: {file_url}")
  164. continue
  165. print(f"正在下载: {file_name} → {file_url}")
  166. # 记录下载前的文件列表
  167. existing_files = set(f.name for f in Path(download_dir).glob('*'))
  168. # 随机点击延迟
  169. time.sleep(random.uniform(1, 3))
  170. download_btn.click()
  171. downloaded_file = wait_for_download_complete(existing_files=existing_files)
  172. year, start_month, month = extract_year_and_month(file_name)
  173. final_path = Path(download_dir) / year / month / f"{file_name}"
  174. if os.path.exists(final_path):
  175. print(f"文件已存在:{file_name} 正在覆盖...")
  176. os.unlink(final_path)
  177. final_dir = Path(download_dir) / year / month
  178. final_dir.mkdir(parents=True, exist_ok=True)
  179. print(f"√ 正在移动文件 {downloaded_file} 至 {final_path}")
  180. downloaded_file.rename(final_path)
  181. print(f"√ 下载成功:{final_path}")
  182. except Exception as e:
  183. print(f"详情页处理异常: {str(e)}")
  184. def extract_year_and_month(file_name):
  185. # 支持两种格式:
  186. # - 2025年1-2月xxx
  187. # - 2025年3月xxx
  188. match = re.search(r"(\d{4})年(\d{1,2})(?:-(\d{1,2}))?月", file_name)
  189. if match:
  190. year = match.group(1)
  191. start_month = match.group(2)
  192. end_month = match.group(3) if match.group(3) else start_month
  193. return year, start_month.zfill(2), end_month.zfill(2)
  194. else:
  195. raise ValueError(f"无法从文件名中提取年份和月份:{file_name}")
  196. def extract_rar(rar_path, extract_to):
  197. """备用解压函数(当 rarfile 失效时使用)"""
  198. winrar_path = r"C:\Program Files\WinRAR\Rar.exe" # 推荐使用 Rar.exe 而非 WinRAR.exe
  199. cmd = [winrar_path, 'x', '-y', rar_path, str(extract_to)]
  200. # 使用 CREATE_NO_WINDOW 防止弹出命令行窗口
  201. creationflags = subprocess.CREATE_NO_WINDOW if os.name == 'nt' else 0
  202. result = subprocess.run(
  203. cmd,
  204. stdout=subprocess.PIPE,
  205. stderr=subprocess.PIPE,
  206. creationflags=creationflags # 关键点:隐藏窗口
  207. )
  208. if result.returncode == 0:
  209. print(f"解压成功: {rar_path} → {extract_to}")
  210. return True
  211. else:
  212. print(f"解压失败: {result.stderr.decode('gbk')}")
  213. return False
  214. def crawl_with_selenium(url):
  215. driver = webdriver.Firefox(options=configure_stealth_options())
  216. base_url = 'http://hangzhou.customs.gov.cn'
  217. try:
  218. # 注入反检测脚本
  219. driver.execute_script("""
  220. Object.defineProperty(navigator, 'webdriver', {
  221. get: () => undefined
  222. });
  223. window.alert = () => {};
  224. """)
  225. # 页面加载策略
  226. driver.get(url)
  227. # 按年份导航
  228. crawl_by_year_tabs(driver, base_url)
  229. finally:
  230. driver.quit()
  231. def wait_for_download_complete(timeout=30, existing_files=None):
  232. """
  233. 监控下载目录,等待文件下载完成并返回新下载的文件。
  234. :param timeout: 超时时间(秒)
  235. :param existing_files: 下载前已存在的文件列表
  236. :return: 新下载的文件路径
  237. """
  238. start_time = time.time()
  239. temp_exts = ('.part', '.crdownload')
  240. if existing_files is None:
  241. existing_files = set(f.name for f in Path(download_dir).glob('*'))
  242. while (time.time() - start_time) < timeout:
  243. # 获取有效文件列表
  244. valid_files = []
  245. for f in Path(download_dir).glob('*'):
  246. if (f.name not in existing_files and
  247. not f.name.endswith(temp_exts) and
  248. f.stat().st_size > 0):
  249. valid_files.append(f)
  250. # 等待最新文件稳定
  251. if valid_files:
  252. return max(valid_files, key=lambda x: x.stat().st_mtime)
  253. time.sleep(2)
  254. raise TimeoutError("文件下载超时")
  255. def hierarchical_traversal(root_path):
  256. """分层遍历:省份->年份->月目录"""
  257. root = Path(root_path)
  258. # 获取所有年份目录
  259. year_dirs = [
  260. item for item in root.iterdir()
  261. if item.is_dir() and base_country_code.YEAR_PATTERN.match(item.name)
  262. ]
  263. # 按年倒序
  264. for year_dir in sorted(year_dirs, key=lambda x: x.name, reverse=True):
  265. # 构造完整的路径:download/shandong/2025/03
  266. print(f"\n年份:{year_dir.name} | 省份:jiangsu")
  267. # 提取月份目录
  268. month_dirs = []
  269. for item in year_dir.iterdir():
  270. if item.is_dir() and base_country_code.MONTH_PATTERN.match(item.name):
  271. month_dirs.append({
  272. "path": item,
  273. "month": int(item.name)
  274. })
  275. # 按月倒序输出
  276. if month_dirs:
  277. for md in sorted(month_dirs, key=lambda x: x["month"], reverse=True):
  278. print(f" 月份:{md['month']:02d} | 路径:{md['path']}")
  279. gov_commodity_zhejiang_import_export.process_folder(md['path'])
  280. gov_commodity_zhejiang_country.process_folder(md['path'])
  281. gov_commodity_zhejiang_city.process_folder(md['path'])
  282. if __name__ == "__main__":
  283. crawl_with_selenium('http://hangzhou.customs.gov.cn/hangzhou_customs/575609/zlbd/575612/575612/6430241/6430315/index.html')
  284. print(f"浙江杭州海关全量数据下载任务完成")
  285. # 等待5s后执行
  286. time.sleep(5)
  287. hierarchical_traversal(base_country_code.download_dir)
  288. print("浙江杭州海关类章、国家、城市所有文件处理完成!")
  289. time.sleep(5)
  290. base_mysql.update_january_yoy('浙江省')
  291. base_mysql.update_shandong_yoy('浙江省')
  292. print("浙江杭州海关城市同比sql处理完成")