<output id="qn6qe"></output>

    1. <output id="qn6qe"><tt id="qn6qe"></tt></output>
    2. <strike id="qn6qe"></strike>

      亚洲 日本 欧洲 欧美 视频,日韩中文字幕有码av,一本一道av中文字幕无码,国产线播放免费人成视频播放,人妻少妇偷人无码视频,日夜啪啪一区二区三区,国产尤物精品自在拍视频首页,久热这里只有精品12

      爬蟲數據保存到csv中

      import json
      import os.path
      import time
      
      from jsonpath import *
      # import jsonpath as jsonpath
      import pandas as pd
      import requests
      
      
      
      # url = "http://www.whggzy.com/front/search/category"
      
      
      
      
      def get_resp(url,name,i):
          headers = {
      
              "Referer": "http://www.whggzy.com/PoliciesAndRegulations/index.html?utm=sites_group_front.26a79a93.0.0.715108e02e0e11ee837be5c5ca3fd993",
              "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/114.0.0.0 Safari/537.36",
              "Accept": "*/*",
              "Content-Type": "application/json",
              "X-Requested-With": "XMLHttpRequest"
          }
          data = {
              "utm":"sites_group_front.26a79a93.0.0.715108e02e0e11ee837be5c5ca3fd993",
              "categoryCode":f"{name}",
              "pageSize":15,
              "pageNo":f"{i}"
          }
          # json = data,json傳參就算將參數轉化為json格式進行傳遞的
          resp = requests.post(url, headers=headers, json=data).json()
      
          return resp
      
      def save_json(content):
          data = json.dumps(content)
          with open("wh_data.json",'w',encoding="utf-8") as w:
              w.write(data)
      
      def get_data(data_list,csv_path,i):
          base_url = 'http://www.whggzy.com/'
          pathName = ''
          for data in data_list:
              pathName = jsonpath(data,'$..pathName')[0] if jsonpath(data,'$..pathName') else None
      
              title = jsonpath(data,'$..title')[0] if jsonpath(data,'$..title') else None
      
              publishDate = jsonpath(data,'$..publishDate')[0] if jsonpath(data,'$..publishDate') else None
              date = time.strftime('%Y-%m-%d',time.localtime(publishDate / 1000))
              attachmentUrl = jsonpath(data,'$..attachmentUrl')[0] if jsonpath(data,'$..attachmentUrl') else None
              url  = base_url + jsonpath(data,'$..url')[0] if jsonpath(data,'$..url') else None
              csv_list = [pathName,title,date,attachmentUrl,url]
              save_csv(csv_list,csv_path)
      
      
          print(f'政策法規-->>{pathName}-->> 第{i}頁下爬取完畢 !!!')
      
      
      
      def judge_csv_file():
          # 當前腳本文件的絕對路徑,_file_代表的是appLogger 這個文件
          current_path = os.path.abspath(__file__)
          # 定義一個類屬性,保存的文件名稱
          csv_path = os.path.join(os.path.abspath(os.path.dirname(current_path)),
                                  'wh_data.csv')
          print(csv_path)
          if not os.path.exists(csv_path):
              head_list = ['項目','標題','日期','附件網址','內容地址']
              tb_head = ",".join(head_list) + '\n'
              with open(csv_path,'w',encoding="utf-8") as wf:
                  wf.write(tb_head)
          return csv_path
      
      
      def save_csv(data_list,csv_path):
          data = pd.DataFrame(data=[data_list])
          # 追加數據,mode = 'a',表示追加,index=False 表示不給每行數據加索引序號,header=False 表示不加標題
          data.to_csv(csv_path,mode='a',index=False,header=False,encoding='utf-8')
      
      def run(url):
          csv_path = judge_csv_file()
          name_list = ["GovernmentProcurement","BidAndEngineerConstruction","LandAndMineralRightsTransaction",
                       "TransactionOfPropertyRights","TransactionOfPublicResources"]
      
          for name in name_list:
              i = 1
              while True:
                  content = get_resp(url,name,i)
                  save_json(content)
                  data_list = content['hits']['hits']
                  if data_list:
                      get_data(data_list,csv_path,i)
                  else:
                      break
                  i += 1
      
      if __name__ == '__main__':
          url = "http://www.whggzy.com/front/search/category"
          run(url)
      

        

      posted @ 2023-07-29 22:11  xingmeng1  閱讀(47)  評論(0)    收藏  舉報
      主站蜘蛛池模板: 久9re热视频这里只有精品免费| 亚洲天堂精品一区二区| 久久精品国产亚洲欧美| 国产欧美一区二区精品性色| 午夜在线不卡| 国产jizzjizz视频| 99精产国品一二三产品香蕉| 亚洲中文字幕亚洲中文精| av深夜免费在线观看| 亚洲男人AV天堂午夜在| 99精品国产在热久久婷婷| 免费人成视频在线| 久久综合伊人77777| 妓院一钑片免看黄大片| 国产粉嫩美女一区二区三| 双柏县| 欧美日韩精品一区二区视频| 亚洲欧美日韩在线码| 国产肉丝袜在线观看| 国产一区二区丰满熟女人妻| 99久久99久久久精品久久| 国产亚洲精品日韩av在| 星座| 国产亚洲精品AA片在线爽| 亚洲欧洲∨国产一区二区三区| 欧美交A欧美精品喷水| 精品亚洲一区二区三区在线观看| 国产爆乳乱码女大生Av| 国产成人精品永久免费视频| 国精产品一区一区三区mba下载| 在线观看成人av天堂不卡| 成人年无码av片在线观看| 免费午夜无码片在线观看影院| AV无码不卡一区二区三区| 午夜精品国产自在| 日韩熟妇| 伊人欧美在线| 啦啦啦视频在线日韩精品| 成人国产精品日本在线观看| 镇江市| 国内自拍偷拍一区二区三区|