wangkun 2 years ago
parent
commit
83b927e046
4 changed files with 6 additions and 1 deletions
  1. BIN
      .DS_Store
  2. 4 0
      README.MD
  3. BIN
      kuaishou/.DS_Store
  4. 2 1
      kuaishou/kuaishou_follow/kuaishou_follow.py

BIN
.DS_Store


+ 4 - 0
README.MD

@@ -37,6 +37,8 @@ ps aux | grep run_weixinzhishu | grep -v grep | awk '{print $2}' | xargs kill -9
 sh ./main/main.sh ./xigua/xigua_main/run_xigua_follow.py --log_type="follow" --crawler="xigua" --strategy="定向爬虫策略" --oss_endpoint="inner" --env="prod" --machine="aliyun" xigua/nohup.log
 本机
 sh ./main/main.sh ./xigua/xigua_main/run_xigua_follow.py --log_type="follow" --crawler="xigua" --strategy="定向爬虫策略" --oss_endpoint="out" --env="prod" --machine="local" xigua/nohup.log
+macpro
+sh ./main/main.sh ./xigua/xigua_main/run_xigua_follow.py --log_type="follow" --crawler="xigua" --strategy="定向爬虫策略" --oss_endpoint="out" --env="prod" --machine="macpro" xigua/nohup.log
 杀进程命令:
 ps aux | grep run_xigua
 ps aux | grep run_xigua | grep -v grep | awk '{print $2}' | xargs kill -9
@@ -48,6 +50,8 @@ ps aux | grep run_xigua | grep -v grep | awk '{print $2}' | xargs kill -9
 sh ./main/main.sh ./kuaishou/kuaishou_main/run_kuaishou_follow.py --log_type="follow" --crawler="kuaishou" --strategy="定向爬虫策略" --oss_endpoint="inner" --env="prod" --machine="aliyun" kuaishou/nohup.log
 本机
 sh ./main/main.sh ./kuaishou/kuaishou_main/run_kuaishou_follow.py --log_type="follow" --crawler="kuaishou" --strategy="定向爬虫策略" --oss_endpoint="out" --env="dev" --machine="local" kuaishou/nohup.log
+macpro
+sh ./main/main.sh ./kuaishou/kuaishou_main/run_kuaishou_follow.py --log_type="follow" --crawler="kuaishou" --strategy="定向爬虫策略" --oss_endpoint="out" --env="prod" --machine="macpro" kuaishou/nohup.log
 杀进程命令:
 ps aux | grep run_kuaishou
 ps aux | grep run_kuaishou | grep -v grep | awk '{print $2}' | xargs kill -9

BIN
xigua/logs/.DS_Store → kuaishou/.DS_Store


+ 2 - 1
kuaishou/kuaishou_follow/kuaishou_follow.py

@@ -336,6 +336,7 @@ class Follow:
                     # 'sec-ch-ua-platform': '"macOS"'
                 }
                 response = requests.post(url=url, headers=headers, data=payload)
+                # Common.logger(log_type, crawler).info(f"get_videoList:{response.text}\n")
                 if response.status_code != 200:
                     Common.logger(log_type, crawler).warning(f"get_videoList_response:{response.text}\n")
                     return
@@ -354,7 +355,7 @@ class Follow:
                 else:
                     feeds = response.json()['data']['visionProfilePhotoList']['feeds']
                     pcursor = response.json()['data']['visionProfilePhotoList']['pcursor']
-                    Common.logger(log_type, crawler).info(f"feeds0: {feeds}\n")
+                    # Common.logger(log_type, crawler).info(f"feeds0: {feeds}\n")
                     for i in range(len(feeds)):
                         if 'photo' not in feeds[i]:
                             Common.logger(log_type, crawler).warning(f"get_videoList:{feeds[i]}\n")