Browse Source

feat:添加评估结果分析脚本

zhaohaipeng 6 months ago
parent
commit
8ff461b327
2 changed files with 14 additions and 64 deletions
  1. 12 63
      ad/01_ad_model_update.sh
  2. 2 1
      ad/25_ad_data_make.sh

+ 12 - 63
ad/01_ad_model_update.sh

@@ -1,13 +1,9 @@
 #!/bin/sh
 set -x
 
-
-
 export PATH=$SPARK_HOME/bin:$PATH
 export HADOOP_CONF_DIR=/etc/taihao-apps/hadoop-conf
 export JAVA_HOME=/usr/lib/jvm/java-1.8.0
-export SPARK_HOME=/opt/apps/SPARK2/spark-2.4.8-hadoop3.2-1.0.8
-
 
 sh_path=$(dirname $0)
 source ${sh_path}/00_common.sh
@@ -17,7 +13,6 @@ source /root/anaconda3/bin/activate py37
 
 # 全局常量
 HADOOP=/opt/apps/HADOOP-COMMON/hadoop-common-current/bin/hadoop
-
 TRAIN_PATH=/dw/recommend/model/31_ad_sample_data_v4_test
 BUCKET_FEATURE_PATH=/dw/recommend/model/33_ad_train_data_v4_test
 MODEL_PATH=/dw/recommend/model/35_ad_model_test
@@ -148,64 +143,18 @@ check_ad_hive() {
   echo "$LOG_PREFIX -- 大数据数据生产校验 -- 大数据数据生产校验通过: 耗时 $elapsed"
 }
 
-make_origin_data() {
-  
-  local step_start_time=$(date +%s)
-
-  /opt/apps/SPARK2/spark-2.4.8-hadoop3.2-1.0.8/bin/spark-class2 org.apache.spark.deploy.SparkSubmit \
-  --class com.aliyun.odps.spark.examples.makedata_ad.v20240718.makedata_ad_31_originData_20240718 \
-  --master yarn --driver-memory 1G --executor-memory 2G --executor-cores 1 --num-executors 16 \
-  ./target/spark-examples-1.0.0-SNAPSHOT-shaded.jar \
-  tablePart:64 repartition:32 \
-  beginStr:${today_early_1}00 endStr:${today_early_1}12 \
-  savePath:${TRAIN_PATH} \
-  table:${TABLE} \
-  filterHours:00,01,02,03,04,05,06,07 \
-  idDefaultValue:0.1 &
-
-  /opt/apps/SPARK2/spark-2.4.8-hadoop3.2-1.0.8/bin/spark-class2 org.apache.spark.deploy.SparkSubmit \
-  --class com.aliyun.odps.spark.examples.makedata_ad.v20240718.makedata_ad_31_originData_20240718 \
-  --master yarn --driver-memory 1G --executor-memory 2G --executor-cores 1 --num-executors 16 \
-  ./target/spark-examples-1.0.0-SNAPSHOT-shaded.jar \
-  tablePart:64 repartition:32 \
-  beginStr:${today_early_1}13 endStr:${today_early_1}18 \
-  savePath:${TRAIN_PATH} \
-  table:${TABLE} \
-  filterHours:00,01,02,03,04,05,06,07 \
-  idDefaultValue:0.1 &
-
-  /opt/apps/SPARK2/spark-2.4.8-hadoop3.2-1.0.8/bin/spark-class2 org.apache.spark.deploy.SparkSubmit \
-  --class com.aliyun.odps.spark.examples.makedata_ad.v20240718.makedata_ad_31_originData_20240718 \
-  --master yarn --driver-memory 1G --executor-memory 2G --executor-cores 1 --num-executors 16 \
-  ./target/spark-examples-1.0.0-SNAPSHOT-shaded.jar \
-  tablePart:64 repartition:32 \
-  beginStr:${today_early_1}19 endStr:${today_early_1}23 \
-  savePath:${TRAIN_PATH} \
-  table:${TABLE} \
-  filterHours:00,01,02,03,04,05,06,07 \
-  idDefaultValue:0.1 &
-
-  wait
-
-  local return_code=$?
-  check_run_status $return_code $step_start_time "spark原始样本生产任务"
+origin_data() {
+  (
+    source ${sh_path}/25_xgb_make_data_origin_bucket.sh
+    make_origin_data
+  )
 }
 
-make_bucket_feature() {
-
-  local step_start_time=$(date +%s)
-  
-  /opt/apps/SPARK2/spark-2.4.8-hadoop3.2-1.0.8/bin/spark-class2 org.apache.spark.deploy.SparkSubmit \
-  --class com.aliyun.odps.spark.examples.makedata_ad.v20240718.makedata_ad_33_bucketData_20240718 \
-  --master yarn --driver-memory 2G --executor-memory 4G --executor-cores 1 --num-executors 16 \
-  ./target/spark-examples-1.0.0-SNAPSHOT-shaded.jar \
-  beginStr:${today_early_1} endStr:${today_early_1} repartition:100 \
-  filterNames:_4h_,_5h_,adid_,targeting_conversion_ \
-  readPath:${TRAIN_PATH} \
-  savePath:${BUCKET_FEATURE_PATH}
-
-  local return_code=$?
-  check_run_status $return_code $step_start_time "spark特征分桶任务"
+bucket_feature() {
+  (
+    source ${sh_path}/25_xgb_make_data_origin_bucket.sh
+    make_bucket_feature
+  )
 }
 
 xgb_train() {
@@ -300,9 +249,9 @@ main() {
 
   check_ad_hive
 
-  make_origin_data
+  origin_data
 
-  make_bucket_feature
+  bucket_feature
 
   xgb_train
 

+ 2 - 1
ad/25_ad_data_make.sh

@@ -4,13 +4,14 @@
 
 set -x 
 
-source /root/anaconda3/bin/activate py37
 
 export SPARK_HOME=/opt/apps/SPARK2/spark-2.4.8-hadoop3.2-1.0.8
 export PATH=$SPARK_HOME/bin:$PATH
 export HADOOP_CONF_DIR=/etc/taihao-apps/hadoop-conf
 export JAVA_HOME=/usr/lib/jvm/java-1.8.0
 
+source /root/anaconda3/bin/activate py37
+
 # 全局常量
 HADOOP=/opt/apps/HADOOP-COMMON/hadoop-common-current/bin/hadoop
 FM_HOME=/root/sunmingze/alphaFM