|
@@ -3,8 +3,7 @@ package com.aliyun.odps.spark.examples.makedata_ad
|
|
|
import com.alibaba.fastjson.{JSON, JSONObject}
|
|
|
import com.aliyun.odps.TableSchema
|
|
|
import com.aliyun.odps.data.Record
|
|
|
-import com.aliyun.odps.spark.common.AdUtil
|
|
|
-import com.aliyun.odps.spark.examples.myUtils.{MyDateUtils, MyHdfsUtils, ParamUtils, env}
|
|
|
+import com.aliyun.odps.spark.examples.myUtils.{DataTimeUtil, MyDateUtils, MyHdfsUtils, ParamUtils, env}
|
|
|
import examples.extractor.RankExtractorFeature_20240530
|
|
|
import org.apache.hadoop.io.compress.GzipCodec
|
|
|
import org.apache.spark.sql.SparkSession
|
|
@@ -93,10 +92,10 @@ object makedata_ad_31_originData_20240718 {
|
|
|
featureMap.put("targeting_conversion_" + b1.getString("targeting_conversion"), idDefaultValue)
|
|
|
}
|
|
|
|
|
|
- val hour = AdUtil.getHourByTimestamp(ts)
|
|
|
+ val hour = DataTimeUtil.getHourByTimestamp(ts)
|
|
|
featureMap.put("hour_" + hour, 0.1)
|
|
|
|
|
|
- val dayOfWeek = AdUtil.getDayOrWeekByTimestamp(ts)
|
|
|
+ val dayOfWeek = DataTimeUtil.getDayOrWeekByTimestamp(ts)
|
|
|
featureMap.put("dayOfWeek_" + dayOfWeek, 0.1);
|
|
|
|
|
|
if (b1.containsKey("cpa")) {
|