|
@@ -6,8 +6,7 @@ import com.aliyun.odps.data.Record
|
|
import com.aliyun.odps.spark.examples.myUtils.env
|
|
import com.aliyun.odps.spark.examples.myUtils.env
|
|
import org.apache.commons.net.util.Base64
|
|
import org.apache.commons.net.util.Base64
|
|
import org.apache.spark.sql.SparkSession
|
|
import org.apache.spark.sql.SparkSession
|
|
-
|
|
|
|
-import scala.collection.JavaConversions.mapAsScalaMap
|
|
|
|
|
|
+import java.util.Base64
|
|
|
|
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -50,8 +49,6 @@ object diff_data_20250319 {
|
|
|
|
|
|
val rdd1: JSONObject = odpsData1.first()
|
|
val rdd1: JSONObject = odpsData1.first()
|
|
|
|
|
|
- println("rdd1")
|
|
|
|
- println(rdd1.toString)
|
|
|
|
|
|
|
|
val rdd2: JSONObject = odpsData2.first()
|
|
val rdd2: JSONObject = odpsData2.first()
|
|
|
|
|
|
@@ -145,13 +142,9 @@ object diff_data_20250319 {
|
|
for (i <- 0 until columns.size()) {
|
|
for (i <- 0 until columns.size()) {
|
|
val column = columns.get(i)
|
|
val column = columns.get(i)
|
|
val name = column.getName
|
|
val name = column.getName
|
|
- val encodedValue = record.get(name)
|
|
|
|
featureMap.put(name, record.get(name))
|
|
featureMap.put(name, record.get(name))
|
|
if (name == "logkey") {
|
|
if (name == "logkey") {
|
|
- val decodedBytes = Base64.decodeBase64(encodedValue.toString)
|
|
|
|
- // 将字节数组转换为字符串
|
|
|
|
- val decodedValue = new String(decodedBytes)
|
|
|
|
- featureMap.put(name, decodedValue)
|
|
|
|
|
|
+ featureMap.put(name, new String(Base64.getDecoder.decode(record.get(name).toString), "UTF-8"))
|
|
}
|
|
}
|
|
}
|
|
}
|
|
featureMap
|
|
featureMap
|