|
@@ -6,14 +6,9 @@ sys.path.append(__dir__)
|
|
|
from utils.oss_client import HangZhouOSSClient
|
|
|
import utils.compress as compress
|
|
|
from utils.my_hdfs_client import MyHDFSClient
|
|
|
-import logging
|
|
|
# 引用 paddle inference 推理库
|
|
|
import paddle.inference as paddle_infer
|
|
|
-
|
|
|
-
|
|
|
-logging.basicConfig(
|
|
|
- format='%(asctime)s - %(levelname)s - %(message)s', level=logging.INFO)
|
|
|
-logger = logging.getLogger(__name__)
|
|
|
+import json
|
|
|
|
|
|
hadoop_home = "/app/env/hadoop-3.2.4" # Hadoop 安装目录
|
|
|
configs = {
|
|
@@ -60,8 +55,8 @@ def main():
|
|
|
|
|
|
input_data2 = {k: input_data[k] for k in list(input_data)[:5]}
|
|
|
# 设置输入
|
|
|
-
|
|
|
|
|
|
+ result = {}
|
|
|
for k,v in input_data2.items():
|
|
|
v2 = np.array([v], dtype=np.float32)
|
|
|
input_handle.copy_from_cpu(v2)
|
|
@@ -69,8 +64,14 @@ def main():
|
|
|
predictor.run()
|
|
|
# 获取输出
|
|
|
output_data = output_handle.copy_to_cpu() # numpy.ndarray类型
|
|
|
- logger.info("input k:{} v:{}".format(k, v))
|
|
|
- logger.info("Output {}".format(output_data))
|
|
|
+ result[k] = output_data
|
|
|
+
|
|
|
+
|
|
|
+ json_data = json.dumps(result, indent=4) # indent参数用于美化输出,使其更易读
|
|
|
+ # 写入文件
|
|
|
+ with open('/app/data.json', 'w') as json_file:
|
|
|
+ json_file.write(json_data)
|
|
|
+
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
main()
|