|
|
@@ -13,16 +13,14 @@ def cow_says():
|
|
|
os.system(f'source {PROJECT_ROOT_PATH}/bin/common/functions.sh')
|
|
|
|
|
|
|
|
|
-# HADOOP_CONF_DIR / HIVE_CONF_DIR:Spark on YARN + Hive metastore 定位需要
|
|
|
-# 缺 HADOOP_CONF_DIR → SparkSubmitArguments 启动校验失败
|
|
|
-# 缺 HIVE_CONF_DIR → enableHiveSupport 回落 in-memory metastore,看不到 HMS 真实库表
|
|
|
-# DataX JVM 不读 classpath 中的 conf,HA 由 ini [hadoop_config] 节显式注入
|
|
|
+# HADOOP_CONF_DIR:spark-submit 启动 YARN 校验需要;DataX JVM 不读 classpath conf,HA 由 ini [hadoop_config] 节显式注入
|
|
|
os.environ['HADOOP_CONF_DIR'] = '/etc/hadoop/conf'
|
|
|
-os.environ['HIVE_CONF_DIR'] = '/etc/hive/conf'
|
|
|
+# SPARK_CONF_DIR:pip pyspark 默认指向自身空 conf/,显式指到集群配置才能加载 hive-site.xml,否则 enableHiveSupport 回落 in-memory metastore
|
|
|
+os.environ.setdefault('SPARK_CONF_DIR', '/etc/spark/conf')
|
|
|
+# os.environ['HIVE_CONF_DIR'] = '/etc/hive/conf'
|
|
|
# os.environ['JAVA_HOME'] = '/usr/local/java'
|
|
|
os.environ["PYSPARK_DRIVER_PYTHON"] = "/usr/bin/python3"
|
|
|
os.environ["PYSPARK_PYTHON"] = "/usr/bin/python3"
|
|
|
-# os.environ['SPARK_HOME'] = '/usr/hdp/3.1.5.0-152/spark2'
|
|
|
os.environ['PYTHONUNBUFFERED'] = 'x'
|
|
|
PROJECT_ROOT_PATH = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))
|
|
|
PROJECT_NAME = os.path.basename(PROJECT_ROOT_PATH)
|