|
|
@@ -11,12 +11,12 @@ from dw_base.utils.env_loader import bootstrap_env
|
|
|
bootstrap_env()
|
|
|
|
|
|
|
|
|
-# HADOOP_CONF_DIR:spark-submit 启动 YARN 校验需要;DataX JVM 不读 classpath conf,HA 由 ini [hadoop_config] 节显式注入
|
|
|
-os.environ['HADOOP_CONF_DIR'] = '/etc/hadoop/conf'
|
|
|
-# SPARK_CONF_DIR:pip pyspark 默认指向自身空 conf/,显式指到集群配置才能加载 hive-site.xml,否则 enableHiveSupport 回落 in-memory metastore
|
|
|
-os.environ.setdefault('SPARK_CONF_DIR', '/etc/spark/conf')
|
|
|
-os.environ["PYSPARK_DRIVER_PYTHON"] = "/usr/bin/python3"
|
|
|
-os.environ["PYSPARK_PYTHON"] = "/usr/bin/python3"
|
|
|
+# HADOOP_CONF_DIR / SPARK_CONF_DIR 由 conf/env.sh 提供默认值经 bootstrap_env 注入(shell 侧 export 优先):
|
|
|
+# - HADOOP_CONF_DIR:spark-submit 启动 YARN 校验需要;DataX JVM 不读 classpath conf,HA 由 ini [hadoop_config] 节显式注入
|
|
|
+# - SPARK_CONF_DIR:pip pyspark 默认指向自身空 conf/,显式指到集群配置才能加载 hive-site.xml,否则 enableHiveSupport 回落 in-memory metastore
|
|
|
+# PYSPARK_*:复用 PYTHON3_PATH,避免与 conf/env.sh 双份硬编码
|
|
|
+os.environ.setdefault('PYSPARK_DRIVER_PYTHON', os.environ['PYTHON3_PATH'])
|
|
|
+os.environ.setdefault('PYSPARK_PYTHON', os.environ['PYTHON3_PATH'])
|
|
|
os.environ['PYTHONUNBUFFERED'] = 'x'
|
|
|
PROJECT_ROOT_PATH = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))
|
|
|
PROJECT_NAME = os.path.basename(PROJECT_ROOT_PATH)
|