wangbowen1024 commented on issue #15352:
URL:
https://github.com/apache/dolphinscheduler/issues/15352#issuecomment-1886770249
> > **I have the same issue with version 3.2** Here is my environment setup
in dolphinscheduler_env.sh
> > ```shell
> > export SPARK_HOME=${SPARK_HOME:-/data/spark3}
> >
> > export
PYTHON_LAUNCHER=${PYTHON_LAUNCHER:-/usr/local/python3/bin/python3}
> >
> > export DATAX_LAUNCHER=${DATAX_LAUNCHER:-/usr/local/datax/bin/datax.py}
> >
> > export
PATH=$HADOOP_HOME/bin:$SPARK_HOME/bin:$PYTHON_LAUNCHER:$JAVA_HOME/bin:$HIVE_HOME/bin:$FLINK_HOME/bin:$DATAX_LAUNCHER:$PATH
> > ```
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > When I pass the task or the result of these variables
> > ```shell
> > echo " ------shell task 1 ------"
> > echo "The python launcher program is ${PYTHON_LAUNCHER} "
> > echo "The spark home is ${SPARK_HOME}"
> > echo "The datax launcher program is ${DATAX_LAUNCHER}"`
> > ```
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > As shown in the log below, they are all null.
> > ```
> > ------shell task 1 ------
> > The python launcher program is
> > The spark home is
> > The datax launcher program is
> > ```
>
> Add environment variable parameters from the management interface
environment management options Then associate the above variables during flow
and project operations, but it still doesn't work.
you can set these env to /etc/profile to sloved
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]