EtachGu opened a new issue #5709:
URL: https://github.com/apache/dolphinscheduler/issues/5709
**For better global communication, Please describe it in English. If you
feel the description in English is not clear, then you can append description
in Chinese(just for Mandarin(CN)), thx! **
**Describe the bug**
A clear and concise description of what the bug is.
A sqoop task create file permission denied
**To Reproduce**
log file, mysql url has been desensitization
```log
[INFO] 2021-06-28 20:06:42.543 - [taskAppId=TASK-3-66-67]:[115] - create
dir success /tmp/dolphinscheduler/exec/process/1/3/66/67
[INFO] 2021-06-28 20:06:42.599 - [taskAppId=TASK-3-66-67]:[115] - create
linux os user : tenet1
[INFO] 2021-06-28 20:06:42.600 - [taskAppId=TASK-3-66-67]:[115] - execute
cmd : sudo useradd -g root
tenet1
[INFO] 2021-06-28 20:06:42.761 - [taskAppId=TASK-3-66-67]:[115] - create
user name success tenet1
[INFO] 2021-06-28 20:06:42.885 - [taskAppId=TASK-3-66-67]:[57] - sqoop task
params
{"customShell":"/opt/cloudera/parcels/CDH-6.3.2-1.cdh6.3.2.p0.1605554/bin/sqoop
import-all-tables \\\n --connect
\"jdbc:mysql://xxx:3306/record?allowMultiQueries=true&useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=GMT%2B8&allowPublicKeyRetrieval=true\"
\\\n --username root \\\n --password xxx \\\n --hive-import
\\\n --hive-database ods_record2 \\\n --hive-overwrite \\\n
--create-hive-table \\\n -m 5","jobType":"CUSTOM","localParams":[]}
[INFO] 2021-06-28 20:06:42.903 - [taskAppId=TASK-3-66-67]:[84] - sqoop
script: /opt/cloudera/parcels/CDH-6.3.2-1.cdh6.3.2.p0.1605554/bin/sqoop
import-all-tables \
--connect
"jdbc:mysql://xxx:3306/platform_record?allowMultiQueries=true&useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=GMT%2B8&allowPublicKeyRetrieval=true"
\
--username root \
--password xxx \
--hive-import \
--hive-database ods_record2 \
--hive-overwrite \
--create-hive-table \
-m 5
[INFO] 2021-06-28 20:06:42.905 - [taskAppId=TASK-3-66-67]:[87] - tenantCode
user:tenet1, task dir:3_66_67
[INFO] 2021-06-28 20:06:42.909 - [taskAppId=TASK-3-66-67]:[92] - create
command file:/tmp/dolphinscheduler/exec/process/1/3/66/67/3_66_67.command
[INFO] 2021-06-28 20:06:42.910 - [taskAppId=TASK-3-66-67]:[111] - command :
#!/bin/sh
BASEDIR=$(cd `dirname $0`; pwd)
cd $BASEDIR
source
/opt/apache-dolphinscheduler-1.3.6-bin/conf/env/dolphinscheduler_env.sh
/opt/cloudera/parcels/CDH-6.3.2-1.cdh6.3.2.p0.1605554/bin/sqoop
import-all-tables \
--connect
"jdbc:mysql://xxx:3306/platform_record?allowMultiQueries=true&useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=GMT%2B8&allowPublicKeyRetrieval=true"
\
--username root \
--password xxx \
--hive-import \
--hive-database ods_record2 \
--hive-overwrite \
--create-hive-table \
-m 5
[INFO] 2021-06-28 20:06:42.920 - [taskAppId=TASK-3-66-67]:[327] - task run
command:
sudo -u tenet1 sh
/tmp/dolphinscheduler/exec/process/1/3/66/67/3_66_67.command
[INFO] 2021-06-28 20:06:42.926 - [taskAppId=TASK-3-66-67]:[208] - process
start, process id is: 153
[INFO] 2021-06-28 20:06:43.934 - [taskAppId=TASK-3-66-67]:[129] - ->
Warning:
/opt/cloudera/parcels/CDH-6.3.2-1.cdh6.3.2.p0.1605554/bin/../lib/sqoop/../accumulo
does not exist! Accumulo imports will fail.
Please set $ACCUMULO_HOME to the root of your Accumulo installation.
WARNING: log4j.properties is not found. HADOOP_CONF_DIR may be
incomplete.
[INFO] 2021-06-28 20:06:44.937 - [taskAppId=TASK-3-66-67]:[129] - ->
WARNING: log4j.properties is not found. HADOOP_CONF_DIR may be incomplete.
2021-06-28 20:06:44,872 INFO [main] sqoop.Sqoop
(Sqoop.java:<init>(94)) - Running Sqoop version: 1.4.7-cdh6.3.2
[INFO] 2021-06-28 20:06:45.939 - [taskAppId=TASK-3-66-67]:[129] - ->
2021-06-28 20:06:45,022 WARN [main] tool.BaseSqoopTool
(BaseSqoopTool.java:applyCredentialsOptions(1179)) - Setting your password on
the command-line is insecure. Consider using -P instead.
2021-06-28 20:06:45,031 INFO [main] tool.BaseSqoopTool
(BaseSqoopTool.java:validateOutputFormatOptions(1550)) - Using Hive-specific
delimiters for output. You can override
2021-06-28 20:06:45,031 INFO [main] tool.BaseSqoopTool
(BaseSqoopTool.java:validateOutputFormatOptions(1551)) - delimiters with
--fields-terminated-by, etc.
2021-06-28 20:06:45,163 INFO [main] manager.MySQLManager
(MySQLManager.java:initOptionDefaults(68)) - Preparing to use a MySQL streaming
resultset.
Loading class `com.mysql.jdbc.Driver'. This is deprecated. The new
driver class is `com.mysql.cj.jdbc.Driver'. The driver is automatically
registered via the SPI and manual loading of the driver class is generally
unnecessary.
2021-06-28 20:06:45,701 INFO [main] tool.CodeGenTool
(CodeGenTool.java:generateORM(91)) - Beginning code generation
2021-06-28 20:06:45,733 INFO [main] manager.SqlManager
(SqlManager.java:execute(775)) - Executing SQL statement: SELECT t.* FROM
`api_time_statistics` AS t LIMIT 1
2021-06-28 20:06:45,785 INFO [main] manager.SqlManager
(SqlManager.java:execute(775)) - Executing SQL statement: SELECT t.* FROM
`api_time_statistics` AS t LIMIT 1
2021-06-28 20:06:45,799 INFO [main] orm.CompilationManager
(CompilationManager.java:findHadoopJars(100)) - HADOOP_MAPRED_HOME is
/opt/cloudera/parcels/CDH/lib/hadoop-mapreduce
[INFO] 2021-06-28 20:06:47.943 - [taskAppId=TASK-3-66-67]:[129] - ->
2021-06-28 20:06:47,874 ERROR [main] orm.CompilationManager
(CompilationManager.java:compile(255)) - Could not rename
/tmp/sqoop-tenet1/compile/c7a2c509fe2a239a24844297b4d03f00/api_time_statistics.java
to /tmp/dolphinscheduler/exec/process/1/3/66/67/./api_time_statistics.java.
Error: /tmp/dolphinscheduler/exec/process/1/3/66/67/./api_time_statistics.java
(Permission denied)
2021-06-28 20:06:47,874 INFO [main] orm.CompilationManager
(CompilationManager.java:jar(363)) - Writing jar file:
/tmp/sqoop-tenet1/compile/c7a2c509fe2a239a24844297b4d03f00/api_time_statistics.jar
2021-06-28 20:06:47,906 WARN [main] manager.MySQLManager
(MySQLManager.java:importTable(113)) - It looks like you are importing from
mysql.
2021-06-28 20:06:47,906 WARN [main] manager.MySQLManager
(MySQLManager.java:importTable(114)) - This transfer can be faster! Use the
--direct
2021-06-28 20:06:47,906 WARN [main] manager.MySQLManager
(MySQLManager.java:importTable(115)) - option to exercise a MySQL-specific fast
path.
2021-06-28 20:06:47,907 INFO [main] manager.MySQLManager
(MySQLManager.java:checkDateTimeBehavior(204)) - Setting zero DATETIME behavior
to convertToNull (mysql)
2021-06-28 20:06:47,919 INFO [main] mapreduce.ImportJobBase
(ImportJobBase.java:runImport(241)) - Beginning import of api_time_statistics
[INFO] 2021-06-28 20:06:48.944 - [taskAppId=TASK-3-66-67]:[129] - ->
2021-06-28 20:06:48,312 INFO [main] Configuration.deprecation
(Configuration.java:logDeprecation(1381)) - mapred.jar is deprecated. Instead,
use mapreduce.job.jar
2021-06-28 20:06:48,627 WARN [main] impl.MetricsConfig
(MetricsConfig.java:loadFirst(133)) - Cannot locate configuration: tried
hadoop-metrics2-jobtracker.properties,hadoop-metrics2.properties
2021-06-28 20:06:48,717 INFO [main] impl.MetricsSystemImpl
(MetricsSystemImpl.java:startTimer(374)) - Scheduled Metric snapshot period at
10 second(s).
2021-06-28 20:06:48,720 INFO [main] impl.MetricsSystemImpl
(MetricsSystemImpl.java:start(191)) - JobTracker metrics system started
2021-06-28 20:06:48,866 INFO [main] db.DBInputFormat
(DBInputFormat.java:setTxIsolation(192)) - Using read commited transaction
isolation
2021-06-28 20:06:48,868 INFO [main] db.DataDrivenDBInputFormat
(DataDrivenDBInputFormat.java:getSplits(167)) - BoundingValsQuery: SELECT
MIN(`id`), MAX(`id`) FROM `api_time_statistics`
2021-06-28 20:06:48,873 INFO [main] db.IntegerSplitter
(IntegerSplitter.java:split(145)) - Split size: 23; Num splits: 5 from: 1 to:
117
2021-06-28 20:06:48,894 INFO [main] mapreduce.JobSubmitter
(JobSubmitter.java:submitJobInternal(202)) - number of splits:5
[INFO] 2021-06-28 20:06:49.946 - [taskAppId=TASK-3-66-67]:[129] - ->
2021-06-28 20:06:49,033 INFO [main] mapreduce.JobSubmitter
(JobSubmitter.java:printTokens(298)) - Submitting tokens for job:
job_local2104280907_0001
2021-06-28 20:06:49,035 INFO [main] mapreduce.JobSubmitter
(JobSubmitter.java:printTokens(299)) - Executing with tokens: []
2021-06-28 20:06:49,307 INFO [main]
mapred.LocalDistributedCacheManager
(LocalDistributedCacheManager.java:symlink(201)) - Creating symlink:
/tmp/hadoop-tenet1/mapred/local/1624882009141/libjars <-
/tmp/dolphinscheduler/exec/process/1/3/66/67/libjars/*
2021-06-28 20:06:49,340 WARN [main] fs.FileUtil
(FileUtil.java:symLink(1054)) - Command 'ln -s
/tmp/hadoop-tenet1/mapred/local/1624882009141/libjars
/tmp/dolphinscheduler/exec/process/1/3/66/67/libjars/*' failed 1 with: ln:
failed to create symbolic link
'/tmp/dolphinscheduler/exec/process/1/3/66/67/libjars/*': No such file or
directory
2021-06-28 20:06:49,341 WARN [main]
mapred.LocalDistributedCacheManager
(LocalDistributedCacheManager.java:symlink(203)) - Failed to create symlink:
/tmp/hadoop-tenet1/mapred/local/1624882009141/libjars <-
/tmp/dolphinscheduler/exec/process/1/3/66/67/libjars/*
2021-06-28 20:06:49,341 INFO [main]
mapred.LocalDistributedCacheManager
(LocalDistributedCacheManager.java:setup(165)) - Localized
file:/tmp/hadoop/mapred/staging/tenet12104280907/.staging/job_local2104280907_0001/libjars
as file:/tmp/hadoop-tenet1/mapred/local/1624882009141/libjars
2021-06-28 20:06:49,483 INFO [main] mapreduce.Job
(Job.java:submit(1574)) - The url to track the job: http://localhost:8080/
2021-06-28 20:06:49,484 INFO [main] mapreduce.Job
(Job.java:monitorAndPrintJob(1619)) - Running job: job_local2104280907_0001
2021-06-28 20:06:49,491 INFO [Thread-6] mapred.LocalJobRunner
(LocalJobRunner.java:createOutputCommitter(501)) - OutputCommitter set in
config null
2021-06-28 20:06:49,506 INFO [Thread-6] output.FileOutputCommitter
(FileOutputCommitter.java:<init>(129)) - File Output Committer Algorithm
version is 2
2021-06-28 20:06:49,506 INFO [Thread-6] output.FileOutputCommitter
(FileOutputCommitter.java:<init>(144)) - FileOutputCommitter skip cleanup
_temporary folders under output directory:false, ignore cleanup failures: false
2021-06-28 20:06:49,507 INFO [Thread-6] mapred.LocalJobRunner
(LocalJobRunner.java:createOutputCommitter(519)) - OutputCommitter is
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2021-06-28 20:06:49,511 ERROR [Thread-6] output.FileOutputCommitter
(FileOutputCommitter.java:setupJob(344)) - Mkdirs failed to create
file:/tmp/dolphinscheduler/exec/process/1/3/66/67/api_time_statistics/_temporary/0
2021-06-28 20:06:49,561 INFO [Thread-6] mapred.LocalJobRunner
(LocalJobRunner.java:runTasks(478)) - Waiting for map tasks
2021-06-28 20:06:49,564 INFO [LocalJobRunner Map Task Executor #0]
mapred.LocalJobRunner (LocalJobRunner.java:run(252)) - Starting task:
attempt_local2104280907_0001_m_000000_0
2021-06-28 20:06:49,602 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(129)) - File Output
Committer Algorithm version is 2
2021-06-28 20:06:49,602 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(144)) -
FileOutputCommitter skip cleanup _temporary folders under output
directory:false, ignore cleanup failures: false
2021-06-28 20:06:49,637 INFO [LocalJobRunner Map Task Executor #0]
mapred.Task (Task.java:initialize(625)) - Using ResourceCalculatorProcessTree
: [ ]
2021-06-28 20:06:49,648 INFO [LocalJobRunner Map Task Executor #0]
db.DBInputFormat (DBInputFormat.java:setTxIsolation(192)) - Using read commited
transaction isolation
2021-06-28 20:06:49,657 INFO [LocalJobRunner Map Task Executor #0]
mapred.MapTask (MapTask.java:runNewMapper(768)) - Processing split: `id` >= 1
AND `id` < 25
2021-06-28 20:06:49,673 INFO [LocalJobRunner Map Task Executor #0]
mapred.LocalJobRunner (LocalJobRunner.java:run(252)) - Starting task:
attempt_local2104280907_0001_m_000001_0
2021-06-28 20:06:49,677 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(129)) - File Output
Committer Algorithm version is 2
2021-06-28 20:06:49,677 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(144)) -
FileOutputCommitter skip cleanup _temporary folders under output
directory:false, ignore cleanup failures: false
2021-06-28 20:06:49,678 INFO [LocalJobRunner Map Task Executor #0]
mapred.Task (Task.java:initialize(625)) - Using ResourceCalculatorProcessTree
: [ ]
2021-06-28 20:06:49,684 INFO [LocalJobRunner Map Task Executor #0]
db.DBInputFormat (DBInputFormat.java:setTxIsolation(192)) - Using read commited
transaction isolation
2021-06-28 20:06:49,687 INFO [LocalJobRunner Map Task Executor #0]
mapred.MapTask (MapTask.java:runNewMapper(768)) - Processing split: `id` >= 25
AND `id` < 48
2021-06-28 20:06:49,689 INFO [LocalJobRunner Map Task Executor #0]
mapred.LocalJobRunner (LocalJobRunner.java:run(252)) - Starting task:
attempt_local2104280907_0001_m_000002_0
2021-06-28 20:06:49,693 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(129)) - File Output
Committer Algorithm version is 2
2021-06-28 20:06:49,694 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(144)) -
FileOutputCommitter skip cleanup _temporary folders under output
directory:false, ignore cleanup failures: false
2021-06-28 20:06:49,694 INFO [LocalJobRunner Map Task Executor #0]
mapred.Task (Task.java:initialize(625)) - Using ResourceCalculatorProcessTree
: [ ]
2021-06-28 20:06:49,702 INFO [LocalJobRunner Map Task Executor #0]
db.DBInputFormat (DBInputFormat.java:setTxIsolation(192)) - Using read commited
transaction isolation
2021-06-28 20:06:49,704 INFO [LocalJobRunner Map Task Executor #0]
mapred.MapTask (MapTask.java:runNewMapper(768)) - Processing split: `id` >= 48
AND `id` < 71
2021-06-28 20:06:49,706 INFO [LocalJobRunner Map Task Executor #0]
mapred.LocalJobRunner (LocalJobRunner.java:run(252)) - Starting task:
attempt_local2104280907_0001_m_000003_0
2021-06-28 20:06:49,708 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(129)) - File Output
Committer Algorithm version is 2
2021-06-28 20:06:49,708 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(144)) -
FileOutputCommitter skip cleanup _temporary folders under output
directory:false, ignore cleanup failures: false
2021-06-28 20:06:49,709 INFO [LocalJobRunner Map Task Executor #0]
mapred.Task (Task.java:initialize(625)) - Using ResourceCalculatorProcessTree
: [ ]
2021-06-28 20:06:49,716 INFO [LocalJobRunner Map Task Executor #0]
db.DBInputFormat (DBInputFormat.java:setTxIsolation(192)) - Using read commited
transaction isolation
2021-06-28 20:06:49,719 INFO [LocalJobRunner Map Task Executor #0]
mapred.MapTask (MapTask.java:runNewMapper(768)) - Processing split: `id` >= 71
AND `id` < 94
2021-06-28 20:06:49,723 INFO [LocalJobRunner Map Task Executor #0]
mapred.LocalJobRunner (LocalJobRunner.java:run(252)) - Starting task:
attempt_local2104280907_0001_m_000004_0
2021-06-28 20:06:49,726 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(129)) - File Output
Committer Algorithm version is 2
2021-06-28 20:06:49,726 INFO [LocalJobRunner Map Task Executor #0]
output.FileOutputCommitter (FileOutputCommitter.java:<init>(144)) -
FileOutputCommitter skip cleanup _temporary folders under output
directory:false, ignore cleanup failures: false
2021-06-28 20:06:49,726 INFO [LocalJobRunner Map Task Executor #0]
mapred.Task (Task.java:initialize(625)) - Using ResourceCalculatorProcessTree
: [ ]
2021-06-28 20:06:49,733 INFO [LocalJobRunner Map Task Executor #0]
db.DBInputFormat (DBInputFormat.java:setTxIsolation(192)) - Using read commited
transaction isolation
2021-06-28 20:06:49,735 INFO [LocalJobRunner Map Task Executor #0]
mapred.MapTask (MapTask.java:runNewMapper(768)) - Processing split: `id` >= 94
AND `id` <= 117
2021-06-28 20:06:49,738 INFO [Thread-6] mapred.LocalJobRunner
(LocalJobRunner.java:runTasks(486)) - map task executor complete.
2021-06-28 20:06:49,739 WARN [Thread-6] mapred.LocalJobRunner
(LocalJobRunner.java:run(590)) - job_local2104280907_0001
java.lang.Exception: java.io.IOException: Mkdirs failed to create
file:/tmp/dolphinscheduler/exec/process/1/3/66/67/api_time_statistics/_temporary/0/_temporary/attempt_local2104280907_0001_m_000000_0
(exists=false, cwd=file:/tmp/dolphinscheduler/exec/process/1/3/66/67)
at
org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:492)
at
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:552)
Caused by: java.io.IOException: Mkdirs failed to create
file:/tmp/dolphinscheduler/exec/process/1/3/66/67/api_time_statistics/_temporary/0/_temporary/attempt_local2104280907_0001_m_000000_0
(exists=false, cwd=file:/tmp/dolphinscheduler/exec/process/1/3/66/67)
at
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:458)
at
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:443)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:1052)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:1032)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:921)
at
org.apache.sqoop.mapreduce.RawKeyTextOutputFormat.getFSDataOutputStream(RawKeyTextOutputFormat.java:44)
at
org.apache.sqoop.mapreduce.RawKeyTextOutputFormat.getOutputStream(RawKeyTextOutputFormat.java:63)
at
org.apache.sqoop.mapreduce.RawKeyTextOutputFormat.getRecordWriter(RawKeyTextOutputFormat.java:74)
at
org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:659)
at
org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:779)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:347)
at
org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:271)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
[INFO] 2021-06-28 20:06:50.605 - [taskAppId=TASK-3-66-67]:[217] - process
has exited, execute path:/tmp/dolphinscheduler/exec/process/1/3/66/67,
processId:153 ,exitStatusCode:0
[INFO] 2021-06-28 20:06:50.949 - [taskAppId=TASK-3-66-67]:[129] - ->
2021-06-28 20:06:50,493 INFO [main] mapreduce.Job
(Job.java:monitorAndPrintJob(1640)) - Job job_local2104280907_0001 running in
uber mode : false
2021-06-28 20:06:50,496 INFO [main] mapreduce.Job
(Job.java:monitorAndPrintJob(1647)) - map 0% reduce 0%
2021-06-28 20:06:50,499 INFO [main] mapreduce.Job
(Job.java:monitorAndPrintJob(1660)) - Job job_local2104280907_0001 failed with
state FAILED due to: NA
2021-06-28 20:06:50,519 INFO [main] mapreduce.Job
(Job.java:monitorAndPrintJob(1665)) - Counters: 0
2021-06-28 20:06:50,528 WARN [main] mapreduce.Counters
(AbstractCounters.java:getGroup(235)) - Group FileSystemCounters is deprecated.
Use org.apache.hadoop.mapreduce.FileSystemCounter instead
2021-06-28 20:06:50,530 INFO [main] mapreduce.ImportJobBase
(ImportJobBase.java:runJob(190)) - Transferred 0 bytes in 1.9577 seconds (0
bytes/sec)
2021-06-28 20:06:50,531 WARN [main] mapreduce.Counters
(AbstractCounters.java:getGroup(235)) - Group
org.apache.hadoop.mapred.Task$Counter is deprecated. Use
org.apache.hadoop.mapreduce.TaskCounter instead
2021-06-28 20:06:50,531 INFO [main] mapreduce.ImportJobBase
(ImportJobBase.java:runJob(192)) - Retrieved 0 records.
2021-06-28 20:06:50,533 ERROR [main] tool.ImportAllTablesTool
(ImportAllTablesTool.java:run(111)) - Error during import: Import job failed!
```
**Expected behavior**
It should work successed
**Which version of Dolphin Scheduler:**
-[1.3.6]
**Additional context**
I tried to add `umask 0002` in `/etc/bash.bashrc`, the root user `umask -p`
is `0002`, but the permission of directory created by taskAppId=TASK-3-66-67
still `drwxr-xr-x`
which cause above error, please tell how to solve this problem,Thank you in
advance
**Requirement or improvement**
- The process directory should created by tenet, the owner should be the
tenet who exec command.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]