[
https://issues.apache.org/jira/browse/HBASE-19841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
stack reopened HBASE-19841:
---------------------------
Reopening. Breaks launching of MR jobs on a cluster. Here is what a good launch
looks like:
{code}
...
18/02/05 17:11:33 INFO impl.YarnClientImpl: Submitted application
application_1517369646236_0009
18/02/05 17:11:33 INFO mapreduce.Job: The url to track the job:
http://ve0524.halxg.cloudera.com:10134/proxy/application_1517369646236_0009/
18/02/05 17:11:33 INFO mapreduce.Job: Running job: job_1517369646236_0009
18/02/05 17:11:40 INFO mapreduce.Job: Job job_1517369646236_0009 running in
uber mode : false
18/02/05 17:11:40 INFO mapreduce.Job: map 0% reduce 0%
18/02/05 17:11:57 INFO mapreduce.Job: map 14% reduce 0%
...
{code}
... but now it does this....
{code}
18/02/05 17:17:54 INFO mapreduce.Job: The url to track the job:
http://ve0524.halxg.cloudera.com:10134/proxy/application_1517369646236_0011/
18/02/05 17:17:54 INFO mapreduce.Job: Running job: job_1517369646236_0011
18/02/05 17:17:56 INFO mapreduce.Job: Job job_1517369646236_0011 running in
uber mode : false
18/02/05 17:17:56 INFO mapreduce.Job: map 0% reduce 0%
18/02/05 17:17:56 INFO mapreduce.Job: Job job_1517369646236_0011 failed with
state FAILED due to: Application application_1517369646236_0011 failed 2 times
due to AM Container for appattempt_1517369646236_0011_000002 exited with
exitCode: -1000
Failing this attempt.Diagnostics: File
file:/tmp/stack/.staging/job_1517369646236_0011/job.splitmetainfo does not exist
java.io.FileNotFoundException: File
file:/tmp/stack/.staging/job_1517369646236_0011/job.splitmetainfo does not exist
at
org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:635)
at
org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:861)
at
org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:625)
at
org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:442)
at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:253)
at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:63)
at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:361)
at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:358)
at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:62)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
For more detailed output, check the application tracking page:
http://ve0524.halxg.cloudera.com:8188/applicationhistory/app/application_1517369646236_0011
Then click on links to logs of each attempt.
. Failing the application.
18/02/05 17:17:56 INFO mapreduce.Job: Counters: 0
{code}
If I revert this patch, the submit runs again.
I'd made the staging dir /tmp/stack and seemed to get further... The job
staging was made in the local fs... but it seems like we are then looking for
it up in hdfs. My guess is our stamping the fs as local until minihdfscluster
starts works for the unit test case but it messes up the inference of fs that
allows the above submission to work.
I'd like to revert this if thats ok.
> Tests against hadoop3 fail with StreamLacksCapabilityException
> --------------------------------------------------------------
>
> Key: HBASE-19841
> URL: https://issues.apache.org/jira/browse/HBASE-19841
> Project: HBase
> Issue Type: Test
> Reporter: Ted Yu
> Assignee: Mike Drob
> Priority: Major
> Fix For: 2.0.0-beta-2
>
> Attachments: 19841.007.patch, 19841.06.patch, 19841.v0.txt,
> 19841.v1.txt, HBASE-19841.v10.patch, HBASE-19841.v11.patch,
> HBASE-19841.v11.patch, HBASE-19841.v2.patch, HBASE-19841.v3.patch,
> HBASE-19841.v4.patch, HBASE-19841.v5.patch, HBASE-19841.v7.patch,
> HBASE-19841.v8.patch, HBASE-19841.v8.patch, HBASE-19841.v8.patch,
> HBASE-19841.v9.patch
>
>
> The following can be observed running against hadoop3:
> {code}
> java.io.IOException: cannot get log writer
> at
> org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.compactingSetUp(TestCompactingMemStore.java:107)
> at
> org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.setUp(TestCompactingMemStore.java:89)
> Caused by:
> org.apache.hadoop.hbase.util.CommonFSUtils$StreamLacksCapabilityException:
> hflush and hsync
> at
> org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.compactingSetUp(TestCompactingMemStore.java:107)
> at
> org.apache.hadoop.hbase.regionserver.TestCompactingMemStore.setUp(TestCompactingMemStore.java:89)
> {code}
> This was due to hbase-server/src/test/resources/hbase-site.xml not being
> picked up by Configuration object. Among the configs from this file, the
> value for "hbase.unsafe.stream.capability.enforce" relaxes check for presence
> of hflush and hsync. Without this config entry,
> StreamLacksCapabilityException is thrown.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)