[ 
https://issues.apache.org/jira/browse/HADOOP-13786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15939015#comment-15939015
 ] 

Steve Loughran commented on HADOOP-13786:
-----------------------------------------

And for the very curious, here's a trace of a test run of the latest code doing 
some simple {{sparkContact.makeRDD(1 to 
count).saveAsTextFile("s3a:///something")}} with the directory committer. 
There's a job committer and task committer in the logs, hence the role log info 
to work out whats happening. fs.s3a mostly at debug, except for 
instrumentation. FS stats exclude all the info on the uploads because they 
(currently) go straight through the s3a client.
{code}
    
-------------------------------------------
2017-03-23 18:42:41,455 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api)
2017-03-23 18:42:42,024 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1905)) - Not Found: 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:42,024 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:delete(1418)) - Couldn't delete 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
 - does not exist
2017-03-23 18:42:42,024 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 s3.S3ANumbersSuiteV2APISuite (CloudLogging.scala:logInfo(56)) - Switching to 
local file:// fs for default FS
2017-03-23 18:42:42,098 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkContext (Logging.scala:logInfo(54)) - Running Spark version 
2.2.0-SNAPSHOT
2017-03-23 18:42:42,133 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkContext (Logging.scala:logInfo(54)) - Submitted application: test
2017-03-23 18:42:42,147 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SecurityManager (Logging.scala:logInfo(54)) - Changing view acls to: 
stevel
2017-03-23 18:42:42,147 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SecurityManager (Logging.scala:logInfo(54)) - Changing modify acls to: 
stevel
2017-03-23 18:42:42,148 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SecurityManager (Logging.scala:logInfo(54)) - Changing view acls groups 
to: 
2017-03-23 18:42:42,149 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SecurityManager (Logging.scala:logInfo(54)) - Changing modify acls 
groups to: 
2017-03-23 18:42:42,150 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SecurityManager (Logging.scala:logInfo(54)) - SecurityManager: 
authentication disabled; ui acls disabled; users  with view permissions: 
Set(stevel); groups with view permissions: Set(); users  with modify 
permissions: Set(stevel); groups with modify permissions: Set()
2017-03-23 18:42:42,369 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 util.Utils (Logging.scala:logInfo(54)) - Successfully started service 
'sparkDriver' on port 54392.
2017-03-23 18:42:42,389 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkEnv (Logging.scala:logInfo(54)) - Registering MapOutputTracker
2017-03-23 18:42:42,402 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkEnv (Logging.scala:logInfo(54)) - Registering BlockManagerMaster
2017-03-23 18:42:42,405 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManagerMasterEndpoint (Logging.scala:logInfo(54)) - Using 
org.apache.spark.storage.DefaultTopologyMapper for getting topology information
2017-03-23 18:42:42,405 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManagerMasterEndpoint (Logging.scala:logInfo(54)) - 
BlockManagerMasterEndpoint up
2017-03-23 18:42:42,412 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.DiskBlockManager (Logging.scala:logInfo(54)) - Created local directory 
at 
/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/blockmgr-9ab575c8-964f-4cb7-96cb-e5c6efc04d60
2017-03-23 18:42:42,429 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 memory.MemoryStore (Logging.scala:logInfo(54)) - MemoryStore started with 
capacity 2004.6 MB
2017-03-23 18:42:42,466 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkEnv (Logging.scala:logInfo(54)) - Registering 
OutputCommitCoordinator
2017-03-23 18:42:42,579 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 util.Utils (Logging.scala:logInfo(54)) - Successfully started service 
'SparkUI' on port 4040.
2017-03-23 18:42:42,606 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 ui.SparkUI (Logging.scala:logInfo(54)) - Bound SparkUI to 0.0.0.0, and started 
at http://192.168.1.38:4040
2017-03-23 18:42:42,679 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 executor.Executor (Logging.scala:logInfo(54)) - Starting executor ID driver on 
host localhost
2017-03-23 18:42:42,694 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 util.Utils (Logging.scala:logInfo(54)) - Successfully started service 
'org.apache.spark.network.netty.NettyBlockTransferService' on port 54394.
2017-03-23 18:42:42,695 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 netty.NettyBlockTransferService (Logging.scala:logInfo(54)) - Server created 
on 192.168.1.38:54394
2017-03-23 18:42:42,696 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManager (Logging.scala:logInfo(54)) - Using 
org.apache.spark.storage.RandomBlockReplicationPolicy for block replication 
policy
2017-03-23 18:42:42,697 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManagerMaster (Logging.scala:logInfo(54)) - Registering 
BlockManager BlockManagerId(driver, 192.168.1.38, 54394, None)
2017-03-23 18:42:42,701 [dispatcher-event-loop-2] INFO  
storage.BlockManagerMasterEndpoint (Logging.scala:logInfo(54)) - Registering 
block manager 192.168.1.38:54394 with 2004.6 MB RAM, BlockManagerId(driver, 
192.168.1.38, 54394, None)
2017-03-23 18:42:42,703 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManagerMaster (Logging.scala:logInfo(54)) - Registered 
BlockManager BlockManagerId(driver, 192.168.1.38, 54394, None)
2017-03-23 18:42:42,703 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManager (Logging.scala:logInfo(54)) - Initialized BlockManager: 
BlockManagerId(driver, 192.168.1.38, 54394, None)
2017-03-23 18:42:42,898 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 s3.S3ANumbersSuiteV2APISuite (CloudLogging.scala:logInfo(56)) - 
Generating output under 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api

2017-03-23 18:42:42,940 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api)
2017-03-23 18:42:43,486 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1905)) - Not Found: 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:43,492 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG output.PathOutputCommitterFactory 
(PathOutputCommitterFactory.java:getOutputCommitterFactory(92)) - Using 
FileOutputCommitter factory class class 
org.apache.hadoop.fs.s3a.commit.staging.DirectoryStagingCommitterFactory
2017-03-23 18:42:43,495 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG commit.AbstractS3GuardCommitter 
(AbstractS3GuardCommitter.java:<init>(85)) - Task committer 
attempt_20170323184243_0000_m_000000_0 instantiated for job "" ID 
job_20170323184243_0000 with destination 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:43,497 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG commit.AbstractS3GuardCommitter 
(AbstractS3GuardCommitter.java:<init>(115)) - Task committer 
attempt_20170323184243_0000_m_000000_0} instantiated for"" ID 
job_20170323184243_0000
2017-03-23 18:42:43,499 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG commit.AbstractS3GuardCommitter 
(AbstractS3GuardCommitter.java:setWorkPath(162)) - Setting work path to 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184243_0000_m_000000_0/_temporary/0/_temporary/attempt_20170323184243_0000_m_000000_0
2017-03-23 18:42:43,501 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 output.FileOutputCommitter (FileOutputCommitter.java:<init>(123)) - File 
Output Committer Algorithm version is 1
2017-03-23 18:42:43,501 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 output.FileOutputCommitter (FileOutputCommitter.java:<init>(138)) - 
FileOutputCommitter skip cleanup _temporary folders under output 
directory:false, ignore cleanup failures: true
2017-03-23 18:42:43,502 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:postCreationActions(168)) - Conflict resolution 
mode: FAIL
2017-03-23 18:42:43,502 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 commit.Abstract3GuardCommitterFactory 
(Abstract3GuardCommitterFactory.java:createOutputCommitter(61)) - Using 
Commmitter StagingS3GuardCommitter{AbstractS3GuardCommitter{role=Task committer 
attempt_20170323184243_0000_m_000000_0outputPath=s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api,
 
workPath=file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184243_0000_m_000000_0/_temporary/0/_temporary/attempt_20170323184243_0000_m_000000_0},
 finalOutputPath=null, conflictResolution=FAIL, 
wrappedCommitter=FileOutputCommitter{outputPath=file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/stevel/job_20170323184243_0000/staging-uploads,
 workPath=null, algorithmVersion=1, skipCleanup=false, 
ignoreCleanupFailures=true}} for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:43,502 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:setupJob(520)) - Task committer 
attempt_20170323184243_0000_m_000000_0, Setting up job job_20170323184242_0001
2017-03-23 18:42:43,539 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:getOutputPath(1057)) - Task committer 
attempt_20170323184243_0000_m_000000_0: final output path is 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:43,539 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api)
2017-03-23 18:42:44,084 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1905)) - Not Found: 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:44,122 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkContext (Logging.scala:logInfo(54)) - Starting job: runJob at 
SparkHadoopMapReduceWriter.scala:88
2017-03-23 18:42:44,131 [dag-scheduler-event-loop] INFO  scheduler.DAGScheduler 
(Logging.scala:logInfo(54)) - Got job 0 (runJob at 
SparkHadoopMapReduceWriter.scala:88) with 1 output partitions
2017-03-23 18:42:44,132 [dag-scheduler-event-loop] INFO  scheduler.DAGScheduler 
(Logging.scala:logInfo(54)) - Final stage: ResultStage 0 (runJob at 
SparkHadoopMapReduceWriter.scala:88)
2017-03-23 18:42:44,132 [dag-scheduler-event-loop] INFO  scheduler.DAGScheduler 
(Logging.scala:logInfo(54)) - Parents of final stage: List()
2017-03-23 18:42:44,133 [dag-scheduler-event-loop] INFO  scheduler.DAGScheduler 
(Logging.scala:logInfo(54)) - Missing parents: List()
2017-03-23 18:42:44,137 [dag-scheduler-event-loop] INFO  scheduler.DAGScheduler 
(Logging.scala:logInfo(54)) - Submitting ResultStage 0 (MapPartitionsRDD[1] at 
map at NumbersRddTests.scala:120), which has no missing parents
2017-03-23 18:42:44,208 [dag-scheduler-event-loop] INFO  memory.MemoryStore 
(Logging.scala:logInfo(54)) - Block broadcast_0 stored as values in memory 
(estimated size 74.9 KB, free 2004.5 MB)
2017-03-23 18:42:44,232 [dag-scheduler-event-loop] INFO  memory.MemoryStore 
(Logging.scala:logInfo(54)) - Block broadcast_0_piece0 stored as bytes in 
memory (estimated size 25.0 KB, free 2004.5 MB)
2017-03-23 18:42:44,234 [dispatcher-event-loop-4] INFO  
storage.BlockManagerInfo (Logging.scala:logInfo(54)) - Added broadcast_0_piece0 
in memory on 192.168.1.38:54394 (size: 25.0 KB, free: 2004.6 MB)
2017-03-23 18:42:44,236 [dag-scheduler-event-loop] INFO  spark.SparkContext 
(Logging.scala:logInfo(54)) - Created broadcast 0 from broadcast at 
DAGScheduler.scala:1002
2017-03-23 18:42:44,251 [dag-scheduler-event-loop] INFO  scheduler.DAGScheduler 
(Logging.scala:logInfo(54)) - Submitting 1 missing tasks from ResultStage 0 
(MapPartitionsRDD[1] at map at NumbersRddTests.scala:120) (first 15 tasks are 
for partitions Vector(0))
2017-03-23 18:42:44,252 [dag-scheduler-event-loop] INFO  
scheduler.TaskSchedulerImpl (Logging.scala:logInfo(54)) - Adding task set 0.0 
with 1 tasks
2017-03-23 18:42:44,281 [dispatcher-event-loop-5] INFO  
scheduler.TaskSetManager (Logging.scala:logInfo(54)) - Starting task 0.0 in 
stage 0.0 (TID 0, localhost, executor driver, partition 0, PROCESS_LOCAL, 7598 
bytes)
2017-03-23 18:42:44,286 [Executor task launch worker for task 0] INFO  
executor.Executor (Logging.scala:logInfo(54)) - Running task 0.0 in stage 0.0 
(TID 0)
2017-03-23 18:42:44,333 [Executor task launch worker for task 0] DEBUG 
output.PathOutputCommitterFactory 
(PathOutputCommitterFactory.java:getOutputCommitterFactory(92)) - Using 
FileOutputCommitter factory class class 
org.apache.hadoop.fs.s3a.commit.staging.DirectoryStagingCommitterFactory
2017-03-23 18:42:44,334 [Executor task launch worker for task 0] DEBUG 
commit.AbstractS3GuardCommitter (AbstractS3GuardCommitter.java:<init>(85)) - 
Task committer attempt_20170323184242_0000_r_000000_0 instantiated for job "" 
ID job_20170323184242_0000 with destination 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:44,335 [Executor task launch worker for task 0] DEBUG 
commit.AbstractS3GuardCommitter (AbstractS3GuardCommitter.java:<init>(115)) - 
Task committer attempt_20170323184242_0000_r_000000_0} instantiated for"" ID 
job_20170323184242_0000
2017-03-23 18:42:44,336 [Executor task launch worker for task 0] DEBUG 
commit.AbstractS3GuardCommitter 
(AbstractS3GuardCommitter.java:setWorkPath(162)) - Setting work path to 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0
2017-03-23 18:42:44,337 [Executor task launch worker for task 0] INFO  
output.FileOutputCommitter (FileOutputCommitter.java:<init>(123)) - File Output 
Committer Algorithm version is 1
2017-03-23 18:42:44,337 [Executor task launch worker for task 0] INFO  
output.FileOutputCommitter (FileOutputCommitter.java:<init>(138)) - 
FileOutputCommitter skip cleanup _temporary folders under output 
directory:false, ignore cleanup failures: true
2017-03-23 18:42:44,337 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:postCreationActions(168)) - Conflict resolution 
mode: FAIL
2017-03-23 18:42:44,338 [Executor task launch worker for task 0] INFO  
commit.Abstract3GuardCommitterFactory 
(Abstract3GuardCommitterFactory.java:createOutputCommitter(61)) - Using 
Commmitter StagingS3GuardCommitter{AbstractS3GuardCommitter{role=Task committer 
attempt_20170323184242_0000_r_000000_0outputPath=s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api,
 
workPath=file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0},
 finalOutputPath=null, conflictResolution=FAIL, 
wrappedCommitter=FileOutputCommitter{outputPath=file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/stevel/job_20170323184243_0000/staging-uploads,
 workPath=null, algorithmVersion=1, skipCleanup=false, 
ignoreCleanupFailures=true}} for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:44,339 [Executor task launch worker for task 0] INFO  
commit.DurationInfo (DurationInfo.java:<init>(33)) - Starting Task committer 
attempt_20170323184242_0000_r_000000_0: task 
attempt_20170323184242_0000_r_000000_0: creating task attempt path 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0
 
2017-03-23 18:42:44,386 [Executor task launch worker for task 0] INFO  
commit.DurationInfo (DurationInfo.java:close(44)) - Task committer 
attempt_20170323184242_0000_r_000000_0: task 
attempt_20170323184242_0000_r_000000_0: creating task attempt path 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0
 : 0:00:047
2017-03-23 18:42:44,388 [Executor task launch worker for task 0] DEBUG 
output.PathOutputCommitterFactory 
(PathOutputCommitterFactory.java:getOutputCommitterFactory(92)) - Using 
FileOutputCommitter factory class class 
org.apache.hadoop.fs.s3a.commit.staging.DirectoryStagingCommitterFactory
2017-03-23 18:42:44,388 [Executor task launch worker for task 0] DEBUG 
commit.AbstractS3GuardCommitter (AbstractS3GuardCommitter.java:<init>(85)) - 
Task committer attempt_20170323184242_0000_r_000000_0 instantiated for job "" 
ID job_20170323184242_0000 with destination 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:44,388 [Executor task launch worker for task 0] DEBUG 
commit.AbstractS3GuardCommitter (AbstractS3GuardCommitter.java:<init>(115)) - 
Task committer attempt_20170323184242_0000_r_000000_0} instantiated for"" ID 
job_20170323184242_0000
2017-03-23 18:42:44,389 [Executor task launch worker for task 0] DEBUG 
commit.AbstractS3GuardCommitter 
(AbstractS3GuardCommitter.java:setWorkPath(162)) - Setting work path to 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0
2017-03-23 18:42:44,390 [Executor task launch worker for task 0] INFO  
output.FileOutputCommitter (FileOutputCommitter.java:<init>(123)) - File Output 
Committer Algorithm version is 1
2017-03-23 18:42:44,390 [Executor task launch worker for task 0] INFO  
output.FileOutputCommitter (FileOutputCommitter.java:<init>(138)) - 
FileOutputCommitter skip cleanup _temporary folders under output 
directory:false, ignore cleanup failures: true
2017-03-23 18:42:44,390 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:postCreationActions(168)) - Conflict resolution 
mode: FAIL
2017-03-23 18:42:44,390 [Executor task launch worker for task 0] INFO  
commit.Abstract3GuardCommitterFactory 
(Abstract3GuardCommitterFactory.java:createOutputCommitter(61)) - Using 
Commmitter StagingS3GuardCommitter{AbstractS3GuardCommitter{role=Task committer 
attempt_20170323184242_0000_r_000000_0outputPath=s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api,
 
workPath=file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0},
 finalOutputPath=null, conflictResolution=FAIL, 
wrappedCommitter=FileOutputCommitter{outputPath=file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/stevel/job_20170323184243_0000/staging-uploads,
 workPath=null, algorithmVersion=1, skipCleanup=false, 
ignoreCleanupFailures=true}} for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:44,448 [Executor task launch worker for task 0] INFO  
commit.DurationInfo (DurationInfo.java:<init>(33)) - Starting Task committer 
attempt_20170323184242_0000_r_000000_0: needsTaskCommit() Task 
attempt_20170323184242_0000_r_000000_0
2017-03-23 18:42:44,454 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:needsTaskCommit(849)) - 1 files to commit
2017-03-23 18:42:44,455 [Executor task launch worker for task 0] INFO  
commit.DurationInfo (DurationInfo.java:close(44)) - Task committer 
attempt_20170323184242_0000_r_000000_0: needsTaskCommit() Task 
attempt_20170323184242_0000_r_000000_0: 0:00:006
2017-03-23 18:42:44,457 [Executor task launch worker for task 0] INFO  
commit.DurationInfo (DurationInfo.java:<init>(33)) - Starting Task committer 
attempt_20170323184242_0000_r_000000_0: commit task 
attempt_20170323184242_0000_r_000000_0
2017-03-23 18:42:44,457 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:getTaskOutput(419)) - Scanning 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0
 for files to commit
2017-03-23 18:42:44,468 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:commitTaskInternal(885)) - Task committer 
attempt_20170323184242_0000_r_000000_0: commitTaskInternal
2017-03-23 18:42:44,468 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:commitTaskInternal(890)) - Task committer 
attempt_20170323184242_0000_r_000000_0: attempt path is 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0
2017-03-23 18:42:44,468 [Executor task launch worker for task 0] INFO  
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:commitTaskInternal(903)) - Task committer 
attempt_20170323184242_0000_r_000000_0: uploading from staging directory to S3
2017-03-23 18:42:44,468 [Executor task launch worker for task 0] INFO  
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:commitTaskInternal(904)) - Task committer 
attempt_20170323184242_0000_r_000000_0: Saving pending data information to 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/stevel/job_20170323184243_0000/staging-uploads/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0
2017-03-23 18:42:44,469 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:getOutputPath(1057)) - Task committer 
attempt_20170323184242_0000_r_000000_0: final output path is 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:44,493 [Executor task launch worker for task 0] DEBUG 
staging.S3Util (S3Util.java:multipartUpload(155)) - Initiating multipart upload 
from 
/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/test/mapred/local/job_20170323184243_0000/0/attempt_20170323184242_0000_r_000000_0/_temporary/0/_temporary/attempt_20170323184242_0000_r_000000_0/part-r-00000
 to 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000
 partition=null
2017-03-23 18:42:44,684 [Executor task launch worker for task 0] DEBUG 
staging.S3Util (S3Util.java:multipartUpload(177)) - File size is 515303, number 
of parts to upload = 1
2017-03-23 18:42:46,352 [Executor task launch worker for task 0] DEBUG 
staging.StagingS3GuardCommitter (StagingS3GuardCommitter.java:run(939)) - Task 
committer attempt_20170323184242_0000_r_000000_0: adding pending commit 
PendingUpload{ to 
's3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000'
 , 
uploadId='_P7n0N2FxTdVYiTBBFNB2Y9TOQgnbuKT1Lt.mwfx4pdmT6aK4E5XlxfpFZNwSUv0viLzNx45qvhJAqDyhSDqajozcfuFesRF2enq.bpkfc0-'}
2017-03-23 18:42:46,360 [Executor task launch worker for task 0] INFO  
output.FileOutputCommitter (FileOutputCommitter.java:commitTask(582)) - Saved 
output of task 'attempt_20170323184242_0000_r_000000_0' to 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/stevel/job_20170323184243_0000/staging-uploads/_temporary/0/task_20170323184242_0000_r_000000
2017-03-23 18:42:46,362 [Executor task launch worker for task 0] INFO  
staging.StagingS3GuardCommitter (StagingS3GuardCommitter.java:commitTask(866)) 
- Task committer attempt_20170323184242_0000_r_000000_0: upload file count: 1
2017-03-23 18:42:46,362 [Executor task launch worker for task 0] INFO  
commit.DurationInfo (DurationInfo.java:close(44)) - Task committer 
attempt_20170323184242_0000_r_000000_0: commit task 
attempt_20170323184242_0000_r_000000_0: 0:01:906
2017-03-23 18:42:46,363 [Executor task launch worker for task 0] INFO  
mapred.SparkHadoopMapRedUtil (Logging.scala:logInfo(54)) - 
attempt_20170323184242_0000_r_000000_0: Committed
2017-03-23 18:42:46,374 [Executor task launch worker for task 0] INFO  
executor.Executor (Logging.scala:logInfo(54)) - Finished task 0.0 in stage 0.0 
(TID 0). 1306 bytes result sent to driver
2017-03-23 18:42:46,380 [task-result-getter-0] INFO  scheduler.TaskSetManager 
(Logging.scala:logInfo(54)) - Finished task 0.0 in stage 0.0 (TID 0) in 2111 ms 
on localhost (executor driver) (1/1)
2017-03-23 18:42:46,381 [task-result-getter-0] INFO  
scheduler.TaskSchedulerImpl (Logging.scala:logInfo(54)) - Removed TaskSet 0.0, 
whose tasks have all completed, from pool 
2017-03-23 18:42:46,384 [dag-scheduler-event-loop] INFO  scheduler.DAGScheduler 
(Logging.scala:logInfo(54)) - ResultStage 0 (runJob at 
SparkHadoopMapReduceWriter.scala:88) finished in 2.124 s
2017-03-23 18:42:46,388 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 scheduler.DAGScheduler (Logging.scala:logInfo(54)) - Job 0 finished: runJob at 
SparkHadoopMapReduceWriter.scala:88, took 2.265104 s
2017-03-23 18:42:46,389 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 commit.DurationInfo (DurationInfo.java:<init>(33)) - Starting Task committer 
attempt_20170323184243_0000_m_000000_0: preparing to commit Job
2017-03-23 18:42:46,390 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:getThreadPool(1094)) - Task committer 
attempt_20170323184243_0000_m_000000_0: creating thread pool of size 8
2017-03-23 18:42:46,393 [s3-committer-pool-0] DEBUG staging.S3Util 
(S3Util.java:readPendingCommits(234)) - Reading pending commits in file 
file:/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/stevel/job_20170323184243_0000/staging-uploads/_temporary/0/task_20170323184242_0000_r_000000
2017-03-23 18:42:46,406 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api)
2017-03-23 18:42:47,006 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1905)) - Not Found: 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
2017-03-23 18:42:47,006 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 commit.DurationInfo (DurationInfo.java:close(44)) - Task committer 
attempt_20170323184243_0000_m_000000_0: preparing to commit Job: 0:00:617
2017-03-23 18:42:47,007 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 commit.DurationInfo (DurationInfo.java:<init>(33)) - Starting Task committer 
attempt_20170323184243_0000_m_000000_0: committing Job job_20170323184242_0001
2017-03-23 18:42:47,007 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG staging.StagingS3GuardCommitter 
(StagingS3GuardCommitter.java:commitJobInternal(649)) - Task committer 
attempt_20170323184243_0000_m_000000_0: committing the output of 1 task(s)
2017-03-23 18:42:47,010 [s3-committer-pool-1] DEBUG staging.S3Util 
(S3Util.java:finishCommit(101)) - Finish PendingUpload{ to 
's3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000'
 , 
uploadId='_P7n0N2FxTdVYiTBBFNB2Y9TOQgnbuKT1Lt.mwfx4pdmT6aK4E5XlxfpFZNwSUv0viLzNx45qvhJAqDyhSDqajozcfuFesRF2enq.bpkfc0-'}
2017-03-23 18:42:47,364 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary)
2017-03-23 18:42:47,941 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1905)) - Not Found: 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary
2017-03-23 18:42:47,942 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:delete(1418)) - Couldn't delete 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary
 - does not exist
2017-03-23 18:42:47,943 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 commit.DurationInfo (DurationInfo.java:close(44)) - Task committer 
attempt_20170323184243_0000_m_000000_0: committing Job job_20170323184242_0001: 
0:00:936
2017-03-23 18:42:47,943 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG commit.FileCommitActions (FileCommitActions.java:touchSuccessMarker(231)) 
- Touching success marker for job 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS
2017-03-23 18:42:47,944 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS)
2017-03-23 18:42:48,494 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1905)) - Not Found: 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS
2017-03-23 18:42:48,495 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG commit.DelayedCommitFSIntegration 
(DelayedCommitFSIntegration.java:getTracker(98)) - Created 
DefaultPutTracker{destKey='spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS'}
2017-03-23 18:42:48,498 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:createWriteOperationHelper(703)) - 
Created WriteOperationHelper {bucket=hwdev-steve-new, 
key='spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS'}
2017-03-23 18:42:48,514 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3ADataBlocks (S3ADataBlocks.java:startUpload(324)) - Start 
datablock[1] upload
2017-03-23 18:42:48,514 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3ADataBlocks (S3ADataBlocks.java:enterState(231)) - 
FileBlock{index=1, 
destFile=/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/s3ablock-0001-4746647579488981641.tmp,
 state=Writing, dataSize=0, limit=8388608}: entering state Upload
2017-03-23 18:42:48,518 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3AFileSystem 
(S3AFileSystem.java:putObjectDirect(1280)) - PUT 0 bytes to 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS
2017-03-23 18:42:48,519 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3AFileSystem 
(S3AFileSystem.java:incrementPutStartStatistics(1321)) - PUT start 0 bytes
2017-03-23 18:42:48,740 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3AFileSystem 
(S3AFileSystem.java:incrementPutCompletedStatistics(1338)) - PUT completed 
success=true; 0 bytes
2017-03-23 18:42:48,740 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3AFileSystem 
(S3AFileSystem.java:finishedWrite(2183)) - Finished write to 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS, len 0
2017-03-23 18:42:48,938 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3ADataBlocks 
(S3ADataBlocks.java:enterState(231)) - FileBlock{index=1, 
destFile=/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/s3ablock-0001-4746647579488981641.tmp,
 state=Upload, dataSize=0, limit=8388608}: entering state Closed
2017-03-23 18:42:48,938 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3ADataBlocks 
(S3ADataBlocks.java:close(346)) - Closed FileBlock{index=1, 
destFile=/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/s3ablock-0001-4746647579488981641.tmp,
 state=Closed, dataSize=0, limit=8388608}
2017-03-23 18:42:48,938 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3ADataBlocks 
(S3ADataBlocks.java:innerClose(886)) - Closing FileBlock{index=1, 
destFile=/Users/stevel/Projects/sparkwork/spark-cloud-examples/cloud-examples/target/tmp/s3ablock-0001-4746647579488981641.tmp,
 state=Closed, dataSize=0, limit=8388608}
2017-03-23 18:42:48,939 [s3a-transfer-shared-pool1-t1] DEBUG s3a.S3ADataBlocks 
(S3ADataBlocks.java:closeBlock(940)) - block[1]: closeBlock()
2017-03-23 18:42:48,939 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:writeSuccessful(2701)) - Successful 
write to 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS, len 0
2017-03-23 18:42:48,941 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary-1
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary-1)
2017-03-23 18:42:50,072 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1905)) - Not Found: 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary-1
2017-03-23 18:42:50,072 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:delete(1418)) - Couldn't delete 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_temporary-1
 - does not exist
2017-03-23 18:42:50,074 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 io.SparkHadoopMapReduceWriter (Logging.scala:logInfo(54)) - Job 
job_20170323184242_0001 committed.
2017-03-23 18:42:50,081 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 ui.SparkUI (Logging.scala:logInfo(54)) - Stopped Spark web UI at 
http://192.168.1.38:4040
2017-03-23 18:42:50,108 [dispatcher-event-loop-2] INFO  
spark.MapOutputTrackerMasterEndpoint (Logging.scala:logInfo(54)) - 
MapOutputTrackerMasterEndpoint stopped!
2017-03-23 18:42:50,117 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 memory.MemoryStore (Logging.scala:logInfo(54)) - MemoryStore cleared
2017-03-23 18:42:50,118 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManager (Logging.scala:logInfo(54)) - BlockManager stopped
2017-03-23 18:42:50,123 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 storage.BlockManagerMaster (Logging.scala:logInfo(54)) - BlockManagerMaster 
stopped
2017-03-23 18:42:50,125 [dispatcher-event-loop-0] INFO  
scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint 
(Logging.scala:logInfo(54)) - OutputCommitCoordinator stopped!
2017-03-23 18:42:50,128 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkContext (Logging.scala:logInfo(54)) - Successfully stopped 
SparkContext
2017-03-23 18:42:50,130 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 s3.S3ANumbersSuiteV2APISuite (CloudLogging.scala:logInfo(56)) - Filesystem 
statistics
 S3AFileSystem
{uri=s3a://hwdev-steve-new, workingDir=s3a://hwdev-steve-new/user/stevel, 
inputPolicy=sequential, partSize=8388608, enableMultiObjectsDelete=true, 
maxKeys=5000, readAhead=65536, blockSize=1048576, 
multiPartThreshold=2147483647, serverSideEncryptionAlgorithm='NONE', 
blockFactory=org.apache.hadoop.fs.s3a.S3ADataBlocks$DiskBlockFactory@6707a4bf, 
metastore=NullMetadataStore, authoritative=false, 
boundedExecutor=BlockingThreadPoolExecutorService
{SemaphoredDelegatingExecutor
{permitCount=25, available=25, waiting=0}, activeCount=0}, 
unboundedExecutor=java.util.concurrent.ThreadPoolExecutor@3fc7c734[Running, 
pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0], 
statistics 
{0 bytes read, 0 bytes written, 7 read ops, 0 large read ops, 3 write ops}, 
metrics 
{
{Context=S3AFileSystem} 
{FileSystemId=58e091c4-0ae9-4626-8ecf-0d1ac3ffdb27-hwdev-steve-new} 
{fsURI=s3a://hwdev-steve-new} 
{files_created=1} 
{files_copied=0} 
{files_copied_bytes=0} 
{files_deleted=0} 
{fake_directories_deleted=3} 
{directories_created=0} 
{directories_deleted=0} 
{ignored_errors=3} 
{op_copy_from_local_file=0} 
{op_exists=3} 
{op_get_file_status=7} 
{op_glob_status=0} 
{op_is_directory=0} 
{op_is_file=0} 
{op_list_files=0} 
{op_list_located_status=0} 
{op_list_status=0} 
{op_mkdirs=0} 
{op_rename=0} 
{object_copy_requests=0} 
{object_delete_requests=1} 
{object_list_requests=7} 
{object_continue_list_requests=0} 
{object_metadata_requests=14} 
{object_multipart_aborted=0} 
{object_put_bytes=0} 
{object_put_requests=1} 
{object_put_requests_completed=1} 
{stream_write_failures=0} 
{stream_write_block_uploads=0} 
{stream_write_block_uploads_committed=0} 
{stream_write_block_uploads_aborted=0} 
{stream_write_total_time=0} 
{stream_write_total_data=0} 
{object_put_requests_active=0} 
{object_put_bytes_pending=0} 
{stream_write_block_uploads_active=0} 
{stream_write_block_uploads_pending=1} 
{stream_write_block_uploads_data_pending=0} 
{stream_read_fully_operations=0} 
{stream_opened=0} 
{stream_bytes_skipped_on_seek=0} 
{stream_closed=0} 
{stream_bytes_backwards_on_seek=0} 
{stream_bytes_read=0} 
{stream_read_operations_incomplete=0} 
{stream_bytes_discarded_in_abort=0} 
{stream_close_operations=0} 
{stream_read_operations=0} 
{stream_aborted=0} 
{stream_forward_seek_operations=0} 
{stream_backward_seek_operations=0} 
{stream_seek_operations=0} 
{stream_bytes_read_in_close=0} 
{stream_read_exceptions=0} }}
2017-03-23 18:42:50,130 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:listFiles(2505)) - 
listFiles(s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api,
 true)
2017-03-23 18:42:50,130 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api
  (spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api)
2017-03-23 18:42:50,761 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1880)) - Found path 
as directory (with /): 0/1
2017-03-23 18:42:50,761 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1885)) - Summary: 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS 0
2017-03-23 18:42:50,763 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:listFiles(2518)) - Requesting all 
entries under spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/ 
with delimiter 'null'
2017-03-23 18:42:50,953 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (Listing.java:buildNextStatusBatch(296)) - 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS:
 spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS size=0
2017-03-23 18:42:50,954 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (Listing.java:buildNextStatusBatch(302)) - Adding: 
S3AFileStatus{path=s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS;
 isDirectory=false; length=0; replication=1; blocksize=1048576; 
modification_time=1490294569000; access_time=0; owner=stevel; group=stevel; 
permission=rw-rw-rw-; isSymlink=false} isEmptyDirectory=FALSE
2017-03-23 18:42:50,954 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (Listing.java:buildNextStatusBatch(296)) - 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000:
 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000
 size=515303
2017-03-23 18:42:50,955 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (Listing.java:buildNextStatusBatch(302)) - Adding: 
S3AFileStatus{path=s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000;
 isDirectory=false; length=515303; replication=1; blocksize=1048576; 
modification_time=1490294565000; access_time=0; owner=stevel; group=stevel; 
permission=rw-rw-rw-; isSymlink=false} isEmptyDirectory=FALSE
2017-03-23 18:42:50,955 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (Listing.java:buildNextStatusBatch(330)) - Added 2 
entries; ignored 0; hasNext=true; hasMoreObjects=false
2017-03-23 18:42:50,957 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 s3.S3ANumbersSuiteV2APISuite (CloudLogging.scala:logInfo(56)) - Contents of 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api:
  
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS
 [0]
  
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000
 [515303]
Dirs: 0, Files: 2, total bytes = 515303
2017-03-23 18:42:50,964 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] INFO 
 spark.SparkContext (Logging.scala:logInfo(54)) - SparkContext already stopped.
2017-03-23 18:42:50,966 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite  
(spark-cloud/S3ANumbersSuiteV2APISuite)
- SaveRDD
+ Cleaning s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite 
2017-03-23 18:42:51,514 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1880)) - Found path 
as directory (with /): 1/0
2017-03-23 18:42:51,514 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1888)) - Prefix: 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/
2017-03-23 18:42:51,514 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerGetFileStatus(1782)) - Getting 
path status for s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite  
(spark-cloud/S3ANumbersSuiteV2APISuite)
2017-03-23 18:42:52,059 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1880)) - Found path 
as directory (with /): 1/0
2017-03-23 18:42:52,059 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1888)) - Prefix: 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/
2017-03-23 18:42:52,060 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerDelete(1444)) - Delete path 
s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite - recursive true
2017-03-23 18:42:52,060 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerDelete(1449)) - delete: Path 
is a directory: s3a://hwdev-steve-new/spark-cloud/S3ANumbersSuiteV2APISuite
2017-03-23 18:42:52,060 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerDelete(1473)) - Getting 
objects for directory prefix spark-cloud/S3ANumbersSuiteV2APISuite/ to delete
2017-03-23 18:42:52,248 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerDelete(1483)) - Got object to 
delete spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/_SUCCESS
2017-03-23 18:42:52,248 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:innerDelete(1483)) - Got object to 
delete 
spark-cloud/S3ANumbersSuiteV2APISuite/numbers_rdd_tests_v2api/part-r-00000-job_20170323184243_0000
2017-03-23 18:42:53,499 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1880)) - Found path 
as directory (with /): 1/0
2017-03-23 18:42:53,500 [ScalaTest-main-running-S3ANumbersSuiteV2APISuite] 
DEBUG s3a.S3AFileSystem (S3AFileSystem.java:s3GetFileStatus(1888)) - Prefix: 
spark-cloud/S3ALineCountWritebackSuite/
Run completed in 15 seconds, 386 milliseconds.
{code}



> Add S3Guard committer for zero-rename commits to consistent S3 endpoints
> ------------------------------------------------------------------------
>
>                 Key: HADOOP-13786
>                 URL: https://issues.apache.org/jira/browse/HADOOP-13786
>             Project: Hadoop Common
>          Issue Type: New Feature
>          Components: fs/s3
>    Affects Versions: HADOOP-13345
>            Reporter: Steve Loughran
>            Assignee: Steve Loughran
>         Attachments: HADOOP-13786-HADOOP-13345-001.patch, 
> HADOOP-13786-HADOOP-13345-002.patch, HADOOP-13786-HADOOP-13345-003.patch, 
> HADOOP-13786-HADOOP-13345-004.patch, HADOOP-13786-HADOOP-13345-005.patch, 
> HADOOP-13786-HADOOP-13345-006.patch, HADOOP-13786-HADOOP-13345-006.patch, 
> HADOOP-13786-HADOOP-13345-007.patch, HADOOP-13786-HADOOP-13345-009.patch, 
> HADOOP-13786-HADOOP-13345-010.patch, HADOOP-13786-HADOOP-13345-011.patch, 
> HADOOP-13786-HADOOP-13345-012.patch, HADOOP-13786-HADOOP-13345-013.patch, 
> HADOOP-13786-HADOOP-13345-015.patch, HADOOP-13786-HADOOP-13345-016.patch, 
> HADOOP-13786-HADOOP-13345-017.patch, HADOOP-13786-HADOOP-13345-018.patch, 
> HADOOP-13786-HADOOP-13345-019.patch, s3committer-master.zip
>
>
> A goal of this code is "support O(1) commits to S3 repositories in the 
> presence of failures". Implement it, including whatever is needed to 
> demonstrate the correctness of the algorithm. (that is, assuming that s3guard 
> provides a consistent view of the presence/absence of blobs, show that we can 
> commit directly).
> I consider ourselves free to expose the blobstore-ness of the s3 output 
> streams (ie. not visible until the close()), if we need to use that to allow 
> us to abort commit operations.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to