[jira] [Created] (HADOOP-18298) Hadoop AWS | Staging committer Multipartupload not implemented properly

2022-06-16 Thread Ayush Goyal (Jira)
Ayush Goyal created HADOOP-18298:


 Summary: Hadoop AWS | Staging committer Multipartupload not 
implemented properly
 Key: HADOOP-18298
 URL: https://issues.apache.org/jira/browse/HADOOP-18298
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs/s3
Affects Versions: 3.3.1
Reporter: Ayush Goyal


In Hadoop aws staging 
committer(org.apache.hadoop.fs.s3a.commit.staging.StagingCommitter), Committer 
uploads files from local to s3(method- commitTaskInternal) which calls 
uploadFileToPendingCommit of CommitOperation to upload file using multipart 
upload.

 

Multipart upload consists of three steps-

1)Initialise multipartupload.

2) Breaks the file to part and upload Parts.

3) Merge all the parts of files and finalize multipart.

 

In the implementation of uploadFileToPendingCommit, first 2 steps are 
implemented. However, 3rd part is missing which leads to uploading the parts 
file but because it is not merged at the end of job no files are there in 
destination directory.

 

S3 logs before implement 3rd steps-

 
{code:java}
2022-05-30T13:49:31:000 [200 OK] s3.NewMultipartUpload 
localhost:9000/minio-feature-testing/spark-job/processed/output-parquet-staging-7/part-0-ce0a965f-622a-4950-bb4b-550470883134-c000-b552fb34-6156-4aa8-9085-679ad14fab6e.snappy.parquet?uploads
  240b:c1d1:123:664f:c5d2:2::               8.677ms      ↑ 137 B ↓ 724 B
2022-05-30T13:49:31:000 [200 OK] s3.PutObjectPart 
localhost:9000/minio-feature-testing/spark-job/processed/output-parquet-staging-7/part-0-ce0a965f-622a-4950-bb4b-550470883134-c000-b552fb34-6156-4aa8-9085-679ad14fab6e.snappy.parquet?uploadId=f3beae8e-3001-48be-9bc4-306b71940e50=1
  240b:c1d1:123:664f:c5d2:2::                443.156ms    ↑ 51 KiB ↓ 325 B
2022-05-30T13:49:32:000 [200 OK] s3.ListObjectsV2 
localhost:9000/minio-feature-testing/?list-type=2=%2F=2=spark-job%2Fprocessed%2Foutput-parquet-staging-7%2F_SUCCESS%2F=false
  240b:c1d1:123:664f:c5d2:2::                3.414ms      ↑ 137 B ↓ 646 B
2022-05-30T13:49:32:000 [200 OK] s3.PutObject 
localhost:9000/minio-feature-testing/spark-job/processed/output-parquet-staging-7/_SUCCESS
 240b:c1d1:123:664f:c5d2:2::                52.734ms     ↑ 8.7 KiB ↓ 380 B
2022-05-30T13:49:32:000 [200 OK] s3.DeleteMultipleObjects 
localhost:9000/minio-feature-testing/?delete  240b:c1d1:123:664f:c5d2:2::       
         73.954ms     ↑ 350 B ↓ 432 B
2022-05-30T13:49:32:000 [404 Not Found] s3.HeadObject 
localhost:9000/minio-feature-testing/spark-job/processed/output-parquet-staging-7/_temporary
 240b:c1d1:123:664f:c5d2:2::                2.658ms      ↑ 137 B ↓ 291 B
2022-05-30T13:49:32:000 [200 OK] s3.ListObjectsV2 
localhost:9000/minio-feature-testing/?list-type=2=%2F=2=spark-job%2Fprocessed%2Foutput-parquet-staging-7%2F_temporary%2F=false
  240b:c1d1:123:664f:c5d2:2::                 4.807ms      ↑ 137 B ↓ 648 B
2022-05-30T13:49:32:000 [200 OK] s3.ListMultipartUploads 
localhost:9000/minio-feature-testing/?uploads=spark-job%2Fprocessed%2Foutput-parquet-staging-7%2F
  240b:c0e0:102:553e:b4c2:2::               1.081ms      ↑ 137 B ↓ 776 B
2022-05-30T13:49:32:000 [404 Not Found] s3.HeadObject 
localhost:9000/minio-feature-testing/spark-job/processed/output-parquet-staging-7/.spark-staging-ce0a965f-622a-4950-bb4b-550470883134
 240b:c1d1:123:664f:c5d2:2::                 5.68ms       ↑ 137 B ↓ 291 B
2022-05-30T13:49:32:000 [200 OK] s3.ListObjectsV2 
localhost:9000/minio-feature-testing/?list-type=2=%2F=2=spark-job%2Fprocessed%2Foutput-parquet-staging-7%2F.spark-staging-ce0a965f-622a-4950-bb4b-550470883134%2F=false
  240b:c1d1:123:664f:c5d2:2::              2.452ms      ↑ 137 B ↓ 689 B
  {code}
Here , After s3.PutObjectPart there is no completeMultipartupload call for 3rd 
step.

 

S3 logs after implement 3rd steps-

 
{code:java}
2022-06-17T10:56:12:000 [200 OK] s3.NewMultipartUpload 
localhost:9000/minio-feature-testing/spark-job/pm-processed/output-parquet-staging-39/day%3D23/hour%3D16/quarter%3D0/part-4-d0b529ca-112f-43f2-a7dd-44de4db6aa7f-dffa7213-d492-48f9-9e6a-fb08bc81ceeb.c000.snappy.parquet?uploads
  240b:c1d1:123:664f:c5d2:2::               9.116ms      ↑ 137 B ↓ 750 B
2022-06-17T10:56:12:000 [200 OK] s3.NewMultipartUpload 
localhost:9000/minio-feature-testing/spark-job/pm-processed/output-parquet-staging-39/day%3D23/hour%3D15/quarter%3D45/part-4-d0b529ca-112f-43f2-a7dd-44de4db6aa7f-dffa7213-d492-48f9-9e6a-fb08bc81ceeb.c000.snappy.parquet?uploads
  240b:c1d1:123:664f:c5d2:2::               9.416ms      ↑ 137 B ↓ 751 B
2022-06-17T10:56:12:000 [200 OK] s3.NewMultipartUpload 
localhost:9000/minio-feature-testing/spark-job/pm-processed/output-parquet-staging-39/day%3D23/hour%3D16/quarter%3D45/part-4-d0b529ca-112f-43f2-a7dd-44de4db6aa7f-dffa7213-d492-48f9-9e6a-fb08bc81ceeb.c000.snappy.parquet?uploads
  240b:c1d1:123:664f:c5d2:2::               8.506ms      ↑ 137 B ↓ 751 B

Apache Hadoop qbt Report: branch-3.3+JDK8 on Linux/x86_64

2022-06-16 Thread Apache Jenkins Server
For more details, see 
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/

[Jun 9, 2022 1:42:04 PM] (Akira Ajisaka) HDFS-16576. Remove unused imports in 
HDFS project (#4389)
[Jun 10, 2022 4:03:25 PM] (Steve Loughran) YARN-11173. remove redeclaration of 
os-maven-plugin.version from yarn-csi (#4417)
[Jun 10, 2022 7:02:47 PM] (Chris Nauroth) HDFS-16623. Avoid 
IllegalArgumentException in LifelineSender (#4409)




-1 overall


The following subsystems voted -1:
blanks pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

XML :

   Parsing Error(s): 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-excerpt.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags2.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-sample-output.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/fair-scheduler-invalid.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/yarn-site-with-invalid-allocation-file-ref.xml
 

Failed junit tests :

   hadoop.hdfs.server.federation.router.TestRouterRpc 
   hadoop.hdfs.server.federation.router.TestRouterRpcMultiDestination 
   hadoop.yarn.server.resourcemanager.TestClientRMTokens 
   hadoop.yarn.client.api.impl.TestAMRMClient 
   hadoop.mapred.TestLocalDistributedCacheManager 
   hadoop.mapreduce.security.TestJHSSecurity 
  

   cc:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/results-compile-cc-root.txt
 [48K]

   javac:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/results-compile-javac-root.txt
 [388K]

   blanks:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/blanks-eol.txt
 [13M]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/blanks-tabs.txt
 [2.0M]

   checkstyle:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/results-checkstyle-root.txt
 [14M]

   pathlen:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/results-pathlen.txt
 [16K]

   pylint:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/results-pylint.txt
 [20K]

   shellcheck:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/results-shellcheck.txt
 [20K]

   xml:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/xml.txt
 [24K]

   javadoc:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/results-javadoc-javadoc-root.txt
 [1.1M]

   unit:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt
 [60K]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 [168K]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt
 [52K]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/patch-unit-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-common.txt
 [32K]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-branch-3.3-java8-linux-x86_64/60/artifact/out/patch-unit-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-jobclient.txt
 [96K]

Powered by Apache Yetus 0.14.0-SNAPSHOT   https://yetus.apache.org

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org

[jira] [Created] (HADOOP-18297) Upgrade dependency-check-maven to 7.1.1

2022-06-16 Thread Ashutosh Gupta (Jira)
Ashutosh Gupta created HADOOP-18297:
---

 Summary: Upgrade dependency-check-maven to 7.1.1
 Key: HADOOP-18297
 URL: https://issues.apache.org/jira/browse/HADOOP-18297
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Affects Versions: 3.3.3
Reporter: Ashutosh Gupta
Assignee: Ashutosh Gupta


The OWASP dependency-check-maven Plugin version has corrected various false 
positives in 7.1.1. We can upgrade to it.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18289) Remove WhiteBox in hadoop-kms module.

2022-06-16 Thread Wei-Chiu Chuang (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang resolved HADOOP-18289.
--
Fix Version/s: 3.4.0
   Resolution: Fixed

Done. Thanks [~slfan1989]

> Remove WhiteBox in hadoop-kms module.
> -
>
> Key: HADOOP-18289
> URL: https://issues.apache.org/jira/browse/HADOOP-18289
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 3.4.0
>Reporter: fanshilun
>Assignee: fanshilun
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> WhiteBox is deprecated, try to remove this method in hadoop-kms.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86_64

2022-06-16 Thread Apache Jenkins Server
For more details, see 
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/

[Jun 15, 2022 9:58:10 AM] (noreply) HDFS-16469. Locate protoc-gen-hrpc across 
platforms (#4434)
[Jun 15, 2022 9:59:55 AM] (noreply) YARN-11078. Set env vars in a cross 
platform compatible way (#4432)
[Jun 15, 2022 1:16:24 PM] (noreply) HDFS-16628 RBF: Correct target directory 
when move to trash for kerberos login user. (#4424). Contributed by Xiping 
Zhang.
[Jun 15, 2022 7:03:26 PM] (noreply) HADOOP-18159. Bump cos_api-bundle to 5.6.69 
to update public-suffix-list.txt (#)




-1 overall


The following subsystems voted -1:
blanks pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

XML :

   Parsing Error(s): 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-excerpt.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags2.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-sample-output.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/fair-scheduler-invalid.xml
 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/yarn-site-with-invalid-allocation-file-ref.xml
 

Failed junit tests :

   
hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistReplicaPlacement 
   hadoop.yarn.server.resourcemanager.TestClientRMTokens 
   hadoop.mapred.TestLocalDistributedCacheManager 
   hadoop.mapreduce.security.TestJHSSecurity 
  

   cc:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/results-compile-cc-root.txt
 [96K]

   javac:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/results-compile-javac-root.txt
 [540K]

   blanks:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/blanks-eol.txt
 [13M]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/blanks-tabs.txt
 [2.0M]

   checkstyle:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/results-checkstyle-root.txt
 [14M]

   pathlen:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/results-pathlen.txt
 [16K]

   pylint:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/results-pylint.txt
 [20K]

   shellcheck:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/results-shellcheck.txt
 [28K]

   xml:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/xml.txt
 [24K]

   javadoc:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/results-javadoc-javadoc-root.txt
 [400K]

   unit:

  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
 [568K]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 [188K]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/patch-unit-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-common.txt
 [48K]
  
https://ci-hadoop.apache.org/job/hadoop-qbt-trunk-java8-linux-x86_64/893/artifact/out/patch-unit-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-jobclient.txt
 [100K]

Powered by Apache Yetus 0.14.0-SNAPSHOT   https://yetus.apache.org

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org

[jira] [Created] (HADOOP-18296) Memory fragmentation in ChecksumFileSystem Vectored IO implementation.

2022-06-16 Thread Mukund Thakur (Jira)
Mukund Thakur created HADOOP-18296:
--

 Summary: Memory fragmentation in ChecksumFileSystem Vectored IO 
implementation.
 Key: HADOOP-18296
 URL: https://issues.apache.org/jira/browse/HADOOP-18296
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: common
Affects Versions: 3.4.0
Reporter: Mukund Thakur


As we have implemented merging of ranges in the ChecksumFSInputChecker 
implementation of vectored IO api, it can lead to memory fragmentation. Let me 
explain by example.

 

Suppose client requests for 3 ranges. 

0-500, 700-1000 and 1200-1500.

Now because of merging, all the above ranges will get merged into one and we 
will allocate a big byte buffer of 0-1500 size but return sliced byte buffers 
for the desired ranges.

Now once the client is done reading all the ranges, it will only be able to 
free the memory for requested ranges and memory of the gaps will never be 
released for eg here (500-700 and 1000-1200).

 

Note this only happens for direct byte buffers.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18107) Vectored IO support for large S3 files.

2022-06-16 Thread Mukund Thakur (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukund Thakur resolved HADOOP-18107.

Resolution: Fixed

Merged in feature branch 
[https://github.com/apache/hadoop/commits/feature-vectored-io] 
 

> Vectored IO support for large S3 files. 
> 
>
> Key: HADOOP-18107
> URL: https://issues.apache.org/jira/browse/HADOOP-18107
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Mukund Thakur
>Assignee: Mukund Thakur
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> This effort would mostly be adding more tests for large files under scale 
> tests and see if any new issue surfaces. 



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18105) Implement a variant of ElasticByteBufferPool which uses weak references for garbage collection.

2022-06-16 Thread Mukund Thakur (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukund Thakur resolved HADOOP-18105.

Resolution: Fixed

Merged in feature branch 
[https://github.com/apache/hadoop/commits/feature-vectored-io] 
 

> Implement a variant of ElasticByteBufferPool which uses weak references for 
> garbage collection.
> ---
>
> Key: HADOOP-18105
> URL: https://issues.apache.org/jira/browse/HADOOP-18105
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: common, fs
>Reporter: Mukund Thakur
>Assignee: Mukund Thakur
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> Currently in hadoop codebase, we have two classes which implements byte 
> buffers pooling.
> One is ElasticByteBufferPool which doesn't use weak references and thus could 
> cause memory leaks in production environment. 
> Other is DirectBufferPool which uses weak references but doesn't support 
> caller's preference for either on-heap or off-heap buffers. 
>  
> The idea is to create an improved version of ElasticByteBufferPool by 
> subclassing it ( as it is marked as public and stable and used widely in hdfs 
> ) with essential functionalities required for effective buffer pooling. This 
> is important for the parent Vectored IO work.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-18295) Add S3A configuration property for `no_proxy` hosts

2022-06-16 Thread Sam Kramer (Jira)
Sam Kramer created HADOOP-18295:
---

 Summary: Add S3A configuration property for `no_proxy` hosts
 Key: HADOOP-18295
 URL: https://issues.apache.org/jira/browse/HADOOP-18295
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs/s3
Reporter: Sam Kramer


Seeing as there are configuration options for proxy host, port, username, and 
password, there should also be an option to be able to provide to the S3 client 
a list of hosts to not use the proxy for (i.e. `no_proxy`)

 

I'm happy to contribute the code, but figured I'd file a ticket first to see if 
this the hadoop community would be open to this idea or have any desire for 
this feature. 



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-18294) Ensure build folder exists before writing checksum file.ProtocRunner#writeChecksums

2022-06-16 Thread Ashutosh Gupta (Jira)
Ashutosh Gupta created HADOOP-18294:
---

 Summary: Ensure build folder exists before writing checksum 
file.ProtocRunner#writeChecksums
 Key: HADOOP-18294
 URL: https://issues.apache.org/jira/browse/HADOOP-18294
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 3.3.3
Reporter: Ashutosh Gupta
Assignee: Ashutosh Gupta


Ensure build folder exists before writing checksum 
file.ProtocRunner#writeChecksums



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Re: Where is aarch64 tgz for 3.3.2 and 3.3.3

2022-06-16 Thread Ayush Saxena
>
> I'm assuming that the arm docker image can do the build, either on my
> macbook or more likely an arm linux vm in someones cloud infrastructure?

Yes, it would be the same release script but should be run on an ARM
machine. The create-release script handles it here:
https://github.com/apache/hadoop/blob/1b25851ad9ca5d1fc12b67089bb5adaa53641245/dev-support/bin/create-release#L207-L210

if that's true, how do i get its artifacts into the release? is the idea I
> just upload a whole new binary .tar.gz file with a suffix to make clear its
> arm, while the rest of release (source, maven artifacts etc) comes from the
> main one?


Uploading the .tar.gz, the signature and the checksum files suffixed with
-aarch64.tar.gz and so should do. Can check this commit, just upload files
in a way that they get recognized by this code:
https://github.com/apache/hadoop-site/commit/7487c909e835c10f6fe24aeb52d9784c4f06bdea


for example
if 3.3.3 had arm release the paths would have looked the way here:
https://github.com/apache/hadoop-site/blob/asf-site/content/releases.html#L214-L216

-Ayush

On Thu, 16 Jun 2022 at 15:03, Steve Loughran  wrote:

> i'm going to cut a 3.3.4 release this week, i will try to do an arm binary
> too.
>
> on that topic: how do i do this?
>
> I'm assuming that the arm docker image can do the build, either on my
> macbook or more likely an arm linux vm in someones cloud infrastructure?
>
> if that's true, how do i get its artifacts into the release? is the idea I
> just upload a whole new binary .tar.gz file with a suffix to make clear its
> arm, while the rest of release (source, maven artifacts etc) comes from the
> main one?
>
> On Thu, 16 Jun 2022 at 09:56, Ayush Saxena  wrote:
>
>> Same question here:
>>
>> https://lists.apache.org/thread/rjn4zhdc30zrorsyr8jcsovmbl5hq6ol
>>
>> -Ayush
>>
>> Sent from my iPhone
>>
>> > On 16-Jun-2022, at 10:11 AM, Cheng Pan  wrote:
>> >
>> > Hi community,
>> >
>> > I'm using Hadoop 3.3.1 to build cross-platform Docker images, and when
>> > I'm trying to upgrade to 3.3.2 or 3.3.3, I found there is no aarch64
>> > tgz like 3.3.1[1]. Where can I get them?
>> >
>> > [1]
>> https://dlcdn.apache.org/hadoop/core/hadoop-3.3.1/hadoop-3.3.1-aarch64.tar.gz
>> >
>> > Thanks,
>> > Cheng Pan
>> >
>> > -
>> > To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
>> > For additional commands, e-mail: common-dev-h...@hadoop.apache.org
>> >
>>
>


[jira] [Created] (HADOOP-18293) Release Hadoop 3.3.4 critical fix update

2022-06-16 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-18293:
---

 Summary: Release Hadoop 3.3.4 critical fix update
 Key: HADOOP-18293
 URL: https://issues.apache.org/jira/browse/HADOOP-18293
 Project: Hadoop Common
  Issue Type: Task
  Components: build
Reporter: Steve Loughran
Assignee: Steve Loughran


Create a new release off the branch-3.3.3 line with a few more changes

* wrap up of security changes
* cut hadoop-cos out of hadoop-cloud-storage as its dependencies break s3a 
client...reinstate once the updated jar is tested




--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Re: Where is aarch64 tgz for 3.3.2 and 3.3.3

2022-06-16 Thread Steve Loughran
i'm going to cut a 3.3.4 release this week, i will try to do an arm binary
too.

on that topic: how do i do this?

I'm assuming that the arm docker image can do the build, either on my
macbook or more likely an arm linux vm in someones cloud infrastructure?

if that's true, how do i get its artifacts into the release? is the idea I
just upload a whole new binary .tar.gz file with a suffix to make clear its
arm, while the rest of release (source, maven artifacts etc) comes from the
main one?

On Thu, 16 Jun 2022 at 09:56, Ayush Saxena  wrote:

> Same question here:
>
> https://lists.apache.org/thread/rjn4zhdc30zrorsyr8jcsovmbl5hq6ol
>
> -Ayush
>
> Sent from my iPhone
>
> > On 16-Jun-2022, at 10:11 AM, Cheng Pan  wrote:
> >
> > Hi community,
> >
> > I'm using Hadoop 3.3.1 to build cross-platform Docker images, and when
> > I'm trying to upgrade to 3.3.2 or 3.3.3, I found there is no aarch64
> > tgz like 3.3.1[1]. Where can I get them?
> >
> > [1]
> https://dlcdn.apache.org/hadoop/core/hadoop-3.3.1/hadoop-3.3.1-aarch64.tar.gz
> >
> > Thanks,
> > Cheng Pan
> >
> > -
> > To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
> > For additional commands, e-mail: common-dev-h...@hadoop.apache.org
> >
>


Re: Where is aarch64 tgz for 3.3.2 and 3.3.3

2022-06-16 Thread Ayush Saxena
Same question here:

https://lists.apache.org/thread/rjn4zhdc30zrorsyr8jcsovmbl5hq6ol

-Ayush

Sent from my iPhone

> On 16-Jun-2022, at 10:11 AM, Cheng Pan  wrote:
> 
> Hi community,
> 
> I'm using Hadoop 3.3.1 to build cross-platform Docker images, and when
> I'm trying to upgrade to 3.3.2 or 3.3.3, I found there is no aarch64
> tgz like 3.3.1[1]. Where can I get them?
> 
> [1] 
> https://dlcdn.apache.org/hadoop/core/hadoop-3.3.1/hadoop-3.3.1-aarch64.tar.gz
> 
> Thanks,
> Cheng Pan
> 
> -
> To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
> For additional commands, e-mail: common-dev-h...@hadoop.apache.org
>