Ruben Van Wanzeele created HBASE-28998: ------------------------------------------
Summary: Backup support for S3 broken by checksum added in HBASE28625 Key: HBASE-28998 URL: https://issues.apache.org/jira/browse/HBASE-28998 Project: HBase Issue Type: Bug Reporter: Ruben Van Wanzeele With the 2.6.1 version, we fail to backup to S3 because of the introduction of the checksum validation (introduced with HBASE-28625). Stacktrace: {code:java} Error: java.io.IOException: Checksum mismatch between hdfs://hdfsns/hbase/hbase/data/SYSTEM/CATALOG/b884434cc05aae3a21c0d0723173ce02/0/43b0e3a7b608441eab7dbce2782511bf and s3a://product-eks-v2-brt-master-574-backup/hbase/backup_1732545852227/SYSTEM/CATALOG/archive/data/SYSTEM/CATALOG/b884434cc05aae3a21c0d0723173ce02/0/43b0e3a7b608441eab7dbce2782511bf. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) {code} I think the solution is to only do the checksum validation if the filesystems are the same. -- This message was sent by Atlassian Jira (v8.20.10#820010)