Hello, i have a 8 TB volume with multiple files with hundreds of GB each. I try to dedupe this because the first hundred GB of many files are identical. With 128KB blocksize with nofiemap and lookup-extends=no option, will take more then a week (only dedupe, previously hashed). So i tryed -b 100M but this returned me an error: "Blocksize is bounded ...".
The reason is that the blocksize is limit to #define MAX_BLOCKSIZE (1024U*1024) But i can't found any description why. -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to [email protected] More majordomo info at http://vger.kernel.org/majordomo-info.html
