Hi, Can you also post the md5 (or sha1, or ...) of the big file. The connection is frequently interrupted, and I cannot rely on the downloaded file without a check.
Jean-Pierre Jelle de Jong wrote: > Hi Jean-Pierre, > > Thank you! > > The reparse-tags.gz file: > https://powermail.nu/nextcloud/index.php/s/fS6Y6bpzoMgPiZ0 > > Generated by running: getfattr -e hex -n system.ntfs_reparse_data -R > /mnt/sr7-sdb2/ 2> /dev/null | grep ntfs_reparse_data | gzip > > /root/reparse-tags.gz > > Kind regards, > > Jelle de Jong > > On 23/02/17 12:07, Jean-Pierre André wrote: >> Hi, >> >> Jelle de Jong wrote: >>> Dear Jean-Pierre, >>> >>> I thought version 1.2.1 of the plug-in was working so I took it further >>> into production, but during backups with rdiff-backup and guestmount it >>> created a 100% cpu load in qemu process that stayed there for days until >>> I killed them, I tested this twice. So I went back to a xpart/mount -t >>> ntfs command and found more "Bad stream for offset" and found that the >>> /sbin/mount.ntfs-3g command was running at 100% cpu load and hanged >>> there. >> >> Too bad. >> >>> I have added the whole Stream directory here: (1.1GB) >>> https://powermail.nu/nextcloud/index.php/s/vbq85qZ2wcVYxrG >>> >>> Separate stream file: stream.data.full.000c0000.00020001.gz >>> https://powermail.nu/nextcloud/index.php/s/QinV51XE4jrAH7a >>> >>> All the commands I used: >>> http://paste.debian.net/plainh/c0ea5950 >>> >>> I do not know how to get the reparse tags of all the files, maybe you >>> can help me how to get all the information you need. >> >> Just use option -R on the base directory : >> >> getfattr -e hex -n system.ntfs_reparse_data -R base-dir >> >> Notes : >> 1) files with no reparse tags (those which are not deduplicated) >> will throw an error >> 2) this will output the file names, which you might not want >> to disclose. Fortunately I do not need them for now. >> >> So you may append to the above command : >> >> 2> /dev/null | grep ntfs_reparse_data | gz > reparse-tags.gz >> >> With that, I will be able to build a configuration similar >> to yours... apart from the files themselves. >> >> Regards >> >> Jean-Pierre >> >>> >>> Thank you for your help! >>> >>> Kind regards, >>> >>> Jelle de Jong >>> >>> On 14/02/17 15:55, Jean-Pierre André wrote: >>>> Hi, >>>> >>>> Jelle de Jong wrote: >>>>> Hi Jean-Pierre, >>>>> >>>>> If we have to switch to Windows 2012 and thereby having an environment >>>>> similar to yours then we can switch to an other Windows version. >>>> >>>> I do not have any Windows Server, and my analysis >>>> and tests are based on an unofficial deduplication >>>> package which was adapted to Windows 10 Pro. >>>> >>>> A few months ago, following a bug report, I had to >>>> make changes for Windows Server 2012 which uses an >>>> older data format, and my only experience about this >>>> format is related to this report. So switching to >>>> Windows 2012 is not guaranteed to make debugging easier. >>>> >>>>> We are running out of disk space here so if switching Windows versions >>>>> makes the process of having data deduplication working easer then me >>>>> know. >>>> >>>> I have not yet analyzed your latest report, but it >>>> would probably be useful I build a full copy of >>>> non-user data from your partition : >>>> - the reparse tags of all your files, >>>> - all the "*.ccc" files in the Stream directory >>>> >>>> Do not do it now, I must first dig into the data you >>>> posted. >>>> >>>> Regards >>>> >>>> Jean-Pierre >>>> >>>> >>>>> Kind regards, >>>>> >>>>> Jelle de Jong >>>>> >>>>> On 09/02/17 13:46, Jelle de Jong wrote: >>>>>> Hi Jean-Pierre, >>>>>> >>>>>> In case you are wondering: >>>>>> >>>>>> I am using data deduplication in Windows 2016 for my test environment >>>>>> iso: >>>>>> SW_DVD9_Win_Svr_STD_Core_and_DataCtr_Core_2016_64Bit_English_-2_MLF_X21-22843.ISO >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Kind regards, >>>>>> >>>>>> Jelle de Jong >>>>>> >>>>>> On 09/02/17 11:41, Jean-Pierre André wrote: >>>>>>> Hi, >>>>>>> >>>>>>> Jelle de Jong wrote: >>>>>>>> Hi Jean-Pierre, >>>>>>>> >>>>>>>> Thank you! >>>>>>>> >>>>>>>> The new plug-in seems to work for now, I am moving it into testing >>>>>>>> phase >>>>>>>> with-in our production back-up scripts. >>>>>>> >>>>>>> Please wait a few hours, I have found a bug which >>>>>>> I have fixed. I am currently inserting your data >>>>>>> into my test base in order to rerun all my tests. >>>>>>> >>>>>>>> Will you release the source code eventually, would like to write a >>>>>>>> blog >>>>>>>> post about how to add the support. >>>>>>> >>>>>>> What exactly do you mean ? If it is about how to >>>>>>> collect the data in a unsupported condition, it is >>>>>>> difficult, because unsupported generally means >>>>>>> unknown territory... >>>>>>> >>>>>>>> What do you think the changes are of the plug-in stop working >>>>>>>> again? >>>>>>> >>>>>>> (assuming a typo changes -> chances) >>>>>>> Your files were in a condition not met before : data >>>>>>> has been relocated according to a logic I do not fully >>>>>>> understand. Maybe this is an intermediate step in the >>>>>>> process of updating the files, anyway this can happen. >>>>>>> >>>>>>> The situation I am facing is that I have a single >>>>>>> example from which it is difficult to derive the rules. >>>>>>> So yes, the plugin may stop working again. >>>>>>> >>>>>>> Note : there are strict consistency checks in the plugin, >>>>>>> so it is unlikely you read invalid data. Moreover if >>>>>>> you only mount read-only you cannot damage the deduplicated >>>>>>> partition. >>>>>>> >>>>>>>> We do not have an automatic test running to verify the back-ups at >>>>>>>> this >>>>>>>> moment _yet_, so if the plug-in stops working, incremental >>>>>>>> file-based >>>>>>>> back-ups with empty files will slowly get in the back-ups this >>>>>>>> way :| >>>>>>> >>>>>>> Usually a deduplicated partition is only used for backups, >>>>>>> and reading from backups is only for recovering former >>>>>>> versions of files (on demand). >>>>>>> >>>>>>> If you access deduplicated files with no human control, >>>>>>> you have to insert your own checks in the process. I >>>>>>> would at least check whether the size of the recovered >>>>>>> file is the same as the deduplicated one (also grep for >>>>>>> messages in the syslog). >>>>>>> >>>>>>> Regards >>>>>>> >>>>>>> Jean-Pierre >>>>>>> >>>>>>>> Again thank you for all your help so far! >>>>>>>> >>>>>>>> Kind regards, >>>>>>>> >>>>>>>> Jelle de Jong >>>>>>>> >>>>>>>> >>>>>>>> On 08/02/17 15:59, Jean-Pierre André wrote: >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> Can you please make a try with : >>>>>>>>> http://jp-andre.pagesperso-orange.fr/dedup120-beta.zip >>>>>>>>> >>>>>>>>> This is experimental and based on assumptions which have >>>>>>>>> to be clarified, but it should work in your environment. >>>>>>>>> >>>>>>>>> Regards >>>>>>>>> >>>>>>>>> Jean-Pierre >> >> > ------------------------------------------------------------------------------ Check out the vibrant tech community on one of the world's most engaging tech sites, SlashDot.org! http://sdm.link/slashdot _______________________________________________ ntfs-3g-devel mailing list ntfs-3g-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ntfs-3g-devel