Martin and team, Statement: "So the only way to solve your problem is to create a tool which parses the dump files and creates a checksum in a defined way so that they are comparable."
Agreed. My thoughts exactly. Thank-you, Luke Perkins -----Original Message----- From: Martin Furter [mailto:mfur...@bluewin.ch] Sent: Tuesday, January 24, 2017 19:56 To: lukeperk...@epicdgs.us Subject: Re: [PATCH] Issue #4668: Fixing the node key order during svnadmin dump On 01/25/2017 03:15 AM, Luke Perkins wrote: > Michael, > > I appreciate everyone's audience on this issue. I have not felt a need to be > directly involved in the subversion system mainly because it works so well. > This is the first time in 10 years I have felt the need to get directly > involved in the SVN development team. > > Statement: " As a bug report alone, this one seems pretty easy: > Closed/INVALID." > > I completely disagree with this statement. I have nearly 300GB of dump files > used as a means of backing up my repositories. Some of these dump files are > 10 years old. The incremental SVN dump file is automatically generated at > each and every commit. After these incremental SVN dump files are created, > they are copied and distributed to offsite locations. That way if my server > farm crashes, I have a means of assured recovery. > > Every month I run sha512sum integrity checks on both the dump files (remotely > located in 3 different locations) and the dump file produced by the > subversion server. Transferring thousands of 128 byte files is a much better > option than transferring thousands of MB dump files over the internet to > remote locations. This method and automated scripts have worked for 10 years. > I have rebuilt my servers from the original dump files on at least 2 > occasions because of computer crashes. This provides me a sanity and > validation methodology so that I can spot problems quickly and rebuild before > things get out of hand. > > Asking me to redistribute 300GB of data to 3 different offsite (and remote) > locations, is not a good option. > > The SVN dump file has always been presented as the ultimate backup tool of > the subversion system. The integrity of the SVN dump file system is of > paramount importance. The whole reason why SVN exists in the first place is > "data integrity and traceability". The code was changed back in 2015, for > better or worse, and we need present solutions to address legacy backups. A stable order of header lines will solve your problem for now. But in the future somebody might add a new feature to subversion and a new header field to the dump files. This will break your checksums again. Back in the pre-1.0 days when I was working on svmdumptool i had the same troubles with changing headers and new fields. So the only way to solve your problem is to create a tool which parses the dump files and creates a checksum in a defined way so that they are comparable. - Martin