Bob, why not use a NAS? I have a Qnap in the office with two 3tb drives. 7 DP users in the office and two on VPN / RDP outside the office. Every input is mirrored on the second drive.
Regards, Gerard van Loenhout 2013/5/12 Bob DeRosier <[email protected]> > > > This approach has me concerned about database coherence as it takes some > time to run. I have reports which export data to another application and > those take at least 12min to run when the machine is not doing anything > else. Even running a t-log file for the whole thing takes several minutes. > I would trust a transaction log file more - though that adds some more > complexity to the mix. If users are making changes to the db- when does the > transaction log file stop ? Also, if the server is caching writes to the > transaction log and gets clobbered before closing it, are the files still > usable, especially if it dies during some middle part of the process before > all the tables have been exported ? > > I admit I am not sure how to dynamically assign file names or manually > create a transaction log from within a report. > > At this point, I am leaning towards the Acronis solution. For either > solution, I want to write to an external disk (can't have a power spike > kill the primary and backup hardware) and possibly put it on a network > attached drive. Even better would be to mirror that to the cloud in some > manner to have it completely offsite. That raises other security and > access issues but does address the dataloss problem. > > Bob > > > At 03:16 PM 05/10/2013, Tim Rude wrote: > > What about writing a report that exports the data to file(s) that could > then be imported to rebuild the database in case of a disaster. Depending > on how ambitious you felt like being, the data could be exported to > multiple files (one for each panel), or even better would be to create a > synthetic transaction log file with all of the data in one big file. You > could have the report dynamically assign the filename(s) based on the > current date and time. Then you could have a scheduled task on one of the > machines run the report periodically, using the command line. > > That should suffice to keep everyone from having to exit from the database > every couple of hours, and provide some protection for the data being > entered. > > The downside is the time it takes to create the report to export all of > the data, and the need to update the report if the database structure > changes. It would be much easier if there was a command-line option to > export the entire database to a T-log file (doing the same thing as > Shift-F9, A), but there isn't that I know of. > > Tim Rude > ----- Original Message ----- > From: Bob DeRosier <[email protected]> > To: [email protected] > Sent: Friday, May 10, 2013 12:49 PM > Subject: [Dataperf] How to do multiple backups during work day ? > > Hi All; > > I have an application that the users would like to change their workflow. > One of the ideas would require multiple backups during the day, just to > avoid losing work. When I backup DP, I usually have everybody off the > system, then copy the files to the backup location and work from there. > Obviously, this would be a bit cumbersome with people working on it all > day, especially if I want to do this multiple times a day. Any thoughts ? > > If this were running on a VM or storage area network or some such with a > SQL server, I would just schedule a job to quiesce the database and take a > snapshot of the files at that time and go from there. I don't know of a > similar command for DP, that is if there are pending writes or the database > is in an incoherent state where the records are out of synch, I don't know > how to force them to be in synch without just throwing everybody out of the > system. As it stands, it is running on XP in a low budget environment, > no SAN, no NAS, no Windows server with VSS, no fancy backup appliance, > just a non-profit with a small database and a need.. > > thanks for any advice. > > Bob > > ------------------------------ > _______________________________________________ > Dataperf mailing list > [email protected] > http://lists.dataperfect.nl/mailman/listinfo/dataperf > > _______________________________________________ > Dataperf mailing list > [email protected] > http://lists.dataperfect.nl/mailman/listinfo/dataperf > > > _______________________________________________ > Dataperf mailing list > [email protected] > http://lists.dataperfect.nl/mailman/listinfo/dataperf > >
_______________________________________________ Dataperf mailing list [email protected] http://lists.dataperfect.nl/mailman/listinfo/dataperf
