This approach has me concerned about database coherence as it takes some time to run. I have reports which export data to another application and those take at least 12min to run when the machine is not doing anything else. Even running a t-log file for the whole thing takes several minutes. I would trust a transaction log file more - though that adds some more complexity to the mix. If users are making changes to the db- when does the transaction log file stop ? Also, if the server is caching writes to the transaction log and gets clobbered before closing it, are the files still usable, especially if it dies during some middle part of the process before all the tables have been exported ?

I admit I am not sure how to dynamically assign file names or manually create a transaction log from within a report.

At this point, I am leaning towards the Acronis solution. For either solution, I want to write to an external disk (can't have a power spike kill the primary and backup hardware) and possibly put it on a network attached drive. Even better would be to mirror that to the cloud in some manner to have it completely offsite. That raises other security and access issues but does address the dataloss problem.

Bob

At 03:16 PM 05/10/2013, Tim Rude wrote:
What about writing a report that exports the data to file(s) that could then be imported to rebuild the database in case of a disaster. Depending on how ambitious you felt like being, the data could be exported to multiple files (one for each panel), or even better would be to create a synthetic transaction log file with all of the data in one big file. You could have the report dynamically assign the filename(s) based on the current date and time. Then you could have a scheduled task on one of the machines run the report periodically, using the command line.

That should suffice to keep everyone from having to exit from the database every couple of hours, and provide some protection for the data being entered.

The downside is the time it takes to create the report to export all of the data, and the need to update the report if the database structure changes. It would be much easier if there was a command-line option to export the entire database to a T-log file (doing the same thing as Shift-F9, A), but there isn't that I know of.

Tim Rude
----- Original Message -----
From: <mailto:[email protected]>Bob DeRosier
To: <mailto:[email protected]>[email protected]
Sent: Friday, May 10, 2013 12:49 PM
Subject: [Dataperf] How to do multiple backups during work day ?

Hi All;

I have an application that the users would like to change their workflow. One of the ideas would require multiple backups during the day, just to avoid losing work. When I backup DP, I usually have everybody off the system, then copy the files to the backup location and work from there. Obviously, this would be a bit cumbersome with people working on it all day, especially if I want to do this multiple times a day. Any thoughts ?

If this were running on a VM or storage area network or some such with a SQL server, I would just schedule a job to quiesce the database and take a snapshot of the files at that time and go from there. I don't know of a similar command for DP, that is if there are pending writes or the database is in an incoherent state where the records are out of synch, I don't know how to force them to be in synch without just throwing everybody out of the system. As it stands, it is running on XP in a low budget environment, no SAN, no NAS, no Windows server with VSS, no fancy backup appliance, just a non-profit with a small database and a need..

thanks for any advice.

Bob


----------
_______________________________________________
Dataperf mailing list
[email protected]
http://lists.dataperfect.nl/mailman/listinfo/dataperf

_______________________________________________
Dataperf mailing list
[email protected]
http://lists.dataperfect.nl/mailman/listinfo/dataperf
_______________________________________________
Dataperf mailing list
[email protected]
http://lists.dataperfect.nl/mailman/listinfo/dataperf

Reply via email to