Steve Had this problem last week and found this Microsoft Technet article. Fixed the problem after a reboot
http://support.microsoft.com/kb/30401 Regards, Allan Allan Mills Technology Operations 02 8835 8035 0422 208 031 Steven Harris <[email protected]> Sent by: "ADSM: Dist Stor Manager" <[email protected]> 01/05/2009 09:39 Please respond to "ADSM: Dist Stor Manager" <[email protected]> To [email protected] cc Subject Re: [ADSM-L] best backup method for millions of small files? Hi Norman Your post worries me, as I'm just implementing an email archive solution that will depend on windows journalling to back up some huge repositories. The particular product fills up "containers" that once filled never change, so the change rate will be low there, but there are also index files that will change often. Have you determined whether the memory issue is related to number of files or number of changes? Regards Steve Steven Harris TSM Admin, Sydney Australia "Gee, Norman" <[email protected] .GOV> To Sent by: "ADSM: [email protected] Dist Stor cc Manager" <[email protected] Subject .EDU> Re: [ADSM-L] best backup method for millions of small files? 01/05/2009 07:12 AM Please respond to "ADSM: Dist Stor Manager" <[email protected] .EDU> What options are there when journaling runs out of memory on a 32 bit Windows server? I have about 10 million files on one server that the journal engine runs out of memory. With memory efficient disk cache method and resource utilization 5, its runs out of memory, resource utilization of 4 runs too long. -----Original Message----- From: ADSM: Dist Stor Manager [mailto:[email protected]] On Behalf Of Huebner,Andy,FORT WORTH,IT Sent: Thursday, April 30, 2009 8:16 AM To: [email protected] Subject: Re: best backup method for millions of small files? You have a disk array copy of the data, is that located close or far? Have you considered a disk array snap shot also? If you perform a journaled file system backup and an image backup then you should be able to restore the image and then update the image with the file system restore. This might take a long time, I have never tried it. What failure are you trying to protect against? In our case we use the disk arrays to protect against a data center loss and a corrupt file system and a TSM file system backup to protect against the loss of a file. Our big ones are in the 10 million file range. Using a 64bit Windows server we can backup the file system in about 6 - 8 hours without journaling. We suspect we could get the time down to around 4 hours if the TSM server was not busy backing up 500 other nodes. To me the important thing is to figure out what you are protecting against with each thing you do. Also be sure and ask what the Recovery Point Objective (RPO) is. If it is less than 24 hours then array based solutions may be the best choice. Over 24 hours then TSM may be the best choice. Andy Huebner -----Original Message----- From: ADSM: Dist Stor Manager [mailto:[email protected]] On Behalf Of Mehdi Salehi Sent: Thursday, April 30, 2009 9:39 AM To: [email protected] Subject: Re: [ADSM-L] best backup method for millions of small files? Hi, None the two methods that you mean in the user's guide are suitable for my case. "Image+normal incremental" that you emphasized in your post means getting full image backups for example every week. For the incremental part, one file-based full backup is needed which is a nightmare for 20 millions. OK, if I accept the initial incremental backup time (that might take for days), what happens in restoration? Naturally, last image backup should be restored first and it will take A minutes. Provided that image backups are weekly, the progressive incremental backups of the week is about 6*20MB=120MB. Now imagine 120MB of 15-20K files are to be restored in filesystem with an incredibly big file address table and system should create an inode-like entry for each. If this step takes B minutes, the total restoration time would be A+B. (A+B/A) ratio is important and I will try to measure and share it with the group. Steven, your solution is excellent for ordinary filesystems with a limited number of files. But I think for millions of files, only backup/restore method that do not care how many files exist in the volume are feasible. Somehing like pure image backup (like Acronis image incremental backup) or the method that FastBack exploites. Your points are welcomed. Regards, Mehdi Salehi This e-mail (including any attachments) is confidential and may be legally privileged. If you are not an intended recipient or an authorized representative of an intended recipient, you are prohibited from using, copying or distributing the information in this e-mail or its attachments. If you have received this e-mail in error, please notify the sender immediately by return e-mail and delete all copies of this message and any attachments. Thank you. _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ All mail is subject to content scanning for possible violation of NSW Police Force Electronic Messaging Policy. All NSW Police Force employees are required to familiarise themselves with the content of the policy, found under Policies on the NSW Police Force Intranet. _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ The information contained in this email is intended for the named recipient(s) only. It may contain private, confidential, copyright or legally privileged information. If you are not the intended recipient or you have received this email by mistake, please reply to the author and delete this email immediately. You must not copy, print, forward or distribute this email, nor place reliance on its contents. This email and any attachment have been virus scanned. However, you are requested to conduct a virus scan as well. No liability is accepted for any loss or damage resulting from a computer virus, or resulting from a delay or defect in transmission of this email or any attached file. This email does not constitute a representation by the NSW Police Force unless the author is legally entitled to do so.
