I do have backups running on my own cron job using a script. However, from the mySQL manual: Normally, you should never have to run REPAIR TABLE <http://dev.mysql.com/doc/refman/5.1/en/repair-table.html>. However, if disaster strikes, this statement is very likely to get back all your data from a MyISAM table. If your tables become corrupted often, you should try to find the reason for it, to eliminate the need to use REPAIR TABLE<http://dev.mysql.com/doc/refman/5.1/en/repair-table.html> .
I would like to know what is causing the issue, since I never had it for over a year and then suddenly occurred. I'm going to try updating to 1.13.4 from 1.13.3 to see if that helps. Any tips would be great and thanks for the help I already got everyone! On Fri, Feb 13, 2009 at 11:01 AM, Henny Savenije < [email protected]> wrote: > Actually there are programs which do that. > > http://www.mysqldumper.de/en/ > > But that's beside the point. There are also programs out there which > optimize databases server wide. Don't know anymore where I found > them, but I have them installed on my servers. > > At 12:32 AM 2/14/2009, you wrote: > >Ha, yeah, sure, I was just giving the structure and did not want to > >specify a time. You would of course need to replace this like you would > >need to replace everything else (e.g. "username"). I just wanted to give > >an idea for research, not a finished solution (which I cannot give, not > >knowing the specifics). I would dump once every day, but on a site with > >very little traffic once a week may be enough, while a site with very > >heavy traffic or a large database might want to dump individual tables > >separately (so as not to have huge dumps or exceed execution time limits) > ... > > > >I, personally, would have the cronjob call an .sh script that gives a > >unique name based on the date to each dump and deletes old dumps after a > >certain time (maybe 90 days). > > > > > > > > >Manfred Kooistra wrote: > > >> And (first of all): Set up a cronjob to dump your database on a > regular > > >> basis so whatever happens you won't loose your data. Kind of like > this: > > >> > > >> * * * * * mysqldump db_name -u username -ppassword > /home/username/ > > >backup.sql > > > > > >Shouldn't at least one of those * be different? As it is, it would be > > >started every minute... > > > > > > > > >_______________________________________________ > > >MediaWiki-l mailing list > > >[email protected] > > >https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > > > > > > >_______________________________________________ > >MediaWiki-l mailing list > >[email protected] > >https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > _ _ > (o) (o) > oOOO----(_)----OOOo--- > Henny (Lee Hae Kang) > ----------------------------- > http://www.henny-savenije.pe.kr Portal to all my sites > http://www.hendrick-hamel.henny-savenije.pe.kr (in English) Feel free > to discover Korea with Hendrick Hamel (1653-1666) > http://www.hendrick-hamel.henny-savenije.pe.kr/indexk2.htm In Korean > http://www.hendrick-hamel.henny-savenije.pe.kr/Dutch In Dutch > http://www.vos.henny-savenije.pe.kr Frits Vos Article about Witsen > and Eibokken and his first Korean-Dutch dictionary > http://www.cartography.henny-savenije.pe.kr (in English) Korea > through Western Cartographic eyes > http://www.hwasong.henny-savenije.pe.kr Hwasong the fortress in Suwon > http://www.oldKorea.henny-savenije.pe.kr Old Korea in pictures > http://www.british.henny-savenije.pe.kr A British encounter in Pusan > (1797) > http://www.genealogy.henny-savenije.pe.kr/ Genealogy > http://www.henny-savenije.pe.kr/phorum Bulletin board for Korean studies > > > > > _______________________________________________ > MediaWiki-l mailing list > [email protected] > https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > _______________________________________________ MediaWiki-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
