At 01:20 PM 4/14/2010, Carsten Pedersen wrote:
Been there, done that. It's a maintenance nightmare.
Why is it a maintenance nightmare? I've been using this technique for a
couple of years to store large amounts of data and it has been working just
fine. I have each table representing one year of data and I can go back and
access 30 years worth of data using either the individual table or a merge
table. Loading data is also faster than trying to store tens of millions of
rows of data into one table because the index may get too unbalanced.
Your method is preferred if there is no way to separate the data into
distinct tables. But with the example provided, it appeared (at least to
me), he was saving just one days worth of data. If that's the case, my
method would take only milliseconds to get rid of the old data. Your
suggestion of flagging the rows as deleted and then deleting them later
requires more work for the database and there needs to be a lull in order
to delete the data. He will also have to optimize the table to get rid of
the deleted rows and this requires the table to be locked.
Like I said, both methods will work. It depends on how the data is
organized and how much down time he can dedicate to the process in order to
determine which process is the best best .
Mike
Another idea: Have a separate "deleted" table with the IDs of the rows
that you consider deleted. Re-write your queries to do a
left-join-not-in-the-other-table agains the "delete" table. Then, either
wait for a maintenance window to delete the rows both the original table
and the "delete" table. Or remove just a few rows at a time.
The "deleted" table can be created with just a read lock on the original
table, and since it's going to be fairly small, the impact of stuffing it
with data is not going to be great.
It's a bit of a hassle to set up, but once done you don't have to worry
about creating and deleting tables every day.
/ Carsten
mos skrev:
It looks like you only want to keep the current data, perhaps the current
day's worth, and delete the old data.
I would store the data in separate MySIAM tables, each table would
represent a date, like D20100413 and D20100414. Your program will decide
which table to insert the data into by creating a current date variable
and now all data gets inserted to the table named by that variable. When
the older data is no longer needed, just drop the table. If you want to
keep the last 7 days of data, create a merge table of the last 7 tables.
When you drop the oldest table, redefine the merge table. You can
accomplish all this in just milliseconds.
Mike
At 08:08 AM 4/14/2010, you wrote:
Hi,
I am using MySQL version 4.1.12-log. All the databases on it are using
MyISAM database engine.
Every day, I delete almost 90000 rows on a table of 3 153 916 rows.
To delete the rows, I use a request like this : "DELETE QUICK FROM [table]
WHERE [column] < '2010-04-13 00:00:00' LIMIT 7500". I execute this request
until all the rows are delete.
This works but when I run the request, I can't access to the database (make
INSERT and SELECT requests) during I do the DELETE.
How can I do a "DELETE" without impact on INSERT and SELECT requests done on
the same time?
Regards,
David.
--
MySQL General Mailing List
For list archives: http://lists.mysql.com/mysql
To unsubscribe: http://lists.mysql.com/mysql?unsub=arch...@jab.org