I'm working on an application where we have a central database server and a bunch of clients that are disconnected most of the time, need to maintain a local copy of the central database. The client databases are based on One$DB since it has to be lightweight. The client does not access the server database directly, but through a server application. Periodically (about once a week) the clients will connect to the central database and retrieve updates made to the central database so that their local database will be up-to-date, as well as send local updates back to the server. A lot of these clients will be connecting through a dial-up connection, and some of the tables can get quite large, so just retrieving all rows in a table is not an option when synchronizing.

This is how I currently have things designed on the server side:

Each table has a revision column, which is set to a sequence value every time a row is inserted or updated (deletes are translated into updates that "hide" the row by setting an "enabled" column to false), as you can see in the following statements:

CREATE TABLE codes (
   id SERIAL PRIMARY KEY,
   created TIMESTAMP DEFAULT CURRENT_TIMESTAMP NOT NULL,
   updated TIMESTAMP,
   revision SERIAL,
   enabled BOOLEAN DEFAULT TRUE NOT NULL,
   name VARCHAR(8) NOT NULL,
   description VARCHAR(50)
) WITHOUT OIDS;

CREATE INDEX codes_revision_idx ON codes (revision);

CREATE FUNCTION row_update_codes() RETURNS trigger AS '
   BEGIN
       NEW.revision = nextval(''codes_revision_seq'');
       NEW.updated = current_timestamp;
       RETURN NEW;
   END;
' LANGUAGE plpgsql;

CREATE TRIGGER update_row BEFORE INSERT OR UPDATE ON codes
   FOR EACH ROW EXECUTE PROCEDURE row_update_codes();

CREATE RULE delete_row AS
   ON DELETE TO codes DO INSTEAD
   UPDATE codes SET enabled = FALSE WHERE id = OLD.id;

To fetch all updates since the last synchronization, the client would calculated a value for $lastrevision by running this query on its local database:
SELECT max(revision) AS lastrevision FROM codes;


It would then fetch all updated rows by running this query against the server:
SELECT * FROM codes WHERE revision > $lastrevision;


This seems to work, except there exists a race condition. Consider the following series of events (in chronological order):

1. Initially, in the codes table there's a row with id=1, revision=1,
and a row with id=2, revision=2
2. Client A begins a transaction, and updates the row with id=1,
resulting in revision=3 for that row
3. Client B begins a transaction, and updates the row with id=2,
resulting in revision=4 for that row
4. Client B commits the transaction
5. Client C (which has $lastrevision=2 in its local database)
synchronizes with the database by doing SELECT * FROM codes WHERE
revision > 2; and retrieves client B's update to the row with
id=2, revision=4 (it doesn't yet see the update from client A)
6. Client A commits the transaction
7. Some time later, Client C synchronizes with the database again. $lastrevision for its database is now 4, so doing SELECT * FROM
codes WHERE revision > 4; does not retrieve any rows. So client C
never sees client A's update to the row with id=1


Essentially, the race condition occurs when the order of clients committing transactions (i.e. the updates becoming visible to other clients) differs from the order of clients generating sequence values. Do you guys have any suggestions on how to avoid this race condition, or maybe a more elegant way to synchronize the clients with the server?

I was thinking about doing the following to solve this problem. I'd create a new table, e.g. codes_active_txns, with a sequence_value column. Before the server application starts any transaction involving an insert/update/delete to codes, it would retrieve nextval('codes_revision_seq') as $seq_val, insert it into codes_active_txns, commit, begin another transaction, and do its inserts/updates, setting revision = $seq_val for each row that it touches. Once it's done with the table updates, it would commit the transaction, and delete the row containing $seq_val from codes_active_txns, and commit that.

The server would calculate $lastrevision with:
SELECT max(revision) FROM codes WHERE revision < (SELECT COALESCE(min(sequence_value), 2147483647) FROM codes_active_txns);
and send this to the client. $lastrevision would contain the highest revision for which there are no active (non-committed) transactions with a revision lower than $lastrevision. The client would save this value, and pass it back to the server during the next synchronization session so that the server can figure out what rows it needs to send to the client with SELECT * FROM codes WHERE revision > $lastrevision;. Occasionally the same row might be sent to the client twice during two synchronization sessions, but that's no big deal because that number should be low and the client can handle these duplicate rows.


Now, this design isn't very elegant for several reasons. One, it requires the server application to manually keep track of active transactions, which is really something that should belong in the database. Also, if the server application crashes midway, and never removes its row from codes_active_txns, the list of duplicate rows being sent will grow larger and larger, so we'll need some way to periodically (e.g. every 5 minutes) clean stale values from codes_active_txns.

So, my questions are:

1. Can you guys think of a simpler and/or more elegant way to send
only recently updated rows to the client?
2. If not:
1. Is there any way to move this logic that maintains
codes_active_txns from the server application to the server
database? E.g. something like a trigger on BEGIN and
COMMIT, which as far as I know does not exist.
2. How would I go about identifying sessions that have died off
but did not remove their row from codes_active_txns? I can
add extra columns to store connection/backend information. I think pg_locks comes close to providing this information
(we'd just store the backend pid in codes_active_txns and
have the cleanup script check pg_locks to make sure that pid
is still connected), except I don't think this would work if
I were to use something like pgpool to connect to the
database. The other option is adding a timestamp column and
removing old rows, but it is possible that occasionally a
transaction runs for a long time, so we don't want to remove
rows for such long-running active transactions.


I've also thought of the following alternatives, but I don't think they are feasible:

  1. Lock the entire codes table when making an update.  Our
     application will have a high level of concurrent inserts and
     updates during peak times, and this would just about kill its
     throughput.
  2. Having the client send a list of all ids and revisions for each
     row in its local database to the server.  This would scale
     poorly.  However, it is OK to send a small (bound) list of
     ids/revisions.

I really appreciate any insights you guys might be able to give me on this problem.

Thanks,

Alex


---------------------------(end of broadcast)--------------------------- TIP 5: Have you checked our extensive FAQ?

http://www.postgresql.org/docs/faq

Reply via email to