During an import from a large table, we want to avoid using too many mappers, as that would put too much load on the database.
However that winds up generating very large transactions, e.g. 30M+ rows per request. Which in turn can cause a transaction timeout, if it takes longer than about 3000 seconds. Is there any way to control both the load (number of parallel requests) and the size of each request? Thanks, -- Ken -------------------------- Ken Krugler +1 530-210-6378 http://bixolabs.com custom big data solutions & training Hadoop, Cascading, Mahout & Solr