Hi again, i think i solved my issue.
I simply use the byte offset of the row currently read by the Mapper as the timestamp for the Put. This is unique for my input file, which contains one triple for each row. So the timestamps are unique.
Regards, Christopher Am 01.10.2011 13:19, schrieb Christopher Dorner:
Hallo, I am reading a File containing RDF triples in a Map-job. the RDF triples then are stored in a table, where columns can have lots of versions. So i need to store many values for one rowKey in the same column. I made the observation, that reading the file is very fast and thus some values are put into the table with the same timestamp and therefore overriding an existing value. How can i avoid that? The timestamps are not necessary for later usage. Could i simply use some sort of custom counter? How would that work in fully distributed mode? I am working on pseudo-distributed-mode for testing purpose right now. Thank You and Regards, Christopher
