For reference:
http://groups.google.com/group/mongodb-user/browse_thread/thread/f137fe0a1a81a3cd

Looks like it is not yet supported.  Ack, shabby work on my part I'm afraid
:(

On Thursday, December 29, 2011, Ayon Sinha <[email protected]> wrote:
> I'm getting IllegalArgumentException while storing a bag of tuples.
>
> 2011-12-29 13:28:32,576 [Thread-42] INFO
 com.mongodb.hadoop.pig.MongoStorage - I: 2
tuple:(hipster,foobar,{(halloween,21),(oprah,17),(motorcycles,15)})
> 2011-12-29 13:28:32,579 [Thread-42] WARN
 org.apache.hadoop.mapred.LocalJobRunner - job_local_0003
> java.lang.IllegalArgumentException: can't serialize class
org.apache.pig.data.BinSedesTuple
> at org.bson.BSONEncoder._putObjectField(BSONEncoder.java:205)
> at org.bson.BSONEncoder.putIterable(BSONEncoder.java:230)
> at org.bson.BSONEncoder._putObjectField(BSONEncoder.java:179)
> at org.bson.BSONEncoder.putObject(BSONEncoder.java:121)
> at org.bson.BSONEncoder.putObject(BSONEncoder.java:67)
> at com.mongodb.DBApiLayer$MyCollection.insert(DBApiLayer.java:231)
> at com.mongodb.DBApiLayer$MyCollection.insert(DBApiLayer.java:197)
> at com.mongodb.DBCollection.insert(DBCollection.java:73)
> at com.mongodb.DBCollection.save(DBCollection.java:524)
> at com.mongodb.DBCollection.save(DBCollection.java:504)
> at
com.mongodb.hadoop.output.MongoRecordWriter.write(MongoRecordWriter.java:98)
> at com.mongodb.hadoop.pig.MongoStorage.putNext(MongoStorage.java:77)
> at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat$PigRecordWriter.write(PigOutputFormat.java:138)
> at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat$PigRecordWriter.write(PigOutputFormat.java:97)
> at
org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:498)
> at
org.apache.hadoop.mapreduce.TaskInputOutputContext.write(TaskInputOutputContext.java:80)
> at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigMapOnly$Map.collect(PigMapOnly.java:48)
> at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigMapBase.runPipeline(PigMapBase.java:239)
> at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigMapBase.map(PigMapBase.java:232)
> at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigMapBase.map(PigMapBase.java:53)
> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:621)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:305)
> at
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:177)
> 2011-12-29 13:28:37,338 [main] INFO
 org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
- job job_local_0003 has failed! Stop running all dependent jobs
>
> -Ayon
> See My Photos on Flickr
> Also check out my Blog for answers to commonly asked questions.
>
>
>
> ________________________________
>  From: Russell Jurney <[email protected]>
> To: [email protected]
> Sent: Thursday, December 22, 2011 1:36 PM
> Subject: Pig and MongoDB
>
> I was impressed by MongoDB's Pig integration enough to write about it.
> Thought it might be of interest to the list:
>
http://datasyndrome.com/post/14631249157/mongodb-is-web-scale-hadoop-mongodb
>
> --
> Russell Jurney
> twitter.com/rjurney
> [email protected]
> datasyndrome.com

-- 
Russell Jurney
twitter.com/rjurney
[email protected]
datasyndrome.com

Reply via email to