if i change to ramdonpartioner
an reload changes i have this error:
java.lang.RuntimeException: InvalidRequestException(why:Start token
sorts after end token)
at
org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.maybeInit(ColumnFamilyRecordReader.java:384)
at
org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.computeNext(ColumnFamilyRecordReader.java:390)
at
org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.computeNext(ColumnFamilyRecordReader.java:313)
at
com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
at
com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
at
org.apache.cassandra.hadoop.ColumnFamilyRecordReader.nextKeyValue(ColumnFamilyRecordReader.java:184)
at
org.apache.cassandra.hadoop.pig.CassandraStorage.getNext(CassandraStorage.java:228)
at
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211)
at
org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:532)
at
org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67)
at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1136)
at org.apache.hadoop.mapred.Child.main(Child.java:249)
Caused by: InvalidRequestException(why:Start token sorts after end token)
at
org.apache.cassandra.thrift.Cassandra$get_range_slices_result.read(Cassandra.java:12916)
at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
at
org.apache.cassandra.thrift.Cassandra$Client.recv_get_range_slices(Cassandra.java:734)
at
org.apache.cassandra.thrift.Cassandra$Client.get_range_slices(Cassandra.java:718)
at
org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.maybeInit(ColumnFamilyRecordReader.java:346)
... 17 more
thanks in advance
note.-i runing script with pig_cassandra and cassandra 1.2.0
2013/4/29 Miguel Angel Martin junquera <[email protected]>
> *hi all:*
>
> *
> *
>
> *i can run pig with cassandra and hadoop in EC2.*
>
> *
> *
>
> *I ,m trying to run pig with cassandra ring and hadoop *
>
> *The ring cassandra have the tasktrackers and datanodes , too. *
>
> *
> *
>
> *and i running pig from another machine where i have intalled the
> namenode-jobtracker.*
>
> *ihave a simple script to load data ffrom pygmalion keyspace adn
> columfalimily account and dump result to test.*
>
> *I installed another simple local cassandra in namenode-job tacker machine
> and i can run pig jobs ok, but when i try to run script in cassandra ring
> config changig the config of envitronment variable PIG_INITIAL_ADDRESS to
> the IP of one of the nodes of cassandra ring i have this error:*
>
>
> *
> *
>
> *---*
>
> *
> *
>
> *
> *
>
> *java.lang.RuntimeException: UnavailableException()
> at
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.maybeInit(ColumnFamilyRecordReader.java:384)
> at
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.computeNext(ColumnFamilyRecordReader.java:390)
> at
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.computeNext(ColumnFamilyRecordReader.java:313)
> at
> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
> at
> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
> at
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader.nextKeyValue(ColumnFamilyRecordReader.java:184)
> at
> org.apache.cassandra.hadoop.pig.CassandraStorage.getNext(CassandraStorage.java:226)
> at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211)
> at
> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:532)
> at
> org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67)
> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
> at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1136)
> at org.apache.hadoop.mapred.Child.main(Child.java:249)
> Caused by: UnavailableException()
> at
> org.apache.cassandra.thrift.Cassandra$get_range_slices_result.read(Cassandra.java:12924)
> at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
> at
> org.apache.cassandra.thrift.Cassandra$Client.recv_get_range_slices(Cassandra.java:734)
> at
> org.apache.cassandra.thrift.Cassandra$Client.get_range_slices(Cassandra.java:718)
> at
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.maybeInit(ColumnFamilyRecordReader.java:346)
> ... 17 more*
>
> *
> *
>
> *
> *
>
> *can anybody help me o have any idea?*
>
> *Thanks in advance*
>
> *pd:*
>
> *1.- the ports **are open in EC2 *
>
> *2 The keyspace and cF are created in the cassandra cluster EC2 too nad
> likey at the name node cassandra installation.*
>
> *3.-i have this bash_profile configuration:*
>
> *# .bash_profile
>
> # Get the aliases and functions
> if [ -f ~/.bashrc ]; then
> . ~/.bashrc
> fi
>
> # User specific environment and startup programs
>
> PATH=$PATH:$HOME/.local/bin:$HOME/bin
> export PATH=$PATH:/usr/lib/jvm/java-1.7.0-openjdk.x86_64/bin
> export JAVA_HOME=/usr/lib/jvm/java-1.7.0-openjdk.x86_64
> export CASSANDRA_HOME=/home/ec2-user/apache-cassandra-1.2.4
> export PIG_HOME=/home/ec2-user/pig-0.11.1-src
> export PIG_INITIAL_ADDRESS=10.210.164.233
> #export PIG_INITIAL_ADDRESS=127.0.0.1
> export PIG_RPC_PORT=9160
> export PIG_CONF_DIR=/home/ec2-user/hadoop-1.1.1/conf
> export PIG_PARTITIONER=org.apache.cassandra.dht.Murmur3Partitioner
> #export PIG_PARTITIONER=org.apache.cassandra.dht.RandomPartitioner*
>
> *
> *
>
> *4.- I export all cassandrasjars in the hadoop-env.sh for all nodes of hadoop*
>
> *5.- i have the same error running PIG in local mode *
>
> *
> *
>
> *
> *
>