ERROR 2998: Unhandled internal error. org/jruby/ext/posix/util/Platform
java.lang.NoClassDefFoundError: org/jruby/ext/posix/util/Platform
at org.python.core.PySystemState.getPath(PySystemState.java:513)
at org.python.core.PySystemState.getPathLazy(PySystemState.java:502)
at org.python.core.util.RelativeFile.<init>(RelativeFile.java:17)
at org.python.core.PyTraceback.getLine(PyTraceback.java:52)
at org.python.core.PyTraceback.tracebackInfo(PyTraceback.java:37)
at org.python.core.PyTraceback.dumpStack(PyTraceback.java:108)
at org.python.core.PyTraceback.dumpStack(PyTraceback.java:119)
at org.python.core.Py.displayException(Py.java:1007)
at org.python.core.PyException.printStackTrace(PyException.java:79)
at org.python.core.PyException.toString(PyException.java:98)
at java.lang.String.valueOf(String.java:2826)
at java.lang.StringBuilder.append(StringBuilder.java:115)
at
org.apache.pig.scripting.jython.JythonFunction.exec(JythonFunction.java:107)
at
org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNext(POUserFunc.java:229)
at
org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNext(POUserFunc.java:273)
at
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.processPlan(POForEach.java:343)
at
org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.getNext(POForEach.java:291)
at
org.apache.pig.pen.DerivedDataVisitor.evaluateOperator(DerivedDataVisitor.java:354)
at
org.apache.pig.pen.DerivedDataVisitor.visit(DerivedDataVisitor.java:232)
at
org.apache.pig.impl.logicalLayer.LOForEach.visit(LOForEach.java:132)
at
org.apache.pig.impl.logicalLayer.LOForEach.visit(LOForEach.java:47)
at
org.apache.pig.impl.plan.DependencyOrderWalker.walk(DependencyOrderWalker.java:70)
at org.apache.pig.impl.plan.PlanVisitor.visit(PlanVisitor.java:51)
at
org.apache.pig.pen.LineageTrimmingVisitor.init(LineageTrimmingVisitor.java:94)
at
org.apache.pig.pen.LineageTrimmingVisitor.<init>(LineageTrimmingVisitor.java:86)
at
org.apache.pig.pen.ExampleGenerator.getExamples(ExampleGenerator.java:105)
at org.apache.pig.PigServer.getExamples(PigServer.java:1144)
at
org.apache.pig.tools.grunt.GruntParser.processIllustrate(GruntParser.java:627)
at
org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:308)
at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165)
at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:141)
at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:76)
at org.apache.pig.Main.run(Main.java:465)
at org.apache.pig.Main.main(Main.java:107)
================================================================================
2011/4/23 pob <[email protected]>
> Hello,
>
> I installed jython + jruby on debian.
>
> export
> PIG_CLASSPATH=/path/cassandra-0.7/contrib/pig:/usr/share/java/jython.jar
>
> /path/cassandra-0.7/contrib/pig - here is my udf, myFunc.py
>
>
> #!/usr/bin/python
> @outputSchema("t:tuple(domain:chararray, spam:int, size:int, time:int)")
> def toTuple(bag):
>
> #{(colname, value), (...), ...,}
>
> for word in bag:
> if word[0] == 'domain':
> domain = word[1]
> elif word[0] == 'spam':
> spam = word[1]
> elif word[0] == 'size':
> size = word[1]
> elif word[0] == 'time':
> time = word[1]
>
> return (domain, spam, size, time)
>
>
> After starting grunt, i type
>
> register '/path/cassandra-0.7/contrib/pig/myFunc.py' using jython as myUDF;
>
> rows = LOAD 'cassandra://emailArchive/messagesMetaData' USING
> CassandraStorage() AS (key, columns: bag {T: tuple(name:chararray,
> value:int)});
> d = foreach rows generate myUDF.toTuple($1);
>
> When I type illustrate / dump d; the error occure:
>
>
> Any idea where problem should be?
>
>
> Thanks.
>
>
>
> 2011-04-23 23:40:13,428 [main] ERROR org.apache.pig.tools.grunt.Grunt -
> ERROR 2998: Unhandled internal error. org/jruby/ext/posix/util/Platform
>
> 2011-04-23 23:26:02,211 [Thread-14] WARN
> org.apache.hadoop.mapred.LocalJobRunner - job_local_0001
> java.lang.NoClassDefFoundError: org/jruby/ext/posix/util/Platform
> at org.python.core.PySystemState.getPath(PySystemState.java:513)
> at org.python.core.PySystemState.getPathLazy(PySystemState.java:502)
> at org.python.core.util.RelativeFile.<init>(RelativeFile.java:17)
> at org.python.core.PyTraceback.getLine(PyTraceback.java:52)
> at org.python.core.PyTraceback.tracebackInfo(PyTraceback.java:37)
> at org.python.core.PyTraceback.dumpStack(PyTraceback.java:108)
> at org.python.core.PyTraceback.dumpStack(PyTraceback.java:119)
> at org.python.core.Py.displayException(Py.java:1007)
> at org.python.core.PyException.printStackTrace(PyException.java:79)
> at org.python.core.PyException.toString(PyException.java:98)
> at java.lang.String.valueOf(String.java:2826)
> at java.lang.StringBuilder.append(StringBuilder.java:115)
> at
> org.apache.pig.scripting.jython.JythonFunction.exec(JythonFunction.java:107)
> at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNext(POUserFunc.java:229)
> at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNext(POUserFunc.java:273)
> at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.processPlan(POForEach.java:343)
> at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.getNext(POForEach.java:291)
> at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigMapBase.runPipeline(PigMapBase.java:236)
> at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigMapBase.map(PigMapBase.java:231)
> at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigMapBase.map(PigMapBase.java:53)
> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:621)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:305)
> at
> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:177)
> 2011-04-23 23:26:02,331 [main] INFO
>
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
> - HadoopJobId: job_local_0001
>
>