I think this looks like the typical LZO error that people get when they
dont install it and try to use the codec it happens because LZO isnt
(cant be) bundled, so you wont have it by default in any canned hadoop
installation.
On Sun, Aug 3, 2014 at 8:29 PM, Eric Friedman
wrote:
> I am clos
I am close to giving up on PySpark on YARN. It simply doesn't work for
straightforward operations and it's quite difficult to understand why.
I would love to be proven wrong, by the way.
Eric Friedman
> On Aug 3, 2014, at 7:03 AM, Rahul Bhojwani
> wrote:
>
> The logs provided in the i
The logs provided in the image may not be enough for help. Here I have
copied the whole logs:
WARNING: Running python applications through ./bin/pyspark is deprecated as
of Spark 1.0.
Use ./bin/spark-submit
14/08/03 11:10:57 INFO SparkConf: Using Spark's default log4j profile:
org/apache/spark/l
Hi,
I used to run spark scripts on local machine. Now i am porting my codes to
EMR and i am facing lots of problem.
The main one now is that the spark script which is running properly on my
local machine is giving error when run on Amazon EMR Cluster.
Here is the error:
[image: Inline image 1]