You can use broadcast variable for passing connection information.
Cheers
> On Sep 21, 2015, at 4:27 AM, Priya Ch wrote:
>
> can i use this sparkContext on executors ??
> In my application, i have scenario of reading from db for certain records in
> rdd. Hence I
build instruction for 2.11 is obsolete? Or there are still
>> some limitations?
>>
>>
>> http://spark.apache.org/docs/latest/building-spark.html#building-for-scala-211
>>
>> On Fri, Sep 11, 2015 at 2:09 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>
>&
Which release are you using ?
>From the line number in ClosureCleaner, it seems you're using 1.4.x
Cheers
On Mon, Sep 21, 2015 at 4:07 PM, Balaji Vijayan
wrote:
> Howdy,
>
> I'm a relative novice at Spark/Scala and I'm puzzled by some behavior that
> I'm seeing in
bq. hadoop-core-0.20.204.0
How come the above got into play - it was from hadoop-1
On Mon, Sep 21, 2015 at 11:34 AM, Ellen Kraffmiller <
ellen.kraffmil...@gmail.com> wrote:
> I am including the Spark core dependency in my maven pom.xml:
>
>
> org.apache.spark
> spark-core_2.10
> 1.5.0
>
>
>
Have you seen this thread:
http://search-hadoop.com/m/q3RTtVJJ3I15OJ251
Cheers
On Sun, Sep 20, 2015 at 6:11 PM, Aaroncq4 <475715...@qq.com> wrote:
> When I used “sbt/sbt assembly" to compile spark code of spark-1.5.0,I got a
> problem and I did not know why.It signs that:
>
> NOTE: The sbt/sbt
Can you tell us how you configured the JVM heap size ?
Which version of Java are you using ?
When I build Spark, I do the following:
export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M
-XX:ReservedCodeCacheSize=512m"
Cheers
On Sat, Sep 19, 2015 at 5:31 AM, Eyal Altshuler
gt;> wrote.
>> My java version is 1.7.0_75.
>> I didn't customized the JVM heap size specifically. Is there an
>> additional configuration I have to run besides the MAVEN_OPTS configutaion?
>>
>> Thanks,
>> Eyal
>>
>> On Sat, Sep 19, 2015 at 5:29 PM
For #2, please see:
examples/src/main/scala//org/apache/spark/examples/HBaseTest.scala
examples/src/main/scala//org/apache/spark/examples/pythonconverters/HBaseConverters.scala
In hbase, there is hbase-spark module which is being polished. Should be
available in hbase 1.3.0 release.
Cheers
On
For #1, see this thread: http://search-hadoop.com/m/q3RTti0Thneenne2
For #2, also see:
examples//src/main/python/hbase_inputformat.py
examples//src/main/python/hbase_outputformat.py
Cheers
On Fri, Sep 18, 2015 at 5:12 PM, Ted Yu <yuzhih...@gmail.com> wrote:
> For #2, please see:
>
Which version of Java are you using ?
And release of Spark, please.
Thanks
On Fri, Sep 18, 2015 at 9:15 AM, swetha wrote:
> Hi,
>
> When I try to recover my Spark Streaming job from a checkpoint directory, I
> get a StackOverFlow Error as shown below. Any idea as to
I assume you don't use Kinesis.
Are you running Spark 1.5.0 ?
If you must use S3, is switching to Kinesis possible ?
Cheers
On Thu, Sep 17, 2015 at 1:09 PM, Michal Čizmazia wrote:
> How to make Write Ahead Logs to work with S3? Any pointers welcome!
>
> It seems as a known
bq. and check if 5 minutes have passed
What if the duration for the window is longer than 5 minutes ?
Cheers
On Wed, Sep 16, 2015 at 1:25 PM, Adrian Tanase wrote:
> If you don't need the counts in betweem the DB writes, you could simply
> use a 5 min window for the
How about using this method:
* Return a new RDD by applying a function to all elements of this RDD.
*/
def mapToDouble[R](f: DoubleFunction[T]): JavaDoubleRDD = {
new JavaDoubleRDD(rdd.map(x => f.call(x).doubleValue()))
On Wed, Sep 16, 2015 at 8:30 PM, Tapan Sharma
Can you tell us which release you were using ?
Thanks
> On Sep 16, 2015, at 7:11 PM, "guoqing0...@yahoo.com.hk"
> wrote:
>
> Hi all,
> I found the table structure was modified when use DataFrameWriter.jdbc to
> save the content of DataFrame ,
>
>
See this thread:
http://search-hadoop.com/m/q3RTtUz0cyiPjYX
On Tue, Sep 15, 2015 at 1:19 PM, Ashish Soni wrote:
> Hi All ,
>
> Are there any framework which can be used to execute workflows with in
> spark or Is it possible to use ML Pipeline for workflow execution but
Have you considered posting on vendor forum ?
FYI
On Mon, Sep 14, 2015 at 6:09 AM, Renu Yadav wrote:
>
> -- Forwarded message --
> From: Renu Yadav
> Date: Mon, Sep 14, 2015 at 4:51 PM
> Subject: Spark job failed
> To:
For #1, there is the following method:
@DeveloperApi
def getExecutorStorageStatus: Array[StorageStatus] = {
assertNotStopped()
You can wrap the call in try block catching IllegalStateException.
Of course, this is just a workaround.
FYI
On Sun, Sep 13, 2015 at 1:48 AM, Ophir Cohen
Can you retrieve log for appattempt_1440495451668_0258_01 and see if
there is some clue there ?
Cheers
On Sun, Sep 13, 2015 at 3:28 AM, Bin Wang wrote:
> There is some error logs in the executor and I don't know if it is related:
>
> 15/09/11 10:54:05 WARN ipc.Client:
Please also see this thread: http://search-hadoop.com/m/q3RTtGpLeLyv97B1
On Sun, Sep 13, 2015 at 9:49 AM, Ted Yu <yuzhih...@gmail.com> wrote:
> For #1, there is the following method:
>
> @DeveloperApi
> def getExecutorStorageStatus: Array[StorageStatus] = {
> asser
You may have seen this:
https://spark.apache.org/docs/latest/sql-programming-guide.html
Please suggest what should be added.
Cheers
On Fri, Sep 11, 2015 at 3:43 AM, vivek bhaskar wrote:
> Hi all,
>
> I am looking for a reference manual for Spark SQL some thing like many
>
Have you looked at:
https://issues.apache.org/jira/browse/SPARK-8013
> On Sep 11, 2015, at 4:53 AM, Petr Novak wrote:
>
> Does it still apply for 1.5.0?
>
> What actual limitation does it mean when I switch to 2.11? No JDBC
> Thriftserver? No JDBC DataSource? No
Have you seen this thread ?
http://search-hadoop.com/m/q3RTtPPuSvBu0rj2
> On Sep 11, 2015, at 3:00 AM, Jagat Singh wrote:
>
> Hi,
>
> We have queries which were running fine on 1.4.1 system.
>
> We are testing upgrade and even simple query like
> val t1=
It has not been ported yet.
On Fri, Sep 11, 2015 at 4:13 PM, LucaMartinetti wrote:
> Hi,
>
> I am trying to use countApproxDistinctByKey in pyspark but cannot find it.
>
>
>
Was your intention that exception from rdd.saveToCassandra() be caught ?
In that case you can place try / catch around that call.
Cheers
On Fri, Sep 11, 2015 at 7:30 AM, Samya wrote:
> Hi Team,
>
> I am facing this issue where in I can't figure out why the exception is
tor.syntactical.StandardTokenParsers.
>
> Thanks,
> -Rick
>
> vivekw...@gmail.com wrote on 09/11/2015 05:05:47 AM:
>
> > From: vivek bhaskar <vivekw...@gmail.com>
> > To: Ted Yu <yuzhih...@gmail.com>
> > Cc: user <user@spark.apache.org>
> >
Please see the following
in sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala :
val TUNGSTEN_ENABLED = booleanConf("spark.sql.tungsten.enabled",
defaultValue = Some(true),
doc = "When true, use the optimized Tungsten physical execution backend
which explicitly " +
if (o1.hits == o2.hits) {
> return 0;
> } else if (o1.hits > o2.hits) {
> return -1;
> } else {
> return 1;
> }
> }
>
> }
>
> ...
>
>
>
> Thanks,
> Ashish
>
> On Wed, Sep 9, 2015 at 5:13 PM, Ted Yu <yuzhih...@gm
f partitions. Fixed that.
>
> Thanks,
> Ashish
>
> On Thu, Sep 10, 2015 at 10:44 AM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> Here is snippet of ExternalSorter.scala where ArrayIndexOutOfBoundsException
>> was thrown:
>>
>> while (iterator.
Have you checked the contents of __app__.jar ?
> On Sep 9, 2015, at 3:28 AM, Tom Seddon wrote:
>
> Thanks for your reply Aniket.
>
> Ok I've done this and I'm still confused. Output from running locally shows:
>
>
Prachicsa:
If the number of EC tokens is high, please consider using a set instead of
array for better lookup performance.
BTW use short, descriptive subject for future emails.
> On Sep 9, 2015, at 3:13 AM, Akhil Das wrote:
>
> Try this:
>
> val tocks =
Please take a look at the example in SPARK-10287
FYI
On Wed, Sep 9, 2015 at 8:50 AM, prachicsa wrote:
>
>
> I am very new to Spark.
>
> I have a very basic question. I read a file in Spark RDD in which each line
> is a JSON. I want to make apply groupBy like
Which release of Spark are you using ?
Can you show skeleton of your partitioner and comparator ?
Thanks
> On Sep 9, 2015, at 4:45 PM, Ashish Shenoy wrote:
>
> Hi,
>
> I am trying to sort a RDD pair using repartitionAndSortWithinPartitions() for
> my key [which
I used your first command with mvn 3.3.3 (without build/)
The build passed.
FYI
On Wed, Sep 9, 2015 at 8:50 PM, stark_summer wrote:
> codeurl: http://d3kbcqa49mib13.cloudfront.net/spark-1.5.0.tgz
> build scripts:
>
> build/mvn -Phadoop-2.3 -Dhadoop.version=2.3.0-cdh5.1.0
See the following announcement:
http://search-hadoop.com/m/q3RTtojAyW1dabFk
On Wed, Sep 9, 2015 at 9:05 PM, Netwaver wrote:
> Hi Spark experts,
> I am trying to migrate my Spark cluster from
> 1.4.1 to latest 1.5.0 , but meet below issues when run
Take a look at the following methods:
* Filters rows using the given condition.
* {{{
* // The following are equivalent:
* peopleDf.filter($"age" > 15)
* peopleDf.where($"age" > 15)
* }}}
* @group dfops
* @since 1.3.0
*/
def filter(condition: Column): DataFrame
Have you tried:
df.rdd.isEmpty
Cheers
On Tue, Sep 8, 2015 at 1:22 PM, Axel Dahl wrote:
> I have a join, that fails when one of the data frames is empty.
>
> To avoid this I am hoping to check if the dataframe is empty or not before
> the join.
>
> The question is
Performance wise, Scala is by far the best choice when you use Spark.
The cost of learning Scala is not negligible but not insurmountable either.
My personal opinion.
On Tue, Sep 8, 2015 at 6:50 AM, Bryan Jeffrey
wrote:
> All,
>
> We're looking at language choice in
/scalac are huge
> resource hogs, since so much of Scala is really implemented in the
> compiler; prepare to update your laptop to develop in Scala on your
> IDE of choice, and start to think about running long-running compile
> servers like we did in the year 2000.
>
> Still net
Do you run Zinc while compiling ?
Cheers
On Tue, Sep 8, 2015 at 7:56 AM, Benjamin Zaitlen wrote:
> I'm still getting errors with 3g. I've increase to 4g and I'll report back
>
> To be clear:
>
> export MAVEN_OPTS="-Xmx4g -XX:MaxPermSize=1024M
>
Use the following command if needed:
df -i /tmp
See
https://wiki.gentoo.org/wiki/Knowledge_Base:No_space_left_on_device_while_there_is_plenty_of_space_available
On Sun, Sep 6, 2015 at 6:15 AM, Shixiong Zhu wrote:
> The folder is in "/tmp" by default. Could you use "df -h" to
What about concurrent access (read / update) to the small file with same
key ?
That can get a bit tricky.
On Thu, Sep 3, 2015 at 2:47 PM, Jörn Franke wrote:
> Well it is the same as in normal hdfs, delete file and put a new one with
> the same name works.
>
> Le jeu. 3
need NOSQL like random update access.
>>
>>
>>
>>
>>
>> On Fri, Sep 4, 2015 at 9:56 AM, Ted Yu <yuzhih...@gmail.com> wrote:
>>
>>> What about concurrent access (read / update) to the small file with same
>>> key ?
>>>
possible to easily process Pig on it
>> directly ?
>>
>> Tks
>> Nicolas
>>
>> - Mail original -
>> De: "Tao Lu" <taolu2...@gmail.com>
>> À: nib...@free.fr
>> Cc: "Ted Yu" <yuzhih...@gmail.com>, "
Ayan:
Please read this:
http://hbase.apache.org/book.html#cp
Cheers
On Thu, Sep 3, 2015 at 2:13 PM, ayan guha wrote:
> Hi
>
> Thanks for your comments. My driving point is instead of loading Hbase
> data entirely I want to process record by record lookup and that is best
>
Instead of storing those messages in HDFS, have you considered storing them
in key-value store (e.g. hbase) ?
Cheers
On Wed, Sep 2, 2015 at 9:07 AM, wrote:
> Hello,
> I'am currently using Spark Streaming to collect small messages (events) ,
> size being <50 KB , volume is high
The following JIRA is close to integration:
HBASE-14181 Add Spark DataFrame DataSource to HBase-Spark Module
after which hbase would provide better support for DataFrame interaction.
On Wed, Sep 2, 2015 at 1:21 PM, ALEX K wrote:
> you can use Phoenix-Spark plugin:
>
Can you utilize the following method in StreamingListener ?
override def onBatchStarted(batchStarted: StreamingListenerBatchStarted) {
Cheers
On Tue, Sep 1, 2015 at 12:36 PM, steve_ash wrote:
> We have some logic that we need to apply while we are processing the events
>> JIRA, but arguably it's something that's nice to just work but isn't
>> >> to do with Spark per se. Or, have a look at others related to the
>> >> closure and shell and you may find this is related to other known
>> >> behavior.
>> >
a:1183)
>
> at
> java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547)
>
> at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508)
>
> at
> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431)
>
> at j
Please see this thread w.r.t. spark.sql.shuffle.partitions :
http://search-hadoop.com/m/q3RTtE7JOv1bDJtY
FYI
On Mon, Aug 31, 2015 at 11:03 AM, unk1102 wrote:
> Hi I have Spark job and its executors hits OOM issue after some time and my
> job hangs because of it followed
> On Fri, Aug 28, 2015 at 9:45 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> What format does your REST server expect ?
>>
>> You may have seen this:
>>
>> https://www.paypal-engineering.com/2014/02/13/hello-newman-a-rest-client-for-scala/
>>
>&
I used the notation on JIRA where bq means quote.
FYI
On Mon, Aug 31, 2015 at 12:34 PM, Ashish Shrowty <ashish.shro...@gmail.com>
wrote:
> Yes .. I am closing the stream.
>
> Not sure what you meant by "bq. and then create rdd"?
>
> -Ashish
>
> On Mon, A
Take a look at the following:
https://github.com/mongodb/mongo-hadoop/blob/master/core/src/main/java/com/mongodb/hadoop/MongoOutputFormat.java
https://jira.mongodb.org/plugins/servlet/mobile#issue/HADOOP-82
> On Aug 31, 2015, at 4:39 AM, Deepesh Maheshwari
>
Pranay:
Please take a look at the Redirector class inside:
./launcher/src/test/java/org/apache/spark/launcher/SparkLauncherSuite.java
Cheers
On Sun, Aug 30, 2015 at 11:25 AM, Pranay Tonpay pranay.ton...@impetus.co.in
wrote:
yes, the context is being closed at the end.
Using Spark shell :
scala import scala.collection.mutable.MutableList
import scala.collection.mutable.MutableList
scala val lst = MutableList[(String,String,Double)]()
lst: scala.collection.mutable.MutableList[(String, String, Double)] =
MutableList()
scala
:55 AM Ted Yu yuzhih...@gmail.com wrote:
Using Spark shell :
scala import scala.collection.mutable.MutableList
import scala.collection.mutable.MutableList
scala val lst = MutableList[(String,String,Double)]()
lst: scala.collection.mutable.MutableList[(String, String, Double)] =
MutableList
Manohar:
See if adding the following dependency to your project helps:
dependency
+groupIdcom.fasterxml.jackson.core/groupId
+artifactIdjackson-databind/artifactId
+version${fasterxml.jackson.version}/version
+ /dependency
+ dependency
+
This is related:
SPARK-10288 Add a rest client for Spark on Yarn
FYI
On Sun, Aug 30, 2015 at 12:12 PM, Dawid Wysakowicz
wysakowicz.da...@gmail.com wrote:
Hi Ajay,
In short story: No, there is no easy way to do that. But if you'd like to
play around this topic a good starting point would be
See
https://hadoop.apache.org/docs/r2.7.0/hadoop-project-dist/hadoop-hdfs/HdfsNfsGateway.html
FYI
On Sat, Aug 29, 2015 at 1:04 AM, Akhil Das ak...@sigmoidanalytics.com
wrote:
You can also mount HDFS through the NFS gateway and access i think.
Thanks
Best Regards
On Tue, Aug 25, 2015 at
What format does your REST server expect ?
You may have seen this:
https://www.paypal-engineering.com/2014/02/13/hello-newman-a-rest-client-for-scala/
On Fri, Aug 28, 2015 at 9:35 PM, Cassa L lcas...@gmail.com wrote:
Hi,
If I have RDD that counts something e.g.:
JavaPairDStreamString,
For the exception w.r.t. ManifestFactory , there is SPARK-6497 which is
Open.
FYI
On Fri, Aug 28, 2015 at 8:25 AM, donhoff_h 165612...@qq.com wrote:
Hi, all
I wrote a spark program which uses the Kryo serialization. When I count a
rdd which type is RDD[(String,String)], it reported an
+1 on Jason's suggestion.
bq. this large variable is broadcast many times during the lifetime
Please consider making this large variable more granular. Meaning, reduce
the amount of data transferred between the key value store and your app
during update.
Cheers
On Fri, Aug 28, 2015 at 12:44
Can you provide a bit more information ?
Are Spark artifacts packaged by you have the same names / paths (in maven
repo) as the ones published by Apache Spark ?
Is Zinc running on the machine where you performed the build ?
Cheers
On Wed, Aug 26, 2015 at 7:56 AM, Muhammad Haseeb Javed
),then eventually I will see OutOfMemory occur
Can you guys try to run it if you have the environment ? I think you may
reproduce it. Thanks!
At 2015-08-26 13:01:34, Ted Yu yuzhih...@gmail.com wrote:
The error in #1 below was not informative.
Are you able to get more detailed error message
Have you run dev/change-version-to-2.11.sh ?
Cheers
On Wed, Aug 26, 2015 at 7:07 AM, Felix Neutatz neut...@googlemail.com
wrote:
Hi everybody,
I tried to build Spark v1.4.1-rc4 with Scala 2.11:
../apache-maven-3.3.3/bin/mvn -Dscala-2.11 -DskipTests clean install
Before running this, I
Looks like it is this PR:
https://github.com/mesos/spark-ec2/pull/133
On Tue, Aug 25, 2015 at 9:52 AM, Shivaram Venkataraman
shiva...@eecs.berkeley.edu wrote:
Yeah thats a know issue and we have a PR out to fix it.
Shivaram
On Tue, Aug 25, 2015 at 7:39 AM, Garry Chen g...@cornell.edu
Corrected a typo in the subject of your email.
What you cited seems to be from worker node startup.
Was there other error you saw ?
Please list the command you used.
Cheers
On Tue, Aug 25, 2015 at 7:39 AM, Garry Chen g...@cornell.edu wrote:
Hi All,
I am trying to lunch a
The error in #1 below was not informative.
Are you able to get more detailed error message ?
Thanks
On Aug 25, 2015, at 6:57 PM, Todd bit1...@163.com wrote:
Thanks Ted Yu.
Following are the error message:
1. The exception that is shown on the UI is :
Exception in thread Thread-113
Looks like you were attaching images to your email which didn't go through.
Consider using third party site for images - or paste error in text.
Cheers
On Tue, Aug 25, 2015 at 4:22 AM, Todd bit1...@163.com wrote:
Hi,
The spark sql perf itself contains benchmark data generation. I am using
Can you show the complete stack trace ?
Which Spark / Kafka release are you using ?
Thanks
On Mon, Aug 24, 2015 at 4:58 PM, Cassa L lcas...@gmail.com wrote:
Hi,
I am storing messages in Kafka using protobuf and reading them into
Spark. I upgraded protobuf version from 2.4.1 to 2.5.0. I got
sc.hadoopConfiguration.set(fs.s3n.awsAccessKeyId, ***)
sc.hadoopConfiguration.set(fs.s3n.awsSecretAccessKey, **)
However, the error still occurs for ORC format.
If I change the format to JSON, although the error does not go, the JSON
files can be saved successfully.
On Sun, Aug 23, 2015 at 5:51 AM, Ted Yu
You may have seen this:
http://search-hadoop.com/m/q3RTtdSyM52urAyI
On Aug 23, 2015, at 1:01 AM, lostrain A donotlikeworkingh...@gmail.com
wrote:
Hi,
I'm trying to save a simple dataframe to S3 in ORC format. The code is as
follows:
val sqlContext = new
On Aug 23, 2015, at 12:49 PM, lostrain A donotlikeworkingh...@gmail.com
wrote:
Ted,
Thanks for the suggestions. Actually I tried both s3n and s3 and the
result remains the same.
On Sun, Aug 23, 2015 at 12:27 PM, Ted Yu yuzhih...@gmail.com wrote:
In your case, I would specify fs.s3
See http://spark.apache.org/community.html
Cheers
On Sat, Aug 22, 2015 at 2:51 AM, Lars Hermes li...@hermes-it-consulting.de
wrote:
subscribe
-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional
to do this with spark 1.3?
such as write the orcfile manually in foreachPartition method?
On Sat, Aug 22, 2015 at 12:19 PM, Ted Yu yuzhih...@gmail.com wrote:
ORC support was added in Spark 1.4
See SPARK-2883
On Fri, Aug 21, 2015 at 7:36 PM, dong.yajun dongt...@gmail.com wrote:
Hi list
Have you read this ?
http://stackoverflow.com/questions/22716346/how-to-use-lz4-compression-in-linux-3-11
On Aug 21, 2015, at 6:57 AM, saif.a.ell...@wellsfargo.com
saif.a.ell...@wellsfargo.com wrote:
Hi all,
I am using pre-compiled spark with hadoop 2.6. LZ4 Codec is not on hadoop’s
Have you considered asking this question on
https://groups.google.com/a/lists.datastax.com/forum/#!forum/spark-connector-user
?
Cheers
On Thu, Aug 20, 2015 at 10:57 PM, Samya samya.ma...@amadeus.com wrote:
Hi All,
I need to write an RDD to Cassandra using the sparkCassandraConnector
from
See also
http://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.sql.package
Cheers
On Thu, Aug 20, 2015 at 7:50 AM, Muhammad Atif muhammadatif...@gmail.com
wrote:
Hi Dawid
The best pace to get started is the Spark SQL Guide from Apache
Saif:
In your example below, the error was due to there is no automatic conversion
from Int to BigDecimal.
Cheers
On Aug 19, 2015, at 6:40 AM, saif.a.ell...@wellsfargo.com
saif.a.ell...@wellsfargo.com wrote:
Hi, thank you all for the asssistance.
It is odd, it works when creating a
See this thread:
http://search-hadoop.com/m/q3RTtdZv0d1btRHl/Spark+build+modulesubj=Building+Spark+Building+just+one+module+
On Aug 19, 2015, at 1:44 AM, canan chen ccn...@gmail.com wrote:
I want to work on one jira, but it is not easy to do unit test, because it
involves different
Do you mind providing a bit more information ?
release of Spark
code snippet of your app
version of Java
Thanks
On Tue, Aug 18, 2015 at 8:57 AM, unk1102 umesh.ka...@gmail.com wrote:
Hi this GC overhead limit error is making me crazy. I have 20 executors
using
25 GB each I dont understand
Normally people would establish maven project with Spark dependencies or,
use sbt.
Can you go with either approach ?
Cheers
On Tue, Aug 18, 2015 at 10:28 AM, Jerry jerry.c...@gmail.com wrote:
Hello,
So I setup Spark to run on my local machine to see if I can reproduce the
issue I'm having
Have you tried adding path to hbase-protocol jar to
spark.driver.extraClassPath and spark.executor.extraClassPath ?
Cheers
On Mon, Aug 17, 2015 at 7:51 PM, stark_summer stark_sum...@qq.com wrote:
spark vesion:1.4.1
java version:1.7
hadoop version:
Hadoop 2.3.0-cdh5.1.0
submit spark job to
I got 404 when trying to access the link.
On Aug 17, 2015, at 5:31 AM, Todd bit1...@163.com wrote:
Hi,
I can't access
http://people.csail.mit.edu/matei/papers/2015/sigmod_spark_sql.pdf.
Could someone help try to see if it is available and reply with it?Thanks!
Thanks Nan.
That is why I always put an extra space between URL and punctuation in my
comments / emails.
On Mon, Aug 17, 2015 at 6:31 AM, Nan Zhu zhunanmcg...@gmail.com wrote:
an extra “,” is at the end
--
Nan Zhu
http://codingcat.me
On Monday, August 17, 2015 at 9:28 AM, Ted Yu wrote
You were building against 1.4.x, right ?
In master branch, switch-to-scala-2.11.sh is gone. There is scala-2.11
profile.
FYI
On Sun, Aug 16, 2015 at 11:12 AM, Stephen Boesch java...@gmail.com wrote:
I am building spark with the following options - most notably the
**scala-2.11**:
.
I tried with master branch and got the following:
http://pastebin.com/2nhtMFjQ
FYI
On Sat, Aug 15, 2015 at 1:03 AM, Kevin Jung itsjb.j...@samsung.com wrote:
Spark shell can't find base directory of class server after running
:reset command.
scala :reset
scala 1
uncaught exception during
path? What's the purpose of prefix and
where do I specify the path if not in prefix?
On Fri, Aug 14, 2015 at 4:36 PM, Ted Yu yuzhih...@gmail.com wrote:
Please take a look at JavaPairDStream.scala:
def saveAsHadoopFiles[F : OutputFormat[_, _]](
prefix: String,
suffix: String
to a local file on the local file system for
verification and I see the data:
$ ls -ltr !$
ls -ltr /tmp/out
-rw-r--r-- 1 yarn yarn 5230 Aug 13 15:45 /tmp/out
On Fri, Aug 14, 2015 at 6:15 AM, Ted Yu yuzhih...@gmail.com wrote:
Which Spark release are you using ?
Can you show us snippet
Which Spark release are you using ?
Can you show us snippet of your code ?
Have you checked namenode log ?
Thanks
On Aug 13, 2015, at 10:21 PM, Mohit Anchlia mohitanch...@gmail.com wrote:
I was able to get this working by using an alternative method however I only
see 0 bytes files in
The code and error didn't go through.
Mind sending again ?
Which Spark release are you using ?
On Thu, Aug 13, 2015 at 6:17 PM, dizzy5112 dave.zee...@gmail.com wrote:
the code below works perfectly on both cluster and local modes
but when i try to create a graph in cluster mode (it works
You can look under Developer Track:
https://spark-summit.org/2015/#day-1
http://www.slideshare.net/jeykottalam/spark-sqlamp-camp2014?related=1 (slightly
old)
Catalyst design:
https://docs.google.com/a/databricks.com/document/d/1Hc_Ehtr0G8SQUg69cmViZsMi55_Kf3tISD9GPGU5M1Y/edit
FYI
On Thu, Aug
I ran your command on Linux which passed.
Are you going to use SparkR ?
If so, consider including the following:
-Psparkr
Cheers
On Wed, Aug 12, 2015 at 3:31 AM, MEETHU MATHEW meethu2...@yahoo.co.in
wrote:
Hi,
I am trying to create a package using the make-distribution.sh script
from the
,
*From:* Ted Yu [mailto:yuzhih...@gmail.com]
*Sent:* Tuesday, August 11, 2015 3:28 PM
*To:* Yan Zhou.sc
*Cc:* Bing Xiao (Bing); d...@spark.apache.org; user@spark.apache.org
*Subject:* Re: 答复: Package Release Annoucement: Spark SQL on HBase Astro
HBase will not have query engine
Pa:
Can you try 1.5.0 SNAPSHOT ?
See SPARK-7075 Project Tungsten (Spark 1.5 Phase 1)
Cheers
On Tue, Aug 11, 2015 at 12:49 AM, jun kit...@126.com wrote:
your detail of log file?
At 2015-08-10 22:02:16, Pa Rö paul.roewer1...@googlemail.com wrote:
hi community,
i have build a spark and
, …, etc., which allows for loosely-coupled query
engines
built on top of it.
Thanks,
发件人: Ted Yu [mailto:yuzhih...@gmail.com]
发送时间: 2015年8月11日 8:54
收件人: Bing Xiao (Bing)
抄送: d...@spark.apache.org; user@spark.apache.org; Yan Zhou.sc
主题: Re: Package Release Annoucement: Spark SQL
What does the following command say ?
mvn -version
Maybe you are using an old maven ?
Cheers
On Tue, Aug 11, 2015 at 7:55 AM, Yakubovich, Alexey
alexey.yakubov...@searshc.com wrote:
I found some discussions online, but it all cpome to advice to use JDF 1.7
(or 1.8).
Well, I use JDK 1.7 on
See first section of http://spark.apache.org/community.html
On Tue, Aug 11, 2015 at 9:47 AM, Michel Robert m...@us.ibm.com wrote:
Michel Robert
Almaden Research Center
EDA - IBM Systems and Technology Group
Phone: (408) 927-2117 T/L 8-457-2117
E-mail: m...@us.ibm.com
What level of logging are you looking at ?
At INFO level, there shouldn't be noticeable difference.
On Tue, Aug 11, 2015 at 12:24 PM, saif.a.ell...@wellsfargo.com wrote:
Hi all,
silly question. Does logging info messages, both print or to file, or
event logging, cause any impact to general
I wonder during recovery from a checkpoint whether we can estimate the size
of the checkpoint and compare with Runtime.getRuntime().freeMemory().
If the size of checkpoint is much bigger than free memory, log warning, etc
Cheers
On Mon, Aug 10, 2015 at 9:34 AM, Dmitry Goldenberg
1001 - 1100 of 1611 matches
Mail list logo