.
Thanks.
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
--
Web: http://alpinenow.com/
this?
Thanks.
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
--
Web: http://alpinenow.com/
On Sun, Mar 2, 2014 at 10:23 AM, Debasish Das debasish.da...@gmail.com wrote:
Hi DB,
1. Could you point to the BFGS repositories used to publish artifacts
.
Is this getting merged to the master or there will be revisions on it ?
https://github.com/apache/spark/pull/53
Thanks.
Deb
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
--
Web: http://alpinenow.com/
to fix the L-BFGS in breeze, and we can get OWL-QN and
L-BFGS-B.
What do you think?
Thanks.
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
--
Web: http://alpinenow.com/
On Mon, Mar 3, 2014 at 3:52 PM, DB Tsai dbt...@alpinenow.com wrote:
Hi Deb
, let's have a design
discussion around this. It may be more effective since we can design a
architecture that have to work for both cases in the codebase, and will
be easier to think about the edge case for it.
Thanks.
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
for you to
investigate the issue? Or do I need to make it as a standalone test?
Will send you the test later today.
Thanks.
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
--
Web: http://alpinenow.com/
Iteration 29: loss 0.30788249908237314, diff 0.23885980452569502
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
--
Web: http://alpinenow.com/
On Wed, Mar 5, 2014 at 2:00 PM, David Hall d...@cs.berkeley.edu wrote:
On Wed, Mar 5, 2014 at 1:57 PM
Hi Xiangrui,
I think it doesn't matter whether we use Fortran/Breeze/RISO for
optimizers since optimization only takes 1% of time. Most of the
time is in gradientSum and lossSum parallel computation.
Sincerely,
DB Tsai
Machine Learning Engineer
Alpine Data Labs
Hi guys,
The latest PR uses Breeze's L-BFGS implement which is introduced by
Xiangrui's sparse input format work in SPARK-1212.
https://github.com/apache/spark/pull/353
Now, it works with the new sparse framework!
Any feedback would be greatly appreciated.
Thanks.
Sincerely,
DB Tsai
stepSize: Double,
var numIterations: Int,
var regParam: Double,
var miniBatchFraction: Double
Xiangrui, what do you think?
For now, you can use my L-BFGS solver by copying and pasting the
LogisticRegressionWithSGD code, and changing the optimizer to L-BFGS.
Sincerely,
DB Tsai
computation per RDD is done on each of the workers...
This miniBatchFraction is also a heuristic which I don't think makes sense
for LogisticRegressionWithBFGS...does it ?
On Tue, Apr 8, 2014 at 3:44 PM, DB Tsai dbt...@stanford.edu wrote:
Hi Debasish,
The L-BFGS solver will be in the master like
I don't experiment it. That's the use-case in theory I could think of. ^^
However, from what I saw, BFGS converges really fast so that I only
need 20~30 iterations in general.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn
/dbtsai/.rbenv/versions/2.1.1/lib/ruby/gems/2.1.0/gems/commander-4.1.6/lib/commander/import.rb:10:in
`block in top (required)'
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Tue, Apr 22
But what doesSKIP_SCALADOC=1 mean? export SKIP_SCALADOC=1?
Thanks.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
Matei, thanks. It works with kramdown.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Tue, Apr 22, 2014 at 11:38 PM, Matei Zaharia matei.zaha...@gmail.comwrote:
Try doing “gem install
result.
Thanks.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
a 1% (or so) sparse dataset to experiment with?
On Apr 23, 2014, at 9:21 PM, DB Tsai dbt...@stanford.edu wrote:
Hi all,
I'm benchmarking Logistic Regression in MLlib using the newly added
optimizer LBFGS and GD. I'm using the same dataset and the same methodology
in this paper, http
The figure showing the Log-Likelihood vs Time can be found here.
https://github.com/dbtsai/spark-lbfgs-benchmark/raw/fd703303fb1c16ef5714901739154728550becf4/result/a9a11M.pdf
Let me know if you can not open it.
Sincerely,
DB Tsai
---
My Blog
ps, it doesn't make sense to have weight and gradient sparse unless
with strong L1 penalty.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Wed, Apr 23, 2014 at 10:17 PM, DB Tsai dbt
Not yet since it's running in the cluster. Will run locally with
profiler. Thanks for help.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Wed, Apr 23, 2014 at 10:22 PM, David Hall d
The figure showing the Log-Likelihood vs Time can be found here.
https://github.com/dbtsai/spark-lbfgs-benchmark/raw/fd703303fb1c16ef5714901739154728550becf4/result/a9a11M.pdf
Let me know if you can not open it. Thanks.
Sincerely,
DB Tsai
.
The difference you saw is actually from dense feature or sparse feature
vector. For LBFGS and GD dense feature, you can see the first iteration
takes the same time. It's true for GD.
I'm going to run rcv1.binary which only has 0.15% non-zero elements to
verify the hypothesis.
Sincerely,
DB Tsai
, Vectors.fromBreeze(gradientSum / miniBatchSize), stepSize,
i, regParam)
weights = update._1
regVal = update._2
timeStamp.append(System.nanoTime() - startTime)
}
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
rcv1.binary is too sparse (0.15% non-zero elements), so dense format will
not run due to out of memory. But sparse format runs really well.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
Also, how many failure of rejection will terminate the optimization
process? How is it related to numberOfImprovementFailures?
Thanks.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
, miniBatchFraction, lbfgs,
miniBatchSize)
val states = lbfgs.iterations(new CachedDiffFunction(costFun),
initialWeights.toBreeze.toDenseVector)
Thanks.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com
straightforward, we wonder if this can
be reviewed and have this in 1.0.
Thanks.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
in Spark before we've deeper
understanding of how stochastic LBFGS works.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Tue, Apr 29, 2014 at 9:50 PM, David Hall d...@cs.berkeley.edu wrote
You could easily achieve this by mapPartition. However, it seems that it
can not be done by using aggregate type of operation. I can see that it's a
general useful operation. For now, you could use mapPartition.
Sincerely,
DB Tsai
---
My Blog
+1 Would be nice that we can use different type in Vector.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Mon, May 5, 2014 at 2:41 PM, Debasish Das debasish.da...@gmail.comwrote:
Hi
Breeze could take any type (Int, Long, Double, and Float) in the matrix
template.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Mon, May 5, 2014 at 2:56 PM, Debasish Das debasish.da
_ = math.log(numerators(math.round(y - 1).toInt) / denominator)
}
(loglike, predicted)
}
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Tue, May 13, 2014 at 4:08 AM, Debasish Das
will not be seen.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
The reflection actually works. But you need to get the loader by `val
loader = Thread.currentThread.getContextClassLoader` which is set by Spark
executor. Our team verified this, and uses it as workaround.
Sincerely,
DB Tsai
---
My Blog
The jars are included in my driver, and I can successfully use them in the
driver. I'm working on a patch, and it's almost working. Will submit a PR
soon.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https
the customClassloader to create a wrapped class, and in this wrapped class,
the classloader is inherited from the customClassloader so that users don't
need to do reflection in the wrapped class. I'm working on this now.
Sincerely,
DB Tsai
---
My Blog: https
: Method =
classOf[URLClassLoader].getDeclaredMethod(addURL, classOf[URL])
method.setAccessible(true)
method.invoke(loader, url)
}
catch {
case t: Throwable = {
throw new IOException(Error, could not add URL to system
classloader)
}
}
}
Sincerely,
DB
the
protected method `addURL` which will not work and throw exception if the
code is wrapped in security manager.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Wed, May 21, 2014 at 1:13 PM, Sandy
the primary jar is not in the system
loader but custom one, so when we reference those jars added dynamically,
we can find it without reflection.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
environment.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Wed, May 21, 2014 at 2:57 PM, Koert Kuipers ko...@tresata.com wrote:
db tsai, i do not think userClassPathFirst is working
Sometimes for this case, I will just standardize without centerization. I
still get good result.
Sent from my Google Nexus 5
On May 28, 2014 7:03 PM, Xiangrui Meng men...@gmail.com wrote:
RowMatrix has a method to compute column summary statistics. There is
a trade-off here because centering
Hi Gang,
This is a bug, and I'm the one who did it :) Just add the comment to your PR.
Thanks.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Tue, Jun 17, 2014 at 7:13 PM, Gang Bai
+1
On Jul 5, 2014 1:39 PM, Michael Armbrust mich...@databricks.com wrote:
+1
I tested sql/hive functionality.
On Sat, Jul 5, 2014 at 9:30 AM, Mark Hamstra m...@clearstorydata.com
wrote:
+1
On Fri, Jul 4, 2014 at 12:40 PM, Patrick Wendell pwend...@gmail.com
wrote:
I'll start
)
at java.lang.reflect.Method.invoke(Method.java:597)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
+1
Tested with my Ubuntu Linux.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Thu, Jul 17, 2014 at 6:36 PM, Matei Zaharia matei.zaha...@gmail.com wrote:
+1
Tested on Mac, verified
I'm working on it with weighted regularization. The problem is that
OWLQN doesn't work nicely with Updater now since all the L1 logic
should be in OWLQN instead of L1Updater.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn
One related question, is mllib jar independent from hadoop version (doesnt
use hadoop api directly)? Can I use mllib jar compile for one version of
hadoop and use it in another version of hadoop?
Sent from my Google Nexus 5
On Aug 6, 2014 8:29 AM, Debasish Das debasish.da...@gmail.com wrote:
Hi
After sbt gen-idea , you can open the intellji project directly without
going through pom.xml
If u want to compile inside intellji, you have to remove one of the messo
jar. This is an open issue, and u can find the detail in JIRA.
Sent from my Google Nexus 5
On Aug 6, 2014 8:54 PM, Ron Gonzalez
.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Sat, Sep 13, 2014 at 2:12 AM, Yanbo Liang yanboha...@gmail.com wrote:
Hi All,
I found that LogisticRegressionWithLBFGS interface
You dont have to include breeze jar which is already in spark assembly jar.
For native one, its optional.
Sent from my Google Nexus 5
On Oct 3, 2014 8:04 PM, Priya Ch learnings.chitt...@gmail.com wrote:
yes. I have included breeze-0.9 in build.sbt file. I ll change this to
0.7. Apart from
,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
On Mon, Sep 29, 2014 at 11:45 AM, Yanbo Liang yanboha...@gmail.com wrote:
Thank you for all your patient response.
I can conclude that if the data
, there are different strategies to do feature
scalling for linear regression
and logistic regression; as a result, we don't want to make it public
api naively without addressing
different use-case.
Sincerely,
DB Tsai
---
My Blog: https
As Marcelo said, CDH5.3 is based on hadoop 2.3, so please try
./make-distribution.sh -Pyarn -Phive -Phadoop-2.3
-Dhadoop.version=2.3.0-cdh5.1.3 -DskipTests
See the detail of how to change the profile at
https://spark.apache.org/docs/latest/building-with-maven.html
Sincerely,
DB Tsai
oh, I meant to say cdh5.1.3 used by Jakub's company is based on 2.3. You
can see it from the first part of the Cloudera's version number - 2.3.0-cdh
5.1.3.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https
Hi Xiangrui,
It seems that it's stateless so will be hard to implement
regularization path. Any suggestion to extend it? Thanks.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https://www.linkedin.com/in/dbtsai
Okay, I got it. In Estimator, fit(dataset: SchemaRDD, paramMaps:
Array[ParamMap]): Seq[M] can be overwritten to implement
regularization path. Correct me if I'm wrong.
Sincerely,
DB Tsai
---
My Blog: https://www.dbtsai.com
LinkedIn: https
are small.
By default, depth 2 is used, so if you have so many partitions of
large vector, this may still cause issue. You can increase the depth
into higher numbers such that in the final reduce in driver, the
number of partitions are very small.
Sincerely,
DB Tsai
I'm working on LinearRegressionWithElasticNet using OWLQN now. This
will do the data standardization internally so it's transparent to
users. With OWLQN, you don't have to manually choose stepSize. Will
send out PR soon next week.
Sincerely,
DB Tsai
Hi Robin,
You can try this PR out. This has built-in features scaling, and has
ElasticNet regularization (L1/L2 mix). This implementation can stably
converge to model from R's glmnet package.
https://github.com/apache/spark/pull/4259
Sincerely,
DB Tsai
It's a bug in breeze's side. Once David fixes it and publishes it to
maven, we can upgrade to breeze 0.11.2. Please file a jira ticket for
this issue. thanks.
Sincerely,
DB Tsai
---
Blog: https://www.dbtsai.com
On Sun, Mar 15, 2015 at 12:45
to avoid the second cache. In this case,
the code will be more complicated, so I will split the code into two
paths. Will be done in another PR.
Sincerely,
DB Tsai
---
Blog: https://www.dbtsai.com
On Wed, Mar 25, 2015 at 11:57 AM, Joseph Bradley
Hi Theodore,
I'm currently working on elastic-net regression in ML framework, and I
decided not to have any extra layer of abstraction for now but focus
on accuracy and performance. We may come out with proper solution
later. Any idea is welcome.
Sincerely,
DB Tsai
.
Sincerely,
DB Tsai
---
Blog: https://www.dbtsai.com
On Tue, Apr 7, 2015 at 3:03 PM, Ulanov, Alexander
alexander.ula...@hp.com wrote:
Hi,
Could anyone elaborate on the regularization in Spark? I've found that L1 and
L2 are implemented
Is your HDP implementation based on distributed gibbs sampling? Thanks.
Sincerely,
DB Tsai
---
Blog: https://www.dbtsai.com
On Wed, Jun 3, 2015 at 8:13 PM, Yang, Yuhao yuhao.y...@intel.com wrote:
Hi Lorenz,
I’m trying to build
I thought LGPL is okay but GPL is not okay for Apache project.
On Saturday, May 23, 2015, Patrick Wendell pwend...@gmail.com wrote:
Yes - spark packages can include non ASF licenses.
On Sat, May 23, 2015 at 6:16 PM, Debasish Das debasish.da...@gmail.com
javascript:; wrote:
Hi,
Is it
There is a JIRA for this. I know Holden is interested in this.
On Thursday, October 22, 2015, YiZhi Liu wrote:
> Would someone mind giving some hint?
>
> 2015-10-20 15:34 GMT+08:00 YiZhi Liu >:
> > Hi all,
> >
> > I noticed that in
shrinkage).
Thanks.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
On Mon, Oct 26, 2015 at 8:37 PM, Meihua Wu <rotationsymmetr...@gmail.com> wrote:
> Hi DB Tsai,
>
> Thank you very much fo
Also, does it support categorical feature?
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
On Mon, Oct 26, 2015 at 4:06 PM, DB Tsai <dbt...@dbtsai.com> wrote:
> Interesting. For feature sub-sampling,
Interesting. For feature sub-sampling, is it per-node or per-tree? Do
you think you can implement generic GBM and have it merged as part of
Spark codebase?
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
On Mon
ear regression, but currently, there is no open source
implementation in Spark.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
On Sun, Nov 1, 2015 at 9:22 AM, Zhiliang Zhu <zchl.j...@yahoo.com> wrote:
> Dear All,
those code to share more.)
Sincerely,
DB Tsai
--
Blog: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
<https://pgp.mit.edu/pks/lookup?search=0x59DF55B8AF08DF8D>
On Mon, Oct 12, 2015 at 1:24 AM, YiZhi Liu <javeli...@gmail.com>
maintenance cost. Once it's getting mature, and
people are asking for them, we will gradually make them public.
Thanks.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
On Sat, Nov 28, 2015 at 5:20 AM, Sasaki Kai
I used reflection initially, but I found it's very slow especially in
a tight loop. Maybe caching the reflection can help which I never try.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
On Mon, Nov 30, 2015
+1 for renaming the jar file.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0xAF08DF8D
On Tue, Apr 5, 2016 at 8:02 PM, Chris Fregly <ch...@fregly.com> wrote:
> perhaps renaming to Spark ML would actually clea
Congrats, Xiao!
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0x9DCC1DBD7FC7BBB2
On Wed, Oct 5, 2016 at 2:36 PM, Fred Reiss <freiss@gmail.com> wrote:
> Congratulations, Xiao!
>
> Fred
>
>
> On
-1
I think that back-porting SPARK-20270
<https://github.com/apache/spark/pull/17577> and SPARK-18555
<https://github.com/apache/spark/pull/15994> are very important since it's
a critical bug that na.fill will mess up the data in Long even the data
isn't null.
Thanks.
Sincere
Congratulations!
On Wed, Oct 4, 2017 at 6:55 PM, Liwei Lin wrote:
> Congratulations!
>
> Cheers,
> Liwei
>
> On Wed, Oct 4, 2017 at 2:27 PM, Yuval Itzchakov wrote:
>>
>> Congratulations and Good luck! :)
>>
>>
>>
>> --
>> Sent from:
+1
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0x5CED8B896A6BDFA0
On Fri, Oct 6, 2017 at 7:46 AM, Felix Cheung <felixcheun...@hotmail.com> wrote:
> Thanks Nick, Hyukjin. Yes this seems to be a longer stand
gt; datatypes?
>>
>>
>> For parquet, this effort is primarily tracked via SPARK-4502 (see
>> https://github.com/apache/spark/pull/16578) and is currently targeted for
>> 2.3.
--
Sincerely,
DB Tsai
--
PGP Key ID: 0x5CED8B896A6BDFA0
-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
a blocker for us to move to newer version of Scala 2.12.x
since the newer version of Scala 2.12.x has the same issue.
In my opinion, Scala should fix the root cause and provide a stable hook for
3rd party developers to initialize their custom code.
DB Tsai | Siri Open Source Technologies
ark context Web UI available at http://192.168.1.169:4040
Spark context available as 'sc' (master = local[*], app id =
local-1528180279528).
Spark session available as 'spark’.
scala>
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
> On Jun 7, 2018, at 5:49 P
4 PM, Holden Karau
>> wrote:
>> > I agree that's a little odd, could we not add the bacspace terminal
>> > character? Regardless even if not, I don't think that should be a
>> blocker
>> > for 2.12 support especially since it doesn't degrade the 2.11
>&g
, and
the result should match R.
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
> On Apr 20, 2018, at 5:56 PM, Weichen Xu <weichen...@databricks.com> wrote:
>
> Right. If regularization item isn't zero, then enable/disable standardization
> will ge
I'll +1 on removing those legacy mllib code. Many users are confused about the
APIs, and some of them have weird behaviors (for example, in gradient descent,
the intercept is regularized which supports not to).
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
are selected simultaneously.
https://issues.apache.org/jira/browse/SPARK-25879
If we decide to not fix it in 2.4, we should at least document it in
the release note to let users know.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID
Given Oracle's new 6-month release model, I think the only realistic option is
to only support and test LTS JDK. I'll send out two separate emails to dev to
facilitate the discussion.
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
> On Nov 6, 2018, at 9:47 AM, sh
to work on bugs and
issues that we may run into.
What do you think?
Thanks,
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
Given Oracle's new 6-month release model, I feel the only realistic option is
to only test and support JDK such as JDK 11 LTS and future LTS release. I would
like to have a discussion on this in Spark community.
Thanks,
DB Tsai | Siri Open Source Technologies [not a contribution
OpenJDK will follow Oracle's release cycle,
https://openjdk.java.net/projects/jdk/
<https://openjdk.java.net/projects/jdk/>, a strict six months model. I'm not
familiar with other non-Oracle VMs and Redhat support.
DB Tsai | Siri Open Source Technologies [not a contribution] | Appl
Ideally, supporting only Scala 2.12 in Spark 3 will be ideal.
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
> On Nov 6, 2018, at 2:55 PM, Felix Cheung wrote:
>
> So to clarify, only scala 2.12 is supported in Spark 3?
>
>
> From: Ryan Blu
agree
with Sean that this can make the decencies really complicated; hence I support
to drop Scala 2.11 in Spark 3.0 directly.
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
> On Nov 6, 2018, at 11:38 AM, Sean Owen wrote:
>
> I think we should make S
if we want to change the alternative Scala version
to 2.13 and drop 2.11 if we just want to support two Scala versions at
one time.
Thanks.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0x5CED8B896A6BDFA0
On Wed, Nov 7, 2018
-1
Agreed with Anton that this bug will potentially corrupt the data
silently. As he is ready to submit a PR, I'll suggest to wait to
include the fix. Thanks!
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0x5CED8B896A6BDFA0
I like the idea of checking only the diff. Even I am sometimes confused
about the right style in Spark since I am working on multiple projects with
slightly different coding styles.
On Wed, Nov 21, 2018 at 1:36 PM Sean Owen wrote:
> I know the PR builder runs SBT, but I presume this would just
+1 on removing Scala 2.11 support for 3.0 given Scala 2.11 is already EOL.
On Tue, Nov 20, 2018 at 2:53 PM Sean Owen wrote:
> PS: pull request at https://github.com/apache/spark/pull/23098
> Not going to merge it until there's clear agreement.
>
> On Tue, Nov 20, 2018 at 10:16 AM Ryan Blue
+1
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 0x5CED8B896A6BDFA0
On Tue, Jan 8, 2019 at 11:14 AM Dongjoon Hyun wrote:
>
> Please vote on releasing the following candidate as Apache Spark version
> 2.2.3.
>
RC9 was just cut. Will send out another thread once the build is finished.
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 42E5B25A8F7A82C1
On Mon, Mar 25, 2019 at 5:10 PM Sean Owen wrote:
>
> That's all merged now. I
typically not hold the
release unless the bug in question is a regression from the previous
release. That being said, if there is something which is a regression
that has not been correctly targeted please ping me or a committer to
help target the issue.
DB Tsai | Siri Open Source Te
Hello Sean,
By looking at SPARK-26961 PR, seems it's ready to go. Do you think we
can merge it into 2.4 branch soon?
Sincerely,
DB Tsai
--
Web: https://www.dbtsai.com
PGP Key ID: 42E5B25A8F7A82C1
On Sat, Mar 23, 2019 at 12:04 PM Sean Owen
.
DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc
-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
1 - 100 of 152 matches
Mail list logo