Hello,

I found it useful to implement the  Hadamard Product
<https://en.wikipedia.org/wiki/Hadamard_product_%28matrices%29http://>   as
a VectorTransformer.  It can be applied to scale (by a constant) a certain
dimension (column) of the data set.  

Since I've already implemented it and am using it, I thought I'd see if
there's interest in this feature going in as Experimental.  I'm not sold on
the name 'Weighter', either.

Here's the current branch with the work (docs, impl, tests).
<https://github.com/ogeagla/spark/compare/spark-mllib-weighting>  

The implementation was heavily inspired by those of StandardScalerModel and
Normalizer.

Thanks
Octavian



--
View this message in context: 
http://apache-spark-developers-list.1001551.n3.nabble.com/Any-interest-in-weighting-VectorTransformer-which-does-component-wise-scaling-tp10265.html
Sent from the Apache Spark Developers List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to