Hi Xiangrui,

That link is currently returning a 503 Over Quota error message.  Would you
mind pinging back out when the page is back up?

Thanks!
Andrew

On Mon, Dec 22, 2014 at 12:37 PM, Xiangrui Meng <men...@gmail.com> wrote:

> Dear Spark users and developers,
>
> I’m happy to announce Spark Packages (http://spark-packages.org), a
> community package index to track the growing number of open source
> packages and libraries that work with Apache Spark. Spark Packages
> makes it easy for users to find, discuss, rate, and install packages
> for any version of Spark, and makes it easy for developers to
> contribute packages.
>
> Spark Packages will feature integrations with various data sources,
> management tools, higher level domain-specific libraries, machine
> learning algorithms, code samples, and other Spark content. Thanks to
> the package authors, the initial listing of packages includes
> scientific computing libraries, a job execution server, a connector
> for importing Avro data, tools for launching Spark on Google Compute
> Engine, and many others.
>
> I’d like to invite you to contribute and use Spark Packages and
> provide feedback! As a disclaimer: Spark Packages is a community index
> maintained by Databricks and (by design) will include packages outside
> of the ASF Spark project. We are excited to help showcase and support
> all of the great work going on in the broader Spark community!
>
> Cheers,
> Xiangrui
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to