nope (what I said :-P)
On Thu, Mar 27, 2014 at 11:05 AM, Pascal Voitot Dev < pascal.voitot....@gmail.com> wrote: > > > > On Thu, Mar 27, 2014 at 10:22 AM, andy petrella > <andy.petre...@gmail.com>wrote: > >> I just mean queries sent at runtime ^^, like for any RDBMS. >> In our project we have such requirement to have a layer to play with the >> data (custom and low level service layer of a lambda arch), and something >> like this is interesting. >> >> > Ok that's what I thought! But for these runtime queries, is a macro useful > for you? > > >> >> >> On Thu, Mar 27, 2014 at 10:15 AM, Pascal Voitot Dev < >> pascal.voitot....@gmail.com> wrote: >> >>> >>> Le 27 mars 2014 09:47, "andy petrella" <andy.petre...@gmail.com> a >>> écrit : >>> >>> > >>> > I hijack the thread, but my2c is that this feature is also important >>> to enable ad-hoc queries which is done at runtime. It doesn't remove >>> interests for such macro for precompiled jobs of course, but it may not be >>> the first use case envisioned with this Spark SQL. >>> > >>> >>> I'm not sure to see what you call "ad- hoc queries"... Any sample? >>> >>> > Again, only my0.2c (ok I divided by 10 after writing my thoughts ^^) >>> > >>> > Andy >>> > >>> > On Thu, Mar 27, 2014 at 9:16 AM, Pascal Voitot Dev < >>> pascal.voitot....@gmail.com> wrote: >>> >> >>> >> Hi, >>> >> Quite interesting! >>> >> >>> >> Suggestion: why not go even fancier & parse SQL queries at >>> compile-time with a macro ? ;) >>> >> >>> >> Pascal >>> >> >>> >> >>> >> >>> >> On Wed, Mar 26, 2014 at 10:58 PM, Michael Armbrust < >>> mich...@databricks.com> wrote: >>> >>> >>> >>> Hey Everyone, >>> >>> >>> >>> This already went out to the dev list, but I wanted to put a pointer >>> here as well to a new feature we are pretty excited about for Spark 1.0. >>> >>> >>> >>> >>> http://databricks.com/blog/2014/03/26/Spark-SQL-manipulating-structured-data-using-Spark.html >>> >>> >>> >>> Michael >>> >> >>> >> >>> > >>> >> >> >