Add back this thread to email list, forgot to reply all. 2015年10月31日 下午7:23,"Michael Armbrust" <mich...@databricks.com>写道:
> Not that I know of. > > On Sat, Oct 31, 2015 at 12:22 PM, Rex Xiong <bycha...@gmail.com> wrote: > >> Good to know that, will have a try. >> So there is no easy way to achieve it in pure hive method? >> 2015年10月31日 下午7:17,"Michael Armbrust" <mich...@databricks.com>写道: >> >>> Yeah, this was rewritten to be faster in Spark 1.5. We use it with >>> 10,000s of partitions. >>> >>> On Sat, Oct 31, 2015 at 7:17 AM, Rex Xiong <bycha...@gmail.com> wrote: >>> >>>> 1.3.1 >>>> It is a lot of improvement in 1.5+? >>>> >>>> 2015-10-30 19:23 GMT+08:00 Michael Armbrust <mich...@databricks.com>: >>>> >>>>> We have tried schema merging feature, but it's too slow, there're >>>>>> hundreds of partitions. >>>>>> >>>>> Which version of Spark? >>>>> >>>> >>>> >>> >