There is no explicit limit but a JVM string cannot be bigger than 2G. It will also at some point run out of memory with too big of a query plan tree or become incredibly slow due to query planning complexity. I've seen queries that are tens of MBs in size.
On Thu, Jul 11, 2019 at 5:01 AM, 李书明 < alemmont...@126.com > wrote: > > I have a question about the limit(biggest) of SQL's length that is > supported in SparkSQL. I can't find the answer in the documents of Spark. > > > Maybe Interger.MAX_VALUE or not ? > > > >