Yes, you can change the number of records. The default value is 1000

On Thu, Jun 25, 2015 at 2:32 PM, Nihal Bhagchandani <
nihal_bhagchand...@yahoo.com> wrote:

> Hi Su,
>
> as per my understanding you can change the limit of 1000record from the
> interpreter section by setting up the value for variable 
> "zeppelin.spark.maxResult",
> moon could you please confirm my understanding?
>
> Regards
> Nihal
>
>
>
>   On Thursday, 25 June 2015 10:00 AM, Su She <suhsheka...@gmail.com>
> wrote:
>
>
> Hello Everyone,
>
> Excited to be making progress, and thanks for the community for providing
> help along the way.This stuff is all really cool.
>
>
> *Questions:*
>
> *1) *I noticed that the limit for the visual representation is 1000
> results. Are there any short term plans to expand the limit? It seemed a
> little on the low side as many of the reasons for working with spark/hadoop
> is to work with large datasets.
>
> *2) *When can I use the %sql function? Is it only on tables that have
> been registered? I have been having trouble registering tables unless I do:
>
> // Apply the schema to the RDD.val peopleSchemaRDD = 
> sqlContext.applySchema(rowRDD, schema)
> // Register the SchemaRDD as a 
> table.peopleSchemaRDD.registerTempTable("people")
>
>
> I am having lots of trouble registering tables through HiveContext or even
> duplicating the Zeppelin tutorial, is this issue mitigated by using
> DataFrames ( I am planning to move to 1.3 very soon)?
>
>
> *Bug:*
>
> When I do this:
> z.show(sqlContext.sql("select * from sensortable limit 100"))
>
> I get the table, but I also get text results in the bottom, please see
> attached image. For some reason, if the image doesn't go through, i
> basically get the table, and everything works well, but the select
> statement also returns text (regardless of its 100 results or all)
>
>
> Thank you !
>
> Best,
>
> Su
>
>
>

Reply via email to