It seems from the attached logs that you have upgraded phoenix to 4.7
version and now you are using old client to connect with it.
"Update statistics" command and guideposts will not work with old client
after upgradation to 4.7, you need to use the new client for such
operations.
On Wed, Mar 16,
Ankit,
We tried a 4.7 client upgrade to use the phoenix spark client as an experiment;
then, rolled back to the sanctioned CDH 5.5 version of 4.5. I had no idea that
someone did an "update stats” during that period, and I didn’t know that there
would be a fundamental change as this. Do you
Hi Swapna,
We don't support custom aggregate functions, only scalar functions
(see PHOENIX-2069). For a custom aggregate function, you'd need to add them
to phoenix-core and rebuild the jar. We're open to adding them to the code
base if they're general enough. That's how FIRST_VALUE, LAST_VALUE,
Sorry Ben, I may not be clear in first comment but I need you to describe
SYSTEM.STATS in some sql client so that I can see the columns present.
And also please scan 'SYSTEM.CATALOG' ,{RAW=>true} in hbase shell and
attach a output here.
On Wed, Mar 16, 2016 at 8:55 PM, Benjamin Kim
Thanks Pari.
The frequency of the job is weekly.
No. of rows is around 10 billion.
Cluster is 13 node.
>From what you have mentioned I see that CsvBulkLoadTool is best option for
my scenario.
I see you have mentioned about increasing the batch size to accommodate
more rows.
Are you talking about