Hi James,
Thanks for the reply.
I was not able to find the exact corner case of the issue when tried with a
sample table.
Can you please let me know if there is something i am doing wrong?
But in our project we are facing this issue when we try to delete the data
from the below table,
Below
Hi,
I used the open source client - phoenix-4.5.2-HBase-1.0-client.jar , and able
to connect, but get time out issue which we faced with phoenix-4.3.1-client.jar
as well (but it is related to MicroStrategy application, which we may have a
temporary work around.)
But I would like to follow up
Hi,
I am able to connect from MicroStrategy (on Linux - 3.10.0-327.3.1.el7.x86_64)
to HBase(CDH-5.5.1) using Phoenix (phoenix-4.3.1-client.jar)
I am unable to connect MicroStrategy (on Linux - 3.10.0-327.3.1.el7.x86_64) to
HBase(CDH-5.5.1) using Phoenix (phoenix-1.2.0-client.jar)
>From java
Cool! I'm using 4.6.0. Will soon try 4.7.0. Thank you, James!
On Thu, Feb 25, 2016 at 2:55 PM, James Taylor
wrote:
> This worked for me in the latest 4.7.0 RC3:
>
> Connected to: Phoenix (version 4.7)
> Driver: PhoenixEmbeddedDriver (version 4.7)
> Autocommit status:
This worked for me in the latest 4.7.0 RC3:
Connected to: Phoenix (version 4.7)
Driver: PhoenixEmbeddedDriver (version 4.7)
Autocommit status: true
Transaction isolation: TRANSACTION_READ_COMMITTED
Building list of tables and columns for tab-completion (set fastconnect to
true to skip)...
83/83
I sort-of ran out of time for the near term, but will look into all these
suggestions at a later date. Thanks!
On Thu, Feb 25, 2016 at 1:01 PM, James Taylor
wrote:
> +1 to Ankit's suggestion. If you haven't altered the table, then you can
> just connect at the timestamp
+1 to Ankit's suggestion. If you haven't altered the table, then you can
just connect at the timestamp of one more than the timestamp at which the
table was created (see [1]), and issue to DROP TABLE command from that
connection. If you have altered the table, then you have to be more careful
as
Hi Steve,
You can do what you want with a view today, but the syntax is just a bit
different than what you tried. You declare your dynamic columns after the
view name, like this:
create MY_VIEW("dynamic_field" varchar) as select * from MY_TABLE
You can also alter a view and dynamically
If I understand what you're attempting to do, I think a snapshot will work. A
snapshot is not a copy of the data and doesn't take much time. It's really just
a copy of the region-> store files mapping at a point in time. It will prevent
the store files from getting deleted so drop the snapshot
I like your outside-the-box thinking. Unfortunately, my end goal was to
convert a table created with dynamic fields into a table or view with all
static fields so that I could avoid having to specify the data type of
every dynamic field in the SQL. And I wanted to avoid having to rewrite
the
Hello Ankit,
Thanks a lot! Cleaning the SYSTEM.STATS table solved the problem instantly.
The exception disappeared.
Jarda
On Thu, Feb 25, 2016 at 9:16 AM, Ankit Singhal
wrote:
> can you try after truncating the SYSTEM.STATS table or deleting records of
> parent
+1 for a view that has dynamic columns. This would make life easier with
dynamic columns.
On Tue, Feb 23, 2016 at 4:00 PM, Steve Terrell
wrote:
> I have a table with many dynamic fields. Works great. However, it's a
> bit of a nuisance to have to supply each dynamic
Hi Pedro,
You may need to configure your split policy
to ConstantSizeRegionSplitPolicy if you want *hbase.hregion.max.filesize to
be adhered strictly * (as default policy is
IncreasingToUpperBoundRegionSplitPolicy which uses Min (R^2 *
“hbase.hregion.memstore.flush.size”,
can you try after truncating the SYSTEM.STATS table or deleting records of
parent table only from SYSTEM.STATS like below.
DELETE * FROM SYSTEM.STATS WHERE PHYSICAL_NAME='media';
Regards,
Ankit Singhal
On Wed, Feb 24, 2016 at 8:16 PM, Jaroslav Šnajdr wrote:
> Hello
14 matches
Mail list logo