Gabriel,I think the load did not run through fully. So yesterday, we tried to 
load the table again. But we ran out of disk space (we had 30TB free).Are there 
any blogs or slideshare that explains how this csvbulkloadtool works ?We need 
to get this to run through to success before we can access/query I guess. 
Regards,
- kiru
      From: Gabriel Reid <[email protected]>
 To: Kiru Pakkirisamy <[email protected]>; [email protected] 
 Sent: Tuesday, May 12, 2015 10:27 PM
 Subject: Re: select w/ limit hanging on large tables
   
Hi Kiru,

How many regions are there on this table?

Could you also share some information on the schema of the table (e.g. how many 
columns are defined)?

Does a "limit 10" query also hang in this table?

Could you also elaborate a bit on the issues you were running into when loading 
data into the table? We're there performance issues, or we're things not 
working at all?

- Gabriel


On Tue, May 12, 2015 at 23:56 Kiru Pakkirisamy <[email protected]> 
wrote:

We are trying to benchmark/test Phoenix with large tables.A 'select * from 
table1 limit 100000' hangs on a 1.4 billion row table (in sqlline.py or 
SQuirreL)The same select of 1million rows works on smaller table (300 
million).Mainly we wanted to create a smaller version of the 1.4 billion table 
and ran into this issue.Any ideas why this is happening ?We had quite a few 
problems crossing the 1 billion mark even when loading (using CsvBulkLoadTool) 
the table.We are also wondering whether our HBase is configured correctly.
Any tips on HBase Configuration for loading/running Phoenix is highly 
appreciated as well.(We are on HBase 0.98.12 and Phoenix 4.3.1) Regards,
- kiru
 Regards,
- kiru




  

Reply via email to