Hi Rajesh,

Here are the steps …

. Generate large CSV file ; It has 125 million rows and is about 12G in size
. Bulk Load it into Table having Local Indexes

Naga

On Jan 30, 2015, at 10:06 AM, Rajeshbabu Chintaguntla 
<[email protected]<mailto:[email protected]>> wrote:

Hi Naga,
can you give me the steps how you are running? and also from the place you are 
running tool, there also you have updated the jars?

Thanks,
Rajeshbabu.

On Fri, Jan 30, 2015 at 9:56 AM, Naga Vijayapuram 
<[email protected]<mailto:[email protected]>> wrote:
I have updated https://issues.apache.org/jira/browse/PHOENIX-1248

On Jan 29, 2015, at 9:14 PM, Naga Vijayapuram 
<[email protected]<mailto:[email protected]>> wrote:

Hi Rajeshbabu,

I picked up the patch and tried on Hortonworks Sandbox (as our cluster is tied 
up with lots of other critical jobs and no scope to test with) and got this …
(ensured phoenix-5.0.0-SNAPSHOT-server.jar is in hbase lib dir) ...

15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client 
environment:java.library.path=:/usr/hdp/2.2.0.0-2041/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.2.0.0-2041/hadoop/lib/native
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client 
environment:java.io.tmpdir=/tmp
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client 
environment:java.compiler=<NA>
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client 
environment:os.name<http://os.name/>=Linux
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:os.arch=amd64
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client 
environment:os.version=2.6.32-504.1.3.el6.x86_64
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client 
environment:user.name<http://user.name/>=root
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:user.home=/root
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client 
environment:user.dir=/root/nv125m
15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Initiating client connection, 
connectString=localhost:2181 sessionTimeout=30000 
watcher=hconnection-0x246e78aa, quorum=localhost:2181, baseZNode=/hbase-unsecure
15/01/30 04:56:46 INFO zookeeper.ClientCnxn: Opening socket connection to 
server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>. Will not 
attempt to authenticate using SASL (unknown error)
15/01/30 04:56:46 INFO zookeeper.ClientCnxn: Socket connection established to 
localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, initiating session
15/01/30 04:56:46 INFO zookeeper.ClientCnxn: Session establishment complete on 
server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, sessionid 
= 0x14b3920fde00026, negotiated timeout = 30000
15/01/30 04:56:48 INFO metrics.Metrics: Initializing metrics system: phoenix
15/01/30 04:56:48 INFO impl.MetricsConfig: loaded properties from 
hadoop-metrics2.properties
15/01/30 04:56:48 INFO impl.MetricsSystemImpl: Scheduled snapshot period at 60 
second(s).
15/01/30 04:56:48 INFO impl.MetricsSystemImpl: phoenix metrics system started
15/01/30 04:56:50 INFO query.ConnectionQueryServicesImpl: Found quorum: 
localhost:2181
15/01/30 04:56:50 INFO zookeeper.RecoverableZooKeeper: Process 
identifier=hconnection-0x2a3709d7 connecting to ZooKeeper 
ensemble=localhost:2181
15/01/30 04:56:50 INFO zookeeper.ZooKeeper: Initiating client connection, 
connectString=localhost:2181 sessionTimeout=30000 
watcher=hconnection-0x2a3709d7, quorum=localhost:2181, baseZNode=/hbase-unsecure
15/01/30 04:56:50 INFO zookeeper.ClientCnxn: Opening socket connection to 
server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>. Will not 
attempt to authenticate using SASL (unknown error)
15/01/30 04:56:50 INFO zookeeper.ClientCnxn: Socket connection established to 
localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, initiating session
15/01/30 04:56:50 INFO zookeeper.ClientCnxn: Session establishment complete on 
server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, sessionid 
= 0x14b3920fde00027, negotiated timeout = 30000
15/01/30 04:56:50 INFO client.HConnectionManager$HConnectionImplementation: 
Closing master protocol: MasterService
15/01/30 04:56:50 INFO client.HConnectionManager$HConnectionImplementation: 
Closing zookeeper sessionid=0x14b3920fde00027
15/01/30 04:56:50 INFO zookeeper.ClientCnxn: EventThread shut down
15/01/30 04:56:50 INFO zookeeper.ZooKeeper: Session: 0x14b3920fde00027 closed
15/01/30 04:56:51 INFO client.HConnectionManager$HConnectionImplementation: 
Closing zookeeper sessionid=0x14b3920fde00026
15/01/30 04:56:51 INFO zookeeper.ZooKeeper: Session: 0x14b3920fde00026 closed
15/01/30 04:56:51 INFO zookeeper.ClientCnxn: EventThread shut down
Exception in thread "main" java.sql.SQLException: ERROR 2006 (INT08): 
Incompatible jars detected between client and server. Ensure that phoenix.jar 
is put on the classpath of HBase in every region server: 
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos$MetaRegionServer.hasState()Z
at 
org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:350)
at 
org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:133)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:966)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:845)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1173)
at 
org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:111)
at 
org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1599)
at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:556)
at 
org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:175)
at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:279)
at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:271)
at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
at 
org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:269)
at 
org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1051)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl$10.call(ConnectionQueryServicesImpl.java:1819)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl$10.call(ConnectionQueryServicesImpl.java:1788)
at 
org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1788)
at 
org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:162)
at 
org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:126)
at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:133)
at java.sql.DriverManager.getConnection(DriverManager.java:571)
at java.sql.DriverManager.getConnection(DriverManager.java:233)
at org.apache.phoenix.mapreduce.CsvBulkLoadTool.run(CsvBulkLoadTool.java:183)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:97)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: java.lang.NoSuchMethodError: 
org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos$MetaRegionServer.hasState()Z
at 
org.apache.hadoop.hbase.zookeeper.MetaRegionTracker.getMetaRegionState(MetaRegionTracker.java:219)
at 
org.apache.hadoop.hbase.zookeeper.MetaRegionTracker.blockUntilAvailable(MetaRegionTracker.java:204)
at 
org.apache.hadoop.hbase.client.ZooKeeperRegistry.getMetaRegionLocation(ZooKeeperRegistry.java:58)
at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1147)
at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1239)
at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1150)
at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1107)
at 
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:948)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:410)
at 
org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:919)
... 30 more

Naga


On Jan 28, 2015, at 4:25 PM, Rajeshbabu Chintaguntla 
<[email protected]<mailto:[email protected]>> wrote:

Hi Naga,

I have uploaded patch at PHOENIX-1248 please apply the patch and try now if 
possible.

Thanks,
Rajeshbabu.

On Thu, Jan 22, 2015 at 8:57 PM, Naga Vijayapuram 
<[email protected]<mailto:[email protected]>> wrote:
Thanks Rajeshbabu !

On Jan 22, 2015, at 8:50 PM, Rajeshbabu Chintaguntla 
<[email protected]<mailto:[email protected]>> wrote:

Hi Naga Vijayapuram,

Sorry I have missed that I can fix it in next week.

Thanks,
Rajeshbabu.

On Fri, Jan 23, 2015 at 9:37 AM, Naga Vijayapuram 
<[email protected]<mailto:[email protected]>> wrote:
I have hit upon https://issues.apache.org/jira/browse/PHOENIX-1248

Any idea when it will be fixed?

Thanks
Naga


On Jan 22, 2015, at 8:01 PM, Naga Vijayapuram 
<[email protected]<mailto:[email protected]>> wrote:

Hello,

Any idea why this exception shows up when running CsvBulkLoadTool?

15/01/22 22:16:50 ERROR mapreduce.CsvBulkLoadTool: Import job on 
table=LOCAL_INDEX_ON_TABLE1 failed due to 
exception:java.lang.IllegalArgumentException: No regions passed

What’s the “No regions passed” message about, and how to overcome?

Thanks
Naga








Reply via email to