Hi Rajesh, Here are the steps …
. Generate large CSV file ; It has 125 million rows and is about 12G in size . Bulk Load it into Table having Local Indexes Naga On Jan 30, 2015, at 10:06 AM, Rajeshbabu Chintaguntla <[email protected]<mailto:[email protected]>> wrote: Hi Naga, can you give me the steps how you are running? and also from the place you are running tool, there also you have updated the jars? Thanks, Rajeshbabu. On Fri, Jan 30, 2015 at 9:56 AM, Naga Vijayapuram <[email protected]<mailto:[email protected]>> wrote: I have updated https://issues.apache.org/jira/browse/PHOENIX-1248 On Jan 29, 2015, at 9:14 PM, Naga Vijayapuram <[email protected]<mailto:[email protected]>> wrote: Hi Rajeshbabu, I picked up the patch and tried on Hortonworks Sandbox (as our cluster is tied up with lots of other critical jobs and no scope to test with) and got this … (ensured phoenix-5.0.0-SNAPSHOT-server.jar is in hbase lib dir) ... 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:java.library.path=:/usr/hdp/2.2.0.0-2041/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.2.0.0-2041/hadoop/lib/native 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:java.compiler=<NA> 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:os.name<http://os.name/>=Linux 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:os.arch=amd64 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:os.version=2.6.32-504.1.3.el6.x86_64 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:user.name<http://user.name/>=root 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:user.home=/root 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Client environment:user.dir=/root/nv125m 15/01/30 04:56:46 INFO zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:2181 sessionTimeout=30000 watcher=hconnection-0x246e78aa, quorum=localhost:2181, baseZNode=/hbase-unsecure 15/01/30 04:56:46 INFO zookeeper.ClientCnxn: Opening socket connection to server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>. Will not attempt to authenticate using SASL (unknown error) 15/01/30 04:56:46 INFO zookeeper.ClientCnxn: Socket connection established to localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, initiating session 15/01/30 04:56:46 INFO zookeeper.ClientCnxn: Session establishment complete on server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, sessionid = 0x14b3920fde00026, negotiated timeout = 30000 15/01/30 04:56:48 INFO metrics.Metrics: Initializing metrics system: phoenix 15/01/30 04:56:48 INFO impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 15/01/30 04:56:48 INFO impl.MetricsSystemImpl: Scheduled snapshot period at 60 second(s). 15/01/30 04:56:48 INFO impl.MetricsSystemImpl: phoenix metrics system started 15/01/30 04:56:50 INFO query.ConnectionQueryServicesImpl: Found quorum: localhost:2181 15/01/30 04:56:50 INFO zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x2a3709d7 connecting to ZooKeeper ensemble=localhost:2181 15/01/30 04:56:50 INFO zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:2181 sessionTimeout=30000 watcher=hconnection-0x2a3709d7, quorum=localhost:2181, baseZNode=/hbase-unsecure 15/01/30 04:56:50 INFO zookeeper.ClientCnxn: Opening socket connection to server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>. Will not attempt to authenticate using SASL (unknown error) 15/01/30 04:56:50 INFO zookeeper.ClientCnxn: Socket connection established to localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, initiating session 15/01/30 04:56:50 INFO zookeeper.ClientCnxn: Session establishment complete on server localhost.localdomain/127.0.0.1:2181<http://127.0.0.1:2181/>, sessionid = 0x14b3920fde00027, negotiated timeout = 30000 15/01/30 04:56:50 INFO client.HConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 15/01/30 04:56:50 INFO client.HConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x14b3920fde00027 15/01/30 04:56:50 INFO zookeeper.ClientCnxn: EventThread shut down 15/01/30 04:56:50 INFO zookeeper.ZooKeeper: Session: 0x14b3920fde00027 closed 15/01/30 04:56:51 INFO client.HConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x14b3920fde00026 15/01/30 04:56:51 INFO zookeeper.ZooKeeper: Session: 0x14b3920fde00026 closed 15/01/30 04:56:51 INFO zookeeper.ClientCnxn: EventThread shut down Exception in thread "main" java.sql.SQLException: ERROR 2006 (INT08): Incompatible jars detected between client and server. Ensure that phoenix.jar is put on the classpath of HBase in every region server: org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos$MetaRegionServer.hasState()Z at org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:350) at org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:133) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:966) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:845) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1173) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:111) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1599) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:556) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:175) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:279) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:271) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:269) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1051) at org.apache.phoenix.query.ConnectionQueryServicesImpl$10.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.query.ConnectionQueryServicesImpl$10.call(ConnectionQueryServicesImpl.java:1788) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1788) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:162) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:126) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:133) at java.sql.DriverManager.getConnection(DriverManager.java:571) at java.sql.DriverManager.getConnection(DriverManager.java:233) at org.apache.phoenix.mapreduce.CsvBulkLoadTool.run(CsvBulkLoadTool.java:183) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) at org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:97) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.util.RunJar.run(RunJar.java:221) at org.apache.hadoop.util.RunJar.main(RunJar.java:136) Caused by: java.lang.NoSuchMethodError: org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos$MetaRegionServer.hasState()Z at org.apache.hadoop.hbase.zookeeper.MetaRegionTracker.getMetaRegionState(MetaRegionTracker.java:219) at org.apache.hadoop.hbase.zookeeper.MetaRegionTracker.blockUntilAvailable(MetaRegionTracker.java:204) at org.apache.hadoop.hbase.client.ZooKeeperRegistry.getMetaRegionLocation(ZooKeeperRegistry.java:58) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1147) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1239) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1150) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1107) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:948) at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:410) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:919) ... 30 more Naga On Jan 28, 2015, at 4:25 PM, Rajeshbabu Chintaguntla <[email protected]<mailto:[email protected]>> wrote: Hi Naga, I have uploaded patch at PHOENIX-1248 please apply the patch and try now if possible. Thanks, Rajeshbabu. On Thu, Jan 22, 2015 at 8:57 PM, Naga Vijayapuram <[email protected]<mailto:[email protected]>> wrote: Thanks Rajeshbabu ! On Jan 22, 2015, at 8:50 PM, Rajeshbabu Chintaguntla <[email protected]<mailto:[email protected]>> wrote: Hi Naga Vijayapuram, Sorry I have missed that I can fix it in next week. Thanks, Rajeshbabu. On Fri, Jan 23, 2015 at 9:37 AM, Naga Vijayapuram <[email protected]<mailto:[email protected]>> wrote: I have hit upon https://issues.apache.org/jira/browse/PHOENIX-1248 Any idea when it will be fixed? Thanks Naga On Jan 22, 2015, at 8:01 PM, Naga Vijayapuram <[email protected]<mailto:[email protected]>> wrote: Hello, Any idea why this exception shows up when running CsvBulkLoadTool? 15/01/22 22:16:50 ERROR mapreduce.CsvBulkLoadTool: Import job on table=LOCAL_INDEX_ON_TABLE1 failed due to exception:java.lang.IllegalArgumentException: No regions passed What’s the “No regions passed” message about, and how to overcome? Thanks Naga
