Gotcha....perhaps I was misunderstanding the tutorial that was suggesting the HADOOP_CLASSPATH="$(hbase mapredcp)" hadoop jar ... approach...Thanks for all the help folks...very appreciated!
On Thu, Nov 29, 2018 at 3:13 PM Josh Elser <els...@apache.org> wrote: > Why didn't it work? > > The hbase-protocol.jar is insufficient to run MapReduce jobs against > HBase; full stop. You're going to get lots of stuff pulled in via the > phoenix-client.jar that you give to `hadoop jar`. That said, I can't > think of a reason that including more jars on the classpath would be > harmful. > > Realistically, you might only need to provide HBASE_CONF_DIR to the > HADOOP_CLASSPATH env variable, so that your mappers and reducers also > get it on their classpath. The rest of the Java classes would be > automatically localized via `hadoop jar`. > > On 11/29/18 1:27 PM, M. Aaron Bossert wrote: > > So, sorry for the super late reply...there is weird lag between the time > > a message is sent or received to this mailing list and when I actually > > see it...But, I have got it working now as follows: > > > > > HADOOP_CLASSPATH=/usr/hdp/3.0.1.0-187/hbase/lib/hbase-protocol.jar:/etc/hbase/ > 3.0.1. > > <http://3.0.0.1/>0-187/0/ hadoop jar ... > > > > using this did not work: > > > > HADOOP_CLASSPATH="$(hbase mapredcp)" hadoop jar ... > > > > > > the output of that command separately is this: > > > > [user@server /somedir $] [mabossert@edge-3 lanl_data]$ hbase mapredcp > > > > > /usr/hdp/3.0.1.0-187/hbase/lib/hbase-shaded-protobuf-2.1.0.jar:/usr/hdp/3.0.1.0-187/zookeeper/zookeeper-3.4.6.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/htrace-core4-4.2.0-incubating.jar:/usr/hdp/3.0.1.0-187/hbase/lib/commons-lang3-3.6.jar:/usr/hdp/ > 3.0.1. > 0-187/hbase/lib/hbase-server-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-protocol-shaded-2.0.0.3.0.1.0-187.jar:/usr/hdp/ > 3.0.1. > 0-187/hbase/lib/hbase-hadoop2-compat-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-mapreduce-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-metrics-api-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/protobuf-java-2.5.0.jar:/usr/hdp/3.0.1.0-187/hbase/lib/metrics-core-3.2.1.jar:/usr/hdp/3.0.1.0-187/hbase/lib/jackson-databind-2.9.5.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-client-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-hadoop-compat-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-protocol-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-shaded-netty-2.1.0.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-shaded-miscellaneous-2.1.0.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-metrics-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-common-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/hbase-zookeeper-2.0.0.3.0.1.0-187.jar:/usr/hdp/3.0.1.0-187/hbase/lib/jackson-annotations-2.9.5.jar:/usr/hdp/3.0.1.0-187/hbase/lib/jackson-core-2.9.5.jar > > > > > > On Tue, Nov 27, 2018 at 4:26 PM Josh Elser <els...@apache.org > > <mailto:els...@apache.org>> wrote: > > > > To add a non-jar file to the classpath of a Java application, you > must > > add the directory containing that file to the classpath. > > > > Thus, the following is wrong: > > > > HADOOP_CLASSPATH=/usr/hdp/3.0.1.0-187/hbase/lib/hbase-protocol.jar:/etc/hbase/ > 3.0.1. > > <http://3.0.1.>0-187/0/hbase-site.xml > > > > And should be: > > > > HADOOP_CLASSPATH=/usr/hdp/3.0.1.0-187/hbase/lib/hbase-protocol.jar:/etc/hbase/ > 3.0.1. > > <http://3.0.1.>0-187/0/ > > > > Most times, including the output of `hbase mapredcp` is sufficient > ala > > > > HADOOP_CLASSPATH="$(hbase mapredcp)" hadoop jar ... > > > > On 11/27/18 10:48 AM, M. Aaron Bossert wrote: > > > Folks, > > > > > > I have, I believe, followed all the directions for turning on > > namespace > > > mapping as well as extra steps to (added classpath) required to > > use the > > > mapreduce bulk load utility, but am still running into this > > error...I am > > > running a Hortonworks cluster with both HDP v 3.0.1 and HDF > > components. > > > Here is what I have tried: > > > > > > * Checked that the proper hbase-site.xml (in my case: > > > /etc/hbase/3.0.1.0-187/0/hbase-site.xml) file is being > referenced > > > when launching the mapreduce utility: > > > > > > > > > ... > > > > > > > > > <property> > > > > > > <name>phoenix.schema.isNamespaceMappingEnabled</name> > > > > > > <value>true</value> > > > > > > </property> > > > > > > <property> > > > > > > <name>phoenix.schema.mapSystemTablesToNamespace</name> > > > > > > <value>true</value> > > > > > > </property> > > > > > > > > > ... > > > > > > * added the appropriate classpath additions to the hadoop jar > > command > > > (zookeeper quorum hostnames changed to remove my corporate > > network > > > info as well as data directory): > > > > > > > > > > HADOOP_CLASSPATH=/usr/hdp/3.0.1.0-187/hbase/lib/hbase-protocol.jar:/etc/hbase/ > 3.0.1. > > <http://3.0.1.>0-187/0/hbase-site.xml > > > hadoop jar > > > /usr/hdp/3.0.1.0-187/phoenix/phoenix-5.0.0.3.0.1.0-187-client.jar > > > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table MYTABLE > --input > > > /ingest/MYCSV -z zk1,zk2,zk3 -g > > > > > > > > > ... > > > > > > > > > 18/11/27 15:31:48 INFO zookeeper.ReadOnlyZKClient: Close zookeeper > > > connection 0x1d58d65f to master-1.punch.datareservoir.net:2181 > > <http://master-1.punch.datareservoir.net:2181> > > > > > <http://master-1.punch.datareservoir.net:2181>, > master-2.punch.datareservoir.net:2181 > > <http://master-2.punch.datareservoir.net:2181> > > > > > <http://master-2.punch.datareservoir.net:2181>, > master-3.punch.datareservoir.net:2181 > > <http://master-3.punch.datareservoir.net:2181> > > > <http://master-3.punch.datareservoir.net:2181> > > > > > > 18/11/27 15:31:48 INFO log.QueryLoggerDisruptor: Shutting down > > > QueryLoggerDisruptor.. > > > > > > Exception in thread "main" java.sql.SQLException: ERROR 726 > > > (43M10):Inconsistent namespace mapping properties. Cannot initiate > > > connection as SYSTEM:CATALOG is found but client does not have > > > phoenix.schema.isNamespaceMappingEnabled enabled > > > > > > at > > > > > > > org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:494) > > > > > > at > > > > > > > org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:150) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1113) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1501) > > > > > > at > > > > > > > org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2740) > > > > > > at > > > > > > > org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:1114) > > > > > > at > > > > > > > org.apache.phoenix.compile.CreateTableCompiler$1.execute(CreateTableCompiler.java:192) > > > > > > at > > > > > > org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:408) > > > > > > at > > > > > > org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:391) > > > > > > at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:390) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:378) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1806) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2569) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2532) > > > > > > at > > > > > > > org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2532) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150) > > > > > > at > > org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221) > > > > > > at java.sql.DriverManager.getConnection(DriverManager.java:664) > > > > > > at java.sql.DriverManager.getConnection(DriverManager.java:208) > > > > > > at > > org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:400) > > > > > > at > > org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:392) > > > > > > at > > > > > > > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(AbstractBulkLoadTool.java:206) > > > > > > at > > > > > > > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(AbstractBulkLoadTool.java:180) > > > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) > > > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90) > > > > > > at > > > > > > org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:109) > > > > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > > > > at > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > > > > > at > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > > > > > at java.lang.reflect.Method.invoke(Method.java:498) > > > > > > at org.apache.hadoop.util.RunJar.run(RunJar.java:318) > > > > > > at org.apache.hadoop.util.RunJar.main(RunJar.java:232) > > > > > > 18/11/27 15:31:48 INFO zookeeper.ZooKeeper: Session: > > 0x3672eebffa800c8 > > > closed > > > > > > 18/11/27 15:31:48 INFO zookeeper.ClientCnxn: EventThread shut down > > > > > > * Also tried the other recommended option: > > > > > > HADOOP_CLASSPATH=$(hbase > > > mapredcp):/etc/hbase/3.0.1.0-187/0/hbase-site.xml hadoop jar > > > /usr/hdp/3.0.1.0-187/phoenix/phoenix-5.0.0.3.0.1.0-187-client.jar > > > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table MYTABLE > --input > > > /ingest/MYCSV -z zk1,zk2,zk3 -g > > > > > > > > > ... > > > > > > > > > 18/11/27 15:31:48 INFO zookeeper.ReadOnlyZKClient: Close zookeeper > > > connection 0x1d58d65f to master-1.punch.datareservoir.net:2181 > > <http://master-1.punch.datareservoir.net:2181> > > > > > <http://master-1.punch.datareservoir.net:2181>, > master-2.punch.datareservoir.net:2181 > > <http://master-2.punch.datareservoir.net:2181> > > > > > <http://master-2.punch.datareservoir.net:2181>, > master-3.punch.datareservoir.net:2181 > > <http://master-3.punch.datareservoir.net:2181> > > > <http://master-3.punch.datareservoir.net:2181> > > > > > > 18/11/27 15:31:48 INFO log.QueryLoggerDisruptor: Shutting down > > > QueryLoggerDisruptor.. > > > > > > Exception in thread "main" java.sql.SQLException: ERROR 726 > > > (43M10):Inconsistent namespace mapping properties. Cannot initiate > > > connection as SYSTEM:CATALOG is found but client does not have > > > phoenix.schema.isNamespaceMappingEnabled enabled > > > > > > at > > > > > > > org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:494) > > > > > > at > > > > > > > org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:150) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1113) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1501) > > > > > > at > > > > > > > org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2740) > > > > > > at > > > > > > > org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:1114) > > > > > > at > > > > > > > org.apache.phoenix.compile.CreateTableCompiler$1.execute(CreateTableCompiler.java:192) > > > > > > at > > > > > > org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:408) > > > > > > at > > > > > > org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:391) > > > > > > at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:390) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:378) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1806) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2569) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2532) > > > > > > at > > > > > > > org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76) > > > > > > at > > > > > > > org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2532) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255) > > > > > > at > > > > > > > org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150) > > > > > > at > > org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221) > > > > > > at java.sql.DriverManager.getConnection(DriverManager.java:664) > > > > > > at java.sql.DriverManager.getConnection(DriverManager.java:208) > > > > > > at > > org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:400) > > > > > > at > > org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:392) > > > > > > at > > > > > > > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(AbstractBulkLoadTool.java:206) > > > > > > at > > > > > > > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(AbstractBulkLoadTool.java:180) > > > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) > > > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90) > > > > > > at > > > > > > org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:109) > > > > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > > > > at > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > > > > > at > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > > > > > at java.lang.reflect.Method.invoke(Method.java:498) > > > > > > at org.apache.hadoop.util.RunJar.run(RunJar.java:318) > > > > > > at org.apache.hadoop.util.RunJar.main(RunJar.java:232) > > > > > > 18/11/27 15:31:48 INFO zookeeper.ZooKeeper: Session: > > 0x3672eebffa800c8 > > > closed > > > > > > 18/11/27 15:31:48 INFO zookeeper.ClientCnxn: EventThread shut down > > > > > > * As well as the recommended approach in the HBase reference > guide > > > linked in the Phoenix docs: > > > > > > HADOOP_CLASSPATH=`${HBASE_HOME}/bin/hbase classpath` hadoop jar > > > /usr/hdp/3.0.1.0-187/phoenix/phoenix-5.0.0.3.0.1.0-187-client.jar > > > org.apache.phoenix.mapreduce.CsvBulkLoadTool --table MYTABLE > --input > > > /ingest/MYCSV -z zk1,zk2,zk3 -g > > > > > > > > > Exception in thread "main" java.lang.NoClassDefFoundError: > > > org/apache/commons/cli/DefaultParser > > > > > > at > > > > > > > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.parseOptions(AbstractBulkLoadTool.java:128) > > > > > > at > > > > > > > org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(AbstractBulkLoadTool.java:176) > > > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) > > > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90) > > > > > > at > > > > > > org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:109) > > > > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > > > > at > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > > > > > at > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > > > > > at java.lang.reflect.Method.invoke(Method.java:498) > > > > > > at org.apache.hadoop.util.RunJar.run(RunJar.java:318) > > > > > > at org.apache.hadoop.util.RunJar.main(RunJar.java:232) > > > > > > Caused by: java.lang.ClassNotFoundException: > > > org.apache.commons.cli.DefaultParser > > > > > > at java.net.URLClassLoader.findClass(URLClassLoader.java:381) > > > > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:424) > > > > > > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331) > > > > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:357) > > > > > > ... 11 more > > > > > > * And finally, here is what the tables look like in both Hbase > > shell > > > and sqlline: > > > > > > hbase shell > > > > > > HBase Shell > > > > > > Use "help" to get list of supported commands. > > > > > > Use "exit" to quit this interactive shell. > > > > > > Version 2.0.0.3.0.1.0-187, > > re9fcf450949102de5069b257a6dee469b8f5aab3, > > > Wed Sep 19 10:16:35 UTC 2018 > > > > > > Took 0.0016 seconds > > > > > > hbase(main):001:0> list > > > > > > TABLE > > > > > > ATLAS_ENTITY_AUDIT_EVENTS > > > > > > MYTABLE > > > > > > SYSTEM:CATALOG > > > > > > SYSTEM:FUNCTION > > > > > > SYSTEM:LOG > > > > > > SYSTEM:MUTEX > > > > > > SYSTEM:SEQUENCE > > > > > > SYSTEM:STATS > > > > > > atlas_janus > > > > > > 9 row(s) > > > > > > Took 0.6114 seconds > > > > > > => ["ATLAS_ENTITY_AUDIT_EVENTS", "MYTABLE", "SYSTEM:CATALOG", > > > "SYSTEM:FUNCTION", "SYSTEM:LOG", "SYSTEM:MUTEX", > "SYSTEM:SEQUENCE", > > > "SYSTEM:STATS", "atlas_janus"] > > > > > > > > > > > > > > > > > > phoenix-sqlline master-1.punch.datareservoir.net > > <http://master-1.punch.datareservoir.net> > > > <http://master-1.punch.datareservoir.net> > > > > > > *Setting property: [incremental, false]* > > > > > > *Setting property: [isolation, TRANSACTION_READ_COMMITTED]* > > > > > > *issuing: !connect jdbc:phoenix:mysrv none none > > > org.apache.phoenix.jdbc.PhoenixDriver* > > > > > > *Connecting to jdbc:phoenix:mysrv* > > > > > > SLF4J: Class path contains multiple SLF4J bindings. > > > > > > SLF4J: Found binding in > > > > > > > [jar:file:/usr/hdp/3.0.1.0-187/phoenix/phoenix-5.0.0.3.0.1.0-187-client.jar!/org/slf4j/impl/StaticLoggerBinder.class] > > > > > > SLF4J: Found binding in > > > > > > > [jar:file:/usr/hdp/3.0.1.0-187/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] > > > > > > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for > an > > > explanation. > > > > > > 18/11/27 15:45:51 WARN util.NativeCodeLoader: Unable to load > > > native-hadoop library for your platform... using builtin-java > > classes > > > where applicable > > > > > > *Connected to: Phoenix (version 5.0)* > > > > > > *Driver: PhoenixEmbeddedDriver (version 5.0)* > > > > > > *Autocommit status: true* > > > > > > *Transaction isolation: TRANSACTION_READ_COMMITTED* > > > > > > Building list of tables and columns for tab-completion (set > > fastconnect > > > to true to skip)... > > > > > > 144/144 (100%) Done > > > > > > Done > > > > > > sqlline version 1.2.0 > > > > > > 0: jdbc:phoenix:mysrv> !tables > > > > > > > > > > *+------------+--------------+-------------+---------------+----------+------------+----------------------------+-----------------+--------------+-----------------+---------------+---------------+-----------------+------------+---------+* > > > > > > *| **TABLE_CAT**| **TABLE_SCHEM**| **TABLE_NAME**|**TABLE_TYPE **| > > > **REMARKS**| **TYPE_NAME**| **SELF_REFERENCING_COL_NAME**| > > > **REF_GENERATION**| **INDEX_STATE**| **IMMUTABLE_ROWS**| > > > **SALT_BUCKETS**| **MULTI_TENANT**| **VIEW_STATEMENT**| > > **VIEW_TYPE**| > > > **INDEX_T**|* > > > > > > > > > > *+------------+--------------+-------------+---------------+----------+------------+----------------------------+-----------------+--------------+-----------------+---------------+---------------+-----------------+------------+---------+* > > > > > > *|**| *SYSTEM *| *CATALOG *| *SYSTEM TABLE*|**|**|**| **|**| > > *false *| > > > *null*| *false *| **|**| **|* > > > > > > *|**| *SYSTEM *| *FUNCTION*| *SYSTEM TABLE*|**|**|**| **|**| > > *false *| > > > *null*| *false *| **|**| **|* > > > > > > *|**| *SYSTEM *| *LOG *| *SYSTEM TABLE*|**|**|**| **|**| *true*| > > *32*| > > > *false *| **|**| **|* > > > > > > *|**| *SYSTEM *| *SEQUENCE*| *SYSTEM TABLE*|**|**|**| **|**| > > *false *| > > > *null*| *false *| **|**| **|* > > > > > > *|**| *SYSTEM *| *STATS *| *SYSTEM TABLE*|**|**|**| **|**| *false > *| > > > *null*| *false *| **|**| **|* > > > > > > *|**|**| *MYTABLE*| *TABLE *|**|**|**| **|**| *false *| *5 *| > > *false *| > > > **|**| **|* > > > > > > > > > > *+------------+--------------+-------------+---------------+----------+------------+----------------------------+-----------------+--------------+-----------------+---------------+---------------+-----------------+------------+---------+* > > > > > >