Looks like database support was fixed by: [SPARK-7943] [SPARK-8105] [SPARK-8435] [SPARK-8714] [SPARK-8561] Fixes multi-database support
On Fri, Mar 25, 2016 at 3:35 PM, Ashok Kumar <ashok34...@yahoo.com> wrote: > 1.5.2 Ted. > > Those two lines I don't know where they come. It finds and gets the table > info OK > > HTH > > > On Friday, 25 March 2016, 22:32, Ted Yu <yuzhih...@gmail.com> wrote: > > > Which release of Spark do you use, Mich ? > > In master branch, the message is more accurate > (sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/NoSuchItemException.scala): > > override def getMessage: String = s"Table $table not found in database > $db" > > > On Fri, Mar 25, 2016 at 3:21 PM, Mich Talebzadeh < > mich.talebza...@gmail.com> wrote: > > You can use DESCRIBE FORMATTED <DATABASE>.<TABLE_NAME> to get that info. > > This is based on the same command in Hive however, it throws two erroneous > error lines as shown below (don't see them in Hive DESCRIBE ...) > > Example > > scala> sql("describe formatted test.t14").collect.foreach(println) > 16/03/25 22:32:38 ERROR Hive: Table test not found: test.test table not > found > 16/03/25 22:32:38 ERROR Hive: Table test not found: test.test table not > found > [# col_name data_type comment ] > [ ] > [invoicenumber int ] > [paymentdate date ] > [net decimal(20,2) ] > [vat decimal(20,2) ] > [total decimal(20,2) ] > [ ] > [# Detailed Table Information ] > [Database: test ] > [Owner: hduser ] > [ > *CreateTime: Fri Mar 25 22:13:44 GMT 2016 ]*[LastAccessTime: > UNKNOWN ] > [Protect Mode: None ] > [Retention: 0 ] > [Location: > hdfs://rhes564:9000/user/hive/warehouse/test.db/t14 ] > [Table Type: MANAGED_TABLE ] > [Table Parameters: ] > [ COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"}] > [ comment from csv file from excel sheet] > [ numFiles 2 ] > [ orc.compress ZLIB ] > [ totalSize 1090 ] > [ transient_lastDdlTime 1458944025 ] > [ ] > [# Storage Information ] > [SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde ] > [InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat ] > [OutputFormat: > org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat ] > [Compressed: No ] > [Num Buckets: -1 ] > [Bucket Columns: [] ] > [Sort Columns: [] ] > [Storage Desc Params: ] > [ serialization.format 1 ] > > HTH > > Dr Mich Talebzadeh > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > http://talebzadehmich.wordpress.com > > > On 25 March 2016 at 22:12, Ashok Kumar <ashok34...@yahoo.com.invalid> > wrote: > > Experts, > > I would like to know when a table was created in Hive database using Spark > shell? > > Thanks > > > > > >