Hi Mark,
Hm. It should work. This is Spark 1.6.1 on Oracle 12c scala> val HiveContext = new org.apache.spark.sql.hive.HiveContext(sc) HiveContext: org.apache.spark.sql.hive.HiveContext = org.apache.spark.sql.hive.HiveContext@70f446c scala> var _ORACLEserver : String = "jdbc:oracle:thin:@rhes564:1521:mydb12" _ORACLEserver: String = jdbc:oracle:thin:@rhes564:1521:mydb12 scala> var _username : String = "sh" _username: String = sh scala> var _password : String = "xxxx" _password: String = sh scala> val c = HiveContext.load("jdbc", | Map("url" -> _ORACLEserver, | "dbtable" -> "(SELECT to_char(CHANNEL_ID) AS CHANNEL_ID, CHANNEL_DESC FROM sh.channels)", | "user" -> _username, | "password" -> _password)) warning: there were 1 deprecation warning(s); re-run with -deprecation for details c: org.apache.spark.sql.DataFrame = [CHANNEL_ID: string, CHANNEL_DESC: string] scala> c.registerTempTable("t_c") scala> c.count res2: Long = 5 scala> HiveContext.sql("select * from t_c").collect.foreach(println) [3,Direct Sales] [9,Tele Sales] [5,Catalog] [4,Internet] [2,Partners] HTH Dr Mich Talebzadeh LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* http://talebzadehmich.wordpress.com *Disclaimer:* Use it at your own risk. Any and all responsibility for any loss, damage or destruction of data or any other property which may arise from relying on this email's technical content is explicitly disclaimed. The author will in no case be liable for any monetary damages arising from such loss, damage or destruction. On 11 July 2016 at 08:25, Mark Vervuurt <m.a.vervu...@gmail.com> wrote: > Hi Mich, > > sorry for bothering did you manage to solve your problem? We have a > similar problem with Spark 1.5.2 using a JDBC connection with a DataFrame > to an Oracle Database. > > Thanks, > Mark > > On 12 Feb 2016, at 11:45, Mich Talebzadeh <m...@peridale.co.uk> wrote: > > Hi, > > I use the following to connect to Oracle DB from Spark shell 1.5.2 > > spark-shell --master spark://50.140.197.217:7077 --driver-class-path > /home/hduser/jars/ojdbc6.jar > > in Scala I do > > scala> val sqlContext = new org.apache.spark.sql.SQLContext(sc) > sqlContext: org.apache.spark.sql.SQLContext = > org.apache.spark.sql.SQLContext@f9d4387 > > scala> val channels = sqlContext.read.format("jdbc").options( > | Map("url" -> "jdbc:oracle:thin:@rhes564:1521:mydb", > | "dbtable" -> "(select * from sh.channels where channel_id = > 14)", > | "user" -> "sh", > | "password" -> "xxxxxxx")).load > channels: org.apache.spark.sql.DataFrame = [CHANNEL_ID: decimal(0,-127), > CHANNEL_DESC: string, CHANNEL_CLASS: string, CHANNEL_CLASS_ID: > decimal(0,-127), CHANNEL_TOTAL: string, CHANNEL_TOTAL_ID: decimal(0,-127)] > > *scala> channels.count()* > > But the latter command keeps hanging? > > Any ideas appreciated > > Thanks, > > Mich Talebzadeh > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > http://talebzadehmich.wordpress.com > > NOTE: The information in this email is proprietary and confidential. This > message is for the designated recipient only, if you are not the intended > recipient, you should destroy it immediately. Any information in this > message shall not be understood as given or endorsed by Peridale Technology > Ltd, its subsidiaries or their employees, unless expressly so stated. It is > the responsibility of the recipient to ensure that this email is virus > free, therefore neither Peridale Technology Ltd, its subsidiaries nor their > employees accept any responsibility. > > > Met vriendelijke groet | Best regards, > _______________________________________________________________________ > > > *Ir. Mark Vervuurt* > Senior Big Data Scientist | Insights & Data > > Capgemini Nederland | Utrecht > Tel.: +31 30 6890978 – Mob.: +31653670390 > www.capgemini.com > > *People matter, results count.* > > __________________________________________________________________ > > >