foreachAsync at RemoteHiveSparkClient.java

2014-12-18 Thread yuemeng1
hi,all i execute a sql on hive on spark,the comand like: select distinct st.sno,sname from student st join score sc on(st.sno=sc.sno) where sc.cno IN(11,12,13) and st.sage 28;(some days ago this sql can work ) but it give me some Info in hive shell : Query Hive on Spark job[0] stages: 0

Hive UDFs?

2014-12-18 Thread Xiaoyong Zhu
Hi experts, I heard that if I want to use Hive UDFs, I must deploy the jars to all the machines that are running Hive (which is painful to me..) After reading thishttps://cwiki.apache.org/confluence/display/Hive/HivePlugins I didn't find any document talking about this - could someone help to

RE: Hive UDFs?

2014-12-18 Thread Xiaoyong Zhu
Thanks! For Q1 - is there a way that Hive helps me to automatically do this (for example, I can register the UDF somewhere and the UDF gets automatically distributed)? Or I need to login to each node to ensure this happens? Xiaoyong -Original Message- From: Rathish A M

RE: Hive UDFs?

2014-12-18 Thread Xiaoyong Zhu
+ Rathish Xiaoyong -Original Message- From: Xiaoyong Zhu Sent: Thursday, December 18, 2014 5:56 PM To: user@hive.apache.org Subject: RE: Hive UDFs? Thanks! For Q1 - is there a way that Hive helps me to automatically do this (for example, I can register the UDF somewhere and the UDF

Re: Row Delimiter in Create table

2014-12-18 Thread Gayathri Swaroop
Thank you. On Wed, Dec 17, 2014 at 8:55 PM, Navis류승우 navis@nexr.com wrote: Afaik, it was restricted by implementation of hadoop. But now hadoop-2 supports custom delimiter, hopefully it also can be implemented in hive. I'm not sure but currently possible way of do that is setting

Re: Hive UDFs?

2014-12-18 Thread Stéphane Verlet
You only need the jar on the computer where you execute the hive command , not on all cluster nodes On Thu, Dec 18, 2014 at 2:55 AM, Xiaoyong Zhu xiaoy...@microsoft.com wrote: Thanks! For Q1 - is there a way that Hive helps me to automatically do this (for example, I can register the UDF

Re: Hive UDFs?

2014-12-18 Thread Jason Dere
Another option is https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-PermanentFunctions, as another user mentioned on this list a few days ago. On Dec 18, 2014, at 5:54 AM, Stéphane Verlet kaweahsoluti...@gmail.com wrote: You only need the jar on the

Re: custom binary format

2014-12-18 Thread Ingo Thon
Hi thanks for the answer so far, however, I still think there must be an easy way. The file format I’m looking at is pretty simple. There is first an header of n bytes, Which can be ignored. After that there is the data. The data consists of rows where ich rows has 9 bytes. First there is a

Re: custom binary format

2014-12-18 Thread Andrew Mains
Hi Ingo, Take a look at https://hadoop.apache.org/docs/r2.3.0/api/org/apache/hadoop/mapred/FixedLengthInputFormat.html--it seems to be designed for use cases very similar to yours. You may need to subclass it to make things work precisely the way you need (in particular, to deal with the

Re: custom binary format

2014-12-18 Thread Ingo Thon
Hello Andrew, this one looks indeed like a good idea. However, there is also another Problem already here. This InputFormat expects that conf.setInt(FixedLengthInputFormat.FIXED_RECORD_LENGTH, recordLength); is set. I haven’t found any way to specify a parameter for a InputFormat. I couldn’t

Re: custom binary format

2014-12-18 Thread Andrew Mains
So in hive you can actually do that via the SET command (documented here https://cwiki.apache.org/confluence/display/Hive/LanguageManual+Cli) as follows: hive SET fixedlengthinputformat.record.length = length This value will be passed through to the JobConf, and the input format ought to

RE: Hive UDFs?

2014-12-18 Thread Xiaoyong Zhu
I see, thanks! Xiaoyong From: Jason Dere [mailto:jd...@hortonworks.com] Sent: Friday, December 19, 2014 3:52 AM To: user@hive.apache.org Subject: Re: Hive UDFs? Another option is https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-PermanentFunctions, as

RE: relationship between a hive query and a tez dag

2014-12-18 Thread Xiaoyong Zhu
Sorry to update this again - but why don't we do a cross query optimization and make the query into 1 DAG (if all the queries in a certain script are linked with each other).. this seems a more optimized way.. Xiaoyong From: Xiaoyong Zhu [mailto:xiaoy...@microsoft.com] Sent: Thursday, December