//localhost:50111/templeton/v1/ddl/database/default?user.name=%3cmyname%3e>
so your URL suggests that you have a database called "testtable"?
I really hope that this helps.
Regards,
Peter Marron
Senior Developer
Trillium Software, A Harte Hanks Company
Theale Court, 1st Floor, 11
iewed as a sensible start?
I'm only just starting and so I'm not really in a position to submit patches yet
but I thought that it would be sensible to see if these sort of changes are
going
to be acceptable.
Regards,
Peter Marron
Senior Developer
Trillium Software, A Harte Hanks Company
Theal
>From https://cwiki.apache.org/confluence/display/Hive/Home
"Hive is not designed for OLTP workloads and does not offer real-time queries
or row-level updates."
As far as I am aware "UPDATE" isn't even in the Hive DML.
Z
Peter Marron
Senior Developer
Trillium Sof
Hi,
Not sure if it is relevant to your problem but I'm just checking
that you know about
hive.optimize.index.filter.compact.minsize
it's set to 5Gbytes by default and if the estimated query size is
less than this then the index won't be used.
HTH.
Regards
Peter Marron
Senior Deve
e as a command line option to the tool.
-- Lefty
On Tue, Jan 28, 2014 at 2:39 AM, Peter Marron
mailto:peter.mar...@trilliumsoftware.com>>
wrote:
Hi,
So I can see from http://hive.apache.org/downloads.html
that I can download versions 11 and 12 and they will work with
Hadoop 1.0.4 which I am cur
If it does work will I have to re-build my metastore?
Any recommendations?
Peter Marron
Office: +44 (0) 118-940-7609
peter.mar...@trilliumsoftware.com<mailto:peter.mar...@trilliumsoftware.com>
Theale Court First Floor, 11-13 High Street, Theale, RG7 5AH, UK
[cid:image009.png@0
have to move to the
latest version and then move to Hadoop 2 and that that will mean that I
will spend some time upgrading my cluster.
Is there anyone who can through any light on my problems? Or suggest
any way forward?
All feedback welcome.
Z
Peter Marron
Office: +44 (0) 118-940-7609
pete
o find out about it.)
Regards,
Peter Marron
Senior Developer, Research & Development
Office: +44 (0) 118-940-7609
peter.mar...@trilliumsoftware.com<mailto:peter.mar...@trilliumsoftware.com>
Theale Court First Floor, 11-13 High Street, Theale, RG7 5AH, UK
[cid:image001.png@01CEF1A7.CCFE
Hi,
I am using Hive 0.11.0 over Hadoop 1.0.4.
Recently I have started investigating the user of Templeton and I have managed
to get
most of the services working. Specifically I can access resources like these:
http://hpcluster1:50111/templeton/v1/version
http://hpcluster1:50111/templeton/v1/stat
": "string" },
{ "name": "Venue", "type": "string" },
{ "name": "Attendance", "type": "string"}
],
"location" : "/user/pmarron/Ex/_output/rows",
"format": {
"rowFormat":
Hi,
(I'm a little bit behind in reading the lists, so apologies if this is a
duplicate question.)
I am running Templeton v1 (?) and HCatalog 0.5.0 with hive 0.11.0 over Hadoop
1.0.4.
I can use something like this:
curl -s -X PUT -HContent-type:application/json -d @createtable.json
http://hpc
Sorry, just caught up with the last couple of day’s email and I feel that this
question
has already been answered fairly comprehensively. Apologies.
Z
From: Peter Marron [mailto:peter.mar...@trilliumsoftware.com]
Sent: 04 July 2013 08:37
To: user@hive.apache.org
Subject: RE: Partition
2:34 AM, Peter Marron
mailto:peter.mar...@trilliumsoftware.com>>
wrote:
Hi Owen,
I’m curious about this advice about partitioning. Is there some fundamental
reason why Hive
is slow when the number of partitions is 10,000 rather than 1,000?
The precise numbers don't matter. I wanted
Hi Owen,
I’m curious about this advice about partitioning. Is there some fundamental
reason why Hive
is slow when the number of partitions is 10,000 rather than 1,000? And the
improvements
that you mention are they going to be in version 12? Is there a JIRA raised so
that I can track them?
(It’
ssage-
From: Navis류승우 [mailto:navis@nexr.com]
Sent: 02 July 2013 08:50
To: user@hive.apache.org
Subject: Re: Override COUNT() function
MetadataOnlyOptimizer changes GBY on partition columns to simple TableScan with
one line dummy.
I think similar things can be done with stats.
2013/6/28
Hi,
I feel sure that someone has asked for this before, but here goes...
In the case where I have the query
SELECT COUNT(*) FROM table;
There are many cases where I can determine the count immediately.
(For example if I have run something like:
ANALYZE TABLE tablename [PARTITIO
er i'm not in a position to suggest anything.
On Thu, Jun 27, 2013 at 3:14 AM, Peter Marron
mailto:peter.mar...@trilliumsoftware.com>>
wrote:
Hi,
If you're suggesting that I use something like
SELECT * FROM data WHERE MyUdf(data. BLOCK__OFFSET__INSIDE__FILE);
rather than
S
e performance should be quite good.
To avoid the full table scan, partitioning is IMHO the best way to speed things
up.
Best regards,
J. Dolinar
On Thu, Jun 27, 2013 at 11:18 AM, Peter Marron
mailto:peter.mar...@trilliumsoftware.com>>
wrote:
Hi,
I have thought about a map-only join, but
2013/05/get-all-extended-hive-tables-with.html
if you want to do it programatically then you will need to look at
HiveMetaStoreClient
If both of these are not what you are looking for then sorry I will need a
little more details on your question
On Thu, Jun 27, 2013 at 2:19 PM, Peter Marron
can help you in this regard as well.
On Thu, Jun 27, 2013 at 2:16 PM, Peter Marron
mailto:peter.mar...@trilliumsoftware.com>>
wrote:
Well, I'm not very good at keeping things brief, unfortunately.
But I'll have a go, trying to keep things simple.
Suppose that I have a data t
Hi,
Hopefully a simple question.
Given that I have a table name (as a String) is there an API call that I can
use to obtain the location of the table? Assume that my code is executing
in a custom InputFormat to define the context. (I'm running "inside" a Hive
query so I assume that there's a way
painful than just hitting yourself in the head with a hammer. Different strokes
for different folks though.
so can we back up? what - briefly if possible - do you want to achieve with a
"wrapper"? (i'm going to regret asking that i know.)
On Tue, Jun 25, 2013 at 7:29 AM, Peter Marron
Hi,
Running Hive 0.11.0 over Hadoop 1.0.4.
I would like to be able to "wrap" a Hive table.
So, if I have table "X" which uses SerDe "s" and InputFormat "i"
then I would like to be able to create a table "Y" which has a
SerDe "ws" which is a wrapper of "s" (and so can encapsulate an instance of "
2013 16:48
To: user@hive.apache.org
Subject: Re: Use of virtual columns in joins
You might be hitting into https://issues.apache.org/jira/browse/HIVE-4033 in
which case its recommended that you upgrade to 0.11 where in this bug is fixed.
On Mon, Jun 10, 2013 at 1:57 AM, Peter Marron
mai
efault.xml
file which refer to Hive are all set to refer to
hive-0.10.0. Isn't this a weird version of hive
to configure given that this is part of release 0.11.0 of Hive?
Peter Marron
Trillium Software UK Limited
Tel : +44 (0) 118 940 7609
Fax : +44 (0) 118 940 7699
E: peter.mar...@trilliumsoftware.com<mailto:roy.willi...@trilliumsoftware.com>
Given that I started the original thread it seems appropriate that I should
point out that I also have a bought and paid for (personal) digital copy.
It's a good book.
Peter Marron
Trillium Software UK Limited
Tel : +44 (0) 118 940 7609
Fax : +44 (0) 118 940 7699
E: pete
Hi,
Using hive 0.10.0 over hadoop 1.0.4
I have a (non-partitioned) table with loads of columns.
I would like to create a partitioned table with the same set of columns.
So the approach that I have been taking is to use "CREATE TABLE copy LIKE
original;"
then I can use ALTER TABLE to change the l
Hi,
I'm using hive 0.10.0 over hadoop 1.0.4.
I have created a couple of test tables and found that various join queries
that refer to virtual columns fail. For example the query:
SELECT * FROM a JOIN b ON b.rownumber = a.number;
works but the following three queries all fail.
SELECT *,a.BLOCK
Hi,
Using Hive 0.10.0 over Hadoop 1.0.4.
I guess that I know that this is a long shot.
Is there any way to access the context from inside a UDF?
Specifically I want to get hold of the value of the virtual
column BLOCK__OFFSET__INSIDE__FILE from inside a
UDF that I'm implementing. Of course I can
me to examine the table properties and
the predicate and to (pre-)filter the records returned?
Or are you saying that what I am trying to do doesn't make sense?
Z
From: Edward Capriolo [mailto:edlinuxg...@gmail.com]
Sent: 28 May 2013 16:45
To: user@hive.apache.org
Cc: Peter Marron
Sub
ion:java.lang.RuntimeException:
java.lang.IllegalArgumentException: Can not create a Path from an empty string
Peter Marron
Trillium Software UK Limited
Tel : +44 (0) 118 940 7609
Fax : +44 (0) 118 940 7699
E: peter.mar...@trilliumsoftware.com<mailto:roy.willi...@trilliumsoftware.com>
r
separate cover.
However I would like to say thanks again. If we ever meet in the real world
I’ll stand you a beer (or equivalent).
Congratulations on version 0.11.0.
Z
aka
Peter Marron
Trillium Software UK Limited
Tel : +44 (0) 118 940 7609
Fax : +44 (0) 118 940 7699
E: peter.mar...@trilliumsoftware.com<mailto:roy.willi...@trilliumsoftware.com>
>>On Wed, May 15, 2013 at 3:38 AM, Peter Marron
>> wrote:
…
>I've started doing similar work for the ORC reader.
I guess that I’m glad that I’m not completely alone here.
>>
>>Firstly although that page mentions InputFormat there doesn’t seem to be any
>
Handler)
in order to get it to work at all.
Is it worth trying to build Hive from source so that I can hack out the tracing
that causes the NPE?
Or is it likely to start failing somewhere else?
Any comments welcome.
Peter Marron
Trillium Software UK Limited
Tel : +44 (0) 118 940 7609
Fax : +44 (0) 118 940 7699
E: peter.mar...@trilliumsoftware.com<mailto:roy.willi...@trilliumsoftware.com>
rmat wrapped class
is not calling my getSplits? And why this only seems to happen if a
Map/Reduce is required? And, most importantly, what do I have to
do to get it to work the way that I expect?
Any help or comments would be welcome.
Peter Marron
Trillium Software UK Limited
T
Hi Nitin,
Can I set these parameters through the CDH management interface?
If not then what file do they need to be set in to make sure that CDH
picks them up?
Peter Marron
Trillium Software UK Limited
Tel : +44 (0) 118 940 7609
Fax : +44 (0) 118 940 7699
E: peter.mar...@trilliumsoftware.com
te(SecureIOUtils.java:172)
... 7 more
Error: GC overhead limit exceeded
Al
If Also when the
If Also when the
If Also when the
If this e-mail shouldn't be here and should only be on
a cloudera mailing list, please re-direct me.
Thanks in advance.
Peter Marron
Trillium Software UK Limited
here it records Hive errors as opposed to Map/Reduce
errors?
Regards,
Peter Marron
From: Dean Wampler [mailto:dean.wamp...@thinkbiganalytics.com]
Sent: 02 November 2012 14:03
To: user@hive.apache.org
Subject: Re: Creating Indexes
Oh, I saw this line in your Hive output and just assumed you w
jar (derby-10.4.2.0.jar) into the
Hadoop directory, where I assume that the reducer would be able to find it.
However I get exactly the same problem as before.
Is there some particular place that I should put the derby.jar to make this
problem go away? Is there anything else that I can try?
Peter Ma
_
Deleted hdfs://localhost/data/warehouse/default__score_bigindex__
Invalid alter operation: Unable to alter index.
FAILED: Execution Error, return code 1 from
org.apache.hadoop.hive.ql.exec.DDLTask
So what have I done wrong, and what am I to do to get this index to build
successfully?
Any help app
his didn't seem to help either.
Maybe this is the wrong list for this question
and I should post to
common-u...@hadoop.apache.org<mailto:common-u...@hadoop.apache.org>?
Any help appreciated.
Peter Marron
2012-10-25 15:55:27,429 INFO org.apache.hadoop.mapred.ReduceTask: In-memory
merg
dexInputFormat;
You will then notice speed up for a query of the form,
select count(*) from tab where indexed_col = some_val
Thanks,
Shreepadma
On Tue, Oct 23, 2012 at 5:44 AM, Peter Marron
mailto:peter.mar...@trilliumsoftware.com>>
wrote:
Hi,
I'm very much a Hive newbie but I'
and developing any Java code of my own.
Is this true? If so, how?
Any help appreciated.
Peter Marron.
43 matches
Mail list logo