I have an Hadoop Streaming program that crawls the web for data items,
processes each retrieved item and then stores the results on S3. For each
processed item a directory on S3 is created to store the results produced by
the processing. At the conclusion of a program run I've been getting a
duplic
Can you look for Exception from jetty in JT logs and report here? That
would tell us the cause for ERROR 500.
Thanks
Amareshwari
Nathan Marz wrote:
Sometimes I am unable to access a job's details and instead only see.
I am seeing this on 0.19.2 branch.
HTTP ERROR: 500
Internal Server Error
On Sun, Mar 22, 2009 at 2:17 PM, nitesh bhatia
wrote:
> Bigtable ???
> Is it opensource ? I am not sure if google has released any code of
> bigtable. So far only 1 research paper is available.
No, Google has never released source code for MapReduce and BigTable.
Hadoop and HBase attempt to fill
On Sun, Mar 22, 2009 at 11:47:35PM +0530, nitesh bhatia wrote:
> Bigtable ???
> Is it opensource ? I am not sure if google has released any code of
> bigtable. So far only 1 research paper is available.
HBase is an implementation of BigTable.
--
Philip
smime.p7s
Description: S/MIME cryptograph
Bigtable ???
Is it opensource ? I am not sure if google has released any code of
bigtable. So far only 1 research paper is available.
--nitesh
On Tue, Mar 17, 2009 at 11:01 AM, Amandeep Khurana wrote:
> AFAIK, Google uses BigTable for pretty much most of their backend stuff.
> The
> thing to n
I'm trying to run the Dirichlet clustering example from
(http://cwiki.apache.org/MAHOUT/syntheticcontroldata.html). The command
line:
$HADOOP_HOME/bin/hadoop jar
$MAHOUT_HOME/examples/target/mahout-examples-0.1.job
org.apache.mahout.clustering.syntheticcontrol.dirichlet.Job
... loads our ex