Yeah, this is where things get a bit tricky. You'll have to experiment with what works for you, but we are using Hive to launch the job with the jar.sh script. This gets the environment straight from the Hive side.

jar_help () {
echo "Used for applications that require Hadoop and Hive classpath and environment."
  echo "./hive --service jar <yourjar> <yourclass> HIVE_OPTS <your_args>"
}

Avery

On 1/17/13 4:49 PM, pradeep kumar wrote:

Hi,

Actually we are trying to use giraph in our project for graph analysis with hive, so far it was good build was successful shortestpath example ran fine but working with hive is been a real issue. we started with command line

hadoop jar giraph-hcatalog-0.2-SNAPSHOT-jar-with-dependencies.jar org.apache.giraph.io.hcatalog.HiveGiraphRunner -db default -vertexClass org.apache.giraph.vertex.Vertex -vertexInputFormatClass org.apache.giraph.io.hcatalog.HCatalogVertexInputFormat -vertexOutputFormatClass org.apache.giraph.io.hcatalog.HCatalogVertexOutputFormat -w 1 -vi testinput -o testoutput -hiveconf javax.jdo.option.ConnectionURL=jdbc:mysql://localhost/metastore -hiveconf javax.jdo.option.ConnectionDriverName=com.mysql.jdbc.Driver -hiveconf javax.jdo.option.ConnectionUserName=root -hiveconf javax.jdo.option.ConnectionPassword=root -hiveconf datanucleus.autoCreateSchema=false -hiveconf datanucleus.fixedDatastore=true

is it a wrong way of doing it.. because we are running into exception while doing so..

and if its wrong,

then any suggestion on how can we proceed will be a great help.

Regards,

Pradeep


Reply via email to