Krishna, You should be able to take the command you are using to start the hadoop job (hadoop jar ..) and paste it into a text file. Then make the file executable and call it as a shell script in a CRON job (crontab -e). To be safe, use absolute paths to reference any files in the command.
Or, I suppose what you crazy kids and your object oriented programming would do is use Quartz. -John On Sat, Jan 5, 2013 at 4:33 PM, Chitresh Deshpande < [email protected]> wrote: > Hi Krishna, > > I dont know what do you mean by Hadoop daemon, but if you mean run when > all the other hadoop daemons like namenode, datanode etc are started, then > you can change start-all file in conf directory. > > Thanks and Regards, > Chitresh Deshpande > > > On Fri, Jan 4, 2013 at 6:40 AM, Krishna Rao <[email protected]>wrote: > >> Hi al, >> >> I have a java application jar that converts some files and writes >> directly into hdfs. >> >> If I want to run the jar I need to run it using "hadoop jar <application >> jar>", so that it can access HDFS (that is running "java -jar <application >> jar> results in a HDFS error"). >> >> Is it possible to run an jar as a hadoop daemon? >> >> Cheers, >> >> Krishna >> > >
