Hi Rich, Yes infact i was too thinking the same but then somehow slipped of my mind. I guess the second option would be really great so i don't even need to build the complex and length commands. The shared folder will be anyways appear as local in vm.
Thanks a lot Rich. Rgds, Anil On Thu, Dec 18, 2014 at 10:03 AM, Rich Haase <[email protected]> wrote: > > Anil, > > You have two main options: > > 1. install the hadoop software on OSX and add the configuration files > appropriate for your sandbox, then do use hdfs dfs –put <local> <remote> > 2. Setup your sandbox VM to share a directory between OS X and Linux. > All virtual machines that I know of support sharing a file system between > the VM and host. This is probably the easiest solution since it will allow > you to see the files you have on OS X in your Linux VM and then you can use > the hdfs/hadoop/yarn commands on linux (which you already have configured). > > Cheers, > > Rich > > *Rich Haase* | Sr. Software Engineer | Pandora > m 303.887.1146 | [email protected] > > From: Anil Jagtap <[email protected]> > Reply-To: "[email protected]" <[email protected]> > Date: Wednesday, December 17, 2014 at 3:58 PM > To: "[email protected]" <[email protected]> > Subject: Re: Copying files to hadoop. > > Yes i can do that but I have connected from my mac os terminal to linux > using ssh. > Now when I run LS command it shows me list of files & folders from Linux > and not from Mac OS. > I have files which I need to put onto Hadoop directly from Mac OS. > So something like below. > > From Mac OS Terminal: > > [root@sandbox ~]#hadoop fs -put <MAC OS FOLDER PATH/FILE> <HADOOP PATH> > > Hope my requirement is clear. > > Rgds, Anil > > > > > On Thu, Dec 18, 2014 at 9:39 AM, johny casanova <[email protected]> > wrote: >> >> Hi Anil, >> >> you can use the hadoop fs put "file" or directory and that should add it >> to your hdfs >> >> ------------------------------ >> Date: Thu, 18 Dec 2014 09:29:34 +1100 >> Subject: Copying files to hadoop. >> From: [email protected] >> To: [email protected] >> >> Dear All, >> >> I'm pretty new to Hadoop technology and Linux environment hence >> struggling even to find solutions for the basic stuff. >> >> For now, Hortonworks Sandbox is working fine for me and i managed to >> connect to it thru SSH. >> >> Now i have some csv files in my mac os folders which i want to copy >> onto Hadoop. As per my knowledge i can copy those files first to Linux and >> then put to Hadoop. But is there a way in which just in one command it will >> copy to Hadoop directly from mac os folder? >> >> Appreciate your advices. >> >> Thank you guys... >> >> Rgds, Anil >> >>
