You're still a bit confused :). John's link might help too to read. The "FileSystem" I referred to originally is a class provided by HDFS. It has multiple implementations (notably the LocalFileSystem and DistributedFileSystem) which are returned when you use FileSystem.get(...) based on the contents of the Configuration object pulled in by the hdfs configuration files (core-site.xml and hdfs-site.xml). A complete programmatic API is available via HDFS -- this is what Accumulo uses.
Regarding your most recent error, you should be providing a directory of RFiles. Not sure what's in the text file you provided, or what you intended it to do... On Tue, Apr 8, 2014 at 12:59 PM, pdread <[email protected]> wrote: > Josh > > I want to thank you...you solved some of my problems, unintentionally but > none the less. > > I added all the hadoop config files to the classpath and bingo, accumulo can > now find the files...the only down side now is my load file must require a > certain extension. I had used ".txt" but it complained with > > 2014-04-08 12:53:47,764 [tableOps.BulkImport] WARN : > hdfs://localhost:9000/data/accu1/hdfs/tmp/bulk/entities/load/entities.txt > does not have a valid extension, ignoring > > So now the hunt is on for the correct RFile extension. > > Thanks > > Paul > > > > -- > View this message in context: > http://apache-accumulo.1065345.n5.nabble.com/bulk-ingest-without-mapred-tp8904p8917.html > Sent from the Users mailing list archive at Nabble.com.
