I'm plagued with this error: Retrying connect to server: localhost/127.0.0.1:9000.
I'm trying to set up hadoop on a new machine, just a basic pseudo-distributed setup. I've done this quite a few times on other machines, but this time I'm kinda stuck. I formatted the namenode without obvious errors and ran start-all.sh with no errors to stdout. However, the logs are full of that error above and if I attempt to access hdfs (ala "hadoop fs -ls /") I get that error again. Obviously, my core-site.xml sets fs.default.name to "hdfs://localhost:9000". I assume something is wrong with /etc/hosts, but I'm not sure how to fix it. If "hostname" returns X and "hostname -f" returns Y, then what are the corresponding entries in /etc/hosts? Thanks for any help. ________________________________________________________________________________ Keith Wiley kwi...@keithwiley.com keithwiley.com music.keithwiley.com "I used to be with it, but then they changed what it was. Now, what I'm with isn't it, and what's it seems weird and scary to me." -- Abe (Grandpa) Simpson ________________________________________________________________________________