Everyone on this list is probably geeky enough to already know about 
what-if.xkcd.com<http://what-if.xkcd.com>, but just in case:

http://what-if.xkcd.com/31/


On Feb 6, 2013, at 10:22 AM, Eric Wolf 
<[email protected]<mailto:[email protected]>>
 wrote:

USGS Eros has a service where you can ship hard drives to them to get bulk 
data, including NED:

http://cumulus.cr.usgs.gov/bulk.php

As a friend once said, "Never underestimate the bandwidth of a station wagon 
filled with backup tapes!"

-Eric

-=--=---=----=----=---=--=-=--=---=----=---=--=-=-
Eric B. Wolf                           720-334-7734




On Wed, Feb 6, 2013 at 7:31 AM, Eric Russell 
<[email protected]<mailto:[email protected]>> wrote:
Hi Ragi,

Here's my Python code for downloading and unzipping the 1/3 arcsecond NED from 
tdds.cr.usgs.gov<http://tdds.cr.usgs.gov/>:

def download_zips ():
    for x in xrange(99, 66, -1):
        for y in xrange(48, 24, -1):
            try:
                url = 
"http://tdds.cr.usgs.gov/ned/13arcsec/grid/grid_zips/n%sw0%s.zip";<http://tdds.cr.usgs.gov/ned/13arcsec/grid/grid_zips/n%sw0%s.zip>
 % (y, x)
                dest = "D:\\ned\\zip\\n%sw%s.zip<smb://ned//zip//n%sw%s.zip>" % 
(y, x)
                opener = urllib.URLopener()
                opener.retrieve(url, dest)
                print "downloaded /n%sw%s.zip" % (y, x)
            except IOError:
                print "skipped /n%sw%s.zip" % (y, x)

def unzip_files ():
    for zip_file in glob.iglob("D:\\ned\\zip\\*.zip<smb://ned//zip//*.zip>"):
        print zip_file
        with zipfile.ZipFile(zip_file, 'r') as zip:
            for f in zip.infolist():
                name_index = f.filename.find('/grd')
                if name_index < 0:
                    continue
                dest = os.path.join('E:', f.filename[name_index:])
                if dest.endswith('/'):
                    os.makedirs(dest)
                else:
                    with open(dest, 'wb') as fout:
                        fout.write(zip.read(f))

Eric


On 2/5/13 11:33 PM, Ragi Burhum wrote:
Hello all,

I am trying to bulk download the entire 1/9 arc and 1/3 arc NED dataset (I have 
the rest). I have a server with gigabit connection and several terabytes of 
space - so space/network speed is not an issue. Any pointers on how to get a 
hold of that dataset? I could write a scraping script, but I would rather not 
go there.

Thanks!

- Ragi

_______________________________________________
Geowanking mailing list
[email protected]
http://geowanking.org/mailman/listinfo/geowanking_geowanking.org

Reply via email to