Thank you Eric W., Eric R. Colin and Nicolas.

All your links have been extremely useful in a particular way.

I ended up tweaking Eric R's script to fit my needs. Great to know that at
least the 1/3 arcsecond files are available in a web accessible folder.
Nevertheless it would be nice if
http://tdds.cr.usgs.gov/ned/13arcsec/grid/grid_zips/ just listed all the
files instead of having to guess the file naming convention.

For those interested, according to this
announcement<http://nationalmap.gov/data_delivery/ned.html>,
the 1/9 arcsecond datasets will be available from there soon.

Thank you all again,

- Ragi

On Wed, Feb 6, 2013 at 7:22 AM, Eric Wolf <[email protected]> wrote:

> USGS Eros has a service where you can ship hard drives to them to get bulk
> data, including NED:
>
> http://cumulus.cr.usgs.gov/bulk.php
>
> As a friend once said, "Never underestimate the bandwidth of a station
> wagon filled with backup tapes!"
>
> -Eric
>
> -=--=---=----=----=---=--=-=--=---=----=---=--=-=-
> Eric B. Wolf                           720-334-7734
>
>
>
>
> On Wed, Feb 6, 2013 at 7:31 AM, Eric Russell <[email protected]>wrote:
>
>>  Hi Ragi,
>>
>> Here's my Python code for downloading and unzipping the 1/3 arcsecond NED
>> from tdds.cr.usgs.gov:
>>
>> def download_zips ():
>>     for x in xrange(99, 66, -1):
>>         for y in xrange(48, 24, -1):
>>             try:
>>                 url =
>> "http://tdds.cr.usgs.gov/ned/13arcsec/grid/grid_zips/n%sw0%s.zip"; % (y,
>> x)
>>                 dest = "D:\\ned\\zip\\n%sw%s.zip" % (y, x)
>>                 opener = urllib.URLopener()
>>                 opener.retrieve(url, dest)
>>                 print "downloaded /n%sw%s.zip" % (y, x)
>>             except IOError:
>>                 print "skipped /n%sw%s.zip" % (y, x)
>>
>> def unzip_files ():
>>     for zip_file in glob.iglob("D:\\ned\\zip\\*.zip"):
>>         print zip_file
>>         with zipfile.ZipFile(zip_file, 'r') as zip:
>>             for f in zip.infolist():
>>                 name_index = f.filename.find('/grd')
>>                 if name_index < 0:
>>                     continue
>>                 dest = os.path.join('E:', f.filename[name_index:])
>>                 if dest.endswith('/'):
>>                     os.makedirs(dest)
>>                 else:
>>                     with open(dest, 'wb') as fout:
>>                         fout.write(zip.read(f))
>>
>> Eric
>>
>>
>> On 2/5/13 11:33 PM, Ragi Burhum wrote:
>>
>> Hello all,
>>
>>  I am trying to bulk download* the entire 1/9 arc and 1/3 arc NED dataset
>> * (I have the rest). I have a server with gigabit connection and several
>> terabytes of space - so space/network speed is not an issue. Any pointers
>> on how to get a hold of that dataset? I could write a scraping script, but
>> I would rather not go there.
>>
>>  Thanks!
>>
>>  - Ragi
>>
>>
>> _______________________________________________
>> Geowanking mailing 
>> [email protected]http://geowanking.org/mailman/listinfo/geowanking_geowanking.org
>>
>>
>>
>> _______________________________________________
>> Geowanking mailing list
>> [email protected]
>> http://geowanking.org/mailman/listinfo/geowanking_geowanking.org
>>
>>
>
_______________________________________________
Geowanking mailing list
[email protected]
http://geowanking.org/mailman/listinfo/geowanking_geowanking.org

Reply via email to