This is an automated email from the ASF dual-hosted git repository. zhreshold pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/incubator-mxnet.git
The following commit(s) were added to refs/heads/master by this push: new bcf4245 Some Python 3 fixes in ./example (#11671) bcf4245 is described below commit bcf4245056d2a3961280a0a072eb4ef92e8f2fc5 Author: cclauss <ccla...@bluewin.ch> AuthorDate: Mon Jul 16 23:36:02 2018 +0200 Some Python 3 fixes in ./example (#11671) * Some Python 3 fixes in ./example * mxnet.base.string_types and xrange() --> range() --- example/gluon/data.py | 7 ++++--- example/kaggle-ndsb1/gen_img_list.py | 5 +---- example/reinforcement-learning/ddpg/ddpg.py | 4 ++-- example/ssd/dataset/pycocotools/coco.py | 12 ++++++------ 4 files changed, 13 insertions(+), 15 deletions(-) diff --git a/example/gluon/data.py b/example/gluon/data.py index 5a7f9e6..6aa5316 100644 --- a/example/gluon/data.py +++ b/example/gluon/data.py @@ -19,6 +19,7 @@ """ data iterator for mnist """ import os import random +import tarfile import logging logging.basicConfig(level=logging.INFO) @@ -111,7 +112,7 @@ def get_caltech101_iterator(batch_size, num_workers, dtype): # center and crop an area of size (224,224) cropped, crop_info = mx.image.center_crop(resized, 224) # transpose the channels to be (3,224,224) - transposed = nd.transpose(cropped, (2, 0, 1)) + transposed = mx.nd.transpose(cropped, (2, 0, 1)) image = mx.nd.cast(image, dtype) return image, label @@ -119,8 +120,8 @@ def get_caltech101_iterator(batch_size, num_workers, dtype): dataset_train = ImageFolderDataset(root=training_path, transform=transform) dataset_test = ImageFolderDataset(root=testing_path, transform=transform) - train_data = gluon.data.DataLoader(dataset_train, batch_size, shuffle=True, num_workers=num_workers) - test_data = gluon.data.DataLoader(dataset_test, batch_size, shuffle=False, num_workers=num_workers) + train_data = mx.gluon.data.DataLoader(dataset_train, batch_size, shuffle=True, num_workers=num_workers) + test_data = mx.gluon.data.DataLoader(dataset_test, batch_size, shuffle=False, num_workers=num_workers) return DataLoaderIter(train_data), DataLoaderIter(test_data) class DummyIter(mx.io.DataIter): diff --git a/example/kaggle-ndsb1/gen_img_list.py b/example/kaggle-ndsb1/gen_img_list.py index adfc4fe..8b27550 100644 --- a/example/kaggle-ndsb1/gen_img_list.py +++ b/example/kaggle-ndsb1/gen_img_list.py @@ -18,7 +18,6 @@ from __future__ import print_function import csv import os -import sys import random import numpy as np import argparse @@ -57,7 +56,7 @@ head = "acantharia_protist_big_center,acantharia_protist_halo,acantharia_protist img_lst = [] cnt = 0 if args.train: - for i in xrange(len(head)): + for i in range(len(head)): path = args.image_folder + head[i] lst = os.listdir(args.image_folder + head[i]) for img in lst: @@ -104,5 +103,3 @@ if args.train: tr_fo.writerow(item) for item in va_lst: va_fo.writerow(item) - - diff --git a/example/reinforcement-learning/ddpg/ddpg.py b/example/reinforcement-learning/ddpg/ddpg.py index aa34e4d..14c1795 100644 --- a/example/reinforcement-learning/ddpg/ddpg.py +++ b/example/reinforcement-learning/ddpg/ddpg.py @@ -190,7 +190,7 @@ class DDPG(object): end = False obs = self.env.reset() - for epoch in xrange(self.n_epochs): + for epoch in range(self.n_epochs): logger.push_prefix("epoch #%d | " % epoch) logger.log("Training started") for epoch_itr in pyprind.prog_bar(range(self.epoch_length)): @@ -220,7 +220,7 @@ class DDPG(object): obs = nxt if memory.size >= self.memory_start_size: - for update_time in xrange(self.n_updates_per_sample): + for update_time in range(self.n_updates_per_sample): batch = memory.get_batch(self.batch_size) self.do_update(itr, batch) diff --git a/example/ssd/dataset/pycocotools/coco.py b/example/ssd/dataset/pycocotools/coco.py index a8939f6..19a7b8b 100755 --- a/example/ssd/dataset/pycocotools/coco.py +++ b/example/ssd/dataset/pycocotools/coco.py @@ -55,12 +55,12 @@ import itertools # from . import mask as maskUtils import os from collections import defaultdict -import sys -PYTHON_VERSION = sys.version_info[0] -if PYTHON_VERSION == 2: - from urllib import urlretrieve -elif PYTHON_VERSION == 3: +from mxnet.base import string_types +try: from urllib.request import urlretrieve +except ImportError: + from urllib import urlretrieve + class COCO: def __init__(self, annotation_file=None): @@ -302,7 +302,7 @@ class COCO: print('Loading and preparing results...') tic = time.time() - if type(resFile) == str or type(resFile) == unicode: + if type(resFile) in string_types: anns = json.load(open(resFile)) elif type(resFile) == np.ndarray: anns = self.loadNumpyAnnotations(resFile)