I followed the example to load the COCO data sets as follows:

    from gluoncv import data, utils
    train_dataset = data.COCODetection(splits=['instances_train2017'])
    val_dataset = data.COCODetection(splits=['instances_val2017'])

The train_dataset is about 117266 images and labels.
Now I am simply trying to loop through the data-set:

    for i in range(len(train_dataset)):
        train_image, train_label = train_dataset[i]
    

This loop simply eats up all of my 15Gb of RAM and I am not even 60% of data. 

How can I make sure that my memory is used properly or how can i fix a batch 
size etc?





---
[Visit 
Topic](https://discuss.mxnet.apache.org/t/out-of-memory-when-using-gluoncv-data-cocodetection/6778/1)
 or reply to this email to respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click 
here](https://discuss.mxnet.apache.org/email/unsubscribe/6373da637b9ad7269b59f4ba0a49689df37efeacb17f0231a0cdc8669232abf4).

Reply via email to