Thank you. On Thursday, May 11, 2017 at 9:22:48 AM UTC+8, Pascal Lamblin wrote: > > Unfortunately, model parallelism using different GPUs in the same > process is still experimental, and not a current focus for development. > > > On Wed, May 10, 2017, T Qri wrote: > > Hi everyone, > > > > In brief, I'm implementing a model with 2 seq2seq components which share > an > > encoder. Each of the 2 components are distributed across different GPU > > devices and the shared encoder is also placed on one of these 2 devices. > > > > When config.optimizer is set to None, the error doesn't exist anymore > but > > the model runs on multiple CPUs quite slowly. > > > > The detailed Theano flags are described in `theano-flags` file. And the > > corresponding error message is in `error-message` file. > > > > I tried to disable constant_folding and elemwise optimization with > > `optimizer_excluding=constant_folding elemwise` set but it didn't work. > > > > I'm using Theano 0.9.0 version on Ubuntu 14.04. There are 4 GPUs > available > > and 2 of them are involved in my model. > > > > Thanks for your help. > > > > > > -- > > > > --- > > You received this message because you are subscribed to the Google > Groups "theano-users" group. > > To unsubscribe from this group and stop receiving emails from it, send > an email to [email protected] <javascript:>. > > For more options, visit https://groups.google.com/d/optout. > > > > > -- > Pascal >
-- --- You received this message because you are subscribed to the Google Groups "theano-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.
