Thank you !  I will try it!

在 2017年11月16日星期四 UTC+8上午4:47:06,Pascal Lamblin写道:
>
> It looks like in the following line: 
>  > ali_to_05[:,1:] - ali_to_05[:,:-1] 
>
> you are subtracting Booleans from Booleans, which is not supported. 
> Maybe cast `ali_to_05` to int8? 
>
> On 2017-11-15 09:03 AM, [email protected] <javascript:> wrote: 
> > I trained the model . But at last it output this error ?  Can anyone 
> > help me? ,thankyou 
> > 
> > 
> -------------------------------------------------------------------------------
>  
>
> > TRAINING HAS BEEN FINISHED: 
> > 
> -------------------------------------------------------------------------------
>  
>
> > .。。。。。。 
> > 
> > Traceback (most recent call last): 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/bin/run.py", 
> > line 154, in <module> 
> >      getattr(lvsr.main, args.pop('func'))(config, **args) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/lvsr/main.py", 
>
> > line 920, in train_multistage 
> >      stage_params, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/lvsr/main.py", 
>
> > line 693, in train 
> >      load_log, fast_start) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/lvsr/main.py", 
>
> > line 325, in initialize_all 
> >      batch=True, prediction=prediction, prediction_mask=prediction_mask) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/lvsr/bricks/recognizer.py",
>  
>
> > line 445, in get_cost_graph 
> >      **inputs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/lvsr/bricks/recognizer.py",
>  
>
> > line 390, in cost 
> >      attended=encoded, attended_mask=encoded_mask) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/sequence_generators.py",
>  
>
> > line 326, in cost_matrix 
> >      return self.evaluate(outputs, mask=mask, **kwargs)[0] 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/sequence_generators.py",
>  
>
> > line 269, in evaluate 
> >      **dict_union(inputs, states, contexts)) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/attention.py",
>  
>
> > line 738, in apply 
> >      preprocessed_attended})) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/recurrent.py",
>  
>
> > line 231, in recurrent_apply 
> >      brick.name, application.application_name)) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/scan_module/scan.py",
>  
>
> > line 773, in scan 
> >      condition, outputs, updates = 
> > scan_utils.get_updates_and_outputs(fn(*args)) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/recurrent.py",
>  
>
> > line 213, in scan_function 
> >      outputs = application(iterate=False, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/recurrent.py",
>  
>
> > line 133, in recurrent_apply 
> >      return application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/attention.py",
>  
>
> > line 703, in do_apply 
> >      self.preprocessed_attended_name: preprocessed_attended})) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/attention.py",
>  
>
> > line 616, in take_glimpses 
> >      **dict_union(states, glimpses_needed)) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 377, in __call__ 
> >      return self.application.apply(self, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/blocks/blocks/bricks/base.py",
>  
>
> > line 312, in apply 
> >      outputs = self.application_function(brick, *args, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/lvsr/bricks/attention.py",
>  
>
> > line 141, in take_glimpses 
> >      ali_median_pos = ali_to_05[:,1:] - ali_to_05[:,:-1] 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/tensor/var.py",
>  
>
> > line 147, in __sub__ 
> >      return theano.tensor.basic.sub(self, other) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/gof/op.py",
>  
>
> > line 615, in __call__ 
> >      node = self.make_node(*inputs, **kwargs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/tensor/elemwise.py",
>  
>
> > line 578, in make_node 
> >      DimShuffle, *inputs) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/tensor/elemwise.py",
>  
>
> > line 520, in get_output_info 
> >      for i in inputs]) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/scalar/basic.py",
>  
>
> > line 1041, in make_node 
> >      for input in inputs])] 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/scalar/basic.py",
>  
>
> > line 1049, in output_types 
> >      variables = self.output_types_preference(*types) 
> >    File 
> > 
> "/nobackup/f1/asr/wulong/tensorflowEnv/attention-lvcsr-master/libs/Theano/theano/scalar/basic.py",
>  
>
> > line 842, in upcast_out_nobool 
> >      raise TypeError("bool output not supported") 
> > TypeError: bool output not supported 
> > 
> > -- 
> > 
> > --- 
> > You received this message because you are subscribed to the Google 
> > Groups "theano-users" group. 
> > To unsubscribe from this group and stop receiving emails from it, send 
> > an email to [email protected] <javascript:> 
> > <mailto:[email protected] <javascript:>>. 
> > For more options, visit https://groups.google.com/d/optout. 
>
> -- 
> Pascal Lamblin 
>

-- 

--- 
You received this message because you are subscribed to the Google Groups 
"theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to