Jeff,

 

Yes, and I thought about that (it's the only way I know of other than
modifying your current voice); however, you give up (or bypass) all the
processing of speech that WE gives you.  This would be everything from their
dictionaries to splitting of mixed case words, or spelling of alphanumeric
words, etc.  And there would be nothing that would coordinate your sapi
voice with any other voice, so if you were relying on your WE voice for most
speech, and then you wanted something else to be spoken in this different
SAPI voice, what would happen would likely be that both voices would speak
their strings at the same time.

 

Chip

 

 

From: Jeff Weiss [mailto:[email protected]] 
Sent: Friday, January 17, 2014 12:01 PM
To: [email protected]
Subject: RE: Using a secondary voice for speech messages

 

You could use SAPI like this:

 

' SapiSpeak

 

wScript.Sleep 2000

 

strText = "Please wait while we get ready for lunch! "

 

Set objVoice = CreateObject("SAPI.SpVoice")

objVoice.Speak strText

 

hth

Jeff Weiss

 

 

From: Chip Orange [mailto:[email protected]] 
Sent: Friday, January 17, 2014 10:41 AM
To: [email protected]
Subject: RE: Using a secondary voice for speech messages

 

Hi David,

 

I absolutely agree with you about the need for this.  I've tried to add it
to a couple of my apps, but not had a lot of luck, because the WE object
model just doesn't support this idea.

 

It does support it just a little, in that the second parameter of the speak
method allows you to specify whether it should be spoken in the screen,
mouse, or keyboard voice settings.  This requires the user to go and alter
these voice settings (which they may not want to do), and the only thing it
allows you to change is the basic pitch or tone of the synthesizer in use,
but you cannot use more than one synthesizer concurrently.

 

So, I tried to do this in my apps without requiring the user to alter the
keyboard, mouse, or screen voices, and even without changing synthes I ran
into the problem Jim brings up: there's some delay introduced even when
staying with the same synthesizer.  If you try to change synthesizers the
delay is intolerable.

 

If you stay with the same synth, and you want this just to happen
automatically without requiring the user to go alter their basic 3 voice
settings, it's very difficult to determine which changes to speech
parameters will actually produce a noticib le change in the sound of the
voice.  You can look at my app named Word Advanced Features for some code
where I try to deal with this automatically, but I've come to believe it may
be better just to ask the user to dedicate one of the mouse or keyboard
voices to this function, and then ask them to make the changes they wish, so
that they will be able to notice the difference in voice sound.

 

Hth,

 

Chip

 

 

 

From: David [mailto:[email protected]] 
Sent: Friday, January 17, 2014 7:56 AM
To: [email protected]
Subject: Using a secondary voice for speech messages

 

Allright, you developers,

Here is another tricky question. Hope someone has a solution, as it would
have improved my projects a bit.

 

Le'ts for samplification say, that a user is operating his computer by
Eloquence Reed. Like many users actually are doing. 

 

My app sends a few messages to the synthesizer, like status messages in
given operations, or resulting messages from other operations. Since they
are all sent to the same synth, it is likely it only will drown in the
amount of other activity the user is performing. And, he may not be able to
discern what the message is about. This could specially be the case, when an
app is performing some activity in the background, or in a separate window,
and the user currently is reading a document. Now, if the status message
from the app pops in, right in the middle of the document reading, either
the user will think it may be part of the document, or he may experience a
break-down of the speech altogether. 

 

It would have many benefits, if the app could have its "internal" messages
sent to another voice, available on the user's system. Now, with the 

    Speak

command, in WE, is there a way to redirect the output to a secondary synth
or voice? Or, is there a "back-door" workaround to this issue?

 

I greatly would welcome any ideas you all may have, and maybe you have
implied any such solution in your apps already, and could please share some
techniques? Thanks alot,

 

 

Reply via email to