On Sat, 2004-05-29 at 09:31, Tim Goetze wrote: > depends on what 'system' we're talking about, which isn't really clear > in the first place (nor is it too important, but here goes anyway ... > :) > > you can see the whole setup as the system, then latency is the time > from keypress to voltage change at the DAC out. or you can just look > at the kernel side as as you do. or you can look at the time from MIDI > interrupt to the audio DAC converting the first affected audio sample. > > all examples of valid 'systems' to look at in this context, depending > on whether you assume the musician's, the kernel- or the audio > application-programmer's view. > Agreed. But the problem that keeps popping up on the lists is that people who are not doing live sound, have cards that do hardware monitoring, and don't need to use a tiny buffer size waste their time trying to get the minimum buffer size because they think they need to get 2-3ms latency. So, they get xruns out the wazoo and wonder how to fix it. In this case he *needs* a small buffer size though since he's doing, essentially, a live application.
Jan ------------------------------------------------------- This SF.Net email is sponsored by: Oracle 10g Get certified on the hottest thing ever to hit the market... Oracle 10g. Take an Oracle 10g class now, and we'll give you the exam FREE. http://ads.osdn.com/?ad_id=3149&alloc_id=8166&op=click _______________________________________________ Alsa-devel mailing list [EMAIL PROTECTED] https://lists.sourceforge.net/lists/listinfo/alsa-devel