On Friday 29 October 2004 11:13, Johannes Erdfelt wrote:
> >      If usbfs instead grabbed a bunch of non-contiguous
> > pages, copied the data into them from user-space, and then sent it (using
> > scatter-gather io), then there is no longer any memory pressure problem.
> > What's more, there would also be no reason not to increase the maximum
> > data size from 16k to something much bigger, say 256k.  Allowing bigger
> > buffers should increase performance.
> 
> Duh, that would be the obvious way to increase performance and
> reliability.
> 
> Not to mention it should be relatively easy to implement.

Yep, and you'd be able to clean up properly after an error in
the middle of your N urbs ... the scatterlist code handles
such things.

Another tweak would be to remove the limit on transfer size.
Use whatever algorithm you want to transfer the first chunk;
for high speed, a scatterlist is *really* desirable.  Then
repeat for successive chunks ... until error, or all done.
(And if you can't get enough buffers to use big chunks, then
you could still make progress using smaller chunks...)

- Dave


-------------------------------------------------------
This SF.Net email is sponsored by:
Sybase ASE Linux Express Edition - download now for FREE
LinuxWorld Reader's Choice Award Winner for best database on Linux.
http://ads.osdn.com/?ad_id=5588&alloc_id=12065&op=click
_______________________________________________
[EMAIL PROTECTED]
To unsubscribe, use the last form field at:
https://lists.sourceforge.net/lists/listinfo/linux-usb-devel

Reply via email to