On 08/07/2017 04:11 PM, Clemens Ladisch wrote:
Oleksandr Andrushchenko wrote:
On 08/07/2017 01:27 PM, Clemens Ladisch wrote:
You have to implement period interrupts (and the .pointer callback)
based on when the samples are actually moved from/to the backend.
Do you think I can implement this in a slightly different way,
without a timer at all, by updating
substream->runtime->hw_ptr_base explicitly in the frontend driver?
As far as I am aware, hw_ptr_base is an internal field that drivers
are not supposed to change.
I know that and always considered not a good solution,
this is why I have timer to emulate things
Just use your own variable, and return it from the .pointer callback.
this can work, but see below
So, that way, whenever I get an ack/response from the backend that it has
successfully played the buffer
That response should come after every period.
How does that interface work?
For the buffer received in .copy_user/.copy_kernel we send
a request to the backend and get response back (async) when it has copied
the bytes into HW/mixer/etc, so the buffer at frontend side can be reused.
So, the amount of bytes in this exchange is not necessarily
a multiply of the period. Also, there is no way to synchronize period
sizes in the front driver and backend to make those equal.
There is also no event from the backend in the
protocol to tell that the period has elapsed, so
sending data in period's size buffers will not probably
help because of possible underruns
Is it possible to change the period size,
or at least to detect what it is?
Unfortunately no, this is not in the protocol.
you can see the protocol at 
Xen-devel mailing list