AudioUnits do not get to control the buffer size delivered via a render call. The host decides this.
On Mon, Nov 30, 2015 at 12:08 AM, Daniel Wilson <[email protected]> wrote: > Does anyone know how to change the frame size when doing the digital > signal processing on an audio unit? Currently my audio unit is set up so > that it receives a single sample, does the signal processing, outputs the > sample, and repeats the process for each sample of the audio signal. I have > created quite a few audio units with this set up but now I want to process > multiple samples at the same time to do the FFT/IFFT, etc. Does anyone know > how to do this? It seems like most people are using audio units for iiOS, > but my audio units are for OS X to be used in programs like Logic Pro. > Don’t know if that makes a difference. > > -Daniel > _______________________________________________ > Do not post admin requests to the list. They will be ignored. > Coreaudio-api mailing list ([email protected]) > Help/Unsubscribe/Update your Subscription: > > https://lists.apple.com/mailman/options/coreaudio-api/paul%40linuxaudiosystems.com > > This email sent to [email protected]
_______________________________________________ Do not post admin requests to the list. They will be ignored. Coreaudio-api mailing list ([email protected]) Help/Unsubscribe/Update your Subscription: https://lists.apple.com/mailman/options/coreaudio-api/archive%40mail-archive.com This email sent to [email protected]
