Skip to content

Latest commit

 

History

History
315 lines (265 loc) · 20 KB

BACKENDS.md

File metadata and controls

315 lines (265 loc) · 20 KB

FlexASIO backends

This document provides background information about what FlexASIO backends are, and the differences between them. It is recommended reading for anyone wanting to optimize their audio pipeline. In particular, this document provides the necessary information to understand what the backend and wasapiExclusiveMode FlexASIO configuration options do.

A backend is just another term for what PortAudio calls the host API. FlexASIO uses the term "backend" to avoid potential confusion with the term "ASIO host".

The following architecture diagram shows the entire audio path, from the ASIO host application (e.g. Cubase) to the hardware audio device (arrows represent the direction of audio streaming in the playback direction):

FlexASIO architecture diagram

The backends are shown in red on the diagram. They lead to various paths that the audio can take through the PortAudio library, and then through the Windows audio subsystem, before reaching the audio hardware driver and device.

Roughly speaking, each backend maps to a single audio API that Windows makes available for applications: MME, DirectSound, WASAPI and Kernel Streaming. The main difference between the backends is therefore which Windows API they use to play and record audio.

In terms of code paths, there are large differences between the various backends, both on the Windows side and on the PortAudio side. Therefore, the choice of backend can have a large impact on the behaviour of the overall audio pipeline. In particular, choice of backend can affect:

  • Reliability: some backends can be more likely to work than others, because of the quality and maturity of the code involved. Some might be more likely to introduce audio discontinuities (glitches).
  • Ease of use: Some backends are more "opinionated" than others about when it comes to which audio formats (i.e. sample rate, number of channels) they will accept. They might refuse to work if the audio format is not exactly the one they expect.
  • Latency: depending on the backend, the audio pipeline might include additional layers of buffering that incur additional latency.
  • Latency reporting: some backends are better than others at reporting the end-to-end latency of the entire audio pipeline. Some backends underestimate their own total latency, resulting in FlexASIO reporting misleading latency numbers.
  • Exclusivity: some backends will seize complete control of the hardware device, preventing any other application from using it.
  • Audio processing: some backends will automatically apply additional processing on audio data. Specifically:
    • Sample rate conversion: some backends will accept any sample rate and then convert to whatever sample rate Windows is configured to use.
    • Sample type conversion: some backends will accept any sample type (16-bit integer, 32-bit float, etc.) and then convert to whatever sample type Windows is configured to use.
    • Downmixing: some backends will accept any channel count and then downmix (or maybe even upmix) to whatever channel count Windows is configured to use.
    • Mixing: if the backend is not exclusive (see above), audio might be mixed with the streams from other applications.
    • APOs: Hardware audio drivers can come bundled with so-called Audio Processing Objects (APOs), which can apply arbitrary pre-mixing (SFX) or post-mixing (MFX/EFX/GFX) processing on audio data. Some backends will bypass this processing, some won't.
  • Automatic device switching: some backends can be set up to track the audio device set as default in the Windows audio control panel. If the default device changes, the stream is automatically and transparently redirected to the new device. Other backends don't support this, and can only select a specific device that cannot change while streaming is taking place.
  • Feature set: some FlexASIO features and options might not work with all backends.

Note: the internal buffer size of the shared Windows audio pipeline has been observed to be 10 ms. This means that only exclusive backends (i.e. WASAPI Exclusive, WDM-KS) can achieve an actual latency below 10 ms.

Note: In addition to APOs, hardware devices can also implement additional audio processing at a low level in the audio driver, or baked into the hardware itself (DSP). Choice of backend cannot affect such processing.

Note: none of the backends seem to be capable of taking the hardware itself into account when reporting latency. So, for example, when playing audio over Bluetooth, the reported latency will not take the Bluetooth stack into account, and will therefore be grossly underestimated.

MME backend

Multimedia Extensions is the first audio API that was introduced in Windows, going all the way back to 1991. Despite its age, it is still widely used because of its ubiquity and simplicity. In PortAudio (but not FlexASIO), it is the default host API used on Windows. It should therefore be expected to be highly mature and extremely reliable; if it doesn't work, it is unlikely any other backend will.

MME goes through the entire Windows audio pipeline, including sample rate conversion, mixing, and APOs. As a result it is extremely permissive when it comes to audio formats. It should be expected to behave just like a typical Windows application would. Its latency should be expected to be mediocre at best, as MME was never designed for low-latency operation. This is compounded by the fact that MME appears to behave very poorly with small buffer sizes.

Latency numbers reported by MME do not seem to take the Windows audio pipeline into account. This means the reported latency is underestimated by at least 10 ms, if not more.

The MME backend exposes "virtual" devices called "Microsoft Sound Mapper - Input" and "Microsoft Sound Mapper - Output". These virtual devices automatically track the Windows audio device selected as "default" in the Windows audio control panel, and will redirect the stream transparently if that setting changes.

Note: the channel count exposed by FlexASIO when using MME can be a bit odd. For example, it might expose 8 channels for a 5.1 output, downmixing the rear channels pairs.

Modern versions of Windows implement the MME API by using WASAPI Shared internally, making this backend a "second-class citizen" compared to WASAPI and WDM-KS.

DirectSound backend

DirectSound was introduced in 1995. It is mainly designed for use in video games.

In practice, DirectSound should be expected to behave somewhat similarly to MME. It will accept most reasonable audio formats. Audio goes through the entire Windows pipeline, converting as necessary.

One would expect latency to be somewhat better than MME, though it's not clear if that's really the case in practice. The DirectSound backend has been observed to behave very poorly with small buffer sizes on the input side, where it appears to enforce an effective minimum buffer size of 31.25 ms, making it a poor choice for low-latency capture use cases.

The DirectSound backend exposes "virtual" devices called "Primary Sound Capture Driver" and "Primary Sound Driver". These virtual devices automatically track the Windows audio device selected as "default" in the Windows audio control panel, and will redirect the stream transparently if that setting changes.

Modern versions of Windows implement the DirectSound API by using WASAPI Shared internally, making this backend a "second-class citizen" compared to WASAPI and WDM-KS.

WASAPI backend

Windows Audio Session API (WASAPI), first introduced in Windows Vista, is the most modern Windows audio API. Microsoft actively supports WASAPI and regularly releases new features for it. PortAudio supports most of these features and makes them available through various options (though, sadly, FlexASIO doesn't provide ways to leverage all these options yet).

WASAPI is the only entry point to the shared Windows audio processing pipeline, which runs in the Windows Audio service (audiosrv). In modern versions of Windows, DirectSound and MME merely forward to WASAPI internally. This means that, in theory, one cannot achieve better performance than WASAPI when using MME or DirectSound. In practice however, the PortAudio side of the WASAPI backend could conceivably have limitations or bugs that these other backends don't have.

WASAPI can be used in two different modes: shared or exclusive. In FlexASIO, the wasapiExclusiveMode option determines which mode is used. The two modes behave very differently; in fact, they should probably be seen as two separate backends entirely.

In shared mode, WASAPI behaves similarly to MME and DirectSound, in that the audio goes through the normal Windows audio processing pipeline, including mixing and APOs. Indeed, in modern versions of Windows, MME and DirectSound are just thin wrappers implemented on top of WASAPI Shared. For this reason it is reasonable to assume that this mode will provide the best possible latency for a shared backend.

In exclusive mode, WASAPI behaves completely differently and bypasses the entirety of the Windows audio pipeline. As a result, PortAudio has a nearly direct path to the audio hardware driver, which makes for an ideal setup for low latency operation - latencies in the single-digit milliseconds have been observed. The lack of APOs in the signal path can also be helpful in applications where fidelity to the original signal is of the utmost importance, and in fact, this mode can be used for "bit-perfect" operation. However, since mixing is bypassed, other applications cannot use the audio device at the same time.

In both modes, the latency numbers reported by WASAPI appear to be more reliable than other backends.

WASAPI provides a feature called loopback recording, which makes it possible to capture whatever audio signal is being played through an output device. This feature is supported by PortAudio (and FlexASIO), which exposes it in the form of extra "virtual" input devices whose names end with [Loopback]. To record the audio flowing through a particular output device, simply use the corresponding loopback device. Note that this feature cannot be used if the output device or loopback device are operating in exclusive mode. The signal is mirrored near the end of the Windows audio engine pipeline, after all APOs (including CAudioLimiter) but before conversion to integer. The mirrored signal is expected to be a bit-perfect reflection of the samples passing through the Windows audio engine as long as: 32-bit floating-point samples are used (which is normally the case if the sampleType option isn't used), and the sample rate and channel count match the Windows settings for the device (this can be enforced by disabling the wasapiAutoConvert option). If the sample type, sample rate, or channel count don't match, then the usual automatic conversions will occur.

The WASAPI backend cannot redirect the stream if the default Windows audio device changes while streaming.

Note: FlexASIO will show channel names (e.g. "Front left") when the WASAPI backend is used. Channel names are not shown when using other backends due to PortAudio limitations.

Windows implements WASAPI using Kernel Streaming internally.

The alternative ASIO2WASAPI universal ASIO driver uses WASAPI.

WDM-KS backend

Kernel Streaming was first introduced in Windows 98 and is the interface between Windows audio device drivers (also known as WDM, which stands for Windows Driver Model), running in kernel mode, and processes running in user mode. Note that, in this context, "Windows audio device driver" refers to Windows hardware drivers, not ASIO drivers - the two concepts are not directly related to each other.

The Windows audio engine itself (WASAPI) uses Kernel Streaming internally to communicate with audio device drivers. It logically follows that any device that behaves as a normal Windows audio device de facto comes with a WDM driver that implements Kernel Streaming (usually not directly but through a PortCls miniport driver). Calls made through any of the standard Windows audio APIs (MME, DirectSound, WASAPI) eventually become Kernel Streaming calls as they cross the boundary into kernel mode and enter the Windows audio device driver.

In the typical case, the only client of the Windows audio device driver is the Windows audio engine. It is, however, technically possible, albeit highly atypical, for an application to issue Kernel Streaming requests directly, bypassing the Windows audio engine and talking to the Windows kernel directly. This is what the WDM-KS PortAudio backend does.

Given the above, Kernel Streaming offers the most direct path to the audio device among all PortAudio backends, but comes with some downsides:

  • Many audio outputs only handle a single stream at a time, because many audio devices do not support hardware mixing. (In Kernel Streaming terms, their pins only support one instance at a time.) These devices can therefore only be used by one KS client at a time, making Kernel Streaming an exclusive backend in this case. Because the Windows audio engine is itself a KS client, it is usually not possible to access an audio device using KS if the Windows audio engine is already using it. Use the device option to select a device that the Windows audio engine is not currently using.
  • Kernel Streaming is a very flexible API, which also makes it quite complicated. Even just enumerating audio devices involves quite a bit of complex, error-prone logic to be implemented in the application (here, in the PortAudio KS backend). Different device drivers implement KS calls in different ways, report different topologies, and even different ways of handling audio buffers. This presents a lot of opportunities for things to go wrong in a variety of different ways depending on the specific audio device used. Presumably this is the reason why most applications do not attempt to use KS directly, and the reason why Microsoft does not recommend this approach.

Note that the list of devices that the PortAudio WDM-KS backend exposes might look a bit different from the list of devices shown in the Windows audio settings. This is because the Windows audio engine generates its own list of devices (or, more specifically, audio endpoint devices) by interpreting information returned by Kernel Streaming. When using KS directly this logic is bypassed, and the PortAudio WDM-KS backend uses its own logic to discover devices. Furthermore, the concept of a device "name" is specific to the Windows audio engine and does not apply to KS, which explains why PortAudio WDM-KS device names do not necessarily match the Windows audio settings.

In principle, similar results should be obtained when using WASAPI Exclusive and Kernel Streaming, since they both offer exclusive access to the hardware. WASAPI is simpler and less likely to cause problems, but Kernel Streaming is more direct and more flexible. For example, Kernel Streaming will typically provide access to underlying hardware buffer sizes that WASAPI Exclusive would not use internally. Furthermore, the PortAudio implementation is very different. Therefore, the WASAPI Exclusive and WDM-KS PortAudio backends might behave somewhat differently depending on the situation.

The WDM-KS backend cannot redirect the stream if the default Windows audio device changes while streaming.

The alternative ASIO4ALL and ASIO2KS universal ASIO drivers use Kernel Streaming.


ASIO is a trademark and software of Steinberg Media Technologies GmbH