/*************************************************************************/ /* audio_driver_wasapi.cpp */ /*************************************************************************/ /* This file is part of: */ /* GODOT ENGINE */ /* https://godotengine.org */ /*************************************************************************/ /* Copyright (c) 2007-2022 Juan Linietsky, Ariel Manzur. */ /* Copyright (c) 2014-2022 Godot Engine contributors (cf. AUTHORS.md). */ /* */ /* Permission is hereby granted, free of charge, to any person obtaining */ /* a copy of this software and associated documentation files (the */ /* "Software"), to deal in the Software without restriction, including */ /* without limitation the rights to use, copy, modify, merge, publish, */ /* distribute, sublicense, and/or sell copies of the Software, and to */ /* permit persons to whom the Software is furnished to do so, subject to */ /* the following conditions: */ /* */ /* The above copyright notice and this permission notice shall be */ /* included in all copies or substantial portions of the Software. */ /* */ /* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */ /* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */ /* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.*/ /* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */ /* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */ /* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */ /* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ /*************************************************************************/ #ifdef WASAPI_ENABLED #include "audio_driver_wasapi.h" #include "core/config/project_settings.h" #include "core/os/os.h" #include // INT32_MAX #include // Define IAudioClient3 if not already defined by MinGW headers #if defined __MINGW32__ || defined __MINGW64__ #ifndef __IAudioClient3_FWD_DEFINED__ #define __IAudioClient3_FWD_DEFINED__ typedef interface IAudioClient3 IAudioClient3; #endif // __IAudioClient3_FWD_DEFINED__ #ifndef __IAudioClient3_INTERFACE_DEFINED__ #define __IAudioClient3_INTERFACE_DEFINED__ MIDL_INTERFACE("7ED4EE07-8E67-4CD4-8C1A-2B7A5987AD42") IAudioClient3 : public IAudioClient2 { public: virtual HRESULT STDMETHODCALLTYPE GetSharedModeEnginePeriod( /* [annotation][in] */ _In_ const WAVEFORMATEX *pFormat, /* [annotation][out] */ _Out_ UINT32 *pDefaultPeriodInFrames, /* [annotation][out] */ _Out_ UINT32 *pFundamentalPeriodInFrames, /* [annotation][out] */ _Out_ UINT32 *pMinPeriodInFrames, /* [annotation][out] */ _Out_ UINT32 *pMaxPeriodInFrames) = 0; virtual HRESULT STDMETHODCALLTYPE GetCurrentSharedModeEnginePeriod( /* [unique][annotation][out] */ _Out_ WAVEFORMATEX * *ppFormat, /* [annotation][out] */ _Out_ UINT32 * pCurrentPeriodInFrames) = 0; virtual HRESULT STDMETHODCALLTYPE InitializeSharedAudioStream( /* [annotation][in] */ _In_ DWORD StreamFlags, /* [annotation][in] */ _In_ UINT32 PeriodInFrames, /* [annotation][in] */ _In_ const WAVEFORMATEX *pFormat, /* [annotation][in] */ _In_opt_ LPCGUID AudioSessionGuid) = 0; }; __CRT_UUID_DECL(IAudioClient3, 0x7ED4EE07, 0x8E67, 0x4CD4, 0x8C, 0x1A, 0x2B, 0x7A, 0x59, 0x87, 0xAD, 0x42) #endif // __IAudioClient3_INTERFACE_DEFINED__ #endif // __MINGW32__ || __MINGW64__ #ifndef PKEY_Device_FriendlyName #undef DEFINE_PROPERTYKEY /* clang-format off */ #define DEFINE_PROPERTYKEY(id, a, b, c, d, e, f, g, h, i, j, k, l) \ const PROPERTYKEY id = { { a, b, c, { d, e, f, g, h, i, j, k, } }, l }; /* clang-format on */ DEFINE_PROPERTYKEY(PKEY_Device_FriendlyName, 0xa45c254e, 0xdf1c, 0x4efd, 0x80, 0x20, 0x67, 0xd1, 0x46, 0xa8, 0x50, 0xe0, 14); #endif const CLSID CLSID_MMDeviceEnumerator = __uuidof(MMDeviceEnumerator); const IID IID_IMMDeviceEnumerator = __uuidof(IMMDeviceEnumerator); const IID IID_IAudioClient = __uuidof(IAudioClient); const IID IID_IAudioClient3 = __uuidof(IAudioClient3); const IID IID_IAudioRenderClient = __uuidof(IAudioRenderClient); const IID IID_IAudioCaptureClient = __uuidof(IAudioCaptureClient); #define SAFE_RELEASE(memory) \ if ((memory) != nullptr) { \ (memory)->Release(); \ (memory) = nullptr; \ } #define REFTIMES_PER_SEC 10000000 #define REFTIMES_PER_MILLISEC 10000 #define CAPTURE_BUFFER_CHANNELS 2 static bool default_render_device_changed = false; static bool default_capture_device_changed = false; // Silence warning due to a COM API weirdness (GH-35194). #if defined(__GNUC__) && !defined(__clang__) #pragma GCC diagnostic push #pragma GCC diagnostic ignored "-Wnon-virtual-dtor" #endif class CMMNotificationClient : public IMMNotificationClient { LONG _cRef = 1; IMMDeviceEnumerator *_pEnumerator = nullptr; public: CMMNotificationClient() {} virtual ~CMMNotificationClient() { if ((_pEnumerator) != nullptr) { (_pEnumerator)->Release(); (_pEnumerator) = nullptr; } } ULONG STDMETHODCALLTYPE AddRef() { return InterlockedIncrement(&_cRef); } ULONG STDMETHODCALLTYPE Release() { ULONG ulRef = InterlockedDecrement(&_cRef); if (0 == ulRef) { delete this; } return ulRef; } HRESULT STDMETHODCALLTYPE QueryInterface(REFIID riid, VOID **ppvInterface) { if (IID_IUnknown == riid) { AddRef(); *ppvInterface = (IUnknown *)this; } else if (__uuidof(IMMNotificationClient) == riid) { AddRef(); *ppvInterface = (IMMNotificationClient *)this; } else { *ppvInterface = nullptr; return E_NOINTERFACE; } return S_OK; } HRESULT STDMETHODCALLTYPE OnDeviceAdded(LPCWSTR pwstrDeviceId) { return S_OK; } HRESULT STDMETHODCALLTYPE OnDeviceRemoved(LPCWSTR pwstrDeviceId) { return S_OK; } HRESULT STDMETHODCALLTYPE OnDeviceStateChanged(LPCWSTR pwstrDeviceId, DWORD dwNewState) { return S_OK; } HRESULT STDMETHODCALLTYPE OnDefaultDeviceChanged(EDataFlow flow, ERole role, LPCWSTR pwstrDeviceId) { if (role == eConsole) { if (flow == eRender) { default_render_device_changed = true; } else if (flow == eCapture) { default_capture_device_changed = true; } } return S_OK; } HRESULT STDMETHODCALLTYPE OnPropertyValueChanged(LPCWSTR pwstrDeviceId, const PROPERTYKEY key) { return S_OK; } }; #if defined(__GNUC__) && !defined(__clang__) #pragma GCC diagnostic pop #endif static CMMNotificationClient notif_client; Error AudioDriverWASAPI::audio_device_init(AudioDeviceWASAPI *p_device, bool p_capture, bool reinit) { WAVEFORMATEX *pwfex; IMMDeviceEnumerator *enumerator = nullptr; IMMDevice *device = nullptr; CoInitialize(nullptr); HRESULT hr = CoCreateInstance(CLSID_MMDeviceEnumerator, nullptr, CLSCTX_ALL, IID_IMMDeviceEnumerator, (void **)&enumerator); ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); if (p_device->device_name == "Default") { hr = enumerator->GetDefaultAudioEndpoint(p_capture ? eCapture : eRender, eConsole, &device); } else { IMMDeviceCollection *devices = nullptr; hr = enumerator->EnumAudioEndpoints(p_capture ? eCapture : eRender, DEVICE_STATE_ACTIVE, &devices); ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); LPWSTR strId = nullptr; bool found = false; UINT count = 0; hr = devices->GetCount(&count); ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); for (ULONG i = 0; i < count && !found; i++) { IMMDevice *tmp_device = nullptr; hr = devices->Item(i, &tmp_device); ERR_BREAK(hr != S_OK); IPropertyStore *props = nullptr; hr = tmp_device->OpenPropertyStore(STGM_READ, &props); ERR_BREAK(hr != S_OK); PROPVARIANT propvar; PropVariantInit(&propvar); hr = props->GetValue(PKEY_Device_FriendlyName, &propvar); ERR_BREAK(hr != S_OK); if (p_device->device_name == String(propvar.pwszVal)) { hr = tmp_device->GetId(&strId); ERR_BREAK(hr != S_OK); found = true; } PropVariantClear(&propvar); props->Release(); tmp_device->Release(); } if (found) { hr = enumerator->GetDevice(strId, &device); } if (strId) { CoTaskMemFree(strId); } if (device == nullptr) { hr = enumerator->GetDefaultAudioEndpoint(p_capture ? eCapture : eRender, eConsole, &device); } } if (reinit) { // In case we're trying to re-initialize the device prevent throwing this error on the console, // otherwise if there is currently no device available this will spam the console. if (hr != S_OK) { return ERR_CANT_OPEN; } } else { ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); } hr = enumerator->RegisterEndpointNotificationCallback(¬if_client); SAFE_RELEASE(enumerator) if (hr != S_OK) { ERR_PRINT("WASAPI: RegisterEndpointNotificationCallback error"); } using_audio_client_3 = !p_capture; // IID_IAudioClient3 is only used for adjustable output latency (not input) if (using_audio_client_3) { hr = device->Activate(IID_IAudioClient3, CLSCTX_ALL, nullptr, (void **)&p_device->audio_client); if (hr != S_OK) { // IID_IAudioClient3 will never activate on OS versions before Windows 10. // Older Windows versions should fall back gracefully. using_audio_client_3 = false; print_verbose("WASAPI: Couldn't activate device with IAudioClient3 interface, falling back to IAudioClient interface"); } else { print_verbose("WASAPI: Activated device using IAudioClient3 interface"); } } if (!using_audio_client_3) { hr = device->Activate(IID_IAudioClient, CLSCTX_ALL, nullptr, (void **)&p_device->audio_client); } SAFE_RELEASE(device) if (reinit) { if (hr != S_OK) { return ERR_CANT_OPEN; } } else { ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); } if (using_audio_client_3) { AudioClientProperties audioProps{}; audioProps.cbSize = sizeof(AudioClientProperties); audioProps.bIsOffload = FALSE; audioProps.eCategory = AudioCategory_GameEffects; hr = ((IAudioClient3 *)p_device->audio_client)->SetClientProperties(&audioProps); ERR_FAIL_COND_V_MSG(hr != S_OK, ERR_CANT_OPEN, "WASAPI: SetClientProperties failed with error 0x" + String::num_uint64(hr, 16) + "."); } hr = p_device->audio_client->GetMixFormat(&pwfex); ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); print_verbose("WASAPI: wFormatTag = " + itos(pwfex->wFormatTag)); print_verbose("WASAPI: nChannels = " + itos(pwfex->nChannels)); print_verbose("WASAPI: nSamplesPerSec = " + itos(pwfex->nSamplesPerSec)); print_verbose("WASAPI: nAvgBytesPerSec = " + itos(pwfex->nAvgBytesPerSec)); print_verbose("WASAPI: nBlockAlign = " + itos(pwfex->nBlockAlign)); print_verbose("WASAPI: wBitsPerSample = " + itos(pwfex->wBitsPerSample)); print_verbose("WASAPI: cbSize = " + itos(pwfex->cbSize)); WAVEFORMATEX *closest = nullptr; hr = p_device->audio_client->IsFormatSupported(AUDCLNT_SHAREMODE_SHARED, pwfex, &closest); if (hr == S_FALSE) { WARN_PRINT("WASAPI: Mix format is not supported by the Device"); if (closest) { print_verbose("WASAPI: closest->wFormatTag = " + itos(closest->wFormatTag)); print_verbose("WASAPI: closest->nChannels = " + itos(closest->nChannels)); print_verbose("WASAPI: closest->nSamplesPerSec = " + itos(closest->nSamplesPerSec)); print_verbose("WASAPI: closest->nAvgBytesPerSec = " + itos(closest->nAvgBytesPerSec)); print_verbose("WASAPI: closest->nBlockAlign = " + itos(closest->nBlockAlign)); print_verbose("WASAPI: closest->wBitsPerSample = " + itos(closest->wBitsPerSample)); print_verbose("WASAPI: closest->cbSize = " + itos(closest->cbSize)); WARN_PRINT("WASAPI: Using closest match instead"); pwfex = closest; } } // Since we're using WASAPI Shared Mode we can't control any of these, we just tag along p_device->channels = pwfex->nChannels; p_device->format_tag = pwfex->wFormatTag; p_device->bits_per_sample = pwfex->wBitsPerSample; p_device->frame_size = (p_device->bits_per_sample / 8) * p_device->channels; if (p_device->format_tag == WAVE_FORMAT_EXTENSIBLE) { WAVEFORMATEXTENSIBLE *wfex = (WAVEFORMATEXTENSIBLE *)pwfex; if (wfex->SubFormat == KSDATAFORMAT_SUBTYPE_PCM) { p_device->format_tag = WAVE_FORMAT_PCM; } else if (wfex->SubFormat == KSDATAFORMAT_SUBTYPE_IEEE_FLOAT) { p_device->format_tag = WAVE_FORMAT_IEEE_FLOAT; } else { ERR_PRINT("WASAPI: Format not supported"); ERR_FAIL_V(ERR_CANT_OPEN); } } else { if (p_device->format_tag != WAVE_FORMAT_PCM && p_device->format_tag != WAVE_FORMAT_IEEE_FLOAT) { ERR_PRINT("WASAPI: Format not supported"); ERR_FAIL_V(ERR_CANT_OPEN); } } if (!using_audio_client_3) { DWORD streamflags = 0; if ((DWORD)mix_rate != pwfex->nSamplesPerSec) { streamflags |= AUDCLNT_STREAMFLAGS_RATEADJUST; pwfex->nSamplesPerSec = mix_rate; pwfex->nAvgBytesPerSec = pwfex->nSamplesPerSec * pwfex->nChannels * (pwfex->wBitsPerSample / 8); } hr = p_device->audio_client->Initialize(AUDCLNT_SHAREMODE_SHARED, streamflags, p_capture ? REFTIMES_PER_SEC : 0, 0, pwfex, nullptr); ERR_FAIL_COND_V_MSG(hr != S_OK, ERR_CANT_OPEN, "WASAPI: Initialize failed with error 0x" + String::num_uint64(hr, 16) + "."); UINT32 max_frames; hr = p_device->audio_client->GetBufferSize(&max_frames); ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); // Due to WASAPI Shared Mode we have no control of the buffer size if (!p_capture) { buffer_frames = max_frames; int64_t latency = 0; audio_output.audio_client->GetStreamLatency(&latency); // WASAPI REFERENCE_TIME units are 100 nanoseconds per unit // https://docs.microsoft.com/en-us/windows/win32/directshow/reference-time // Convert REFTIME to seconds as godot uses for latency real_latency = (float)latency / (float)REFTIMES_PER_SEC; } } else { IAudioClient3 *device_audio_client_3 = (IAudioClient3 *)p_device->audio_client; // AUDCLNT_STREAMFLAGS_RATEADJUST is an invalid flag with IAudioClient3, therefore we have to use // the closest supported mix rate supported by the audio driver. mix_rate = pwfex->nSamplesPerSec; print_verbose("WASAPI: mix_rate = " + itos(mix_rate)); UINT32 default_period_frames, fundamental_period_frames, min_period_frames, max_period_frames; hr = device_audio_client_3->GetSharedModeEnginePeriod( pwfex, &default_period_frames, &fundamental_period_frames, &min_period_frames, &max_period_frames); ERR_FAIL_COND_V_MSG(hr != S_OK, ERR_CANT_OPEN, "WASAPI: GetSharedModeEnginePeriod failed with error 0x" + String::num_uint64(hr, 16) + "."); // Period frames must be an integral multiple of fundamental_period_frames or IAudioClient3 initialization will fail, // so we need to select the closest multiple to the user-specified latency. UINT32 desired_period_frames = target_latency_ms * mix_rate / 1000; UINT32 period_frames = (desired_period_frames / fundamental_period_frames) * fundamental_period_frames; if (ABS((int64_t)period_frames - (int64_t)desired_period_frames) > ABS((int64_t)(period_frames + fundamental_period_frames) - (int64_t)desired_period_frames)) { period_frames = period_frames + fundamental_period_frames; } period_frames = CLAMP(period_frames, min_period_frames, max_period_frames); print_verbose("WASAPI: fundamental_period_frames = " + itos(fundamental_period_frames)); print_verbose("WASAPI: min_period_frames = " + itos(min_period_frames)); print_verbose("WASAPI: max_period_frames = " + itos(max_period_frames)); print_verbose("WASAPI: selected a period frame size of " + itos(period_frames)); buffer_frames = period_frames; hr = device_audio_client_3->InitializeSharedAudioStream(0, period_frames, pwfex, nullptr); ERR_FAIL_COND_V_MSG(hr != S_OK, ERR_CANT_OPEN, "WASAPI: InitializeSharedAudioStream failed with error 0x" + String::num_uint64(hr, 16) + "."); uint32_t output_latency_in_frames; WAVEFORMATEX *current_pwfex; device_audio_client_3->GetCurrentSharedModeEnginePeriod(¤t_pwfex, &output_latency_in_frames); real_latency = (float)output_latency_in_frames / (float)current_pwfex->nSamplesPerSec; CoTaskMemFree(current_pwfex); } if (p_capture) { hr = p_device->audio_client->GetService(IID_IAudioCaptureClient, (void **)&p_device->capture_client); } else { hr = p_device->audio_client->GetService(IID_IAudioRenderClient, (void **)&p_device->render_client); } ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); // Free memory CoTaskMemFree(pwfex); SAFE_RELEASE(device) return OK; } Error AudioDriverWASAPI::init_render_device(bool reinit) { Error err = audio_device_init(&audio_output, false, reinit); if (err != OK) { return err; } switch (audio_output.channels) { case 2: // Stereo case 4: // Surround 3.1 case 6: // Surround 5.1 case 8: // Surround 7.1 channels = audio_output.channels; break; default: WARN_PRINT("WASAPI: Unsupported number of channels: " + itos(audio_output.channels)); channels = 2; break; } // Sample rate is independent of channels (ref: https://stackoverflow.com/questions/11048825/audio-sample-frequency-rely-on-channels) samples_in.resize(buffer_frames * channels); input_position = 0; input_size = 0; print_verbose("WASAPI: detected " + itos(channels) + " channels"); print_verbose("WASAPI: audio buffer frames: " + itos(buffer_frames) + " calculated latency: " + itos(buffer_frames * 1000 / mix_rate) + "ms"); return OK; } Error AudioDriverWASAPI::init_capture_device(bool reinit) { Error err = audio_device_init(&audio_input, true, reinit); if (err != OK) { return err; } // Get the max frames UINT32 max_frames; HRESULT hr = audio_input.audio_client->GetBufferSize(&max_frames); ERR_FAIL_COND_V(hr != S_OK, ERR_CANT_OPEN); input_buffer_init(max_frames); return OK; } Error AudioDriverWASAPI::audio_device_finish(AudioDeviceWASAPI *p_device) { if (p_device->active) { if (p_device->audio_client) { p_device->audio_client->Stop(); } p_device->active = false; } SAFE_RELEASE(p_device->audio_client) SAFE_RELEASE(p_device->render_client) SAFE_RELEASE(p_device->capture_client) return OK; } Error AudioDriverWASAPI::finish_render_device() { return audio_device_finish(&audio_output); } Error AudioDriverWASAPI::finish_capture_device() { return audio_device_finish(&audio_input); } Error AudioDriverWASAPI::init() { mix_rate = GLOBAL_GET("audio/driver/mix_rate"); target_latency_ms = GLOBAL_GET("audio/driver/output_latency"); Error err = init_render_device(); if (err != OK) { ERR_PRINT("WASAPI: init_render_device error"); } exit_thread = false; thread_exited = false; thread.start(thread_func, this); return OK; } int AudioDriverWASAPI::get_mix_rate() const { return mix_rate; } float AudioDriverWASAPI::get_latency() { return real_latency; } AudioDriver::SpeakerMode AudioDriverWASAPI::get_speaker_mode() const { return get_speaker_mode_by_total_channels(channels); } Array AudioDriverWASAPI::audio_device_get_list(bool p_capture) { Array list; IMMDeviceCollection *devices = nullptr; IMMDeviceEnumerator *enumerator = nullptr; list.push_back(String("Default")); CoInitialize(nullptr); HRESULT hr = CoCreateInstance(CLSID_MMDeviceEnumerator, nullptr, CLSCTX_ALL, IID_IMMDeviceEnumerator, (void **)&enumerator); ERR_FAIL_COND_V(hr != S_OK, Array()); hr = enumerator->EnumAudioEndpoints(p_capture ? eCapture : eRender, DEVICE_STATE_ACTIVE, &devices); ERR_FAIL_COND_V(hr != S_OK, Array()); UINT count = 0; hr = devices->GetCount(&count); ERR_FAIL_COND_V(hr != S_OK, Array()); for (ULONG i = 0; i < count; i++) { IMMDevice *device = nullptr; hr = devices->Item(i, &device); ERR_BREAK(hr != S_OK); IPropertyStore *props = nullptr; hr = device->OpenPropertyStore(STGM_READ, &props); ERR_BREAK(hr != S_OK); PROPVARIANT propvar; PropVariantInit(&propvar); hr = props->GetValue(PKEY_Device_FriendlyName, &propvar); ERR_BREAK(hr != S_OK); list.push_back(String(propvar.pwszVal)); PropVariantClear(&propvar); props->Release(); device->Release(); } devices->Release(); enumerator->Release(); return list; } Array AudioDriverWASAPI::get_device_list() { return audio_device_get_list(false); } String AudioDriverWASAPI::get_device() { lock(); String name = audio_output.device_name; unlock(); return name; } void AudioDriverWASAPI::set_device(String device) { lock(); audio_output.new_device = device; unlock(); } int32_t AudioDriverWASAPI::read_sample(WORD format_tag, int bits_per_sample, BYTE *buffer, int i) { if (format_tag == WAVE_FORMAT_PCM) { int32_t sample = 0; switch (bits_per_sample) { case 8: sample = int32_t(((int8_t *)buffer)[i]) << 24; break; case 16: sample = int32_t(((int16_t *)buffer)[i]) << 16; break; case 24: sample |= int32_t(((int8_t *)buffer)[i * 3 + 2]) << 24; sample |= int32_t(((int8_t *)buffer)[i * 3 + 1]) << 16; sample |= int32_t(((int8_t *)buffer)[i * 3 + 0]) << 8; break; case 32: sample = ((int32_t *)buffer)[i]; break; } return sample; } else if (format_tag == WAVE_FORMAT_IEEE_FLOAT) { return int32_t(((float *)buffer)[i] * 32768.0) << 16; } else { ERR_PRINT("WASAPI: Unknown format tag"); } return 0; } void AudioDriverWASAPI::write_sample(WORD format_tag, int bits_per_sample, BYTE *buffer, int i, int32_t sample) { if (format_tag == WAVE_FORMAT_PCM) { switch (bits_per_sample) { case 8: ((int8_t *)buffer)[i] = sample >> 24; break; case 16: ((int16_t *)buffer)[i] = sample >> 16; break; case 24: ((int8_t *)buffer)[i * 3 + 2] = sample >> 24; ((int8_t *)buffer)[i * 3 + 1] = sample >> 16; ((int8_t *)buffer)[i * 3 + 0] = sample >> 8; break; case 32: ((int32_t *)buffer)[i] = sample; break; } } else if (format_tag == WAVE_FORMAT_IEEE_FLOAT) { ((float *)buffer)[i] = (sample >> 16) / 32768.f; } else { ERR_PRINT("WASAPI: Unknown format tag"); } } void AudioDriverWASAPI::thread_func(void *p_udata) { AudioDriverWASAPI *ad = (AudioDriverWASAPI *)p_udata; uint32_t avail_frames = 0; uint32_t write_ofs = 0; while (!ad->exit_thread) { uint32_t read_frames = 0; uint32_t written_frames = 0; if (avail_frames == 0) { ad->lock(); ad->start_counting_ticks(); if (ad->audio_output.active) { ad->audio_server_process(ad->buffer_frames, ad->samples_in.ptrw()); } else { for (int i = 0; i < ad->samples_in.size(); i++) { ad->samples_in.write[i] = 0; } } avail_frames = ad->buffer_frames; write_ofs = 0; ad->stop_counting_ticks(); ad->unlock(); } ad->lock(); ad->start_counting_ticks(); if (avail_frames > 0 && ad->audio_output.audio_client) { UINT32 cur_frames; bool invalidated = false; HRESULT hr = ad->audio_output.audio_client->GetCurrentPadding(&cur_frames); if (hr == S_OK) { // Check how much frames are available on the WASAPI buffer UINT32 write_frames = MIN(ad->buffer_frames - cur_frames, avail_frames); if (write_frames > 0) { BYTE *buffer = nullptr; hr = ad->audio_output.render_client->GetBuffer(write_frames, &buffer); if (hr == S_OK) { // We're using WASAPI Shared Mode so we must convert the buffer if (ad->channels == ad->audio_output.channels) { for (unsigned int i = 0; i < write_frames * ad->channels; i++) { ad->write_sample(ad->audio_output.format_tag, ad->audio_output.bits_per_sample, buffer, i, ad->samples_in.write[write_ofs++]); } } else { for (unsigned int i = 0; i < write_frames; i++) { for (unsigned int j = 0; j < MIN(ad->channels, ad->audio_output.channels); j++) { ad->write_sample(ad->audio_output.format_tag, ad->audio_output.bits_per_sample, buffer, i * ad->audio_output.channels + j, ad->samples_in.write[write_ofs++]); } if (ad->audio_output.channels > ad->channels) { for (unsigned int j = ad->channels; j < ad->audio_output.channels; j++) { ad->write_sample(ad->audio_output.format_tag, ad->audio_output.bits_per_sample, buffer, i * ad->audio_output.channels + j, 0); } } } } hr = ad->audio_output.render_client->ReleaseBuffer(write_frames, 0); if (hr != S_OK) { ERR_PRINT("WASAPI: Release buffer error"); } avail_frames -= write_frames; written_frames += write_frames; } else if (hr == AUDCLNT_E_DEVICE_INVALIDATED) { // Device is not valid anymore, reopen it Error err = ad->finish_render_device(); if (err != OK) { ERR_PRINT("WASAPI: finish_render_device error"); } else { // We reopened the device and samples_in may have resized, so invalidate the current avail_frames avail_frames = 0; } } else { ERR_PRINT("WASAPI: Get buffer error"); ad->exit_thread = true; } } } else if (hr == AUDCLNT_E_DEVICE_INVALIDATED) { invalidated = true; } else { ERR_PRINT("WASAPI: GetCurrentPadding error"); } if (invalidated) { // Device is not valid anymore WARN_PRINT("WASAPI: Current device invalidated, closing device"); Error err = ad->finish_render_device(); if (err != OK) { ERR_PRINT("WASAPI: finish_render_device error"); } } } // If we're using the Default device and it changed finish it so we'll re-init the device if (ad->audio_output.device_name == "Default" && default_render_device_changed) { Error err = ad->finish_render_device(); if (err != OK) { ERR_PRINT("WASAPI: finish_render_device error"); } default_render_device_changed = false; } // User selected a new device, finish the current one so we'll init the new device if (ad->audio_output.device_name != ad->audio_output.new_device) { ad->audio_output.device_name = ad->audio_output.new_device; Error err = ad->finish_render_device(); if (err != OK) { ERR_PRINT("WASAPI: finish_render_device error"); } } if (!ad->audio_output.audio_client) { Error err = ad->init_render_device(true); if (err == OK) { ad->start(); } avail_frames = 0; write_ofs = 0; } if (ad->audio_input.active) { UINT32 packet_length = 0; BYTE *data; UINT32 num_frames_available; DWORD flags; HRESULT hr = ad->audio_input.capture_client->GetNextPacketSize(&packet_length); if (hr == S_OK) { while (packet_length != 0) { hr = ad->audio_input.capture_client->GetBuffer(&data, &num_frames_available, &flags, nullptr, nullptr); ERR_BREAK(hr != S_OK); // fixme: Only works for floating point atm for (UINT32 j = 0; j < num_frames_available; j++) { int32_t l, r; if (flags & AUDCLNT_BUFFERFLAGS_SILENT) { l = r = 0; } else { if (ad->audio_input.channels == 2) { l = read_sample(ad->audio_input.format_tag, ad->audio_input.bits_per_sample, data, j * 2); r = read_sample(ad->audio_input.format_tag, ad->audio_input.bits_per_sample, data, j * 2 + 1); } else if (ad->audio_input.channels == 1) { l = r = read_sample(ad->audio_input.format_tag, ad->audio_input.bits_per_sample, data, j); } else { l = r = 0; ERR_PRINT("WASAPI: unsupported channel count in microphone!"); } } ad->input_buffer_write(l); ad->input_buffer_write(r); } read_frames += num_frames_available; hr = ad->audio_input.capture_client->ReleaseBuffer(num_frames_available); ERR_BREAK(hr != S_OK); hr = ad->audio_input.capture_client->GetNextPacketSize(&packet_length); ERR_BREAK(hr != S_OK); } } // If we're using the Default device and it changed finish it so we'll re-init the device if (ad->audio_input.device_name == "Default" && default_capture_device_changed) { Error err = ad->finish_capture_device(); if (err != OK) { ERR_PRINT("WASAPI: finish_capture_device error"); } default_capture_device_changed = false; } // User selected a new device, finish the current one so we'll init the new device if (ad->audio_input.device_name != ad->audio_input.new_device) { ad->audio_input.device_name = ad->audio_input.new_device; Error err = ad->finish_capture_device(); if (err != OK) { ERR_PRINT("WASAPI: finish_capture_device error"); } } if (!ad->audio_input.audio_client) { Error err = ad->init_capture_device(true); if (err == OK) { ad->capture_start(); } } } ad->stop_counting_ticks(); ad->unlock(); // Let the thread rest a while if we haven't read or write anything if (written_frames == 0 && read_frames == 0) { OS::get_singleton()->delay_usec(1000); } } ad->thread_exited = true; } void AudioDriverWASAPI::start() { if (audio_output.audio_client) { HRESULT hr = audio_output.audio_client->Start(); if (hr != S_OK) { ERR_PRINT("WASAPI: Start failed"); } else { audio_output.active = true; } } } void AudioDriverWASAPI::lock() { mutex.lock(); } void AudioDriverWASAPI::unlock() { mutex.unlock(); } void AudioDriverWASAPI::finish() { exit_thread = true; thread.wait_to_finish(); finish_capture_device(); finish_render_device(); } Error AudioDriverWASAPI::capture_start() { Error err = init_capture_device(); if (err != OK) { ERR_PRINT("WASAPI: init_capture_device error"); return err; } if (audio_input.active) { return FAILED; } audio_input.audio_client->Start(); audio_input.active = true; return OK; } Error AudioDriverWASAPI::capture_stop() { if (audio_input.active) { audio_input.audio_client->Stop(); audio_input.active = false; return OK; } return FAILED; } void AudioDriverWASAPI::capture_set_device(const String &p_name) { lock(); audio_input.new_device = p_name; unlock(); } Array AudioDriverWASAPI::capture_get_device_list() { return audio_device_get_list(true); } String AudioDriverWASAPI::capture_get_device() { lock(); String name = audio_input.device_name; unlock(); return name; } AudioDriverWASAPI::AudioDriverWASAPI() { samples_in.clear(); } #endif // WASAPI_ENABLED