952 lines
29 KiB
C++
952 lines
29 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "modules/audio_device/audio_device_impl.h"
|
|
|
|
#include <stddef.h>
|
|
|
|
#include "api/scoped_refptr.h"
|
|
#include "modules/audio_device/audio_device_config.h" // IWYU pragma: keep
|
|
#include "modules/audio_device/audio_device_generic.h"
|
|
#include "rtc_base/checks.h"
|
|
#include "rtc_base/logging.h"
|
|
#include "rtc_base/ref_counted_object.h"
|
|
#include "system_wrappers/include/metrics.h"
|
|
|
|
#if defined(_WIN32)
|
|
#if defined(WEBRTC_WINDOWS_CORE_AUDIO_BUILD)
|
|
#include "modules/audio_device/win/audio_device_core_win.h"
|
|
#endif
|
|
#elif defined(WEBRTC_ANDROID)
|
|
#include <stdlib.h>
|
|
#if defined(WEBRTC_AUDIO_DEVICE_INCLUDE_ANDROID_AAUDIO)
|
|
#include "modules/audio_device/android/aaudio_player.h"
|
|
#include "modules/audio_device/android/aaudio_recorder.h"
|
|
#endif
|
|
#include "modules/audio_device/android/audio_device_template.h"
|
|
#include "modules/audio_device/android/audio_manager.h"
|
|
#include "modules/audio_device/android/audio_record_jni.h"
|
|
#include "modules/audio_device/android/audio_track_jni.h"
|
|
#include "modules/audio_device/android/opensles_player.h"
|
|
#include "modules/audio_device/android/opensles_recorder.h"
|
|
#elif defined(WEBRTC_LINUX)
|
|
#if defined(WEBRTC_ENABLE_LINUX_ALSA)
|
|
#include "modules/audio_device/linux/audio_device_alsa_linux.h"
|
|
#endif
|
|
#if defined(WEBRTC_ENABLE_LINUX_PULSE)
|
|
#include "modules/audio_device/linux/audio_device_pulse_linux.h"
|
|
#endif
|
|
#elif defined(WEBRTC_IOS)
|
|
#include "sdk/objc/native/src/audio/audio_device_ios.h"
|
|
#elif defined(WEBRTC_MAC)
|
|
#include "modules/audio_device/mac/audio_device_mac.h"
|
|
#endif
|
|
#if defined(WEBRTC_DUMMY_FILE_DEVICES)
|
|
#include "modules/audio_device/dummy/file_audio_device.h"
|
|
#include "modules/audio_device/dummy/file_audio_device_factory.h"
|
|
#endif
|
|
#include "modules/audio_device/dummy/audio_device_dummy.h"
|
|
|
|
#define CHECKinitialized_() \
|
|
{ \
|
|
if (!initialized_) { \
|
|
return -1; \
|
|
} \
|
|
}
|
|
|
|
#define CHECKinitialized__BOOL() \
|
|
{ \
|
|
if (!initialized_) { \
|
|
return false; \
|
|
} \
|
|
}
|
|
|
|
namespace webrtc {
|
|
|
|
rtc::scoped_refptr<AudioDeviceModule> AudioDeviceModule::Create(
|
|
AudioLayer audio_layer,
|
|
TaskQueueFactory* task_queue_factory) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
return AudioDeviceModule::CreateForTest(audio_layer, task_queue_factory);
|
|
}
|
|
|
|
// static
|
|
rtc::scoped_refptr<AudioDeviceModuleForTest> AudioDeviceModule::CreateForTest(
|
|
AudioLayer audio_layer,
|
|
TaskQueueFactory* task_queue_factory) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
|
|
// The "AudioDeviceModule::kWindowsCoreAudio2" audio layer has its own
|
|
// dedicated factory method which should be used instead.
|
|
if (audio_layer == AudioDeviceModule::kWindowsCoreAudio2) {
|
|
RTC_LOG(LS_ERROR) << "Use the CreateWindowsCoreAudioAudioDeviceModule() "
|
|
"factory method instead for this option.";
|
|
return nullptr;
|
|
}
|
|
|
|
// Create the generic reference counted (platform independent) implementation.
|
|
rtc::scoped_refptr<AudioDeviceModuleImpl> audioDevice(
|
|
new rtc::RefCountedObject<AudioDeviceModuleImpl>(audio_layer,
|
|
task_queue_factory));
|
|
|
|
// Ensure that the current platform is supported.
|
|
if (audioDevice->CheckPlatform() == -1) {
|
|
return nullptr;
|
|
}
|
|
|
|
// Create the platform-dependent implementation.
|
|
if (audioDevice->CreatePlatformSpecificObjects() == -1) {
|
|
return nullptr;
|
|
}
|
|
|
|
// Ensure that the generic audio buffer can communicate with the platform
|
|
// specific parts.
|
|
if (audioDevice->AttachAudioBuffer() == -1) {
|
|
return nullptr;
|
|
}
|
|
|
|
return audioDevice;
|
|
}
|
|
|
|
AudioDeviceModuleImpl::AudioDeviceModuleImpl(
|
|
AudioLayer audio_layer,
|
|
TaskQueueFactory* task_queue_factory)
|
|
: audio_layer_(audio_layer), audio_device_buffer_(task_queue_factory) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::CheckPlatform() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
// Ensure that the current platform is supported
|
|
PlatformType platform(kPlatformNotSupported);
|
|
#if defined(_WIN32)
|
|
platform = kPlatformWin32;
|
|
RTC_LOG(INFO) << "current platform is Win32";
|
|
#elif defined(WEBRTC_ANDROID)
|
|
platform = kPlatformAndroid;
|
|
RTC_LOG(INFO) << "current platform is Android";
|
|
#elif defined(WEBRTC_LINUX)
|
|
platform = kPlatformLinux;
|
|
RTC_LOG(INFO) << "current platform is Linux";
|
|
#elif defined(WEBRTC_IOS)
|
|
platform = kPlatformIOS;
|
|
RTC_LOG(INFO) << "current platform is IOS";
|
|
#elif defined(WEBRTC_MAC)
|
|
platform = kPlatformMac;
|
|
RTC_LOG(INFO) << "current platform is Mac";
|
|
#endif
|
|
if (platform == kPlatformNotSupported) {
|
|
RTC_LOG(LERROR)
|
|
<< "current platform is not supported => this module will self "
|
|
"destruct!";
|
|
return -1;
|
|
}
|
|
platform_type_ = platform;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::CreatePlatformSpecificObjects() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
// Dummy ADM implementations if build flags are set.
|
|
#if defined(WEBRTC_DUMMY_AUDIO_BUILD)
|
|
audio_device_.reset(new AudioDeviceDummy());
|
|
RTC_LOG(INFO) << "Dummy Audio APIs will be utilized";
|
|
#elif defined(WEBRTC_DUMMY_FILE_DEVICES)
|
|
audio_device_.reset(FileAudioDeviceFactory::CreateFileAudioDevice());
|
|
if (audio_device_) {
|
|
RTC_LOG(INFO) << "Will use file-playing dummy device.";
|
|
} else {
|
|
// Create a dummy device instead.
|
|
audio_device_.reset(new AudioDeviceDummy());
|
|
RTC_LOG(INFO) << "Dummy Audio APIs will be utilized";
|
|
}
|
|
|
|
// Real (non-dummy) ADM implementations.
|
|
#else
|
|
AudioLayer audio_layer(PlatformAudioLayer());
|
|
// Windows ADM implementation.
|
|
#if defined(WEBRTC_WINDOWS_CORE_AUDIO_BUILD)
|
|
if ((audio_layer == kWindowsCoreAudio) ||
|
|
(audio_layer == kPlatformDefaultAudio)) {
|
|
RTC_LOG(INFO) << "Attempting to use the Windows Core Audio APIs...";
|
|
if (AudioDeviceWindowsCore::CoreAudioIsSupported()) {
|
|
audio_device_.reset(new AudioDeviceWindowsCore());
|
|
RTC_LOG(INFO) << "Windows Core Audio APIs will be utilized";
|
|
}
|
|
}
|
|
#endif // defined(WEBRTC_WINDOWS_CORE_AUDIO_BUILD)
|
|
|
|
#if defined(WEBRTC_ANDROID)
|
|
// Create an Android audio manager.
|
|
audio_manager_android_.reset(new AudioManager());
|
|
// Select best possible combination of audio layers.
|
|
if (audio_layer == kPlatformDefaultAudio) {
|
|
if (audio_manager_android_->IsAAudioSupported()) {
|
|
// Use of AAudio for both playout and recording has highest priority.
|
|
audio_layer = kAndroidAAudioAudio;
|
|
} else if (audio_manager_android_->IsLowLatencyPlayoutSupported() &&
|
|
audio_manager_android_->IsLowLatencyRecordSupported()) {
|
|
// Use OpenSL ES for both playout and recording.
|
|
audio_layer = kAndroidOpenSLESAudio;
|
|
} else if (audio_manager_android_->IsLowLatencyPlayoutSupported() &&
|
|
!audio_manager_android_->IsLowLatencyRecordSupported()) {
|
|
// Use OpenSL ES for output on devices that only supports the
|
|
// low-latency output audio path.
|
|
audio_layer = kAndroidJavaInputAndOpenSLESOutputAudio;
|
|
} else {
|
|
// Use Java-based audio in both directions when low-latency output is
|
|
// not supported.
|
|
audio_layer = kAndroidJavaAudio;
|
|
}
|
|
}
|
|
AudioManager* audio_manager = audio_manager_android_.get();
|
|
if (audio_layer == kAndroidJavaAudio) {
|
|
// Java audio for both input and output audio.
|
|
audio_device_.reset(new AudioDeviceTemplate<AudioRecordJni, AudioTrackJni>(
|
|
audio_layer, audio_manager));
|
|
} else if (audio_layer == kAndroidOpenSLESAudio) {
|
|
// OpenSL ES based audio for both input and output audio.
|
|
audio_device_.reset(
|
|
new AudioDeviceTemplate<OpenSLESRecorder, OpenSLESPlayer>(
|
|
audio_layer, audio_manager));
|
|
} else if (audio_layer == kAndroidJavaInputAndOpenSLESOutputAudio) {
|
|
// Java audio for input and OpenSL ES for output audio (i.e. mixed APIs).
|
|
// This combination provides low-latency output audio and at the same
|
|
// time support for HW AEC using the AudioRecord Java API.
|
|
audio_device_.reset(new AudioDeviceTemplate<AudioRecordJni, OpenSLESPlayer>(
|
|
audio_layer, audio_manager));
|
|
} else if (audio_layer == kAndroidAAudioAudio) {
|
|
#if defined(WEBRTC_AUDIO_DEVICE_INCLUDE_ANDROID_AAUDIO)
|
|
// AAudio based audio for both input and output.
|
|
audio_device_.reset(new AudioDeviceTemplate<AAudioRecorder, AAudioPlayer>(
|
|
audio_layer, audio_manager));
|
|
#endif
|
|
} else if (audio_layer == kAndroidJavaInputAndAAudioOutputAudio) {
|
|
#if defined(WEBRTC_AUDIO_DEVICE_INCLUDE_ANDROID_AAUDIO)
|
|
// Java audio for input and AAudio for output audio (i.e. mixed APIs).
|
|
audio_device_.reset(new AudioDeviceTemplate<AudioRecordJni, AAudioPlayer>(
|
|
audio_layer, audio_manager));
|
|
#endif
|
|
} else {
|
|
RTC_LOG(LS_ERROR) << "The requested audio layer is not supported";
|
|
audio_device_.reset(nullptr);
|
|
}
|
|
// END #if defined(WEBRTC_ANDROID)
|
|
|
|
// Linux ADM implementation.
|
|
// Note that, WEBRTC_ENABLE_LINUX_ALSA is always defined by default when
|
|
// WEBRTC_LINUX is defined. WEBRTC_ENABLE_LINUX_PULSE depends on the
|
|
// 'rtc_include_pulse_audio' build flag.
|
|
// TODO(bugs.webrtc.org/9127): improve support and make it more clear that
|
|
// PulseAudio is the default selection.
|
|
#elif defined(WEBRTC_LINUX)
|
|
#if !defined(WEBRTC_ENABLE_LINUX_PULSE)
|
|
// Build flag 'rtc_include_pulse_audio' is set to false. In this mode:
|
|
// - kPlatformDefaultAudio => ALSA, and
|
|
// - kLinuxAlsaAudio => ALSA, and
|
|
// - kLinuxPulseAudio => Invalid selection.
|
|
RTC_LOG(WARNING) << "PulseAudio is disabled using build flag.";
|
|
if ((audio_layer == kLinuxAlsaAudio) ||
|
|
(audio_layer == kPlatformDefaultAudio)) {
|
|
audio_device_.reset(new AudioDeviceLinuxALSA());
|
|
RTC_LOG(INFO) << "Linux ALSA APIs will be utilized.";
|
|
}
|
|
#else
|
|
// Build flag 'rtc_include_pulse_audio' is set to true (default). In this
|
|
// mode:
|
|
// - kPlatformDefaultAudio => PulseAudio, and
|
|
// - kLinuxPulseAudio => PulseAudio, and
|
|
// - kLinuxAlsaAudio => ALSA (supported but not default).
|
|
RTC_LOG(INFO) << "PulseAudio support is enabled.";
|
|
if ((audio_layer == kLinuxPulseAudio) ||
|
|
(audio_layer == kPlatformDefaultAudio)) {
|
|
// Linux PulseAudio implementation is default.
|
|
audio_device_.reset(new AudioDeviceLinuxPulse());
|
|
RTC_LOG(INFO) << "Linux PulseAudio APIs will be utilized";
|
|
} else if (audio_layer == kLinuxAlsaAudio) {
|
|
audio_device_.reset(new AudioDeviceLinuxALSA());
|
|
RTC_LOG(WARNING) << "Linux ALSA APIs will be utilized.";
|
|
}
|
|
#endif // #if !defined(WEBRTC_ENABLE_LINUX_PULSE)
|
|
#endif // #if defined(WEBRTC_LINUX)
|
|
|
|
// iOS ADM implementation.
|
|
#if defined(WEBRTC_IOS)
|
|
if (audio_layer == kPlatformDefaultAudio) {
|
|
audio_device_.reset(new ios_adm::AudioDeviceIOS());
|
|
RTC_LOG(INFO) << "iPhone Audio APIs will be utilized.";
|
|
}
|
|
// END #if defined(WEBRTC_IOS)
|
|
|
|
// Mac OS X ADM implementation.
|
|
#elif defined(WEBRTC_MAC)
|
|
if (audio_layer == kPlatformDefaultAudio) {
|
|
audio_device_.reset(new AudioDeviceMac());
|
|
RTC_LOG(INFO) << "Mac OS X Audio APIs will be utilized.";
|
|
}
|
|
#endif // WEBRTC_MAC
|
|
|
|
// Dummy ADM implementation.
|
|
if (audio_layer == kDummyAudio) {
|
|
audio_device_.reset(new AudioDeviceDummy());
|
|
RTC_LOG(INFO) << "Dummy Audio APIs will be utilized.";
|
|
}
|
|
#endif // if defined(WEBRTC_DUMMY_AUDIO_BUILD)
|
|
|
|
if (!audio_device_) {
|
|
RTC_LOG(LS_ERROR)
|
|
<< "Failed to create the platform specific ADM implementation.";
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::AttachAudioBuffer() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
audio_device_->AttachAudioBuffer(&audio_device_buffer_);
|
|
return 0;
|
|
}
|
|
|
|
AudioDeviceModuleImpl::~AudioDeviceModuleImpl() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::ActiveAudioLayer(AudioLayer* audioLayer) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
AudioLayer activeAudio;
|
|
if (audio_device_->ActiveAudioLayer(activeAudio) == -1) {
|
|
return -1;
|
|
}
|
|
*audioLayer = activeAudio;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::Init() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
if (initialized_)
|
|
return 0;
|
|
RTC_CHECK(audio_device_);
|
|
AudioDeviceGeneric::InitStatus status = audio_device_->Init();
|
|
RTC_HISTOGRAM_ENUMERATION(
|
|
"WebRTC.Audio.InitializationResult", static_cast<int>(status),
|
|
static_cast<int>(AudioDeviceGeneric::InitStatus::NUM_STATUSES));
|
|
if (status != AudioDeviceGeneric::InitStatus::OK) {
|
|
RTC_LOG(LS_ERROR) << "Audio device initialization failed.";
|
|
return -1;
|
|
}
|
|
initialized_ = true;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::Terminate() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
if (!initialized_)
|
|
return 0;
|
|
if (audio_device_->Terminate() == -1) {
|
|
return -1;
|
|
}
|
|
initialized_ = false;
|
|
return 0;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::Initialized() const {
|
|
RTC_LOG(INFO) << __FUNCTION__ << ": " << initialized_;
|
|
return initialized_;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::InitSpeaker() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
return audio_device_->InitSpeaker();
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::InitMicrophone() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
return audio_device_->InitMicrophone();
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SpeakerVolumeIsAvailable(bool* available) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->SpeakerVolumeIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetSpeakerVolume(uint32_t volume) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << volume << ")";
|
|
CHECKinitialized_();
|
|
return audio_device_->SetSpeakerVolume(volume);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SpeakerVolume(uint32_t* volume) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
uint32_t level = 0;
|
|
if (audio_device_->SpeakerVolume(level) == -1) {
|
|
return -1;
|
|
}
|
|
*volume = level;
|
|
RTC_LOG(INFO) << "output: " << *volume;
|
|
return 0;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::SpeakerIsInitialized() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
bool isInitialized = audio_device_->SpeakerIsInitialized();
|
|
RTC_LOG(INFO) << "output: " << isInitialized;
|
|
return isInitialized;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::MicrophoneIsInitialized() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
bool isInitialized = audio_device_->MicrophoneIsInitialized();
|
|
RTC_LOG(INFO) << "output: " << isInitialized;
|
|
return isInitialized;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MaxSpeakerVolume(uint32_t* maxVolume) const {
|
|
CHECKinitialized_();
|
|
uint32_t maxVol = 0;
|
|
if (audio_device_->MaxSpeakerVolume(maxVol) == -1) {
|
|
return -1;
|
|
}
|
|
*maxVolume = maxVol;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MinSpeakerVolume(uint32_t* minVolume) const {
|
|
CHECKinitialized_();
|
|
uint32_t minVol = 0;
|
|
if (audio_device_->MinSpeakerVolume(minVol) == -1) {
|
|
return -1;
|
|
}
|
|
*minVolume = minVol;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SpeakerMuteIsAvailable(bool* available) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->SpeakerMuteIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetSpeakerMute(bool enable) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << enable << ")";
|
|
CHECKinitialized_();
|
|
return audio_device_->SetSpeakerMute(enable);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SpeakerMute(bool* enabled) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool muted = false;
|
|
if (audio_device_->SpeakerMute(muted) == -1) {
|
|
return -1;
|
|
}
|
|
*enabled = muted;
|
|
RTC_LOG(INFO) << "output: " << muted;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MicrophoneMuteIsAvailable(bool* available) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->MicrophoneMuteIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetMicrophoneMute(bool enable) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << enable << ")";
|
|
CHECKinitialized_();
|
|
return (audio_device_->SetMicrophoneMute(enable));
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MicrophoneMute(bool* enabled) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool muted = false;
|
|
if (audio_device_->MicrophoneMute(muted) == -1) {
|
|
return -1;
|
|
}
|
|
*enabled = muted;
|
|
RTC_LOG(INFO) << "output: " << muted;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MicrophoneVolumeIsAvailable(bool* available) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->MicrophoneVolumeIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetMicrophoneVolume(uint32_t volume) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << volume << ")";
|
|
CHECKinitialized_();
|
|
return (audio_device_->SetMicrophoneVolume(volume));
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MicrophoneVolume(uint32_t* volume) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
uint32_t level = 0;
|
|
if (audio_device_->MicrophoneVolume(level) == -1) {
|
|
return -1;
|
|
}
|
|
*volume = level;
|
|
RTC_LOG(INFO) << "output: " << *volume;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StereoRecordingIsAvailable(
|
|
bool* available) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->StereoRecordingIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetStereoRecording(bool enable) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << enable << ")";
|
|
CHECKinitialized_();
|
|
if (audio_device_->RecordingIsInitialized()) {
|
|
RTC_LOG(LERROR)
|
|
<< "unable to set stereo mode after recording is initialized";
|
|
return -1;
|
|
}
|
|
if (audio_device_->SetStereoRecording(enable) == -1) {
|
|
if (enable) {
|
|
RTC_LOG(WARNING) << "failed to enable stereo recording";
|
|
}
|
|
return -1;
|
|
}
|
|
int8_t nChannels(1);
|
|
if (enable) {
|
|
nChannels = 2;
|
|
}
|
|
audio_device_buffer_.SetRecordingChannels(nChannels);
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StereoRecording(bool* enabled) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool stereo = false;
|
|
if (audio_device_->StereoRecording(stereo) == -1) {
|
|
return -1;
|
|
}
|
|
*enabled = stereo;
|
|
RTC_LOG(INFO) << "output: " << stereo;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StereoPlayoutIsAvailable(bool* available) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->StereoPlayoutIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetStereoPlayout(bool enable) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << enable << ")";
|
|
CHECKinitialized_();
|
|
if (audio_device_->PlayoutIsInitialized()) {
|
|
RTC_LOG(LERROR)
|
|
<< "unable to set stereo mode while playing side is initialized";
|
|
return -1;
|
|
}
|
|
if (audio_device_->SetStereoPlayout(enable)) {
|
|
RTC_LOG(WARNING) << "stereo playout is not supported";
|
|
return -1;
|
|
}
|
|
int8_t nChannels(1);
|
|
if (enable) {
|
|
nChannels = 2;
|
|
}
|
|
audio_device_buffer_.SetPlayoutChannels(nChannels);
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StereoPlayout(bool* enabled) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool stereo = false;
|
|
if (audio_device_->StereoPlayout(stereo) == -1) {
|
|
return -1;
|
|
}
|
|
*enabled = stereo;
|
|
RTC_LOG(INFO) << "output: " << stereo;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::PlayoutIsAvailable(bool* available) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->PlayoutIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::RecordingIsAvailable(bool* available) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
bool isAvailable = false;
|
|
if (audio_device_->RecordingIsAvailable(isAvailable) == -1) {
|
|
return -1;
|
|
}
|
|
*available = isAvailable;
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MaxMicrophoneVolume(uint32_t* maxVolume) const {
|
|
CHECKinitialized_();
|
|
uint32_t maxVol(0);
|
|
if (audio_device_->MaxMicrophoneVolume(maxVol) == -1) {
|
|
return -1;
|
|
}
|
|
*maxVolume = maxVol;
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::MinMicrophoneVolume(uint32_t* minVolume) const {
|
|
CHECKinitialized_();
|
|
uint32_t minVol(0);
|
|
if (audio_device_->MinMicrophoneVolume(minVol) == -1) {
|
|
return -1;
|
|
}
|
|
*minVolume = minVol;
|
|
return 0;
|
|
}
|
|
|
|
int16_t AudioDeviceModuleImpl::PlayoutDevices() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
uint16_t nPlayoutDevices = audio_device_->PlayoutDevices();
|
|
RTC_LOG(INFO) << "output: " << nPlayoutDevices;
|
|
return (int16_t)(nPlayoutDevices);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetPlayoutDevice(uint16_t index) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << index << ")";
|
|
CHECKinitialized_();
|
|
return audio_device_->SetPlayoutDevice(index);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetPlayoutDevice(WindowsDeviceType device) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
return audio_device_->SetPlayoutDevice(device);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::PlayoutDeviceName(
|
|
uint16_t index,
|
|
char name[kAdmMaxDeviceNameSize],
|
|
char guid[kAdmMaxGuidSize]) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << index << ", ...)";
|
|
CHECKinitialized_();
|
|
if (name == NULL) {
|
|
return -1;
|
|
}
|
|
if (audio_device_->PlayoutDeviceName(index, name, guid) == -1) {
|
|
return -1;
|
|
}
|
|
if (name != NULL) {
|
|
RTC_LOG(INFO) << "output: name = " << name;
|
|
}
|
|
if (guid != NULL) {
|
|
RTC_LOG(INFO) << "output: guid = " << guid;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::RecordingDeviceName(
|
|
uint16_t index,
|
|
char name[kAdmMaxDeviceNameSize],
|
|
char guid[kAdmMaxGuidSize]) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << index << ", ...)";
|
|
CHECKinitialized_();
|
|
if (name == NULL) {
|
|
return -1;
|
|
}
|
|
if (audio_device_->RecordingDeviceName(index, name, guid) == -1) {
|
|
return -1;
|
|
}
|
|
if (name != NULL) {
|
|
RTC_LOG(INFO) << "output: name = " << name;
|
|
}
|
|
if (guid != NULL) {
|
|
RTC_LOG(INFO) << "output: guid = " << guid;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int16_t AudioDeviceModuleImpl::RecordingDevices() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
uint16_t nRecordingDevices = audio_device_->RecordingDevices();
|
|
RTC_LOG(INFO) << "output: " << nRecordingDevices;
|
|
return (int16_t)nRecordingDevices;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetRecordingDevice(uint16_t index) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << index << ")";
|
|
CHECKinitialized_();
|
|
return audio_device_->SetRecordingDevice(index);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::SetRecordingDevice(WindowsDeviceType device) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
return audio_device_->SetRecordingDevice(device);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::InitPlayout() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
if (PlayoutIsInitialized()) {
|
|
return 0;
|
|
}
|
|
int32_t result = audio_device_->InitPlayout();
|
|
RTC_LOG(INFO) << "output: " << result;
|
|
RTC_HISTOGRAM_BOOLEAN("WebRTC.Audio.InitPlayoutSuccess",
|
|
static_cast<int>(result == 0));
|
|
return result;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::InitRecording() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
if (RecordingIsInitialized()) {
|
|
return 0;
|
|
}
|
|
int32_t result = audio_device_->InitRecording();
|
|
RTC_LOG(INFO) << "output: " << result;
|
|
RTC_HISTOGRAM_BOOLEAN("WebRTC.Audio.InitRecordingSuccess",
|
|
static_cast<int>(result == 0));
|
|
return result;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::PlayoutIsInitialized() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
return audio_device_->PlayoutIsInitialized();
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::RecordingIsInitialized() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
return audio_device_->RecordingIsInitialized();
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StartPlayout() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
if (Playing()) {
|
|
return 0;
|
|
}
|
|
audio_device_buffer_.StartPlayout();
|
|
int32_t result = audio_device_->StartPlayout();
|
|
RTC_LOG(INFO) << "output: " << result;
|
|
RTC_HISTOGRAM_BOOLEAN("WebRTC.Audio.StartPlayoutSuccess",
|
|
static_cast<int>(result == 0));
|
|
return result;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StopPlayout() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
int32_t result = audio_device_->StopPlayout();
|
|
audio_device_buffer_.StopPlayout();
|
|
RTC_LOG(INFO) << "output: " << result;
|
|
RTC_HISTOGRAM_BOOLEAN("WebRTC.Audio.StopPlayoutSuccess",
|
|
static_cast<int>(result == 0));
|
|
return result;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::Playing() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
return audio_device_->Playing();
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StartRecording() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
if (Recording()) {
|
|
return 0;
|
|
}
|
|
audio_device_buffer_.StartRecording();
|
|
int32_t result = audio_device_->StartRecording();
|
|
RTC_LOG(INFO) << "output: " << result;
|
|
RTC_HISTOGRAM_BOOLEAN("WebRTC.Audio.StartRecordingSuccess",
|
|
static_cast<int>(result == 0));
|
|
return result;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::StopRecording() {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
int32_t result = audio_device_->StopRecording();
|
|
audio_device_buffer_.StopRecording();
|
|
RTC_LOG(INFO) << "output: " << result;
|
|
RTC_HISTOGRAM_BOOLEAN("WebRTC.Audio.StopRecordingSuccess",
|
|
static_cast<int>(result == 0));
|
|
return result;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::Recording() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
return audio_device_->Recording();
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::RegisterAudioCallback(
|
|
AudioTransport* audioCallback) {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
return audio_device_buffer_.RegisterAudioCallback(audioCallback);
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::PlayoutDelay(uint16_t* delayMS) const {
|
|
CHECKinitialized_();
|
|
uint16_t delay = 0;
|
|
if (audio_device_->PlayoutDelay(delay) == -1) {
|
|
RTC_LOG(LERROR) << "failed to retrieve the playout delay";
|
|
return -1;
|
|
}
|
|
*delayMS = delay;
|
|
return 0;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::BuiltInAECIsAvailable() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
bool isAvailable = audio_device_->BuiltInAECIsAvailable();
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return isAvailable;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::EnableBuiltInAEC(bool enable) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << enable << ")";
|
|
CHECKinitialized_();
|
|
int32_t ok = audio_device_->EnableBuiltInAEC(enable);
|
|
RTC_LOG(INFO) << "output: " << ok;
|
|
return ok;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::BuiltInAGCIsAvailable() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
bool isAvailable = audio_device_->BuiltInAGCIsAvailable();
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return isAvailable;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::EnableBuiltInAGC(bool enable) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << enable << ")";
|
|
CHECKinitialized_();
|
|
int32_t ok = audio_device_->EnableBuiltInAGC(enable);
|
|
RTC_LOG(INFO) << "output: " << ok;
|
|
return ok;
|
|
}
|
|
|
|
bool AudioDeviceModuleImpl::BuiltInNSIsAvailable() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized__BOOL();
|
|
bool isAvailable = audio_device_->BuiltInNSIsAvailable();
|
|
RTC_LOG(INFO) << "output: " << isAvailable;
|
|
return isAvailable;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::EnableBuiltInNS(bool enable) {
|
|
RTC_LOG(INFO) << __FUNCTION__ << "(" << enable << ")";
|
|
CHECKinitialized_();
|
|
int32_t ok = audio_device_->EnableBuiltInNS(enable);
|
|
RTC_LOG(INFO) << "output: " << ok;
|
|
return ok;
|
|
}
|
|
|
|
int32_t AudioDeviceModuleImpl::GetPlayoutUnderrunCount() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
CHECKinitialized_();
|
|
int32_t underrunCount = audio_device_->GetPlayoutUnderrunCount();
|
|
RTC_LOG(INFO) << "output: " << underrunCount;
|
|
return underrunCount;
|
|
}
|
|
|
|
#if defined(WEBRTC_IOS)
|
|
int AudioDeviceModuleImpl::GetPlayoutAudioParameters(
|
|
AudioParameters* params) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
int r = audio_device_->GetPlayoutAudioParameters(params);
|
|
RTC_LOG(INFO) << "output: " << r;
|
|
return r;
|
|
}
|
|
|
|
int AudioDeviceModuleImpl::GetRecordAudioParameters(
|
|
AudioParameters* params) const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
int r = audio_device_->GetRecordAudioParameters(params);
|
|
RTC_LOG(INFO) << "output: " << r;
|
|
return r;
|
|
}
|
|
#endif // WEBRTC_IOS
|
|
|
|
AudioDeviceModuleImpl::PlatformType AudioDeviceModuleImpl::Platform() const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
return platform_type_;
|
|
}
|
|
|
|
AudioDeviceModule::AudioLayer AudioDeviceModuleImpl::PlatformAudioLayer()
|
|
const {
|
|
RTC_LOG(INFO) << __FUNCTION__;
|
|
return audio_layer_;
|
|
}
|
|
|
|
} // namespace webrtc
|