gstreamer/subprojects/gst-plugins-good/sys/osxaudio/gstosxcoreaudio.c
Piotr Brzeziński 477beab403 osxaudio: Avoid using private APIs on iOS
Turns out AudioConvertHostTimeToNanos and AudioGetCurrentHostTime are macOS-only APIs, which prevents apps using
GStreamer on iOS from being accepted into App Store.

This commit replaces those functions with a manual version of what they do - mach_absolute_time() for the current time,
and data from mach_timebase_info() at the beginning to convert host timestamps to nanoseconds.

Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/6789>
2024-05-22 08:58:24 +00:00

867 lines
28 KiB
C

/*
* GStreamer
* Copyright (C) 2012-2013 Fluendo S.A. <support@fluendo.com>
* Authors: Josep Torra Vallès <josep@fluendo.com>
* Andoni Morales Alastruey <amorales@fluendo.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
* Boston, MA 02111-1307, USA.
*
*/
#include "gstosxcoreaudio.h"
#include "gstosxcoreaudiocommon.h"
GST_DEBUG_CATEGORY (osx_coreaudio_debug);
#define GST_CAT_DEFAULT osx_coreaudio_debug
G_DEFINE_TYPE (GstCoreAudio, gst_core_audio, G_TYPE_OBJECT);
#ifdef HAVE_IOS
#include "gstosxcoreaudioremoteio.c"
#else
#include "gstosxcoreaudiohal.c"
#endif
static void
gst_core_audio_finalize (GObject * object)
{
GstCoreAudio *core_audio = GST_CORE_AUDIO (object);
g_mutex_clear (&core_audio->timing_lock);
G_OBJECT_CLASS (gst_core_audio_parent_class)->finalize (object);
}
static void
gst_core_audio_class_init (GstCoreAudioClass * klass)
{
GObjectClass *object_klass = G_OBJECT_CLASS (klass);
object_klass->finalize = gst_core_audio_finalize;
}
static void
gst_core_audio_init (GstCoreAudio * core_audio)
{
core_audio->is_passthrough = FALSE;
core_audio->device_id = kAudioDeviceUnknown;
core_audio->is_src = FALSE;
core_audio->audiounit = NULL;
core_audio->cached_caps = NULL;
core_audio->cached_caps_valid = FALSE;
#ifndef HAVE_IOS
core_audio->hog_pid = -1;
core_audio->disabled_mixing = FALSE;
#endif
mach_timebase_info (&core_audio->timebase);
g_mutex_init (&core_audio->timing_lock);
}
static gboolean
_is_outer_scope (AudioUnitScope scope, AudioUnitElement element)
{
return
(scope == kAudioUnitScope_Input && element == 1) ||
(scope == kAudioUnitScope_Output && element == 0);
}
static void
_audio_unit_property_listener (void *inRefCon, AudioUnit inUnit,
AudioUnitPropertyID inID, AudioUnitScope inScope,
AudioUnitElement inElement)
{
GstCoreAudio *core_audio;
core_audio = GST_CORE_AUDIO (inRefCon);
g_assert (inUnit == core_audio->audiounit);
switch (inID) {
case kAudioUnitProperty_AudioChannelLayout:
case kAudioUnitProperty_StreamFormat:
if (_is_outer_scope (inScope, inElement)) {
/* We don't push gst_event_new_caps here (for src),
* nor gst_event_new_reconfigure (for sink), since Core Audio continues
* to happily function with the old format, doing conversion/resampling
* as needed.
* This merely "refreshes" our PREFERRED caps. */
/* This function is called either from a Core Audio thread
* or as a result of a Core Audio API (e.g. AudioUnitInitialize)
* from our own thread. In the latter case, osxbuf can be
* already locked (GStreamer's mutex is not recursive).
* For this reason we use a boolean flag instead of nullifying
* cached_caps. */
core_audio->cached_caps_valid = FALSE;
}
break;
}
}
static GstClockTime
_current_time_ns (GstCoreAudio * core_audio)
{
guint64 mach_t = mach_absolute_time ();
return gst_util_uint64_scale (mach_t, core_audio->timebase.numer,
core_audio->timebase.denom);
}
static GstClockTime
_host_time_to_ns (GstCoreAudio * core_audio, uint64_t host_time)
{
return gst_util_uint64_scale (host_time, core_audio->timebase.numer,
core_audio->timebase.denom);
}
/**************************
* Public API *
*************************/
GstCoreAudio *
gst_core_audio_new (GstObject * osxbuf)
{
GstCoreAudio *core_audio;
core_audio = g_object_new (GST_TYPE_CORE_AUDIO, NULL);
core_audio->osxbuf = osxbuf;
core_audio->cached_caps = NULL;
return core_audio;
}
gboolean
gst_core_audio_close (GstCoreAudio * core_audio)
{
OSStatus status;
/* Uninitialize the AudioUnit */
status = AudioUnitUninitialize (core_audio->audiounit);
if (status) {
GST_ERROR_OBJECT (core_audio, "Failed to uninitialize AudioUnit: %d",
(int) status);
return FALSE;
}
AudioUnitRemovePropertyListenerWithUserData (core_audio->audiounit,
kAudioUnitProperty_AudioChannelLayout, _audio_unit_property_listener,
core_audio);
AudioUnitRemovePropertyListenerWithUserData (core_audio->audiounit,
kAudioUnitProperty_StreamFormat, _audio_unit_property_listener,
core_audio);
/* core_audio->osxbuf is already locked at this point */
core_audio->cached_caps_valid = FALSE;
gst_caps_replace (&core_audio->cached_caps, NULL);
AudioComponentInstanceDispose (core_audio->audiounit);
core_audio->audiounit = NULL;
return TRUE;
}
gboolean
gst_core_audio_open (GstCoreAudio * core_audio)
{
OSStatus status;
/* core_audio->osxbuf is already locked at this point */
core_audio->cached_caps_valid = FALSE;
gst_caps_replace (&core_audio->cached_caps, NULL);
if (!gst_core_audio_open_impl (core_audio))
return FALSE;
/* Add property listener */
status = AudioUnitAddPropertyListener (core_audio->audiounit,
kAudioUnitProperty_AudioChannelLayout, _audio_unit_property_listener,
core_audio);
if (status != noErr) {
GST_ERROR_OBJECT (core_audio, "Failed to add audio channel layout property "
"listener for AudioUnit: %d", (int) status);
}
status = AudioUnitAddPropertyListener (core_audio->audiounit,
kAudioUnitProperty_StreamFormat, _audio_unit_property_listener,
core_audio);
if (status != noErr) {
GST_ERROR_OBJECT (core_audio, "Failed to add stream format property "
"listener for AudioUnit: %d", (int) status);
}
/* Initialize the AudioUnit. We keep the audio unit initialized early so that
* we can probe the underlying device. */
status = AudioUnitInitialize (core_audio->audiounit);
if (status) {
GST_ERROR_OBJECT (core_audio, "Failed to initialize AudioUnit: %d",
(int) status);
return FALSE;
}
return TRUE;
}
gboolean
gst_core_audio_start_processing (GstCoreAudio * core_audio)
{
return gst_core_audio_start_processing_impl (core_audio);
}
gboolean
gst_core_audio_pause_processing (GstCoreAudio * core_audio)
{
return gst_core_audio_pause_processing_impl (core_audio);
}
gboolean
gst_core_audio_stop_processing (GstCoreAudio * core_audio)
{
return gst_core_audio_stop_processing_impl (core_audio);
}
gboolean
gst_core_audio_get_samples_and_latency (GstCoreAudio * core_audio,
gdouble rate, guint * samples, gdouble * latency)
{
uint64_t now_ns = _current_time_ns (core_audio);
gboolean ret = gst_core_audio_get_samples_and_latency_impl (core_audio, rate,
samples, latency);
if (!ret)
return FALSE;
CORE_AUDIO_TIMING_LOCK (core_audio);
uint32_t samples_remain = 0;
uint64_t anchor_ns = core_audio->anchor_hosttime_ns;
if (core_audio->is_src) {
int64_t captured_ns =
core_audio->rate_scalar * (int64_t) (now_ns - anchor_ns);
/* src, the anchor time is the timestamp of the first sample in the last
* packet received, and we increment up from there, unless the device gets stopped. */
if (captured_ns > 0) {
if (core_audio->io_proc_active) {
samples_remain = (uint32_t) (captured_ns * rate / GST_SECOND);
} else {
samples_remain = core_audio->anchor_pend_samples;
}
} else {
/* Time went backward. This shouldn't happen for sources, but report something anyway */
samples_remain =
(uint32_t) (-captured_ns * rate / GST_SECOND) +
core_audio->anchor_pend_samples;
}
GST_DEBUG_OBJECT (core_audio,
"now_ns %" G_GUINT64_FORMAT " anchor %" G_GUINT64_FORMAT " elapsed ns %"
G_GINT64_FORMAT " rate %f captured_ns %" G_GINT64_FORMAT
" anchor_pend_samples %u samples_remain %u", now_ns, anchor_ns,
now_ns - anchor_ns, rate, captured_ns, core_audio->anchor_pend_samples,
samples_remain);
} else {
/* Sink, the anchor time is the time the most recent buffer will commence play out,
* and we count down to 0 for unplayed samples beyond that */
int64_t unplayed_ns =
core_audio->rate_scalar * (int64_t) (anchor_ns - now_ns);
if (unplayed_ns > 0) {
samples_remain =
(uint32_t) (unplayed_ns * rate / GST_SECOND) +
core_audio->anchor_pend_samples;
} else {
uint32_t samples_played = (uint32_t) (-unplayed_ns * rate / GST_SECOND);
if (samples_played < core_audio->anchor_pend_samples) {
samples_remain = core_audio->anchor_pend_samples - samples_played;
}
}
GST_DEBUG_OBJECT (core_audio,
"now_ns %" G_GUINT64_FORMAT " anchor %" G_GUINT64_FORMAT " elapsed ns %"
G_GINT64_FORMAT " rate %f unplayed_ns %" G_GINT64_FORMAT
" anchor_pend_samples %u", now_ns, anchor_ns, now_ns - anchor_ns, rate,
unplayed_ns, core_audio->anchor_pend_samples);
}
CORE_AUDIO_TIMING_UNLOCK (core_audio);
GST_DEBUG_OBJECT (core_audio, "samples = %u latency %f", samples_remain,
*latency);
*samples = samples_remain;
return TRUE;
}
void
gst_core_audio_update_timing (GstCoreAudio * core_audio,
const AudioTimeStamp * inTimeStamp, unsigned int inNumberFrames)
{
AudioTimeStampFlags target_flags =
kAudioTimeStampSampleHostTimeValid | kAudioTimeStampRateScalarValid;
if ((inTimeStamp->mFlags & target_flags) == target_flags) {
core_audio->anchor_hosttime_ns =
_host_time_to_ns (core_audio, inTimeStamp->mHostTime);
core_audio->anchor_pend_samples = inNumberFrames;
core_audio->rate_scalar = inTimeStamp->mRateScalar;
GST_DEBUG_OBJECT (core_audio,
"anchor hosttime_ns %" G_GUINT64_FORMAT
" scalar_rate %f anchor_pend_samples %u",
core_audio->anchor_hosttime_ns,
core_audio->rate_scalar, core_audio->anchor_pend_samples);
}
}
gboolean
gst_core_audio_initialize (GstCoreAudio * core_audio,
AudioStreamBasicDescription format, GstCaps * caps,
guint32 frames_per_packet, gboolean is_passthrough)
{
GST_DEBUG_OBJECT (core_audio,
"Initializing: passthrough:%d caps:%" GST_PTR_FORMAT, is_passthrough,
caps);
if (!gst_core_audio_initialize_impl (core_audio, format, caps,
is_passthrough, &frames_per_packet)) {
return FALSE;
}
if (core_audio->is_src) {
/* create AudioBufferList needed for recording */
core_audio->recBufferSize = frames_per_packet * format.mBytesPerFrame;
GST_DEBUG_OBJECT (core_audio,
"Allocating record buffers %u bytes %u frames",
core_audio->recBufferSize, frames_per_packet);
core_audio->recBufferList =
buffer_list_alloc (format.mChannelsPerFrame, core_audio->recBufferSize,
/* Currently always TRUE (i.e. interleaved) */
!(format.mFormatFlags & kAudioFormatFlagIsNonInterleaved));
}
return TRUE;
}
void
gst_core_audio_uninitialize (GstCoreAudio * core_audio)
{
buffer_list_free (core_audio->recBufferList);
core_audio->recBufferList = NULL;
}
void
gst_core_audio_set_volume (GstCoreAudio * core_audio, gfloat volume)
{
AudioUnitSetParameter (core_audio->audiounit, kHALOutputParam_Volume,
kAudioUnitScope_Global, 0, (float) volume, 0);
}
gboolean
gst_core_audio_select_device (GstCoreAudio * core_audio)
{
return gst_core_audio_select_device_impl (core_audio);
}
void
gst_core_audio_init_debug (void)
{
GST_DEBUG_CATEGORY_INIT (osx_coreaudio_debug, "osxaudio", 0,
"OSX Audio Elements");
}
gboolean
gst_core_audio_audio_device_is_spdif_avail (AudioDeviceID device_id)
{
return gst_core_audio_audio_device_is_spdif_avail_impl (device_id);
}
/* Does the channel have at least one positioned channel?
* (GStreamer is more strict than Core Audio, in that it requires either
* all channels to be positioned, or all unpositioned.) */
static gboolean
_is_core_audio_layout_positioned (AudioChannelLayout * layout)
{
guint i;
g_assert (layout->mChannelLayoutTag ==
kAudioChannelLayoutTag_UseChannelDescriptions);
for (i = 0; i < layout->mNumberChannelDescriptions; ++i) {
GstAudioChannelPosition p =
gst_core_audio_channel_label_to_gst
(layout->mChannelDescriptions[i].mChannelLabel, i, FALSE);
if (p >= 0) /* not special positition */
return TRUE;
}
return FALSE;
}
static void
_core_audio_parse_channel_descriptions (AudioChannelLayout * layout,
guint * channels, guint64 * channel_mask, GstAudioChannelPosition * pos)
{
gboolean positioned;
guint i;
g_assert (layout->mChannelLayoutTag ==
kAudioChannelLayoutTag_UseChannelDescriptions);
positioned = _is_core_audio_layout_positioned (layout);
*channel_mask = 0;
/* Go over all labels, either taking only positioned or only
* unpositioned channels, up to GST_OSX_AUDIO_MAX_CHANNEL channels.
*
* The resulting 'pos' array will contain either:
* - only regular (>= 0) positions
* - only GST_AUDIO_CHANNEL_POSITION_NONE positions
* in a compact form, skipping over all unsupported positions.
*/
*channels = 0;
for (i = 0; i < layout->mNumberChannelDescriptions; ++i) {
GstAudioChannelPosition p =
gst_core_audio_channel_label_to_gst
(layout->mChannelDescriptions[i].mChannelLabel, i, TRUE);
/* In positioned layouts, skip all unpositioned channels.
* In unpositioned layouts, skip all invalid channels. */
if ((positioned && p >= 0) ||
(!positioned && p == GST_AUDIO_CHANNEL_POSITION_NONE)) {
if (pos)
pos[*channels] = p;
*channel_mask |= G_GUINT64_CONSTANT (1) << p;
++(*channels);
if (*channels == GST_OSX_AUDIO_MAX_CHANNEL)
break; /* not to overflow */
}
}
}
gboolean
gst_core_audio_parse_channel_layout (AudioChannelLayout * layout,
guint * channels, guint64 * channel_mask, GstAudioChannelPosition * pos)
{
g_assert (channels != NULL);
g_assert (channel_mask != NULL);
g_assert (layout != NULL);
if (layout->mChannelLayoutTag ==
kAudioChannelLayoutTag_UseChannelDescriptions) {
switch (layout->mNumberChannelDescriptions) {
case 0:
if (pos)
pos[0] = GST_AUDIO_CHANNEL_POSITION_NONE;
*channels = 0;
*channel_mask = 0;
return TRUE;
case 1:
if (pos)
pos[0] = GST_AUDIO_CHANNEL_POSITION_MONO;
*channels = 1;
*channel_mask = 0;
return TRUE;
case 2:
if (pos) {
pos[0] = GST_AUDIO_CHANNEL_POSITION_FRONT_LEFT;
pos[1] = GST_AUDIO_CHANNEL_POSITION_FRONT_RIGHT;
}
*channels = 2;
*channel_mask =
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_RIGHT);
return TRUE;
default:
_core_audio_parse_channel_descriptions (layout, channels, channel_mask,
pos);
return TRUE;
}
} else if (layout->mChannelLayoutTag == kAudioChannelLayoutTag_Mono) {
if (pos)
pos[0] = GST_AUDIO_CHANNEL_POSITION_MONO;
*channels = 1;
*channel_mask = 0;
return TRUE;
} else if (layout->mChannelLayoutTag == kAudioChannelLayoutTag_Stereo ||
layout->mChannelLayoutTag == kAudioChannelLayoutTag_StereoHeadphones ||
layout->mChannelLayoutTag == kAudioChannelLayoutTag_Binaural) {
if (pos) {
pos[0] = GST_AUDIO_CHANNEL_POSITION_FRONT_LEFT;
pos[1] = GST_AUDIO_CHANNEL_POSITION_FRONT_RIGHT;
}
*channels = 2;
*channel_mask =
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_RIGHT);
return TRUE;
} else if (layout->mChannelLayoutTag == kAudioChannelLayoutTag_Quadraphonic) {
if (pos) {
pos[0] = GST_AUDIO_CHANNEL_POSITION_FRONT_LEFT;
pos[1] = GST_AUDIO_CHANNEL_POSITION_FRONT_RIGHT;
pos[2] = GST_AUDIO_CHANNEL_POSITION_SURROUND_LEFT;
pos[3] = GST_AUDIO_CHANNEL_POSITION_SURROUND_RIGHT;
}
*channels = 4;
*channel_mask =
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_RIGHT) |
GST_AUDIO_CHANNEL_POSITION_MASK (SURROUND_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (SURROUND_RIGHT);
return TRUE;
} else if (layout->mChannelLayoutTag == kAudioChannelLayoutTag_Pentagonal) {
if (pos) {
pos[0] = GST_AUDIO_CHANNEL_POSITION_FRONT_LEFT;
pos[1] = GST_AUDIO_CHANNEL_POSITION_FRONT_RIGHT;
pos[2] = GST_AUDIO_CHANNEL_POSITION_SURROUND_LEFT;
pos[3] = GST_AUDIO_CHANNEL_POSITION_SURROUND_RIGHT;
pos[4] = GST_AUDIO_CHANNEL_POSITION_FRONT_CENTER;
}
*channels = 5;
*channel_mask =
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_RIGHT) |
GST_AUDIO_CHANNEL_POSITION_MASK (SURROUND_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (SURROUND_RIGHT) |
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_CENTER);
return TRUE;
} else if (layout->mChannelLayoutTag == kAudioChannelLayoutTag_Cube) {
if (pos) {
pos[0] = GST_AUDIO_CHANNEL_POSITION_FRONT_LEFT;
pos[1] = GST_AUDIO_CHANNEL_POSITION_FRONT_RIGHT;
pos[2] = GST_AUDIO_CHANNEL_POSITION_REAR_LEFT;
pos[3] = GST_AUDIO_CHANNEL_POSITION_REAR_RIGHT;
pos[4] = GST_AUDIO_CHANNEL_POSITION_TOP_FRONT_LEFT;
pos[5] = GST_AUDIO_CHANNEL_POSITION_TOP_FRONT_RIGHT;
pos[6] = GST_AUDIO_CHANNEL_POSITION_TOP_REAR_LEFT;
pos[7] = GST_AUDIO_CHANNEL_POSITION_TOP_REAR_RIGHT;
}
*channels = 8;
*channel_mask =
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_RIGHT) |
GST_AUDIO_CHANNEL_POSITION_MASK (REAR_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (REAR_RIGHT) |
GST_AUDIO_CHANNEL_POSITION_MASK (TOP_FRONT_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (TOP_FRONT_RIGHT) |
GST_AUDIO_CHANNEL_POSITION_MASK (TOP_REAR_LEFT) |
GST_AUDIO_CHANNEL_POSITION_MASK (TOP_REAR_RIGHT);
return TRUE;
} else {
GST_WARNING
("AudioChannelLayoutTag: %u not yet supported",
layout->mChannelLayoutTag);
*channels = 0;
*channel_mask = 0;
return FALSE;
}
}
/* Converts an AudioStreamBasicDescription to preferred caps.
*
* These caps will indicate the AU element's canonical format, which won't
* make Core Audio resample nor convert.
*
* NOTE ON MULTI-CHANNEL AUDIO:
*
* If layout is not NULL, resulting caps will only include the subset
* of channels supported by GStreamer. If the Core Audio layout contained
* ANY positioned channels, then ONLY positioned channels will be included
* in the resulting caps. Otherwise, resulting caps will be unpositioned,
* and include only unpositioned channels.
* (Channels with unsupported AudioChannelLabel will be skipped either way.)
*
* Naturally, the number of channels indicated by 'channels' can be lower
* than the AU element's total number of channels.
*/
GstCaps *
gst_core_audio_asbd_to_caps (AudioStreamBasicDescription * asbd,
AudioChannelLayout * layout)
{
GstAudioInfo info;
GstAudioFormat format = GST_AUDIO_FORMAT_UNKNOWN;
guint rate, channels, bps, endianness;
guint64 channel_mask;
gboolean sign;
GstAudioChannelPosition pos[GST_OSX_AUDIO_MAX_CHANNEL];
if (asbd->mFormatID != kAudioFormatLinearPCM) {
GST_WARNING ("Only linear PCM is supported");
goto error;
}
if (!(asbd->mFormatFlags & kAudioFormatFlagIsPacked)) {
GST_WARNING ("Only packed formats supported");
goto error;
}
if (asbd->mFormatFlags & kLinearPCMFormatFlagsSampleFractionMask) {
GST_WARNING ("Fixed point audio is unsupported");
goto error;
}
rate = asbd->mSampleRate;
if (rate == kAudioStreamAnyRate) {
GST_WARNING ("No sample rate");
goto error;
}
bps = asbd->mBitsPerChannel;
endianness = asbd->mFormatFlags & kAudioFormatFlagIsBigEndian ?
G_BIG_ENDIAN : G_LITTLE_ENDIAN;
sign = asbd->mFormatFlags & kAudioFormatFlagIsSignedInteger ? TRUE : FALSE;
if (asbd->mFormatFlags & kAudioFormatFlagIsFloat) {
if (bps == 32) {
if (endianness == G_LITTLE_ENDIAN)
format = GST_AUDIO_FORMAT_F32LE;
else
format = GST_AUDIO_FORMAT_F32BE;
} else if (bps == 64) {
if (endianness == G_LITTLE_ENDIAN)
format = GST_AUDIO_FORMAT_F64LE;
else
format = GST_AUDIO_FORMAT_F64BE;
}
} else {
format = gst_audio_format_build_integer (sign, endianness, bps, bps);
}
if (format == GST_AUDIO_FORMAT_UNKNOWN) {
GST_WARNING ("Unsupported sample format");
goto error;
}
if (layout) {
if (!gst_core_audio_parse_channel_layout (layout, &channels, &channel_mask,
pos)) {
GST_WARNING
("Failed to parse channel layout, best effort channels layout mapping will be used");
layout = NULL;
}
}
if (layout) {
/* The AU can have arbitrary channel order, but we're using GstAudioInfo
* which supports only the GStreamer channel order.
* Also, we're eventually producing caps, which only have channel-mask
* (whose implied order is the GStreamer channel order). */
gst_audio_channel_positions_to_valid_order (pos, channels);
gst_audio_info_set_format (&info, format, rate, channels, pos);
} else {
channels = MIN (asbd->mChannelsPerFrame, GST_OSX_AUDIO_MAX_CHANNEL);
gst_audio_info_set_format (&info, format, rate, channels, NULL);
}
return gst_audio_info_to_caps (&info);
error:
return NULL;
}
static gboolean
_core_audio_get_property (GstCoreAudio * core_audio, gboolean outer,
AudioUnitPropertyID inID, void *inData, UInt32 * inDataSize)
{
OSStatus status;
AudioUnitScope scope;
AudioUnitElement element;
scope = outer ?
CORE_AUDIO_OUTER_SCOPE (core_audio) : CORE_AUDIO_INNER_SCOPE (core_audio);
element = CORE_AUDIO_ELEMENT (core_audio);
status =
AudioUnitGetProperty (core_audio->audiounit, inID, scope, element, inData,
inDataSize);
return status == noErr;
}
static gboolean
_core_audio_get_stream_format (GstCoreAudio * core_audio,
AudioStreamBasicDescription * asbd, gboolean outer)
{
UInt32 size;
size = sizeof (AudioStreamBasicDescription);
return _core_audio_get_property (core_audio, outer,
kAudioUnitProperty_StreamFormat, asbd, &size);
}
AudioChannelLayout *
gst_core_audio_get_channel_layout (GstCoreAudio * core_audio, gboolean outer)
{
UInt32 size;
AudioChannelLayout *layout;
if (core_audio->is_src) {
GST_WARNING_OBJECT (core_audio,
"gst_core_audio_get_channel_layout not supported on source.");
return NULL;
}
if (!_core_audio_get_property (core_audio, outer,
kAudioUnitProperty_AudioChannelLayout, NULL, &size)) {
GST_WARNING_OBJECT (core_audio, "unable to get channel layout");
return NULL;
}
layout = g_malloc (size);
if (!_core_audio_get_property (core_audio, outer,
kAudioUnitProperty_AudioChannelLayout, layout, &size)) {
GST_WARNING_OBJECT (core_audio, "unable to get channel layout");
g_free (layout);
return NULL;
}
return layout;
}
#define STEREO_CHANNEL_MASK \
(GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_LEFT) | \
GST_AUDIO_CHANNEL_POSITION_MASK (FRONT_RIGHT))
GstCaps *
gst_core_audio_probe_caps (GstCoreAudio * core_audio, GstCaps * in_caps)
{
guint i, channels;
gboolean spdif_allowed;
AudioChannelLayout *layout;
AudioStreamBasicDescription outer_asbd;
gboolean got_outer_asbd;
GstCaps *caps = NULL;
guint64 channel_mask;
/* Get the ASBD of the outer scope (i.e. input scope of Input,
* output scope of Output).
* This ASBD indicates the hardware format. */
got_outer_asbd =
_core_audio_get_stream_format (core_audio, &outer_asbd, TRUE);
/* Collect info about the HW capabilities and preferences */
spdif_allowed =
gst_core_audio_audio_device_is_spdif_avail (core_audio->device_id);
if (!core_audio->is_src)
layout = gst_core_audio_get_channel_layout (core_audio, TRUE);
else
layout = NULL; /* no supported for sources */
GST_DEBUG_OBJECT (core_audio, "Selected device ID: %u SPDIF allowed: %d",
(unsigned) core_audio->device_id, spdif_allowed);
if (layout) {
if (!gst_core_audio_parse_channel_layout (layout, &channels, &channel_mask,
NULL)) {
GST_WARNING_OBJECT (core_audio, "Failed to parse channel layout");
channel_mask = 0;
}
/* If available, start with the preferred caps. */
if (got_outer_asbd)
caps = gst_core_audio_asbd_to_caps (&outer_asbd, layout);
g_free (layout);
} else if (got_outer_asbd) {
channels = outer_asbd.mChannelsPerFrame;
channel_mask = 0;
/* If available, start with the preferred caps */
caps = gst_core_audio_asbd_to_caps (&outer_asbd, NULL);
} else {
GST_ERROR_OBJECT (core_audio,
"Unable to get any information about hardware");
return NULL;
}
/* Append the allowed subset based on the template caps */
if (!caps)
caps = gst_caps_new_empty ();
for (i = 0; i < gst_caps_get_size (in_caps); i++) {
GstStructure *in_s;
in_s = gst_caps_get_structure (in_caps, i);
if (gst_structure_has_name (in_s, "audio/x-ac3") ||
gst_structure_has_name (in_s, "audio/x-dts")) {
if (spdif_allowed) {
gst_caps_append_structure (caps, gst_structure_copy (in_s));
}
} else {
GstStructure *out_s;
out_s = gst_structure_copy (in_s);
gst_structure_set (out_s, "channels", G_TYPE_INT, channels, NULL);
if (channel_mask != 0) {
/* positioned layout */
gst_structure_set (out_s,
"channel-mask", GST_TYPE_BITMASK, channel_mask, NULL);
} else {
/* unpositioned layout */
gst_structure_remove_field (out_s, "channel-mask");
}
#ifndef HAVE_IOS
if (core_audio->is_src && got_outer_asbd
&& outer_asbd.mSampleRate != kAudioStreamAnyRate) {
/* According to Core Audio engineer, AUHAL does not support sample rate conversion.
* on sources. Therefore, we fixate the sample rate.
*
* "You definitely cannot do rate conversion as part of getting input from AUHAL.
* That's the most common cause of those "cannot do in current context" errors."
* http://lists.apple.com/archives/coreaudio-api/2006/Sep/msg00088.html
*/
gst_structure_set (out_s, "rate", G_TYPE_INT,
(gint) outer_asbd.mSampleRate, NULL);
}
#endif
/* Special cases for upmixing and downmixing.
* Other than that, the AUs don't upmix or downmix multi-channel audio,
* e.g. if you push 5.1-surround audio to a stereo configuration,
* the left and right channels will be played accordingly,
* and the rest will be dropped. */
if (channels == 1) {
/* If have mono, then also offer stereo since CoreAudio downmixes to it */
GstStructure *stereo = gst_structure_copy (out_s);
gst_structure_remove_field (out_s, "channel-mask");
gst_structure_set (stereo, "channels", G_TYPE_INT, 2,
"channel-mask", GST_TYPE_BITMASK, STEREO_CHANNEL_MASK, NULL);
gst_caps_append_structure (caps, stereo);
gst_caps_append_structure (caps, out_s);
} else if (channels == 2 && (channel_mask == 0
|| channel_mask == STEREO_CHANNEL_MASK)) {
/* If have stereo channels, then also offer mono since CoreAudio
* upmixes it. */
GstStructure *mono = gst_structure_copy (out_s);
gst_structure_set (mono, "channels", G_TYPE_INT, 1, NULL);
gst_structure_remove_field (mono, "channel-mask");
gst_structure_set (out_s, "channel-mask", GST_TYPE_BITMASK,
STEREO_CHANNEL_MASK, NULL);
gst_caps_append_structure (caps, out_s);
gst_caps_append_structure (caps, mono);
} else {
/* Otherwise just add the caps */
gst_caps_append_structure (caps, out_s);
}
}
}
GST_DEBUG_OBJECT (core_audio, "Probed caps:%" GST_PTR_FORMAT, caps);
return caps;
}