// Copyright (c) 2012 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifndef CONTENT_RENDERER_MEDIA_WEBRTC_AUDIO_RENDERER_H_
#define CONTENT_RENDERER_MEDIA_WEBRTC_AUDIO_RENDERER_H_
#include "base/memory/ref_counted.h"
#include "base/synchronization/lock.h"
#include "base/threading/non_thread_safe.h"
#include "base/threading/thread_checker.h"
#include "content/renderer/media/media_stream_audio_renderer.h"
#include "content/renderer/media/webrtc_audio_device_impl.h"
#include "media/base/audio_decoder.h"
#include "media/base/audio_pull_fifo.h"
#include "media/base/audio_renderer_sink.h"
#include "media/base/channel_layout.h"
namespace media {
class AudioOutputDevice;
} // namespace media
namespace webrtc {
class AudioSourceInterface;
class MediaStreamInterface;
} // namespace webrtc
namespace content {
class WebRtcAudioRendererSource;
// This renderer handles calls from the pipeline and WebRtc ADM. It is used
// for connecting WebRtc MediaStream with the audio pipeline.
class CONTENT_EXPORT WebRtcAudioRenderer
: NON_EXPORTED_BASE(public media::AudioRendererSink::RenderCallback),
NON_EXPORTED_BASE(public MediaStreamAudioRenderer) {
public:
// This is a little utility class that holds the configured state of an audio
// stream.
// It is used by both WebRtcAudioRenderer and SharedAudioRenderer (see cc
// file) so a part of why it exists is to avoid code duplication and track
// the state in the same way in WebRtcAudioRenderer and SharedAudioRenderer.
class PlayingState : public base::NonThreadSafe {
public:
PlayingState() : playing_(false), volume_(1.0f) {}
bool playing() const {
DCHECK(CalledOnValidThread());
return playing_;
}
void set_playing(bool playing) {
DCHECK(CalledOnValidThread());
playing_ = playing;
}
float volume() const {
DCHECK(CalledOnValidThread());
return volume_;
}
void set_volume(float volume) {
DCHECK(CalledOnValidThread());
volume_ = volume;
}
private:
bool playing_;
float volume_;
};
WebRtcAudioRenderer(
const scoped_refptr<webrtc::MediaStreamInterface>& media_stream,
int source_render_view_id,
int source_render_frame_id,
int session_id,
int sample_rate,
int frames_per_buffer);
// Initialize function called by clients like WebRtcAudioDeviceImpl.
// Stop() has to be called before |source| is deleted.
bool Initialize(WebRtcAudioRendererSource* source);
// When sharing a single instance of WebRtcAudioRenderer between multiple
// users (e.g. WebMediaPlayerMS), call this method to create a proxy object
// that maintains the Play and Stop states per caller.
// The wrapper ensures that Play() won't be called when the caller's state
// is "playing", Pause() won't be called when the state already is "paused"
// etc and similarly maintains the same state for Stop().
// When Stop() is called or when the proxy goes out of scope, the proxy
// will ensure that Pause() is called followed by a call to Stop(), which
// is the usage pattern that WebRtcAudioRenderer requires.
scoped_refptr<MediaStreamAudioRenderer> CreateSharedAudioRendererProxy(
const scoped_refptr<webrtc::MediaStreamInterface>& media_stream);
// Used to DCHECK on the expected state.
bool IsStarted() const;
// Accessors to the sink audio parameters.
int channels() const { return sink_params_.channels(); }
int sample_rate() const { return sink_params_.sample_rate(); }
private:
// MediaStreamAudioRenderer implementation. This is private since we want
// callers to use proxy objects.
// TODO(tommi): Make the MediaStreamAudioRenderer implementation a pimpl?
virtual void Start() OVERRIDE;
virtual void Play() OVERRIDE;
virtual void Pause() OVERRIDE;
virtual void Stop() OVERRIDE;
virtual void SetVolume(float volume) OVERRIDE;
virtual base::TimeDelta GetCurrentRenderTime() const OVERRIDE;
virtual bool IsLocalRenderer() const OVERRIDE;
// Called when an audio renderer, either the main or a proxy, starts playing.
// Here we maintain a reference count of how many renderers are currently
// playing so that the shared play state of all the streams can be reflected
// correctly.
void EnterPlayState();
// Called when an audio renderer, either the main or a proxy, is paused.
// See EnterPlayState for more details.
void EnterPauseState();
protected:
virtual ~WebRtcAudioRenderer();
private:
enum State {
UNINITIALIZED,
PLAYING,
PAUSED,
};
// Holds raw pointers to PlaingState objects. Ownership is managed outside
// of this type.
typedef std::vector<PlayingState*> PlayingStates;
// Maps an audio source to a list of playing states that collectively hold
// volume information for that source.
typedef std::map<webrtc::AudioSourceInterface*, PlayingStates>
SourcePlayingStates;
// Used to DCHECK that we are called on the correct thread.
base::ThreadChecker thread_checker_;
// Flag to keep track the state of the renderer.
State state_;
// media::AudioRendererSink::RenderCallback implementation.
// These two methods are called on the AudioOutputDevice worker thread.
virtual int Render(media::AudioBus* audio_bus,
int audio_delay_milliseconds) OVERRIDE;
virtual void OnRenderError() OVERRIDE;
// Called by AudioPullFifo when more data is necessary.
// This method is called on the AudioOutputDevice worker thread.
void SourceCallback(int fifo_frame_delay, media::AudioBus* audio_bus);
// Goes through all renderers for the |source| and applies the proper
// volume scaling for the source based on the volume(s) of the renderer(s).
void UpdateSourceVolume(webrtc::AudioSourceInterface* source);
// Tracks a playing state. The state must be playing when this method
// is called.
// Returns true if the state was added, false if it was already being tracked.
bool AddPlayingState(webrtc::AudioSourceInterface* source,
PlayingState* state);
// Removes a playing state for an audio source.
// Returns true if the state was removed from the internal map, false if
// it had already been removed or if the source isn't being rendered.
bool RemovePlayingState(webrtc::AudioSourceInterface* source,
PlayingState* state);
// Called whenever the Play/Pause state changes of any of the renderers
// or if the volume of any of them is changed.
// Here we update the shared Play state and apply volume scaling to all audio
// sources associated with the |media_stream| based on the collective volume
// of playing renderers.
void OnPlayStateChanged(
const scoped_refptr<webrtc::MediaStreamInterface>& media_stream,
PlayingState* state);
// The render view and frame in which the audio is rendered into |sink_|.
const int source_render_view_id_;
const int source_render_frame_id_;
const int session_id_;
// The sink (destination) for rendered audio.
scoped_refptr<media::AudioOutputDevice> sink_;
// The media stream that holds the audio tracks that this renderer renders.
const scoped_refptr<webrtc::MediaStreamInterface> media_stream_;
// Audio data source from the browser process.
WebRtcAudioRendererSource* source_;
// Protects access to |state_|, |source_|, |sink_| and |current_time_|.
mutable base::Lock lock_;
// Ref count for the MediaPlayers which are playing audio.
int play_ref_count_;
// Ref count for the MediaPlayers which have called Start() but not Stop().
int start_ref_count_;
// Used to buffer data between the client and the output device in cases where
// the client buffer size is not the same as the output device buffer size.
scoped_ptr<media::AudioPullFifo> audio_fifo_;
// Contains the accumulated delay estimate which is provided to the WebRTC
// AEC.
int audio_delay_milliseconds_;
// Delay due to the FIFO in milliseconds.
int fifo_delay_milliseconds_;
base::TimeDelta current_time_;
// Saved volume and playing state of the root renderer.
PlayingState playing_state_;
// Audio params used by the sink of the renderer.
media::AudioParameters sink_params_;
// Maps audio sources to a list of active audio renderers.
// Pointers to PlayingState objects are only kept in this map while the
// associated renderer is actually playing the stream. Ownership of the
// state objects lies with the renderers and they must leave the playing state
// before being destructed (PlayingState object goes out of scope).
SourcePlayingStates source_playing_states_;
DISALLOW_IMPLICIT_CONSTRUCTORS(WebRtcAudioRenderer);
};
} // namespace content
#endif // CONTENT_RENDERER_MEDIA_WEBRTC_AUDIO_RENDERER_H_