| // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef THIRD_PARTY_BLINK_RENDERER_MODULES_WEBRTC_WEBRTC_AUDIO_RENDERER_H_ |
| #define THIRD_PARTY_BLINK_RENDERER_MODULES_WEBRTC_WEBRTC_AUDIO_RENDERER_H_ |
| |
| #include <stdint.h> |
| |
| #include <atomic> |
| #include <map> |
| #include <memory> |
| #include <string> |
| #include <vector> |
| |
| #include "base/macros.h" |
| #include "base/memory/scoped_refptr.h" |
| #include "base/memory/weak_ptr.h" |
| #include "base/optional.h" |
| #include "base/sequence_checker.h" |
| #include "base/single_thread_task_runner.h" |
| #include "base/synchronization/lock.h" |
| #include "base/threading/thread_checker.h" |
| #include "base/time/time.h" |
| #include "base/unguessable_token.h" |
| #include "media/base/audio_decoder.h" |
| #include "media/base/audio_power_monitor.h" |
| #include "media/base/audio_pull_fifo.h" |
| #include "media/base/audio_renderer_sink.h" |
| #include "media/base/channel_layout.h" |
| #include "third_party/blink/public/platform/modules/mediastream/web_media_stream_audio_renderer.h" |
| #include "third_party/blink/renderer/modules/modules_export.h" |
| #include "third_party/blink/renderer/platform/mediastream/media_stream_descriptor.h" |
| #include "third_party/blink/renderer/platform/timer.h" |
| #include "third_party/blink/renderer/platform/webrtc/webrtc_source.h" |
| #include "third_party/blink/renderer/platform/wtf/text/wtf_string.h" |
| |
| namespace media { |
| class SpeechRecognitionClient; |
| } // namespace media |
| |
| namespace webrtc { |
| class AudioSourceInterface; |
| } // namespace webrtc |
| |
| namespace blink { |
| |
| class WebLocalFrame; |
| class WebRtcAudioRendererSource; |
| |
| // This renderer handles calls from the pipeline and WebRtc ADM. It is used |
| // for connecting WebRtc MediaStream with the audio pipeline. |
| class MODULES_EXPORT WebRtcAudioRenderer |
| : public media::AudioRendererSink::RenderCallback, |
| public blink::WebMediaStreamAudioRenderer { |
| public: |
| // Send the audio to the speech recognition service for caption transcription. |
| using TranscribeAudioCallback = base::RepeatingCallback< |
| void(std::unique_ptr<media::AudioBus>, int, media::ChannelLayout)>; |
| |
| // This is a little utility class that holds the configured state of an audio |
| // stream. |
| // It is used by both WebRtcAudioRenderer and SharedAudioRenderer (see cc |
| // file) so a part of why it exists is to avoid code duplication and track |
| // the state in the same way in WebRtcAudioRenderer and SharedAudioRenderer. |
| class PlayingState { |
| public: |
| PlayingState() : playing_(false), volume_(1.0f) {} |
| |
| ~PlayingState() { DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_); } |
| |
| bool playing() const { |
| DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_); |
| return playing_; |
| } |
| |
| void set_playing(bool playing) { |
| DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_); |
| playing_ = playing; |
| } |
| |
| float volume() const { |
| DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_); |
| return volume_; |
| } |
| |
| void set_volume(float volume) { |
| DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_); |
| volume_ = volume; |
| } |
| |
| private: |
| bool playing_; |
| float volume_; |
| |
| SEQUENCE_CHECKER(sequence_checker_); |
| }; |
| |
| enum State { |
| UNINITIALIZED, |
| PLAYING, |
| PAUSED, |
| }; |
| |
| WebRtcAudioRenderer( |
| const scoped_refptr<base::SingleThreadTaskRunner>& signaling_thread, |
| MediaStreamDescriptor* media_stream_descriptor, |
| WebLocalFrame* web_frame, |
| const base::UnguessableToken& session_id, |
| const String& device_id, |
| base::RepeatingCallback<void()> on_render_error_callback); |
| |
| // Initialize function called by clients like WebRtcAudioDeviceImpl. |
| // Stop() has to be called before |source| is deleted. |
| bool Initialize(WebRtcAudioRendererSource* source); |
| |
| // When sharing a single instance of WebRtcAudioRenderer between multiple |
| // users (e.g. WebMediaPlayerMS), call this method to create a proxy object |
| // that maintains the Play and Stop states per caller. |
| // The wrapper ensures that Play() won't be called when the caller's state |
| // is "playing", Pause() won't be called when the state already is "paused" |
| // etc and similarly maintains the same state for Stop(). |
| // When Stop() is called or when the proxy goes out of scope, the proxy |
| // will ensure that Pause() is called followed by a call to Stop(), which |
| // is the usage pattern that WebRtcAudioRenderer requires. |
| scoped_refptr<blink::WebMediaStreamAudioRenderer> |
| CreateSharedAudioRendererProxy( |
| MediaStreamDescriptor* media_stream_descriptor); |
| |
| // Used to DCHECK on the expected state. |
| bool IsStarted() const; |
| |
| // Accessors to the sink audio parameters. |
| int channels() const { return sink_params_.channels(); } |
| int sample_rate() const { return sink_params_.sample_rate(); } |
| int frames_per_buffer() const { return sink_params_.frames_per_buffer(); } |
| |
| // Returns true if called on rendering thread, otherwise false. |
| bool CurrentThreadIsRenderingThread(); |
| |
| private: |
| // blink::WebMediaStreamAudioRenderer implementation. This is private since |
| // we want callers to use proxy objects. |
| // TODO(tommi): Make the blink::WebMediaStreamAudioRenderer implementation a |
| // pimpl? |
| void Start() override; |
| void Play() override; |
| void Pause() override; |
| void Stop() override; |
| void SetVolume(float volume) override; |
| base::TimeDelta GetCurrentRenderTime() override; |
| bool IsLocalRenderer() override; |
| void SwitchOutputDevice(const std::string& device_id, |
| media::OutputDeviceStatusCB callback) override; |
| |
| // Private utility class which keeps track of the state and duration of |
| // playing (rendered on an HTML5 audio tag) audio streams. Mainly intended |
| // for logging purposes to track down "can't hear" type of issues. |
| class AudioStreamTracker { |
| public: |
| // The internal on-shot timer will use the same |task_runner| as the outer |
| // class (OC) who creates this object. |renderer| must outlive the |
| // AudioStreamTracker. See comments for |AudioStreamTracker::renderer_| why |
| // it is safe to use a raw pointer here. |sample_rate| is the current sample |
| // rate used by the audio sink (see WebRtcAudioRenderer::sink_params_). |
| explicit AudioStreamTracker( |
| scoped_refptr<base::SingleThreadTaskRunner> task_runner, |
| WebRtcAudioRenderer* renderer, |
| int sample_rate); |
| |
| // Note: the destructor takes care of logging of the duration of the stream. |
| ~AudioStreamTracker(); |
| |
| // This function should be called from the audio device callback thread, |
| // i.e., the so-called render thread. |
| void OnRenderCallbackCalled(); |
| |
| // Scans the provided audio samples and updates a power measurement. The |
| // "average power" is a running average calculated by using a first-order |
| // low-pass filter over the square of the samples scanned. |
| // Called from the audio render thread and it is safe. See comments in |
| // AudioPowerMonitor::Scan() for more details. |
| void MeasurePower(const media::AudioBus& buffer, int frames); |
| |
| private: |
| // Called by the timer when it fires once after a delay of ~5 seconds from |
| // start. Reads the state of atomic |render_callbacks_started_|. |
| void CheckAlive(TimerBase*); |
| |
| void LogAudioPowerLevel(); |
| |
| // Task runner of outer class (the creating WebRtcAudioRenderer). |
| const scoped_refptr<base::SingleThreadTaskRunner> task_runner_; |
| |
| // Using a raw pointer is safe since the OC instance will outlive this |
| // object. |
| WebRtcAudioRenderer* const renderer_; |
| |
| // Stores when the timer starts. Used to calculate the stream duration. |
| const base::TimeTicks start_time_; |
| |
| // Set to true in first render callback by the high-priority audio thread. |
| // Use an atomic variable since it can also be read by the outer class (OC) |
| // thread once to verify that callbacks started as intended. See comments |
| // for CheckAlive(). |
| std::atomic_bool render_callbacks_started_; |
| |
| // One-shot timer that fires ~5 seconds after rendering should start and |
| // calls the calls CheckAlive() which checks if |render_callbacks_started_| |
| // has been set to true or not. A message is logged to track this state. |
| // The timer uses the same task runner as the OC. Hence, the only writer of |
| // |render_callbacks_started_| is the render thread and the only reader is |
| // the OC thread. DCHECKs are used to confirm this. |
| TaskRunnerTimer<AudioStreamTracker> check_alive_timer_; |
| |
| // Scans audio samples from Render() as input to compute power levels. |
| media::AudioPowerMonitor power_monitor_; |
| |
| // Updated each time a power measurement is logged. |
| base::TimeTicks last_audio_level_log_time_; |
| |
| base::WeakPtr<AudioStreamTracker> weak_this_; |
| base::WeakPtrFactory<AudioStreamTracker> weak_factory_{this}; |
| }; |
| |
| // Called when an audio renderer, either the main or a proxy, starts playing. |
| // Here we maintain a reference count of how many renderers are currently |
| // playing so that the shared play state of all the streams can be reflected |
| // correctly. |
| void EnterPlayState(); |
| |
| // Called when an audio renderer, either the main or a proxy, is paused. |
| // See EnterPlayState for more details. |
| void EnterPauseState(); |
| |
| protected: |
| ~WebRtcAudioRenderer() override; |
| |
| private: |
| // Holds raw pointers to PlaingState objects. Ownership is managed outside |
| // of this type. |
| typedef std::vector<PlayingState*> PlayingStates; |
| // Maps an audio source to a list of playing states that collectively hold |
| // volume information for that source. |
| typedef std::map<webrtc::AudioSourceInterface*, PlayingStates> |
| SourcePlayingStates; |
| |
| // Used to DCHECK that we are called on the correct thread. |
| THREAD_CHECKER(thread_checker_); |
| |
| // Task runner of the creating thread. |
| const scoped_refptr<base::SingleThreadTaskRunner> task_runner_; |
| |
| // Flag to keep track the state of the renderer. |
| State state_; |
| |
| void TranscribeAudio(std::unique_ptr<media::AudioBus> audio_bus, |
| int sample_rate, |
| media::ChannelLayout channel_layout); |
| |
| // media::AudioRendererSink::RenderCallback implementation. |
| // These two methods are called on the AudioOutputDevice worker thread. |
| int Render(base::TimeDelta delay, |
| base::TimeTicks delay_timestamp, |
| int prior_frames_skipped, |
| media::AudioBus* audio_bus) override; |
| void OnRenderError() override; |
| |
| void OnRenderErrorCrossThread(); |
| |
| // Called by AudioPullFifo when more data is necessary. |
| // This method is called on the AudioOutputDevice worker thread. |
| void SourceCallback(int fifo_frame_delay, media::AudioBus* audio_bus); |
| |
| // Goes through all renderers for the |source| and applies the proper |
| // volume scaling for the source based on the volume(s) of the renderer(s). |
| void UpdateSourceVolume(webrtc::AudioSourceInterface* source); |
| |
| // Tracks a playing state. The state must be playing when this method |
| // is called. |
| // Returns true if the state was added, false if it was already being tracked. |
| bool AddPlayingState(webrtc::AudioSourceInterface* source, |
| PlayingState* state); |
| // Removes a playing state for an audio source. |
| // Returns true if the state was removed from the internal map, false if |
| // it had already been removed or if the source isn't being rendered. |
| bool RemovePlayingState(webrtc::AudioSourceInterface* source, |
| PlayingState* state); |
| |
| // Called whenever the Play/Pause state changes of any of the renderers |
| // or if the volume of any of them is changed. |
| // Here we update the shared Play state and apply volume scaling to all audio |
| // sources associated with the |media_stream_descriptor| based on the |
| // collective volume of playing renderers. |
| void OnPlayStateChanged(MediaStreamDescriptor* media_stream_descriptor, |
| PlayingState* state); |
| |
| // Called when |state| is about to be destructed. |
| void OnPlayStateRemoved(PlayingState* state); |
| |
| // Updates |sink_params_| and |audio_fifo_| based on |sink_|, and initializes |
| // |sink_|. |
| void PrepareSink(); |
| |
| void SendLogMessage(const WTF::String& message); |
| |
| void EnableSpeechRecognition(); |
| |
| // The WebLocalFrame in which the audio is rendered into |sink_|. |
| // |
| // TODO(crbug.com/704136): Replace |source_internal_frame_| with regular |
| // fields once this header file moves to blink/renderer. |
| class InternalFrame; |
| std::unique_ptr<InternalFrame> source_internal_frame_; |
| |
| const base::UnguessableToken session_id_; |
| |
| const scoped_refptr<base::SingleThreadTaskRunner> signaling_thread_; |
| |
| // The sink (destination) for rendered audio. |
| scoped_refptr<media::AudioRendererSink> sink_; |
| |
| // The media stream that holds the audio tracks that this renderer renders. |
| Persistent<MediaStreamDescriptor> media_stream_descriptor_; |
| |
| // Contains a copy the unique id of the media stream. By taking a copy at |
| // construction, we can convert the id from a WebString to an WTF::string |
| // once and that saves resources when |media_stream_descriptor_id_| is added |
| // to log messages. |
| String media_stream_descriptor_id_; |
| |
| // Audio data source from the browser process. |
| // |
| // TODO(crbug.com/704136): Make it a Member. |
| WebRtcAudioRendererSource* source_; |
| |
| // Protects access to |state_|, |source_|, |audio_fifo_|, |
| // |audio_delay_milliseconds_|, |fifo_delay_milliseconds_|, |current_time_|, |
| // |sink_params_|, |render_callback_count_|, |max_render_time_| and |
| // |audio_stream_tracker_|. |
| mutable base::Lock lock_; |
| |
| // Ref count for the MediaPlayers which are playing audio. |
| int play_ref_count_; |
| |
| // Ref count for the MediaPlayers which have called Start() but not Stop(). |
| int start_ref_count_; |
| |
| // Used to buffer data between the client and the output device in cases where |
| // the client buffer size is not the same as the output device buffer size. |
| std::unique_ptr<media::AudioPullFifo> audio_fifo_; |
| |
| // Contains the accumulated delay estimate which is provided to the WebRTC |
| // AEC. |
| base::TimeDelta audio_delay_; |
| |
| base::TimeDelta current_time_; |
| |
| // Saved volume and playing state of the root renderer. |
| PlayingState playing_state_; |
| |
| // Audio params used by the sink of the renderer. |
| media::AudioParameters sink_params_; |
| |
| // The preferred device id of the output device or empty for the default |
| // output device. Can change as a result of a SetSinkId() call. |
| String output_device_id_; |
| |
| // Maps audio sources to a list of active audio renderers. |
| // Pointers to PlayingState objects are only kept in this map while the |
| // associated renderer is actually playing the stream. Ownership of the |
| // state objects lies with the renderers and they must leave the playing state |
| // before being destructed (PlayingState object goes out of scope). |
| SourcePlayingStates source_playing_states_; |
| |
| // Stores the maximum time spent waiting for render data from the source. Used |
| // for logging UMA data. Logged and reset when Stop() is called. |
| base::TimeDelta max_render_time_; |
| |
| // Used for keeping track of and logging stats for playing audio streams. |
| // Created when a stream starts and destroyed when a stream stops. |
| // See comments for AudioStreamTracker for more details. |
| base::Optional<AudioStreamTracker> audio_stream_tracker_; |
| |
| base::RepeatingCallback<void()> on_render_error_callback_; |
| |
| std::unique_ptr<media::SpeechRecognitionClient> speech_recognition_client_; |
| TranscribeAudioCallback transcribe_audio_callback_; |
| |
| base::WeakPtrFactory<WebRtcAudioRenderer> weak_factory_{this}; |
| |
| DISALLOW_IMPLICIT_CONSTRUCTORS(WebRtcAudioRenderer); |
| }; |
| |
| } // namespace blink |
| |
| #endif // THIRD_PARTY_BLINK_RENDERER_MODULES_WEBRTC_WEBRTC_AUDIO_RENDERER_H_ |