// Copyright 2013 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "content/renderer/media/webmediaplayer_impl.h"
#include <algorithm>
#include <limits>
#include <string>
#include <vector>
#include "base/bind.h"
#include "base/callback.h"
#include "base/command_line.h"
#include "base/debug/crash_logging.h"
#include "base/debug/trace_event.h"
#include "base/message_loop/message_loop_proxy.h"
#include "base/metrics/histogram.h"
#include "base/strings/string_number_conversions.h"
#include "base/synchronization/waitable_event.h"
#include "cc/layers/video_layer.h"
#include "content/public/common/content_switches.h"
#include "content/renderer/media/buffered_data_source.h"
#include "content/renderer/media/crypto/key_systems.h"
#include "content/renderer/media/texttrack_impl.h"
#include "content/renderer/media/webaudiosourceprovider_impl.h"
#include "content/renderer/media/webinbandtexttrack_impl.h"
#include "content/renderer/media/webmediaplayer_delegate.h"
#include "content/renderer/media/webmediaplayer_params.h"
#include "content/renderer/media/webmediaplayer_util.h"
#include "content/renderer/media/webmediasource_impl.h"
#include "content/renderer/pepper/pepper_webplugin_impl.h"
#include "gpu/GLES2/gl2extchromium.h"
#include "media/audio/null_audio_sink.h"
#include "media/base/bind_to_loop.h"
#include "media/base/filter_collection.h"
#include "media/base/limits.h"
#include "media/base/media_log.h"
#include "media/base/media_switches.h"
#include "media/base/pipeline.h"
#include "media/base/text_renderer.h"
#include "media/base/video_frame.h"
#include "media/filters/audio_renderer_impl.h"
#include "media/filters/chunk_demuxer.h"
#include "media/filters/ffmpeg_audio_decoder.h"
#include "media/filters/ffmpeg_demuxer.h"
#include "media/filters/ffmpeg_video_decoder.h"
#include "media/filters/gpu_video_accelerator_factories.h"
#include "media/filters/gpu_video_decoder.h"
#include "media/filters/opus_audio_decoder.h"
#include "media/filters/video_renderer_impl.h"
#include "media/filters/vpx_video_decoder.h"
#include "third_party/WebKit/public/platform/WebMediaSource.h"
#include "third_party/WebKit/public/platform/WebRect.h"
#include "third_party/WebKit/public/platform/WebSize.h"
#include "third_party/WebKit/public/platform/WebString.h"
#include "third_party/WebKit/public/platform/WebURL.h"
#include "third_party/WebKit/public/web/WebDocument.h"
#include "third_party/WebKit/public/web/WebRuntimeFeatures.h"
#include "third_party/WebKit/public/web/WebView.h"
#include "v8/include/v8.h"
#include "webkit/renderer/compositor_bindings/web_layer_impl.h"
using blink::WebCanvas;
using blink::WebMediaPlayer;
using blink::WebRect;
using blink::WebSize;
using blink::WebString;
using media::PipelineStatus;
namespace {
// Amount of extra memory used by each player instance reported to V8.
// It is not exact number -- first, it differs on different platforms,
// and second, it is very hard to calculate. Instead, use some arbitrary
// value that will cause garbage collection from time to time. We don't want
// it to happen on every allocation, but don't want 5k players to sit in memory
// either. Looks that chosen constant achieves both goals, at least for audio
// objects. (Do not worry about video objects yet, JS programs do not create
// thousands of them...)
const int kPlayerExtraMemory = 1024 * 1024;
// Limits the range of playback rate.
//
// TODO(kylep): Revisit these.
//
// Vista has substantially lower performance than XP or Windows7. If you speed
// up a video too much, it can't keep up, and rendering stops updating except on
// the time bar. For really high speeds, audio becomes a bottleneck and we just
// use up the data we have, which may not achieve the speed requested, but will
// not crash the tab.
//
// A very slow speed, ie 0.00000001x, causes the machine to lock up. (It seems
// like a busy loop). It gets unresponsive, although its not completely dead.
//
// Also our timers are not very accurate (especially for ogg), which becomes
// evident at low speeds and on Vista. Since other speeds are risky and outside
// the norms, we think 1/16x to 16x is a safe and useful range for now.
const double kMinRate = 0.0625;
const double kMaxRate = 16.0;
// Prefix for histograms related to Encrypted Media Extensions.
const char* kMediaEme = "Media.EME.";
} // namespace
namespace content {
#define COMPILE_ASSERT_MATCHING_ENUM(name) \
COMPILE_ASSERT(static_cast<int>(WebMediaPlayer::CORSMode ## name) == \
static_cast<int>(BufferedResourceLoader::k ## name), \
mismatching_enums)
COMPILE_ASSERT_MATCHING_ENUM(Unspecified);
COMPILE_ASSERT_MATCHING_ENUM(Anonymous);
COMPILE_ASSERT_MATCHING_ENUM(UseCredentials);
#undef COMPILE_ASSERT_MATCHING_ENUM
#define BIND_TO_RENDER_LOOP(function) \
media::BindToLoop(main_loop_, base::Bind(function, AsWeakPtr()))
#define BIND_TO_RENDER_LOOP_1(function, arg1) \
media::BindToLoop(main_loop_, base::Bind(function, AsWeakPtr(), arg1))
#define BIND_TO_RENDER_LOOP_2(function, arg1, arg2) \
media::BindToLoop(main_loop_, base::Bind(function, AsWeakPtr(), arg1, arg2))
static void LogMediaSourceError(const scoped_refptr<media::MediaLog>& media_log,
const std::string& error) {
media_log->AddEvent(media_log->CreateMediaSourceErrorEvent(error));
}
WebMediaPlayerImpl::WebMediaPlayerImpl(
content::RenderView* render_view,
blink::WebFrame* frame,
blink::WebMediaPlayerClient* client,
base::WeakPtr<WebMediaPlayerDelegate> delegate,
const WebMediaPlayerParams& params)
: content::RenderViewObserver(render_view),
frame_(frame),
network_state_(WebMediaPlayer::NetworkStateEmpty),
ready_state_(WebMediaPlayer::ReadyStateHaveNothing),
main_loop_(base::MessageLoopProxy::current()),
media_loop_(params.message_loop_proxy()),
paused_(true),
seeking_(false),
playback_rate_(0.0f),
pending_seek_(false),
pending_seek_seconds_(0.0f),
client_(client),
delegate_(delegate),
defer_load_cb_(params.defer_load_cb()),
media_log_(params.media_log()),
accelerated_compositing_reported_(false),
incremented_externally_allocated_memory_(false),
gpu_factories_(params.gpu_factories()),
is_local_source_(false),
supports_save_(true),
starting_(false),
chunk_demuxer_(NULL),
current_frame_painted_(false),
frames_dropped_before_paint_(0),
pending_repaint_(false),
pending_size_change_(false),
video_frame_provider_client_(NULL),
text_track_index_(0) {
media_log_->AddEvent(
media_log_->CreateEvent(media::MediaLogEvent::WEBMEDIAPLAYER_CREATED));
pipeline_.reset(new media::Pipeline(media_loop_, media_log_.get()));
// |gpu_factories_| requires that its entry points be called on its
// |GetMessageLoop()|. Since |pipeline_| will own decoders created from the
// factories, require that their message loops are identical.
DCHECK(!gpu_factories_ || (gpu_factories_->GetMessageLoop() == media_loop_));
// Let V8 know we started new thread if we did not do it yet.
// Made separate task to avoid deletion of player currently being created.
// Also, delaying GC until after player starts gets rid of starting lag --
// collection happens in parallel with playing.
//
// TODO(enal): remove when we get rid of per-audio-stream thread.
main_loop_->PostTask(
FROM_HERE,
base::Bind(&WebMediaPlayerImpl::IncrementExternallyAllocatedMemory,
AsWeakPtr()));
if (blink::WebRuntimeFeatures::isPrefixedEncryptedMediaEnabled()) {
decryptor_.reset(new ProxyDecryptor(
#if defined(ENABLE_PEPPER_CDMS)
client,
frame,
#endif
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnKeyAdded),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnKeyError),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnKeyMessage)));
}
// Use the null sink if no sink was provided.
audio_source_provider_ = new WebAudioSourceProviderImpl(
params.audio_renderer_sink().get()
? params.audio_renderer_sink()
: new media::NullAudioSink(media_loop_));
}
WebMediaPlayerImpl::~WebMediaPlayerImpl() {
SetVideoFrameProviderClient(NULL);
GetClient()->setWebLayer(NULL);
DCHECK(main_loop_->BelongsToCurrentThread());
media_log_->AddEvent(
media_log_->CreateEvent(media::MediaLogEvent::WEBMEDIAPLAYER_DESTROYED));
if (delegate_.get())
delegate_->PlayerGone(this);
Destroy();
}
namespace {
// Helper enum for reporting scheme histograms.
enum URLSchemeForHistogram {
kUnknownURLScheme,
kMissingURLScheme,
kHttpURLScheme,
kHttpsURLScheme,
kFtpURLScheme,
kChromeExtensionURLScheme,
kJavascriptURLScheme,
kFileURLScheme,
kBlobURLScheme,
kDataURLScheme,
kFileSystemScheme,
kMaxURLScheme = kFileSystemScheme // Must be equal to highest enum value.
};
URLSchemeForHistogram URLScheme(const GURL& url) {
if (!url.has_scheme()) return kMissingURLScheme;
if (url.SchemeIs("http")) return kHttpURLScheme;
if (url.SchemeIs("https")) return kHttpsURLScheme;
if (url.SchemeIs("ftp")) return kFtpURLScheme;
if (url.SchemeIs("chrome-extension")) return kChromeExtensionURLScheme;
if (url.SchemeIs("javascript")) return kJavascriptURLScheme;
if (url.SchemeIs("file")) return kFileURLScheme;
if (url.SchemeIs("blob")) return kBlobURLScheme;
if (url.SchemeIs("data")) return kDataURLScheme;
if (url.SchemeIs("filesystem")) return kFileSystemScheme;
return kUnknownURLScheme;
}
} // anonymous namespace
void WebMediaPlayerImpl::load(LoadType load_type, const blink::WebURL& url,
CORSMode cors_mode) {
if (!defer_load_cb_.is_null()) {
defer_load_cb_.Run(base::Bind(
&WebMediaPlayerImpl::DoLoad, AsWeakPtr(), load_type, url, cors_mode));
return;
}
DoLoad(load_type, url, cors_mode);
}
void WebMediaPlayerImpl::DoLoad(LoadType load_type,
const blink::WebURL& url,
CORSMode cors_mode) {
DCHECK(main_loop_->BelongsToCurrentThread());
GURL gurl(url);
UMA_HISTOGRAM_ENUMERATION("Media.URLScheme", URLScheme(gurl), kMaxURLScheme);
// Set subresource URL for crash reporting.
base::debug::SetCrashKeyValue("subresource_url", gurl.spec());
load_type_ = load_type;
// Handle any volume/preload changes that occurred before load().
setVolume(GetClient()->volume());
setPreload(GetClient()->preload());
SetNetworkState(WebMediaPlayer::NetworkStateLoading);
SetReadyState(WebMediaPlayer::ReadyStateHaveNothing);
media_log_->AddEvent(media_log_->CreateLoadEvent(url.spec()));
// Media source pipelines can start immediately.
if (load_type == LoadTypeMediaSource) {
supports_save_ = false;
StartPipeline();
return;
}
// Otherwise it's a regular request which requires resolving the URL first.
data_source_.reset(new BufferedDataSource(
main_loop_,
frame_,
media_log_.get(),
base::Bind(&WebMediaPlayerImpl::NotifyDownloading, AsWeakPtr())));
data_source_->Initialize(
url, static_cast<BufferedResourceLoader::CORSMode>(cors_mode),
base::Bind(
&WebMediaPlayerImpl::DataSourceInitialized,
AsWeakPtr(), gurl));
is_local_source_ = !gurl.SchemeIsHTTPOrHTTPS();
}
void WebMediaPlayerImpl::play() {
DCHECK(main_loop_->BelongsToCurrentThread());
paused_ = false;
pipeline_->SetPlaybackRate(playback_rate_);
if (data_source_)
data_source_->MediaIsPlaying();
media_log_->AddEvent(media_log_->CreateEvent(media::MediaLogEvent::PLAY));
if (delegate_.get())
delegate_->DidPlay(this);
}
void WebMediaPlayerImpl::pause() {
DCHECK(main_loop_->BelongsToCurrentThread());
paused_ = true;
pipeline_->SetPlaybackRate(0.0f);
if (data_source_)
data_source_->MediaIsPaused();
paused_time_ = pipeline_->GetMediaTime();
media_log_->AddEvent(media_log_->CreateEvent(media::MediaLogEvent::PAUSE));
if (delegate_.get())
delegate_->DidPause(this);
}
bool WebMediaPlayerImpl::supportsFullscreen() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return true;
}
bool WebMediaPlayerImpl::supportsSave() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return supports_save_;
}
void WebMediaPlayerImpl::seek(double seconds) {
DCHECK(main_loop_->BelongsToCurrentThread());
if (ready_state_ > WebMediaPlayer::ReadyStateHaveMetadata)
SetReadyState(WebMediaPlayer::ReadyStateHaveMetadata);
base::TimeDelta seek_time = ConvertSecondsToTimestamp(seconds);
if (starting_ || seeking_) {
pending_seek_ = true;
pending_seek_seconds_ = seconds;
if (chunk_demuxer_)
chunk_demuxer_->CancelPendingSeek(seek_time);
return;
}
media_log_->AddEvent(media_log_->CreateSeekEvent(seconds));
// Update our paused time.
if (paused_)
paused_time_ = seek_time;
seeking_ = true;
if (chunk_demuxer_)
chunk_demuxer_->StartWaitingForSeek(seek_time);
// Kick off the asynchronous seek!
pipeline_->Seek(
seek_time,
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnPipelineSeek));
}
void WebMediaPlayerImpl::setRate(double rate) {
DCHECK(main_loop_->BelongsToCurrentThread());
// TODO(kylep): Remove when support for negatives is added. Also, modify the
// following checks so rewind uses reasonable values also.
if (rate < 0.0)
return;
// Limit rates to reasonable values by clamping.
if (rate != 0.0) {
if (rate < kMinRate)
rate = kMinRate;
else if (rate > kMaxRate)
rate = kMaxRate;
}
playback_rate_ = rate;
if (!paused_) {
pipeline_->SetPlaybackRate(rate);
if (data_source_)
data_source_->MediaPlaybackRateChanged(rate);
}
}
void WebMediaPlayerImpl::setVolume(double volume) {
DCHECK(main_loop_->BelongsToCurrentThread());
pipeline_->SetVolume(volume);
}
#define COMPILE_ASSERT_MATCHING_ENUM(webkit_name, chromium_name) \
COMPILE_ASSERT(static_cast<int>(WebMediaPlayer::webkit_name) == \
static_cast<int>(content::chromium_name), \
mismatching_enums)
COMPILE_ASSERT_MATCHING_ENUM(PreloadNone, NONE);
COMPILE_ASSERT_MATCHING_ENUM(PreloadMetaData, METADATA);
COMPILE_ASSERT_MATCHING_ENUM(PreloadAuto, AUTO);
#undef COMPILE_ASSERT_MATCHING_ENUM
void WebMediaPlayerImpl::setPreload(WebMediaPlayer::Preload preload) {
DCHECK(main_loop_->BelongsToCurrentThread());
if (data_source_)
data_source_->SetPreload(static_cast<content::Preload>(preload));
}
bool WebMediaPlayerImpl::hasVideo() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return pipeline_->HasVideo();
}
bool WebMediaPlayerImpl::hasAudio() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return pipeline_->HasAudio();
}
blink::WebSize WebMediaPlayerImpl::naturalSize() const {
DCHECK(main_loop_->BelongsToCurrentThread());
gfx::Size size;
pipeline_->GetNaturalVideoSize(&size);
return blink::WebSize(size);
}
bool WebMediaPlayerImpl::paused() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return pipeline_->GetPlaybackRate() == 0.0f;
}
bool WebMediaPlayerImpl::seeking() const {
DCHECK(main_loop_->BelongsToCurrentThread());
if (ready_state_ == WebMediaPlayer::ReadyStateHaveNothing)
return false;
return seeking_;
}
double WebMediaPlayerImpl::duration() const {
DCHECK(main_loop_->BelongsToCurrentThread());
if (ready_state_ == WebMediaPlayer::ReadyStateHaveNothing)
return std::numeric_limits<double>::quiet_NaN();
return GetPipelineDuration();
}
double WebMediaPlayerImpl::currentTime() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return (paused_ ? paused_time_ : pipeline_->GetMediaTime()).InSecondsF();
}
WebMediaPlayer::NetworkState WebMediaPlayerImpl::networkState() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return network_state_;
}
WebMediaPlayer::ReadyState WebMediaPlayerImpl::readyState() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return ready_state_;
}
const blink::WebTimeRanges& WebMediaPlayerImpl::buffered() {
DCHECK(main_loop_->BelongsToCurrentThread());
blink::WebTimeRanges web_ranges(
ConvertToWebTimeRanges(pipeline_->GetBufferedTimeRanges()));
buffered_.swap(web_ranges);
return buffered_;
}
double WebMediaPlayerImpl::maxTimeSeekable() const {
DCHECK(main_loop_->BelongsToCurrentThread());
// If we haven't even gotten to ReadyStateHaveMetadata yet then just
// return 0 so that the seekable range is empty.
if (ready_state_ < WebMediaPlayer::ReadyStateHaveMetadata)
return 0.0;
// We don't support seeking in streaming media.
if (data_source_ && data_source_->IsStreaming())
return 0.0;
return duration();
}
bool WebMediaPlayerImpl::didLoadingProgress() const {
DCHECK(main_loop_->BelongsToCurrentThread());
return pipeline_->DidLoadingProgress();
}
void WebMediaPlayerImpl::paint(WebCanvas* canvas,
const WebRect& rect,
unsigned char alpha) {
DCHECK(main_loop_->BelongsToCurrentThread());
if (!accelerated_compositing_reported_) {
accelerated_compositing_reported_ = true;
// Normally paint() is only called in non-accelerated rendering, but there
// are exceptions such as webgl where compositing is used in the WebView but
// video frames are still rendered to a canvas.
UMA_HISTOGRAM_BOOLEAN(
"Media.AcceleratedCompositingActive",
frame_->view()->isAcceleratedCompositingActive());
}
// Avoid locking and potentially blocking the video rendering thread while
// painting in software.
scoped_refptr<media::VideoFrame> video_frame;
{
base::AutoLock auto_lock(lock_);
DoneWaitingForPaint(true);
video_frame = current_frame_;
}
TRACE_EVENT0("media", "WebMediaPlayerImpl:paint");
gfx::Rect gfx_rect(rect);
skcanvas_video_renderer_.Paint(video_frame.get(), canvas, gfx_rect, alpha);
}
bool WebMediaPlayerImpl::hasSingleSecurityOrigin() const {
if (data_source_)
return data_source_->HasSingleOrigin();
return true;
}
bool WebMediaPlayerImpl::didPassCORSAccessCheck() const {
if (data_source_)
return data_source_->DidPassCORSAccessCheck();
return false;
}
double WebMediaPlayerImpl::mediaTimeForTimeValue(double timeValue) const {
return ConvertSecondsToTimestamp(timeValue).InSecondsF();
}
unsigned WebMediaPlayerImpl::decodedFrameCount() const {
DCHECK(main_loop_->BelongsToCurrentThread());
media::PipelineStatistics stats = pipeline_->GetStatistics();
return stats.video_frames_decoded;
}
unsigned WebMediaPlayerImpl::droppedFrameCount() const {
DCHECK(main_loop_->BelongsToCurrentThread());
media::PipelineStatistics stats = pipeline_->GetStatistics();
base::AutoLock auto_lock(lock_);
unsigned frames_dropped =
stats.video_frames_dropped + frames_dropped_before_paint_;
DCHECK_LE(frames_dropped, stats.video_frames_decoded);
return frames_dropped;
}
unsigned WebMediaPlayerImpl::audioDecodedByteCount() const {
DCHECK(main_loop_->BelongsToCurrentThread());
media::PipelineStatistics stats = pipeline_->GetStatistics();
return stats.audio_bytes_decoded;
}
unsigned WebMediaPlayerImpl::videoDecodedByteCount() const {
DCHECK(main_loop_->BelongsToCurrentThread());
media::PipelineStatistics stats = pipeline_->GetStatistics();
return stats.video_bytes_decoded;
}
void WebMediaPlayerImpl::SetVideoFrameProviderClient(
cc::VideoFrameProvider::Client* client) {
// This is called from both the main renderer thread and the compositor
// thread (when the main thread is blocked).
if (video_frame_provider_client_)
video_frame_provider_client_->StopUsingProvider();
video_frame_provider_client_ = client;
}
scoped_refptr<media::VideoFrame> WebMediaPlayerImpl::GetCurrentFrame() {
base::AutoLock auto_lock(lock_);
DoneWaitingForPaint(true);
TRACE_EVENT_ASYNC_BEGIN0(
"media", "WebMediaPlayerImpl:compositing", this);
return current_frame_;
}
void WebMediaPlayerImpl::PutCurrentFrame(
const scoped_refptr<media::VideoFrame>& frame) {
if (!accelerated_compositing_reported_) {
accelerated_compositing_reported_ = true;
DCHECK(frame_->view()->isAcceleratedCompositingActive());
UMA_HISTOGRAM_BOOLEAN("Media.AcceleratedCompositingActive", true);
}
TRACE_EVENT_ASYNC_END0("media", "WebMediaPlayerImpl:compositing", this);
}
bool WebMediaPlayerImpl::copyVideoTextureToPlatformTexture(
blink::WebGraphicsContext3D* web_graphics_context,
unsigned int texture,
unsigned int level,
unsigned int internal_format,
unsigned int type,
bool premultiply_alpha,
bool flip_y) {
scoped_refptr<media::VideoFrame> video_frame;
{
base::AutoLock auto_lock(lock_);
video_frame = current_frame_;
}
TRACE_EVENT0("media", "WebMediaPlayerImpl:copyVideoTextureToPlatformTexture");
if (!video_frame)
return false;
if (video_frame->format() != media::VideoFrame::NATIVE_TEXTURE)
return false;
if (video_frame->texture_target() != GL_TEXTURE_2D)
return false;
// Since this method changes which texture is bound to the TEXTURE_2D target,
// ideally it would restore the currently-bound texture before returning.
// The cost of getIntegerv is sufficiently high, however, that we want to
// avoid it in user builds. As a result assume (below) that |texture| is
// bound when this method is called, and only verify this fact when
// DCHECK_IS_ON.
if (DCHECK_IS_ON()) {
GLint bound_texture = 0;
web_graphics_context->getIntegerv(GL_TEXTURE_BINDING_2D, &bound_texture);
DCHECK_EQ(static_cast<GLuint>(bound_texture), texture);
}
media::VideoFrame::MailboxHolder* mailbox_holder =
video_frame->texture_mailbox();
uint32 source_texture = web_graphics_context->createTexture();
web_graphics_context->waitSyncPoint(mailbox_holder->sync_point());
web_graphics_context->bindTexture(GL_TEXTURE_2D, source_texture);
web_graphics_context->consumeTextureCHROMIUM(GL_TEXTURE_2D,
mailbox_holder->mailbox().name);
// The video is stored in a unmultiplied format, so premultiply
// if necessary.
web_graphics_context->pixelStorei(GL_UNPACK_PREMULTIPLY_ALPHA_CHROMIUM,
premultiply_alpha);
// Application itself needs to take care of setting the right flip_y
// value down to get the expected result.
// flip_y==true means to reverse the video orientation while
// flip_y==false means to keep the intrinsic orientation.
web_graphics_context->pixelStorei(GL_UNPACK_FLIP_Y_CHROMIUM, flip_y);
web_graphics_context->copyTextureCHROMIUM(GL_TEXTURE_2D,
source_texture,
texture,
level,
internal_format,
type);
web_graphics_context->pixelStorei(GL_UNPACK_FLIP_Y_CHROMIUM, false);
web_graphics_context->pixelStorei(GL_UNPACK_PREMULTIPLY_ALPHA_CHROMIUM,
false);
// Restore the state for TEXTURE_2D binding point as mentioned above.
web_graphics_context->bindTexture(GL_TEXTURE_2D, texture);
web_graphics_context->deleteTexture(source_texture);
// The flush() operation is not necessary here. It is kept since the
// performance will be better when it is added than not.
web_graphics_context->flush();
return true;
}
// Helper functions to report media EME related stats to UMA. They follow the
// convention of more commonly used macros UMA_HISTOGRAM_ENUMERATION and
// UMA_HISTOGRAM_COUNTS. The reason that we cannot use those macros directly is
// that UMA_* macros require the names to be constant throughout the process'
// lifetime.
static void EmeUMAHistogramEnumeration(const blink::WebString& key_system,
const std::string& method,
int sample,
int boundary_value) {
base::LinearHistogram::FactoryGet(
kMediaEme + KeySystemNameForUMA(key_system) + "." + method,
1, boundary_value, boundary_value + 1,
base::Histogram::kUmaTargetedHistogramFlag)->Add(sample);
}
static void EmeUMAHistogramCounts(const blink::WebString& key_system,
const std::string& method,
int sample) {
// Use the same parameters as UMA_HISTOGRAM_COUNTS.
base::Histogram::FactoryGet(
kMediaEme + KeySystemNameForUMA(key_system) + "." + method,
1, 1000000, 50, base::Histogram::kUmaTargetedHistogramFlag)->Add(sample);
}
// Helper enum for reporting generateKeyRequest/addKey histograms.
enum MediaKeyException {
kUnknownResultId,
kSuccess,
kKeySystemNotSupported,
kInvalidPlayerState,
kMaxMediaKeyException
};
static MediaKeyException MediaKeyExceptionForUMA(
WebMediaPlayer::MediaKeyException e) {
switch (e) {
case WebMediaPlayer::MediaKeyExceptionKeySystemNotSupported:
return kKeySystemNotSupported;
case WebMediaPlayer::MediaKeyExceptionInvalidPlayerState:
return kInvalidPlayerState;
case WebMediaPlayer::MediaKeyExceptionNoError:
return kSuccess;
default:
return kUnknownResultId;
}
}
// Helper for converting |key_system| name and exception |e| to a pair of enum
// values from above, for reporting to UMA.
static void ReportMediaKeyExceptionToUMA(
const std::string& method,
const WebString& key_system,
WebMediaPlayer::MediaKeyException e) {
MediaKeyException result_id = MediaKeyExceptionForUMA(e);
DCHECK_NE(result_id, kUnknownResultId) << e;
EmeUMAHistogramEnumeration(
key_system, method, result_id, kMaxMediaKeyException);
}
WebMediaPlayer::MediaKeyException
WebMediaPlayerImpl::generateKeyRequest(const WebString& key_system,
const unsigned char* init_data,
unsigned init_data_length) {
WebMediaPlayer::MediaKeyException e =
GenerateKeyRequestInternal(key_system, init_data, init_data_length);
ReportMediaKeyExceptionToUMA("generateKeyRequest", key_system, e);
return e;
}
WebMediaPlayer::MediaKeyException
WebMediaPlayerImpl::GenerateKeyRequestInternal(
const WebString& key_system,
const unsigned char* init_data,
unsigned init_data_length) {
DVLOG(1) << "generateKeyRequest: " << key_system.utf8().data() << ": "
<< std::string(reinterpret_cast<const char*>(init_data),
static_cast<size_t>(init_data_length));
if (!IsConcreteSupportedKeySystem(key_system))
return WebMediaPlayer::MediaKeyExceptionKeySystemNotSupported;
// We do not support run-time switching between key systems for now.
if (current_key_system_.isEmpty()) {
if (!decryptor_->InitializeCDM(key_system.utf8(), frame_->document().url()))
return WebMediaPlayer::MediaKeyExceptionKeySystemNotSupported;
current_key_system_ = key_system;
}
else if (key_system != current_key_system_) {
return WebMediaPlayer::MediaKeyExceptionInvalidPlayerState;
}
// TODO(xhwang): We assume all streams are from the same container (thus have
// the same "type") for now. In the future, the "type" should be passed down
// from the application.
if (!decryptor_->GenerateKeyRequest(init_data_type_,
init_data, init_data_length)) {
current_key_system_.reset();
return WebMediaPlayer::MediaKeyExceptionKeySystemNotSupported;
}
return WebMediaPlayer::MediaKeyExceptionNoError;
}
WebMediaPlayer::MediaKeyException WebMediaPlayerImpl::addKey(
const WebString& key_system,
const unsigned char* key,
unsigned key_length,
const unsigned char* init_data,
unsigned init_data_length,
const WebString& session_id) {
WebMediaPlayer::MediaKeyException e = AddKeyInternal(
key_system, key, key_length, init_data, init_data_length, session_id);
ReportMediaKeyExceptionToUMA("addKey", key_system, e);
return e;
}
WebMediaPlayer::MediaKeyException WebMediaPlayerImpl::AddKeyInternal(
const WebString& key_system,
const unsigned char* key,
unsigned key_length,
const unsigned char* init_data,
unsigned init_data_length,
const WebString& session_id) {
DCHECK(key);
DCHECK_GT(key_length, 0u);
DVLOG(1) << "addKey: " << key_system.utf8().data() << ": "
<< std::string(reinterpret_cast<const char*>(key),
static_cast<size_t>(key_length)) << ", "
<< std::string(reinterpret_cast<const char*>(init_data),
static_cast<size_t>(init_data_length))
<< " [" << session_id.utf8().data() << "]";
if (!IsConcreteSupportedKeySystem(key_system))
return WebMediaPlayer::MediaKeyExceptionKeySystemNotSupported;
if (current_key_system_.isEmpty() || key_system != current_key_system_)
return WebMediaPlayer::MediaKeyExceptionInvalidPlayerState;
decryptor_->AddKey(key, key_length,
init_data, init_data_length, session_id.utf8());
return WebMediaPlayer::MediaKeyExceptionNoError;
}
WebMediaPlayer::MediaKeyException WebMediaPlayerImpl::cancelKeyRequest(
const WebString& key_system,
const WebString& session_id) {
WebMediaPlayer::MediaKeyException e =
CancelKeyRequestInternal(key_system, session_id);
ReportMediaKeyExceptionToUMA("cancelKeyRequest", key_system, e);
return e;
}
WebMediaPlayer::MediaKeyException
WebMediaPlayerImpl::CancelKeyRequestInternal(
const WebString& key_system,
const WebString& session_id) {
if (!IsConcreteSupportedKeySystem(key_system))
return WebMediaPlayer::MediaKeyExceptionKeySystemNotSupported;
if (current_key_system_.isEmpty() || key_system != current_key_system_)
return WebMediaPlayer::MediaKeyExceptionInvalidPlayerState;
decryptor_->CancelKeyRequest(session_id.utf8());
return WebMediaPlayer::MediaKeyExceptionNoError;
}
void WebMediaPlayerImpl::OnDestruct() {
Destroy();
}
void WebMediaPlayerImpl::Repaint() {
DCHECK(main_loop_->BelongsToCurrentThread());
TRACE_EVENT0("media", "WebMediaPlayerImpl:repaint");
bool size_changed = false;
{
base::AutoLock auto_lock(lock_);
std::swap(pending_size_change_, size_changed);
if (pending_repaint_) {
TRACE_EVENT_ASYNC_END0(
"media", "WebMediaPlayerImpl:repaintPending", this);
pending_repaint_ = false;
}
}
if (size_changed) {
TRACE_EVENT0("media", "WebMediaPlayerImpl:clientSizeChanged");
GetClient()->sizeChanged();
}
TRACE_EVENT0("media", "WebMediaPlayerImpl:clientRepaint");
GetClient()->repaint();
}
void WebMediaPlayerImpl::OnPipelineSeek(PipelineStatus status) {
DCHECK(main_loop_->BelongsToCurrentThread());
starting_ = false;
seeking_ = false;
if (pending_seek_) {
pending_seek_ = false;
seek(pending_seek_seconds_);
return;
}
if (status != media::PIPELINE_OK) {
OnPipelineError(status);
return;
}
// Update our paused time.
if (paused_)
paused_time_ = pipeline_->GetMediaTime();
GetClient()->timeChanged();
}
void WebMediaPlayerImpl::OnPipelineEnded() {
DCHECK(main_loop_->BelongsToCurrentThread());
GetClient()->timeChanged();
}
void WebMediaPlayerImpl::OnPipelineError(PipelineStatus error) {
DCHECK(main_loop_->BelongsToCurrentThread());
DCHECK_NE(error, media::PIPELINE_OK);
if (ready_state_ == WebMediaPlayer::ReadyStateHaveNothing) {
// Any error that occurs before reaching ReadyStateHaveMetadata should
// be considered a format error.
SetNetworkState(WebMediaPlayer::NetworkStateFormatError);
Repaint();
return;
}
SetNetworkState(PipelineErrorToNetworkState(error));
if (error == media::PIPELINE_ERROR_DECRYPT)
EmeUMAHistogramCounts(current_key_system_, "DecryptError", 1);
// Repaint to trigger UI update.
Repaint();
}
void WebMediaPlayerImpl::OnPipelineBufferingState(
media::Pipeline::BufferingState buffering_state) {
DVLOG(1) << "OnPipelineBufferingState(" << buffering_state << ")";
switch (buffering_state) {
case media::Pipeline::kHaveMetadata:
SetReadyState(WebMediaPlayer::ReadyStateHaveMetadata);
if (hasVideo() && GetClient()->needsWebLayerForVideo()) {
DCHECK(!video_weblayer_);
video_weblayer_.reset(
new webkit::WebLayerImpl(cc::VideoLayer::Create(this)));
GetClient()->setWebLayer(video_weblayer_.get());
}
break;
case media::Pipeline::kPrerollCompleted:
// Only transition to ReadyStateHaveEnoughData if we don't have
// any pending seeks because the transition can cause Blink to
// report that the most recent seek has completed.
if (!pending_seek_)
SetReadyState(WebMediaPlayer::ReadyStateHaveEnoughData);
break;
}
// Repaint to trigger UI update.
Repaint();
}
void WebMediaPlayerImpl::OnDemuxerOpened() {
DCHECK(main_loop_->BelongsToCurrentThread());
GetClient()->mediaSourceOpened(new WebMediaSourceImpl(
chunk_demuxer_, base::Bind(&LogMediaSourceError, media_log_)));
}
void WebMediaPlayerImpl::OnKeyAdded(const std::string& session_id) {
DCHECK(main_loop_->BelongsToCurrentThread());
EmeUMAHistogramCounts(current_key_system_, "KeyAdded", 1);
GetClient()->keyAdded(current_key_system_,
WebString::fromUTF8(session_id));
}
void WebMediaPlayerImpl::OnNeedKey(const std::string& type,
const std::vector<uint8>& init_data) {
DCHECK(main_loop_->BelongsToCurrentThread());
// Do not fire NeedKey event if encrypted media is not enabled.
if (!decryptor_)
return;
UMA_HISTOGRAM_COUNTS(kMediaEme + std::string("NeedKey"), 1);
DCHECK(init_data_type_.empty() || type.empty() || type == init_data_type_);
if (init_data_type_.empty())
init_data_type_ = type;
const uint8* init_data_ptr = init_data.empty() ? NULL : &init_data[0];
GetClient()->keyNeeded(WebString(),
WebString(),
init_data_ptr,
init_data.size());
}
void WebMediaPlayerImpl::OnAddTextTrack(
const media::TextTrackConfig& config,
const media::AddTextTrackDoneCB& done_cb) {
DCHECK(main_loop_->BelongsToCurrentThread());
const WebInbandTextTrackImpl::Kind web_kind =
static_cast<WebInbandTextTrackImpl::Kind>(config.kind());
const blink::WebString web_label =
blink::WebString::fromUTF8(config.label());
const blink::WebString web_language =
blink::WebString::fromUTF8(config.language());
const blink::WebString web_id =
blink::WebString::fromUTF8(config.id());
scoped_ptr<WebInbandTextTrackImpl> web_inband_text_track(
new WebInbandTextTrackImpl(web_kind, web_label, web_language, web_id,
text_track_index_++));
scoped_ptr<media::TextTrack> text_track(
new TextTrackImpl(main_loop_, GetClient(), web_inband_text_track.Pass()));
done_cb.Run(text_track.Pass());
}
void WebMediaPlayerImpl::OnKeyError(const std::string& session_id,
media::MediaKeys::KeyError error_code,
int system_code) {
DCHECK(main_loop_->BelongsToCurrentThread());
EmeUMAHistogramEnumeration(current_key_system_, "KeyError",
error_code, media::MediaKeys::kMaxKeyError);
GetClient()->keyError(
current_key_system_,
WebString::fromUTF8(session_id),
static_cast<blink::WebMediaPlayerClient::MediaKeyErrorCode>(error_code),
system_code);
}
void WebMediaPlayerImpl::OnKeyMessage(const std::string& session_id,
const std::vector<uint8>& message,
const std::string& default_url) {
DCHECK(main_loop_->BelongsToCurrentThread());
const GURL default_url_gurl(default_url);
DLOG_IF(WARNING, !default_url.empty() && !default_url_gurl.is_valid())
<< "Invalid URL in default_url: " << default_url;
GetClient()->keyMessage(current_key_system_,
WebString::fromUTF8(session_id),
message.empty() ? NULL : &message[0],
message.size(),
default_url_gurl);
}
void WebMediaPlayerImpl::SetOpaque(bool opaque) {
DCHECK(main_loop_->BelongsToCurrentThread());
GetClient()->setOpaque(opaque);
}
void WebMediaPlayerImpl::DataSourceInitialized(const GURL& gurl, bool success) {
DCHECK(main_loop_->BelongsToCurrentThread());
if (!success) {
SetNetworkState(WebMediaPlayer::NetworkStateFormatError);
Repaint();
return;
}
StartPipeline();
}
void WebMediaPlayerImpl::NotifyDownloading(bool is_downloading) {
if (!is_downloading && network_state_ == WebMediaPlayer::NetworkStateLoading)
SetNetworkState(WebMediaPlayer::NetworkStateIdle);
else if (is_downloading && network_state_ == WebMediaPlayer::NetworkStateIdle)
SetNetworkState(WebMediaPlayer::NetworkStateLoading);
media_log_->AddEvent(
media_log_->CreateBooleanEvent(
media::MediaLogEvent::NETWORK_ACTIVITY_SET,
"is_downloading_data", is_downloading));
}
void WebMediaPlayerImpl::StartPipeline() {
const CommandLine* cmd_line = CommandLine::ForCurrentProcess();
// Keep track if this is a MSE or non-MSE playback.
UMA_HISTOGRAM_BOOLEAN("Media.MSE.Playback",
(load_type_ == LoadTypeMediaSource));
// Figure out which demuxer to use.
if (load_type_ != LoadTypeMediaSource) {
DCHECK(!chunk_demuxer_);
DCHECK(data_source_);
demuxer_.reset(new media::FFmpegDemuxer(
media_loop_, data_source_.get(),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnNeedKey),
media_log_));
} else {
DCHECK(!chunk_demuxer_);
DCHECK(!data_source_);
chunk_demuxer_ = new media::ChunkDemuxer(
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnDemuxerOpened),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnNeedKey),
base::Bind(&LogMediaSourceError, media_log_));
demuxer_.reset(chunk_demuxer_);
}
scoped_ptr<media::FilterCollection> filter_collection(
new media::FilterCollection());
filter_collection->SetDemuxer(demuxer_.get());
// Figure out if EME is enabled.
media::SetDecryptorReadyCB set_decryptor_ready_cb;
if (decryptor_) {
set_decryptor_ready_cb = base::Bind(&ProxyDecryptor::SetDecryptorReadyCB,
base::Unretained(decryptor_.get()));
}
// Create our audio decoders and renderer.
ScopedVector<media::AudioDecoder> audio_decoders;
audio_decoders.push_back(new media::FFmpegAudioDecoder(media_loop_));
if (!cmd_line->HasSwitch(switches::kDisableOpusPlayback)) {
audio_decoders.push_back(new media::OpusAudioDecoder(media_loop_));
}
scoped_ptr<media::AudioRenderer> audio_renderer(
new media::AudioRendererImpl(media_loop_,
audio_source_provider_.get(),
audio_decoders.Pass(),
set_decryptor_ready_cb));
filter_collection->SetAudioRenderer(audio_renderer.Pass());
// Create our video decoders and renderer.
ScopedVector<media::VideoDecoder> video_decoders;
if (gpu_factories_.get()) {
video_decoders.push_back(
new media::GpuVideoDecoder(gpu_factories_, media_log_));
}
#if !defined(MEDIA_DISABLE_LIBVPX)
video_decoders.push_back(new media::VpxVideoDecoder(media_loop_));
#endif // !defined(MEDIA_DISABLE_LIBVPX)
video_decoders.push_back(new media::FFmpegVideoDecoder(media_loop_));
scoped_ptr<media::VideoRenderer> video_renderer(
new media::VideoRendererImpl(
media_loop_,
video_decoders.Pass(),
set_decryptor_ready_cb,
base::Bind(&WebMediaPlayerImpl::FrameReady, base::Unretained(this)),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::SetOpaque),
true));
filter_collection->SetVideoRenderer(video_renderer.Pass());
if (cmd_line->HasSwitch(switches::kEnableInbandTextTracks)) {
scoped_ptr<media::TextRenderer> text_renderer(
new media::TextRenderer(
media_loop_,
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnAddTextTrack)));
filter_collection->SetTextRenderer(text_renderer.Pass());
}
// ... and we're ready to go!
starting_ = true;
pipeline_->Start(
filter_collection.Pass(),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnPipelineEnded),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnPipelineError),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnPipelineSeek),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnPipelineBufferingState),
BIND_TO_RENDER_LOOP(&WebMediaPlayerImpl::OnDurationChange));
}
void WebMediaPlayerImpl::SetNetworkState(WebMediaPlayer::NetworkState state) {
DCHECK(main_loop_->BelongsToCurrentThread());
DVLOG(1) << "SetNetworkState: " << state;
network_state_ = state;
// Always notify to ensure client has the latest value.
GetClient()->networkStateChanged();
}
void WebMediaPlayerImpl::SetReadyState(WebMediaPlayer::ReadyState state) {
DCHECK(main_loop_->BelongsToCurrentThread());
DVLOG(1) << "SetReadyState: " << state;
if (state == WebMediaPlayer::ReadyStateHaveEnoughData &&
is_local_source_ &&
network_state_ == WebMediaPlayer::NetworkStateLoading)
SetNetworkState(WebMediaPlayer::NetworkStateLoaded);
ready_state_ = state;
// Always notify to ensure client has the latest value.
GetClient()->readyStateChanged();
}
void WebMediaPlayerImpl::Destroy() {
DCHECK(main_loop_->BelongsToCurrentThread());
// Abort any pending IO so stopping the pipeline doesn't get blocked.
if (data_source_)
data_source_->Abort();
if (chunk_demuxer_) {
chunk_demuxer_->Shutdown();
chunk_demuxer_ = NULL;
}
if (gpu_factories_.get()) {
gpu_factories_->Abort();
gpu_factories_ = NULL;
}
// Make sure to kill the pipeline so there's no more media threads running.
// Note: stopping the pipeline might block for a long time.
base::WaitableEvent waiter(false, false);
pipeline_->Stop(base::Bind(
&base::WaitableEvent::Signal, base::Unretained(&waiter)));
waiter.Wait();
// Let V8 know we are not using extra resources anymore.
if (incremented_externally_allocated_memory_) {
v8::Isolate::GetCurrent()->AdjustAmountOfExternalAllocatedMemory(
-kPlayerExtraMemory);
incremented_externally_allocated_memory_ = false;
}
// Release any final references now that everything has stopped.
pipeline_.reset();
demuxer_.reset();
data_source_.reset();
}
blink::WebMediaPlayerClient* WebMediaPlayerImpl::GetClient() {
DCHECK(main_loop_->BelongsToCurrentThread());
DCHECK(client_);
return client_;
}
blink::WebAudioSourceProvider* WebMediaPlayerImpl::audioSourceProvider() {
return audio_source_provider_.get();
}
void WebMediaPlayerImpl::IncrementExternallyAllocatedMemory() {
DCHECK(main_loop_->BelongsToCurrentThread());
incremented_externally_allocated_memory_ = true;
v8::Isolate::GetCurrent()->AdjustAmountOfExternalAllocatedMemory(
kPlayerExtraMemory);
}
double WebMediaPlayerImpl::GetPipelineDuration() const {
base::TimeDelta duration = pipeline_->GetMediaDuration();
// Return positive infinity if the resource is unbounded.
// http://www.whatwg.org/specs/web-apps/current-work/multipage/video.html#dom-media-duration
if (duration == media::kInfiniteDuration())
return std::numeric_limits<double>::infinity();
return duration.InSecondsF();
}
void WebMediaPlayerImpl::OnDurationChange() {
if (ready_state_ == WebMediaPlayer::ReadyStateHaveNothing)
return;
GetClient()->durationChanged();
}
void WebMediaPlayerImpl::FrameReady(
const scoped_refptr<media::VideoFrame>& frame) {
base::AutoLock auto_lock(lock_);
if (current_frame_ &&
current_frame_->natural_size() != frame->natural_size() &&
!pending_size_change_) {
pending_size_change_ = true;
}
DoneWaitingForPaint(false);
current_frame_ = frame;
current_frame_painted_ = false;
TRACE_EVENT_FLOW_BEGIN0("media", "WebMediaPlayerImpl:waitingForPaint", this);
if (pending_repaint_)
return;
TRACE_EVENT_ASYNC_BEGIN0("media", "WebMediaPlayerImpl:repaintPending", this);
pending_repaint_ = true;
main_loop_->PostTask(FROM_HERE, base::Bind(
&WebMediaPlayerImpl::Repaint, AsWeakPtr()));
}
void WebMediaPlayerImpl::DoneWaitingForPaint(bool painting_frame) {
lock_.AssertAcquired();
if (!current_frame_ || current_frame_painted_)
return;
TRACE_EVENT_FLOW_END0("media", "WebMediaPlayerImpl:waitingForPaint", this);
if (painting_frame) {
current_frame_painted_ = true;
return;
}
// The frame wasn't painted, but we aren't waiting for a Repaint() call so
// assume that the frame wasn't painted because the video wasn't visible.
if (!pending_repaint_)
return;
// The |current_frame_| wasn't painted, it is being replaced, and we haven't
// even gotten the chance to request a repaint for it yet. Mark it as dropped.
TRACE_EVENT0("media", "WebMediaPlayerImpl:frameDropped");
DVLOG(1) << "Frame dropped before being painted: "
<< current_frame_->GetTimestamp().InSecondsF();
if (frames_dropped_before_paint_ < kuint32max)
frames_dropped_before_paint_++;
}
} // namespace content