// Copyright (c) 2011 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. //------------------------------------------------------------------------------ // Description of the life cycle of a instance of MetricsService. // // OVERVIEW // // A MetricsService instance is typically created at application startup. It // is the central controller for the acquisition of log data, and the automatic // transmission of that log data to an external server. Its major job is to // manage logs, grouping them for transmission, and transmitting them. As part // of its grouping, MS finalizes logs by including some just-in-time gathered // memory statistics, snapshotting the current stats of numerous histograms, // closing the logs, translating to XML text, and compressing the results for // transmission. Transmission includes submitting a compressed log as data in a // URL-post, and retransmitting (or retaining at process termination) if the // attempted transmission failed. Retention across process terminations is done // using the the PrefServices facilities. The retained logs (the ones that never // got transmitted) are compressed and base64-encoded before being persisted. // // Logs fall into one of two categories: "initial logs," and "ongoing logs." // There is at most one initial log sent for each complete run of the chromium // product (from startup, to browser shutdown). An initial log is generally // transmitted some short time (1 minute?) after startup, and includes stats // such as recent crash info, the number and types of plugins, etc. The // external server's response to the initial log conceptually tells this MS if // it should continue transmitting logs (during this session). The server // response can actually be much more detailed, and always includes (at a // minimum) how often additional ongoing logs should be sent. // // After the above initial log, a series of ongoing logs will be transmitted. // The first ongoing log actually begins to accumulate information stating when // the MS was first constructed. Note that even though the initial log is // commonly sent a full minute after startup, the initial log does not include // much in the way of user stats. The most common interlog period (delay) // is 20 minutes. That time period starts when the first user action causes a // logging event. This means that if there is no user action, there may be long // periods without any (ongoing) log transmissions. Ongoing logs typically // contain very detailed records of user activities (ex: opened tab, closed // tab, fetched URL, maximized window, etc.) In addition, just before an // ongoing log is closed out, a call is made to gather memory statistics. Those // memory statistics are deposited into a histogram, and the log finalization // code is then called. In the finalization, a call to a Histogram server // acquires a list of all local histograms that have been flagged for upload // to the UMA server. The finalization also acquires a the most recent number // of page loads, along with any counts of renderer or plugin crashes. // // When the browser shuts down, there will typically be a fragment of an ongoing // log that has not yet been transmitted. At shutdown time, that fragment // is closed (including snapshotting histograms), and converted to text. Note // that memory stats are not gathered during shutdown, as gathering *might* be // too time consuming. The textual representation of the fragment of the // ongoing log is then stored persistently as a string in the PrefServices, for // potential transmission during a future run of the product. // // There are two slightly abnormal shutdown conditions. There is a // "disconnected scenario," and a "really fast startup and shutdown" scenario. // In the "never connected" situation, the user has (during the running of the // process) never established an internet connection. As a result, attempts to // transmit the initial log have failed, and a lot(?) of data has accumulated in // the ongoing log (which didn't yet get closed, because there was never even a // contemplation of sending it). There is also a kindred "lost connection" // situation, where a loss of connection prevented an ongoing log from being // transmitted, and a (still open) log was stuck accumulating a lot(?) of data, // while the earlier log retried its transmission. In both of these // disconnected situations, two logs need to be, and are, persistently stored // for future transmission. // // The other unusual shutdown condition, termed "really fast startup and // shutdown," involves the deliberate user termination of the process before // the initial log is even formed or transmitted. In that situation, no logging // is done, but the historical crash statistics remain (unlogged) for inclusion // in a future run's initial log. (i.e., we don't lose crash stats). // // With the above overview, we can now describe the state machine's various // stats, based on the State enum specified in the state_ member. Those states // are: // // INITIALIZED, // Constructor was called. // INIT_TASK_SCHEDULED, // Waiting for deferred init tasks to complete. // INIT_TASK_DONE, // Waiting for timer to send initial log. // INITIAL_LOG_READY, // Initial log generated, and waiting for reply. // SEND_OLD_INITIAL_LOGS, // Sending unsent logs from previous session. // SENDING_OLD_LOGS, // Sending unsent logs from previous session. // SENDING_CURRENT_LOGS, // Sending standard current logs as they accrue. // // In more detail, we have: // // INITIALIZED, // Constructor was called. // The MS has been constructed, but has taken no actions to compose the // initial log. // // INIT_TASK_SCHEDULED, // Waiting for deferred init tasks to complete. // Typically about 30 seconds after startup, a task is sent to a second thread // (the file thread) to perform deferred (lower priority and slower) // initialization steps such as getting the list of plugins. That task will // (when complete) make an async callback (via a Task) to indicate the // completion. // // INIT_TASK_DONE, // Waiting for timer to send initial log. // The callback has arrived, and it is now possible for an initial log to be // created. This callback typically arrives back less than one second after // the deferred init task is dispatched. // // INITIAL_LOG_READY, // Initial log generated, and waiting for reply. // This state is entered only after an initial log has been composed, and // prepared for transmission. It is also the case that any previously unsent // logs have been loaded into instance variables for possible transmission. // // SEND_OLD_INITIAL_LOGS, // Sending unsent logs from previous session. // This state indicates that the initial log for this session has been // successfully sent and it is now time to send any "initial logs" that were // saved from previous sessions. Most commonly, there are none, but all old // logs that were "initial logs" must be sent before this state is exited. // // SENDING_OLD_LOGS, // Sending unsent logs from previous session. // This state indicates that there are no more unsent initial logs, and now any // ongoing logs from previous sessions should be transmitted. All such logs // will be transmitted before exiting this state, and proceeding with ongoing // logs from the current session (see next state). // // SENDING_CURRENT_LOGS, // Sending standard current logs as they accrue. // Current logs are being accumulated. Typically every 20 minutes a log is // closed and finalized for transmission, at the same time as a new log is // started. // // The progression through the above states is simple, and sequential, in the // most common use cases. States proceed from INITIAL to SENDING_CURRENT_LOGS, // and remain in the latter until shutdown. // // The one unusual case is when the user asks that we stop logging. When that // happens, any pending (transmission in progress) log is pushed into the list // of old unsent logs (the appropriate list, depending on whether it is an // initial log, or an ongoing log). An addition, any log that is currently // accumulating is also finalized, and pushed into the unsent log list. With // those pushes performed, we regress back to the SEND_OLD_INITIAL_LOGS state in // case the user enables log recording again during this session. This way // anything we have "pushed back" will be sent automatically if/when we progress // back to SENDING_CURRENT_LOG state. // // Also note that whenever the member variables containing unsent logs are // modified (i.e., when we send an old log), we mirror the list of logs into // the PrefServices. This ensures that IF we crash, we won't start up and // retransmit our old logs again. // // Due to race conditions, it is always possible that a log file could be sent // twice. For example, if a log file is sent, but not yet acknowledged by // the external server, and the user shuts down, then a copy of the log may be // saved for re-transmission. These duplicates could be filtered out server // side, but are not expected to be a significant problem. // // //------------------------------------------------------------------------------ #include "chrome/browser/metrics/metrics_service.h" #include "base/base64.h" #include "base/command_line.h" #include "base/md5.h" #include "base/metrics/histogram.h" #include "base/string_number_conversions.h" #include "base/threading/platform_thread.h" #include "base/threading/thread.h" #include "base/utf_string_conversions.h" #include "base/values.h" #include "chrome/browser/bookmarks/bookmark_model.h" #include "chrome/browser/browser_process.h" #include "chrome/browser/load_notification_details.h" #include "chrome/browser/memory_details.h" #include "chrome/browser/metrics/histogram_synchronizer.h" #include "chrome/browser/metrics/metrics_log.h" #include "chrome/browser/prefs/pref_service.h" #include "chrome/browser/prefs/scoped_user_pref_update.h" #include "chrome/browser/profiles/profile.h" #include "chrome/browser/search_engines/template_url_model.h" #include "chrome/browser/ui/browser_list.h" #include "chrome/common/child_process_logging.h" #include "chrome/common/chrome_switches.h" #include "chrome/common/guid.h" #include "chrome/common/pref_names.h" #include "chrome/common/render_messages.h" #include "content/browser/renderer_host/render_process_host.h" #include "content/common/child_process_info.h" #include "content/common/notification_service.h" #include "libxml/xmlwriter.h" #include "webkit/plugins/npapi/plugin_list.h" #include "webkit/plugins/npapi/webplugininfo.h" // TODO(port): port browser_distribution.h. #if !defined(OS_POSIX) #include "chrome/installer/util/browser_distribution.h" #endif #if defined(OS_CHROMEOS) #include "chrome/browser/chromeos/cros/cros_library.h" #include "chrome/browser/chromeos/external_metrics.h" #include "chrome/browser/chromeos/system_access.h" #endif namespace { MetricsService::LogRecallStatus MakeRecallStatusHistogram( MetricsService::LogRecallStatus status) { UMA_HISTOGRAM_ENUMERATION("PrefService.PersistentLogRecall", status, MetricsService::END_RECALL_STATUS); return status; } // TODO(ziadh): Remove this when done with experiment. void MakeStoreStatusHistogram(MetricsService::LogStoreStatus status) { UMA_HISTOGRAM_ENUMERATION("PrefService.PersistentLogStore2", status, MetricsService::END_STORE_STATUS); } } // namespace using base::Time; using base::TimeDelta; // Check to see that we're being called on only one thread. static bool IsSingleThreaded(); static const char kMetricsType[] = "application/vnd.mozilla.metrics.bz2"; // The delay, in seconds, after startup before sending the first log message. static const int kInitialInterlogDuration = 60; // one minute // This specifies the amount of time to wait for all renderers to send their // data. static const int kMaxHistogramGatheringWaitDuration = 60000; // 60 seconds. // The maximum number of events in a log uploaded to the UMA server. static const int kEventLimit = 2400; // If an upload fails, and the transmission was over this byte count, then we // will discard the log, and not try to retransmit it. We also don't persist // the log to the prefs for transmission during the next chrome session if this // limit is exceeded. static const int kUploadLogAvoidRetransmitSize = 50000; // When we have logs from previous Chrome sessions to send, how long should we // delay (in seconds) between each log transmission. static const int kUnsentLogDelay = 15; // 15 seconds // Minimum time a log typically exists before sending, in seconds. // This number is supplied by the server, but until we parse it out of a server // response, we use this duration to specify how long we should wait before // sending the next log. If the channel is busy, such as when there is a // failure during an attempt to transmit a previous log, then a log may wait // (and continue to accrue new log entries) for a much greater period of time. static const int kMinSecondsPerLog = 30 * 60; // Thirty minutes. // When we don't succeed at transmitting a log to a server, we progressively // wait longer and longer before sending the next log. This backoff process // help reduce load on the server, and makes the amount of backoff vary between // clients so that a collision (server overload?) on retransmit is less likely. // The following is the constant we use to expand that inter-log duration. static const double kBackoff = 1.1; // We limit the maximum backoff to be no greater than some multiple of the // default kMinSecondsPerLog. The following is that maximum ratio. static const int kMaxBackoff = 10; // Interval, in seconds, between state saves. static const int kSaveStateInterval = 5 * 60; // five minutes // The number of "initial" logs we're willing to save, and hope to send during // a future Chrome session. Initial logs contain crash stats, and are pretty // small. static const size_t kMaxInitialLogsPersisted = 20; // The number of ongoing logs we're willing to save persistently, and hope to // send during a this or future sessions. Note that each log may be pretty // large, as presumably the related "initial" log wasn't sent (probably nothing // was, as the user was probably off-line). As a result, the log probably kept // accumulating while the "initial" log was stalled (pending_), and couldn't be // sent. As a result, we don't want to save too many of these mega-logs. // A "standard shutdown" will create a small log, including just the data that // was not yet been transmitted, and that is normal (to have exactly one // ongoing_log_ at startup). static const size_t kMaxOngoingLogsPersisted = 8; // We append (2) more elements to persisted lists: the size of the list and a // checksum of the elements. static const size_t kChecksumEntryCount = 2; // This is used to quickly log stats from child process related notifications in // MetricsService::child_stats_buffer_. The buffer's contents are transferred // out when Local State is periodically saved. The information is then // reported to the UMA server on next launch. struct MetricsService::ChildProcessStats { public: explicit ChildProcessStats(ChildProcessInfo::ProcessType type) : process_launches(0), process_crashes(0), instances(0), process_type(type) {} // This constructor is only used by the map to return some default value for // an index for which no value has been assigned. ChildProcessStats() : process_launches(0), process_crashes(0), instances(0), process_type(ChildProcessInfo::UNKNOWN_PROCESS) {} // The number of times that the given child process has been launched int process_launches; // The number of times that the given child process has crashed int process_crashes; // The number of instances of this child process that have been created. // An instance is a DOM object rendered by this child process during a page // load. int instances; ChildProcessInfo::ProcessType process_type; }; // Handles asynchronous fetching of memory details. // Will run the provided task after finished. class MetricsMemoryDetails : public MemoryDetails { public: explicit MetricsMemoryDetails(Task* completion) : completion_(completion) {} virtual void OnDetailsAvailable() { MessageLoop::current()->PostTask(FROM_HERE, completion_); } private: ~MetricsMemoryDetails() {} Task* completion_; DISALLOW_COPY_AND_ASSIGN(MetricsMemoryDetails); }; class MetricsService::InitTaskComplete : public Task { public: explicit InitTaskComplete( const std::string& hardware_class, const std::vector<webkit::npapi::WebPluginInfo>& plugins) : hardware_class_(hardware_class), plugins_(plugins) {} virtual void Run() { g_browser_process->metrics_service()->OnInitTaskComplete( hardware_class_, plugins_); } private: std::string hardware_class_; std::vector<webkit::npapi::WebPluginInfo> plugins_; }; class MetricsService::InitTask : public Task { public: explicit InitTask(MessageLoop* callback_loop) : callback_loop_(callback_loop) {} virtual void Run() { std::vector<webkit::npapi::WebPluginInfo> plugins; webkit::npapi::PluginList::Singleton()->GetPlugins(false, &plugins); std::string hardware_class; // Empty string by default. #if defined(OS_CHROMEOS) chromeos::SystemAccess::GetInstance()->GetMachineStatistic( "hardware_class", &hardware_class); #endif // OS_CHROMEOS callback_loop_->PostTask(FROM_HERE, new InitTaskComplete( hardware_class, plugins)); } private: MessageLoop* callback_loop_; }; // static void MetricsService::RegisterPrefs(PrefService* local_state) { DCHECK(IsSingleThreaded()); local_state->RegisterStringPref(prefs::kMetricsClientID, ""); local_state->RegisterInt64Pref(prefs::kMetricsClientIDTimestamp, 0); local_state->RegisterInt64Pref(prefs::kStabilityLaunchTimeSec, 0); local_state->RegisterInt64Pref(prefs::kStabilityLastTimestampSec, 0); local_state->RegisterStringPref(prefs::kStabilityStatsVersion, ""); local_state->RegisterInt64Pref(prefs::kStabilityStatsBuildTime, 0); local_state->RegisterBooleanPref(prefs::kStabilityExitedCleanly, true); local_state->RegisterBooleanPref(prefs::kStabilitySessionEndCompleted, true); local_state->RegisterIntegerPref(prefs::kMetricsSessionID, -1); local_state->RegisterIntegerPref(prefs::kStabilityLaunchCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityCrashCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityIncompleteSessionEndCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityPageLoadCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityRendererCrashCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityExtensionRendererCrashCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityRendererHangCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityChildProcessCrashCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityBreakpadRegistrationFail, 0); local_state->RegisterIntegerPref(prefs::kStabilityBreakpadRegistrationSuccess, 0); local_state->RegisterIntegerPref(prefs::kStabilityDebuggerPresent, 0); local_state->RegisterIntegerPref(prefs::kStabilityDebuggerNotPresent, 0); #if defined(OS_CHROMEOS) local_state->RegisterIntegerPref(prefs::kStabilityOtherUserCrashCount, 0); local_state->RegisterIntegerPref(prefs::kStabilityKernelCrashCount, 0); local_state->RegisterIntegerPref(prefs::kStabilitySystemUncleanShutdownCount, 0); #endif // OS_CHROMEOS local_state->RegisterDictionaryPref(prefs::kProfileMetrics); local_state->RegisterIntegerPref(prefs::kNumBookmarksOnBookmarkBar, 0); local_state->RegisterIntegerPref(prefs::kNumFoldersOnBookmarkBar, 0); local_state->RegisterIntegerPref(prefs::kNumBookmarksInOtherBookmarkFolder, 0); local_state->RegisterIntegerPref(prefs::kNumFoldersInOtherBookmarkFolder, 0); local_state->RegisterIntegerPref(prefs::kNumKeywords, 0); local_state->RegisterListPref(prefs::kMetricsInitialLogs); local_state->RegisterListPref(prefs::kMetricsOngoingLogs); local_state->RegisterInt64Pref(prefs::kUninstallMetricsPageLoadCount, 0); local_state->RegisterInt64Pref(prefs::kUninstallLaunchCount, 0); local_state->RegisterInt64Pref(prefs::kUninstallMetricsInstallDate, 0); local_state->RegisterInt64Pref(prefs::kUninstallMetricsUptimeSec, 0); local_state->RegisterInt64Pref(prefs::kUninstallLastLaunchTimeSec, 0); local_state->RegisterInt64Pref(prefs::kUninstallLastObservedRunTimeSec, 0); } // static void MetricsService::DiscardOldStabilityStats(PrefService* local_state) { local_state->SetBoolean(prefs::kStabilityExitedCleanly, true); local_state->SetBoolean(prefs::kStabilitySessionEndCompleted, true); local_state->SetInteger(prefs::kStabilityIncompleteSessionEndCount, 0); local_state->SetInteger(prefs::kStabilityBreakpadRegistrationSuccess, 0); local_state->SetInteger(prefs::kStabilityBreakpadRegistrationFail, 0); local_state->SetInteger(prefs::kStabilityDebuggerPresent, 0); local_state->SetInteger(prefs::kStabilityDebuggerNotPresent, 0); local_state->SetInteger(prefs::kStabilityLaunchCount, 0); local_state->SetInteger(prefs::kStabilityCrashCount, 0); local_state->SetInteger(prefs::kStabilityPageLoadCount, 0); local_state->SetInteger(prefs::kStabilityRendererCrashCount, 0); local_state->SetInteger(prefs::kStabilityRendererHangCount, 0); local_state->SetInt64(prefs::kStabilityLaunchTimeSec, 0); local_state->SetInt64(prefs::kStabilityLastTimestampSec, 0); local_state->ClearPref(prefs::kStabilityPluginStats); local_state->ClearPref(prefs::kMetricsInitialLogs); local_state->ClearPref(prefs::kMetricsOngoingLogs); } MetricsService::MetricsService() : recording_active_(false), reporting_active_(false), server_permits_upload_(true), state_(INITIALIZED), current_fetch_(NULL), idle_since_last_transmission_(false), next_window_id_(0), ALLOW_THIS_IN_INITIALIZER_LIST(log_sender_factory_(this)), ALLOW_THIS_IN_INITIALIZER_LIST(state_saver_factory_(this)), interlog_duration_(TimeDelta::FromSeconds(kInitialInterlogDuration)), timer_pending_(false) { DCHECK(IsSingleThreaded()); InitializeMetricsState(); } MetricsService::~MetricsService() { SetRecording(false); } void MetricsService::Start() { HandleIdleSinceLastTransmission(false); SetRecording(true); SetReporting(true); } void MetricsService::StartRecordingOnly() { SetRecording(true); SetReporting(false); } void MetricsService::Stop() { HandleIdleSinceLastTransmission(false); SetReporting(false); SetRecording(false); } void MetricsService::SetRecording(bool enabled) { DCHECK(IsSingleThreaded()); if (enabled == recording_active_) return; if (enabled) { if (client_id_.empty()) { PrefService* pref = g_browser_process->local_state(); DCHECK(pref); client_id_ = pref->GetString(prefs::kMetricsClientID); if (client_id_.empty()) { client_id_ = GenerateClientID(); pref->SetString(prefs::kMetricsClientID, client_id_); // Might as well make a note of how long this ID has existed pref->SetString(prefs::kMetricsClientIDTimestamp, base::Int64ToString(Time::Now().ToTimeT())); } } child_process_logging::SetClientId(client_id_); StartRecording(); SetUpNotifications(®istrar_, this); } else { registrar_.RemoveAll(); PushPendingLogsToUnsentLists(); DCHECK(!pending_log()); if (state_ > INITIAL_LOG_READY && unsent_logs()) state_ = SEND_OLD_INITIAL_LOGS; } recording_active_ = enabled; } bool MetricsService::recording_active() const { DCHECK(IsSingleThreaded()); return recording_active_; } void MetricsService::SetReporting(bool enable) { if (reporting_active_ != enable) { reporting_active_ = enable; if (reporting_active_) StartLogTransmissionTimer(); } } bool MetricsService::reporting_active() const { DCHECK(IsSingleThreaded()); return reporting_active_; } // static void MetricsService::SetUpNotifications(NotificationRegistrar* registrar, NotificationObserver* observer) { registrar->Add(observer, NotificationType::BROWSER_OPENED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::BROWSER_CLOSED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::USER_ACTION, NotificationService::AllSources()); registrar->Add(observer, NotificationType::TAB_PARENTED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::TAB_CLOSING, NotificationService::AllSources()); registrar->Add(observer, NotificationType::LOAD_START, NotificationService::AllSources()); registrar->Add(observer, NotificationType::LOAD_STOP, NotificationService::AllSources()); registrar->Add(observer, NotificationType::RENDERER_PROCESS_CLOSED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::RENDERER_PROCESS_HANG, NotificationService::AllSources()); registrar->Add(observer, NotificationType::CHILD_PROCESS_HOST_CONNECTED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::CHILD_INSTANCE_CREATED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::CHILD_PROCESS_CRASHED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::TEMPLATE_URL_MODEL_LOADED, NotificationService::AllSources()); registrar->Add(observer, NotificationType::OMNIBOX_OPENED_URL, NotificationService::AllSources()); registrar->Add(observer, NotificationType::BOOKMARK_MODEL_LOADED, NotificationService::AllSources()); } void MetricsService::Observe(NotificationType type, const NotificationSource& source, const NotificationDetails& details) { DCHECK(current_log_); DCHECK(IsSingleThreaded()); if (!CanLogNotification(type, source, details)) return; switch (type.value) { case NotificationType::USER_ACTION: current_log_->RecordUserAction(*Details<const char*>(details).ptr()); break; case NotificationType::BROWSER_OPENED: case NotificationType::BROWSER_CLOSED: LogWindowChange(type, source, details); break; case NotificationType::TAB_PARENTED: case NotificationType::TAB_CLOSING: LogWindowChange(type, source, details); break; case NotificationType::LOAD_STOP: LogLoadComplete(type, source, details); break; case NotificationType::LOAD_START: LogLoadStarted(); break; case NotificationType::RENDERER_PROCESS_CLOSED: { RenderProcessHost::RendererClosedDetails* process_details = Details<RenderProcessHost::RendererClosedDetails>(details).ptr(); if (process_details->status == base::TERMINATION_STATUS_PROCESS_CRASHED || process_details->status == base::TERMINATION_STATUS_ABNORMAL_TERMINATION) { if (process_details->was_extension_renderer) { LogExtensionRendererCrash(); } else { LogRendererCrash(); } } } break; case NotificationType::RENDERER_PROCESS_HANG: LogRendererHang(); break; case NotificationType::CHILD_PROCESS_HOST_CONNECTED: case NotificationType::CHILD_PROCESS_CRASHED: case NotificationType::CHILD_INSTANCE_CREATED: LogChildProcessChange(type, source, details); break; case NotificationType::TEMPLATE_URL_MODEL_LOADED: LogKeywords(Source<TemplateURLModel>(source).ptr()); break; case NotificationType::OMNIBOX_OPENED_URL: { MetricsLog* current_log = current_log_->AsMetricsLog(); DCHECK(current_log); current_log->RecordOmniboxOpenedURL( *Details<AutocompleteLog>(details).ptr()); break; } case NotificationType::BOOKMARK_MODEL_LOADED: { Profile* p = Source<Profile>(source).ptr(); if (p) LogBookmarks(p->GetBookmarkModel()); break; } default: NOTREACHED(); break; } HandleIdleSinceLastTransmission(false); if (current_log_) DVLOG(1) << "METRICS: NUMBER OF EVENTS = " << current_log_->num_events(); } void MetricsService::HandleIdleSinceLastTransmission(bool in_idle) { // If there wasn't a lot of action, maybe the computer was asleep, in which // case, the log transmissions should have stopped. Here we start them up // again. if (!in_idle && idle_since_last_transmission_) StartLogTransmissionTimer(); idle_since_last_transmission_ = in_idle; } void MetricsService::RecordStartOfSessionEnd() { LogCleanShutdown(); RecordBooleanPrefValue(prefs::kStabilitySessionEndCompleted, false); } void MetricsService::RecordCompletedSessionEnd() { LogCleanShutdown(); RecordBooleanPrefValue(prefs::kStabilitySessionEndCompleted, true); } void MetricsService:: RecordBreakpadRegistration(bool success) { if (!success) IncrementPrefValue(prefs::kStabilityBreakpadRegistrationFail); else IncrementPrefValue(prefs::kStabilityBreakpadRegistrationSuccess); } void MetricsService::RecordBreakpadHasDebugger(bool has_debugger) { if (!has_debugger) IncrementPrefValue(prefs::kStabilityDebuggerNotPresent); else IncrementPrefValue(prefs::kStabilityDebuggerPresent); } //------------------------------------------------------------------------------ // private methods //------------------------------------------------------------------------------ //------------------------------------------------------------------------------ // Initialization methods void MetricsService::InitializeMetricsState() { #if defined(OS_POSIX) server_url_ = L"https://clients4.google.com/firefox/metrics/collect"; #else BrowserDistribution* dist = BrowserDistribution::GetDistribution(); server_url_ = dist->GetStatsServerURL(); #endif PrefService* pref = g_browser_process->local_state(); DCHECK(pref); if ((pref->GetInt64(prefs::kStabilityStatsBuildTime) != MetricsLog::GetBuildTime()) || (pref->GetString(prefs::kStabilityStatsVersion) != MetricsLog::GetVersionString())) { // This is a new version, so we don't want to confuse the stats about the // old version with info that we upload. DiscardOldStabilityStats(pref); pref->SetString(prefs::kStabilityStatsVersion, MetricsLog::GetVersionString()); pref->SetInt64(prefs::kStabilityStatsBuildTime, MetricsLog::GetBuildTime()); } // Update session ID session_id_ = pref->GetInteger(prefs::kMetricsSessionID); ++session_id_; pref->SetInteger(prefs::kMetricsSessionID, session_id_); // Stability bookkeeping IncrementPrefValue(prefs::kStabilityLaunchCount); if (!pref->GetBoolean(prefs::kStabilityExitedCleanly)) { IncrementPrefValue(prefs::kStabilityCrashCount); } // This will be set to 'true' if we exit cleanly. pref->SetBoolean(prefs::kStabilityExitedCleanly, false); if (!pref->GetBoolean(prefs::kStabilitySessionEndCompleted)) { IncrementPrefValue(prefs::kStabilityIncompleteSessionEndCount); // This is marked false when we get a WM_ENDSESSION. pref->SetBoolean(prefs::kStabilitySessionEndCompleted, true); } // Initialize uptime counters. int64 startup_uptime = MetricsLog::GetIncrementalUptime(pref); DCHECK_EQ(0, startup_uptime); // For backwards compatibility, leave this intact in case Omaha is checking // them. prefs::kStabilityLastTimestampSec may also be useless now. // TODO(jar): Delete these if they have no uses. pref->SetInt64(prefs::kStabilityLaunchTimeSec, Time::Now().ToTimeT()); // Bookkeeping for the uninstall metrics. IncrementLongPrefsValue(prefs::kUninstallLaunchCount); // Save profile metrics. PrefService* prefs = g_browser_process->local_state(); if (prefs) { // Remove the current dictionary and store it for use when sending data to // server. By removing the value we prune potentially dead profiles // (and keys). All valid values are added back once services startup. const DictionaryValue* profile_dictionary = prefs->GetDictionary(prefs::kProfileMetrics); if (profile_dictionary) { // Do a deep copy of profile_dictionary since ClearPref will delete it. profile_dictionary_.reset(static_cast<DictionaryValue*>( profile_dictionary->DeepCopy())); prefs->ClearPref(prefs::kProfileMetrics); } } // Get stats on use of command line. const CommandLine* command_line(CommandLine::ForCurrentProcess()); size_t common_commands = 0; if (command_line->HasSwitch(switches::kUserDataDir)) { ++common_commands; UMA_HISTOGRAM_COUNTS_100("Chrome.CommandLineDatDirCount", 1); } if (command_line->HasSwitch(switches::kApp)) { ++common_commands; UMA_HISTOGRAM_COUNTS_100("Chrome.CommandLineAppModeCount", 1); } UMA_HISTOGRAM_COUNTS_100("Chrome.CommandLineFlagCount", command_line->GetSwitchCount()); UMA_HISTOGRAM_COUNTS_100("Chrome.CommandLineUncommonFlagCount", command_line->GetSwitchCount() - common_commands); // Kick off the process of saving the state (so the uptime numbers keep // getting updated) every n minutes. ScheduleNextStateSave(); } void MetricsService::OnInitTaskComplete( const std::string& hardware_class, const std::vector<webkit::npapi::WebPluginInfo>& plugins) { DCHECK(state_ == INIT_TASK_SCHEDULED); hardware_class_ = hardware_class; plugins_ = plugins; if (state_ == INIT_TASK_SCHEDULED) state_ = INIT_TASK_DONE; } std::string MetricsService::GenerateClientID() { return guid::GenerateGUID(); } //------------------------------------------------------------------------------ // State save methods void MetricsService::ScheduleNextStateSave() { state_saver_factory_.RevokeAll(); MessageLoop::current()->PostDelayedTask(FROM_HERE, state_saver_factory_.NewRunnableMethod(&MetricsService::SaveLocalState), kSaveStateInterval * 1000); } void MetricsService::SaveLocalState() { PrefService* pref = g_browser_process->local_state(); if (!pref) { NOTREACHED(); return; } RecordCurrentState(pref); pref->ScheduleSavePersistentPrefs(); // TODO(jar): Does this run down the batteries???? ScheduleNextStateSave(); } //------------------------------------------------------------------------------ // Recording control methods void MetricsService::StartRecording() { if (current_log_) return; current_log_ = new MetricsLog(client_id_, session_id_); if (state_ == INITIALIZED) { // We only need to schedule that run once. state_ = INIT_TASK_SCHEDULED; // Schedules a task on the file thread for execution of slower // initialization steps (such as plugin list generation) necessary // for sending the initial log. This avoids blocking the main UI // thread. g_browser_process->file_thread()->message_loop()->PostDelayedTask(FROM_HERE, new InitTask(MessageLoop::current()), kInitialInterlogDuration * 1000 / 2); } } void MetricsService::StopRecording(MetricsLogBase** log) { if (!current_log_) return; current_log_->set_hardware_class(hardware_class_); // Adds to ongoing logs. // TODO(jar): Integrate bounds on log recording more consistently, so that we // can stop recording logs that are too big much sooner. if (current_log_->num_events() > kEventLimit) { UMA_HISTOGRAM_COUNTS("UMA.Discarded Log Events", current_log_->num_events()); current_log_->CloseLog(); delete current_log_; current_log_ = NULL; StartRecording(); // Start trivial log to hold our histograms. } // Put incremental data (histogram deltas, and realtime stats deltas) at the // end of all log transmissions (initial log handles this separately). // Don't bother if we're going to discard current_log_. if (log) { // RecordIncrementalStabilityElements only exists on the derived // MetricsLog class. MetricsLog* current_log = current_log_->AsMetricsLog(); DCHECK(current_log); current_log->RecordIncrementalStabilityElements(); RecordCurrentHistograms(); } current_log_->CloseLog(); if (log) *log = current_log_; else delete current_log_; current_log_ = NULL; } void MetricsService::PushPendingLogsToUnsentLists() { if (state_ < INITIAL_LOG_READY) return; // We didn't and still don't have time to get plugin list etc. if (pending_log()) { PreparePendingLogText(); if (state_ == INITIAL_LOG_READY) { // We may race here, and send second copy of initial log later. unsent_initial_logs_.push_back(compressed_log_); state_ = SEND_OLD_INITIAL_LOGS; } else { // TODO(jar): Verify correctness in other states, including sending unsent // initial logs. PushPendingLogTextToUnsentOngoingLogs(); } DiscardPendingLog(); } DCHECK(!pending_log()); StopRecording(&pending_log_); PreparePendingLogText(); PushPendingLogTextToUnsentOngoingLogs(); DiscardPendingLog(); StoreUnsentLogs(); } void MetricsService::PushPendingLogTextToUnsentOngoingLogs() { // If UMA response told us not to upload, there's no need to save the pending // log. It wasn't supposed to be uploaded anyway. if (!server_permits_upload_) return; if (compressed_log_.length() > static_cast<size_t>(kUploadLogAvoidRetransmitSize)) { UMA_HISTOGRAM_COUNTS("UMA.Large Accumulated Log Not Persisted", static_cast<int>(compressed_log_.length())); return; } unsent_ongoing_logs_.push_back(compressed_log_); } //------------------------------------------------------------------------------ // Transmission of logs methods void MetricsService::StartLogTransmissionTimer() { // If we're not reporting, there's no point in starting a log transmission // timer. if (!reporting_active()) return; if (!current_log_) return; // Recorder is shutdown. // If there is already a timer running, we leave it running. // If timer_pending is true because the fetch is waiting for a response, // we return for now and let the response handler start the timer. if (timer_pending_) return; // Before starting the timer, set timer_pending_ to true. timer_pending_ = true; // Right before the UMA transmission gets started, there's one more thing we'd // like to record: the histogram of memory usage, so we spawn a task to // collect the memory details and when that task is finished, it will call // OnMemoryDetailCollectionDone, which will call HistogramSynchronization to // collect histograms from all renderers and then we will call // OnHistogramSynchronizationDone to continue processing. MessageLoop::current()->PostDelayedTask(FROM_HERE, log_sender_factory_. NewRunnableMethod(&MetricsService::LogTransmissionTimerDone), interlog_duration_.InMilliseconds()); } void MetricsService::LogTransmissionTimerDone() { Task* task = log_sender_factory_. NewRunnableMethod(&MetricsService::OnMemoryDetailCollectionDone); scoped_refptr<MetricsMemoryDetails> details(new MetricsMemoryDetails(task)); details->StartFetch(); // Collect WebCore cache information to put into a histogram. for (RenderProcessHost::iterator i(RenderProcessHost::AllHostsIterator()); !i.IsAtEnd(); i.Advance()) i.GetCurrentValue()->Send(new ViewMsg_GetCacheResourceStats()); } void MetricsService::OnMemoryDetailCollectionDone() { DCHECK(IsSingleThreaded()); // HistogramSynchronizer will Collect histograms from all renderers and it // will call OnHistogramSynchronizationDone (if wait time elapses before it // heard from all renderers, then also it will call // OnHistogramSynchronizationDone). // Create a callback_task for OnHistogramSynchronizationDone. Task* callback_task = log_sender_factory_.NewRunnableMethod( &MetricsService::OnHistogramSynchronizationDone); // Set up the callback to task to call after we receive histograms from all // renderer processes. Wait time specifies how long to wait before absolutely // calling us back on the task. HistogramSynchronizer::FetchRendererHistogramsAsynchronously( MessageLoop::current(), callback_task, kMaxHistogramGatheringWaitDuration); } void MetricsService::OnHistogramSynchronizationDone() { DCHECK(IsSingleThreaded()); // This function should only be called via timer, so timer_pending_ // should be true. DCHECK(timer_pending_); timer_pending_ = false; DCHECK(!current_fetch_.get()); // If we're getting no notifications, then the log won't have much in it, and // it's possible the computer is about to go to sleep, so don't upload and // don't restart the transmission timer. if (idle_since_last_transmission_) return; // If somehow there is a fetch in progress, we return setting timer_pending_ // to true and hope things work out. if (current_fetch_.get()) { timer_pending_ = true; return; } // If uploads are forbidden by UMA response, there's no point in keeping // the current_log_, and the more often we delete it, the less likely it is // to expand forever. if (!server_permits_upload_ && current_log_) { StopRecording(NULL); StartRecording(); } if (!current_log_) return; // Logging was disabled. if (!reporting_active()) return; // Don't do work if we're not going to send anything now. MakePendingLog(); // MakePendingLog should have put something in the pending log, if it didn't, // we start the timer again, return and hope things work out. if (!pending_log()) { StartLogTransmissionTimer(); return; } // If we're not supposed to upload any UMA data because the response said so, // cancel the upload at this point, but start the timer. if (!ServerPermitsTransmission()) { DiscardPendingLog(); StartLogTransmissionTimer(); return; } PrepareFetchWithPendingLog(); if (!current_fetch_.get()) { // Compression failed, and log discarded :-/. DiscardPendingLog(); StartLogTransmissionTimer(); // Maybe we'll do better next time // TODO(jar): If compression failed, we should have created a tiny log and // compressed that, so that we can signal that we're losing logs. return; } DCHECK(!timer_pending_); // The URL fetch is a like timer in that after a while we get called back // so we set timer_pending_ true just as we start the url fetch. timer_pending_ = true; current_fetch_->Start(); HandleIdleSinceLastTransmission(true); } void MetricsService::MakePendingLog() { if (pending_log()) return; switch (state_) { case INITIALIZED: case INIT_TASK_SCHEDULED: // We should be further along by now. DCHECK(false); return; case INIT_TASK_DONE: // We need to wait for the initial log to be ready before sending // anything, because the server will tell us whether it wants to hear // from us. PrepareInitialLog(); DCHECK(state_ == INIT_TASK_DONE); RecallUnsentLogs(); state_ = INITIAL_LOG_READY; break; case SEND_OLD_INITIAL_LOGS: if (!unsent_initial_logs_.empty()) { compressed_log_ = unsent_initial_logs_.back(); break; } state_ = SENDING_OLD_LOGS; // Fall through. case SENDING_OLD_LOGS: if (!unsent_ongoing_logs_.empty()) { compressed_log_ = unsent_ongoing_logs_.back(); break; } state_ = SENDING_CURRENT_LOGS; // Fall through. case SENDING_CURRENT_LOGS: StopRecording(&pending_log_); StartRecording(); break; default: NOTREACHED(); return; } DCHECK(pending_log()); } bool MetricsService::ServerPermitsTransmission() const { // If the server forbids uploading, we take that to mean it forbids current // logs, but we still send up the inital logs and any old logs. if (server_permits_upload_) return true; switch (state_) { case INITIAL_LOG_READY: case SEND_OLD_INITIAL_LOGS: case SENDING_OLD_LOGS: return true; case SENDING_CURRENT_LOGS: default: return false; } } void MetricsService::PrepareInitialLog() { DCHECK(state_ == INIT_TASK_DONE); MetricsLog* log = new MetricsLog(client_id_, session_id_); log->set_hardware_class(hardware_class_); // Adds to initial log. log->RecordEnvironment(plugins_, profile_dictionary_.get()); // Histograms only get written to current_log_, so setup for the write. MetricsLogBase* save_log = current_log_; current_log_ = log; RecordCurrentHistograms(); // Into current_log_... which is really log. current_log_ = save_log; log->CloseLog(); DCHECK(!pending_log()); pending_log_ = log; } // static MetricsService::LogRecallStatus MetricsService::RecallUnsentLogsHelper( const ListValue& list, std::vector<std::string>* local_list) { DCHECK(local_list->empty()); if (list.GetSize() == 0) return MakeRecallStatusHistogram(LIST_EMPTY); if (list.GetSize() < 3) return MakeRecallStatusHistogram(LIST_SIZE_TOO_SMALL); // The size is stored at the beginning of the list. int size; bool valid = (*list.begin())->GetAsInteger(&size); if (!valid) return MakeRecallStatusHistogram(LIST_SIZE_MISSING); // Account for checksum and size included in the list. if (static_cast<unsigned int>(size) != list.GetSize() - kChecksumEntryCount) return MakeRecallStatusHistogram(LIST_SIZE_CORRUPTION); MD5Context ctx; MD5Init(&ctx); std::string encoded_log; std::string decoded_log; for (ListValue::const_iterator it = list.begin() + 1; it != list.end() - 1; ++it) { // Last element is the checksum. valid = (*it)->GetAsString(&encoded_log); if (!valid) { local_list->clear(); return MakeRecallStatusHistogram(LOG_STRING_CORRUPTION); } MD5Update(&ctx, encoded_log.data(), encoded_log.length()); if (!base::Base64Decode(encoded_log, &decoded_log)) { local_list->clear(); return MakeRecallStatusHistogram(DECODE_FAIL); } local_list->push_back(decoded_log); } // Verify checksum. MD5Digest digest; MD5Final(&digest, &ctx); std::string recovered_md5; // We store the hash at the end of the list. valid = (*(list.end() - 1))->GetAsString(&recovered_md5); if (!valid) { local_list->clear(); return MakeRecallStatusHistogram(CHECKSUM_STRING_CORRUPTION); } if (recovered_md5 != MD5DigestToBase16(digest)) { local_list->clear(); return MakeRecallStatusHistogram(CHECKSUM_CORRUPTION); } return MakeRecallStatusHistogram(RECALL_SUCCESS); } void MetricsService::RecallUnsentLogs() { PrefService* local_state = g_browser_process->local_state(); DCHECK(local_state); const ListValue* unsent_initial_logs = local_state->GetList( prefs::kMetricsInitialLogs); RecallUnsentLogsHelper(*unsent_initial_logs, &unsent_initial_logs_); const ListValue* unsent_ongoing_logs = local_state->GetList( prefs::kMetricsOngoingLogs); RecallUnsentLogsHelper(*unsent_ongoing_logs, &unsent_ongoing_logs_); } // static void MetricsService::StoreUnsentLogsHelper( const std::vector<std::string>& local_list, const size_t kMaxLocalListSize, ListValue* list) { list->Clear(); size_t start = 0; if (local_list.size() > kMaxLocalListSize) start = local_list.size() - kMaxLocalListSize; DCHECK(start <= local_list.size()); if (local_list.size() == start) return; // Store size at the beginning of the list. list->Append(Value::CreateIntegerValue(local_list.size() - start)); MD5Context ctx; MD5Init(&ctx); std::string encoded_log; for (std::vector<std::string>::const_iterator it = local_list.begin() + start; it != local_list.end(); ++it) { // We encode the compressed log as Value::CreateStringValue() expects to // take a valid UTF8 string. if (!base::Base64Encode(*it, &encoded_log)) { MakeStoreStatusHistogram(ENCODE_FAIL); list->Clear(); return; } MD5Update(&ctx, encoded_log.data(), encoded_log.length()); list->Append(Value::CreateStringValue(encoded_log)); } // Append hash to the end of the list. MD5Digest digest; MD5Final(&digest, &ctx); list->Append(Value::CreateStringValue(MD5DigestToBase16(digest))); DCHECK(list->GetSize() >= 3); // Minimum of 3 elements (size, data, hash). MakeStoreStatusHistogram(STORE_SUCCESS); } void MetricsService::StoreUnsentLogs() { if (state_ < INITIAL_LOG_READY) return; // We never Recalled the prior unsent logs. PrefService* local_state = g_browser_process->local_state(); DCHECK(local_state); { ListPrefUpdate update(local_state, prefs::kMetricsInitialLogs); ListValue* unsent_initial_logs = update.Get(); StoreUnsentLogsHelper(unsent_initial_logs_, kMaxInitialLogsPersisted, unsent_initial_logs); } { ListPrefUpdate update(local_state, prefs::kMetricsOngoingLogs); ListValue* unsent_ongoing_logs = update.Get(); StoreUnsentLogsHelper(unsent_ongoing_logs_, kMaxOngoingLogsPersisted, unsent_ongoing_logs); } } void MetricsService::PreparePendingLogText() { DCHECK(pending_log()); if (!compressed_log_.empty()) return; int text_size = pending_log_->GetEncodedLogSize(); std::string pending_log_text; // Leave room for the NULL terminator. pending_log_->GetEncodedLog(WriteInto(&pending_log_text, text_size + 1), text_size); if (Bzip2Compress(pending_log_text, &compressed_log_)) { // Allow security conscious users to see all metrics logs that we send. VLOG(1) << "COMPRESSED FOLLOWING METRICS LOG: " << pending_log_text; } else { LOG(DFATAL) << "Failed to compress log for transmission."; // We can't discard the logs as other caller functions expect that // |compressed_log_| not be empty. We can detect this failure at the server // after we transmit. compressed_log_ = "Unable to compress!"; MakeStoreStatusHistogram(COMPRESS_FAIL); return; } } void MetricsService::PrepareFetchWithPendingLog() { DCHECK(pending_log()); DCHECK(!current_fetch_.get()); PreparePendingLogText(); DCHECK(!compressed_log_.empty()); current_fetch_.reset(new URLFetcher(GURL(WideToUTF16(server_url_)), URLFetcher::POST, this)); current_fetch_->set_request_context(Profile::GetDefaultRequestContext()); current_fetch_->set_upload_data(kMetricsType, compressed_log_); } static const char* StatusToString(const net::URLRequestStatus& status) { switch (status.status()) { case net::URLRequestStatus::SUCCESS: return "SUCCESS"; case net::URLRequestStatus::IO_PENDING: return "IO_PENDING"; case net::URLRequestStatus::HANDLED_EXTERNALLY: return "HANDLED_EXTERNALLY"; case net::URLRequestStatus::CANCELED: return "CANCELED"; case net::URLRequestStatus::FAILED: return "FAILED"; default: NOTREACHED(); return "Unknown"; } } void MetricsService::OnURLFetchComplete(const URLFetcher* source, const GURL& url, const net::URLRequestStatus& status, int response_code, const ResponseCookies& cookies, const std::string& data) { DCHECK(timer_pending_); timer_pending_ = false; DCHECK(current_fetch_.get()); current_fetch_.reset(NULL); // We're not allowed to re-use it. // Confirm send so that we can move on. VLOG(1) << "METRICS RESPONSE CODE: " << response_code << " status=" << StatusToString(status); // Provide boolean for error recovery (allow us to ignore response_code). bool discard_log = false; if (response_code != 200 && (compressed_log_.length() > static_cast<size_t>(kUploadLogAvoidRetransmitSize))) { UMA_HISTOGRAM_COUNTS("UMA.Large Rejected Log was Discarded", static_cast<int>(compressed_log_.length())); discard_log = true; } else if (response_code == 400) { // Bad syntax. Retransmission won't work. UMA_HISTOGRAM_COUNTS("UMA.Unacceptable_Log_Discarded", state_); discard_log = true; } if (response_code != 200 && !discard_log) { VLOG(1) << "METRICS: transmission attempt returned a failure code: " << response_code << ". Verify network connectivity"; HandleBadResponseCode(); } else { // Successful receipt (or we are discarding log). VLOG(1) << "METRICS RESPONSE DATA: " << data; switch (state_) { case INITIAL_LOG_READY: state_ = SEND_OLD_INITIAL_LOGS; break; case SEND_OLD_INITIAL_LOGS: DCHECK(!unsent_initial_logs_.empty()); unsent_initial_logs_.pop_back(); StoreUnsentLogs(); break; case SENDING_OLD_LOGS: DCHECK(!unsent_ongoing_logs_.empty()); unsent_ongoing_logs_.pop_back(); StoreUnsentLogs(); break; case SENDING_CURRENT_LOGS: break; default: NOTREACHED(); break; } DiscardPendingLog(); // Since we sent a log, make sure our in-memory state is recorded to disk. PrefService* local_state = g_browser_process->local_state(); DCHECK(local_state); if (local_state) local_state->ScheduleSavePersistentPrefs(); // Override usual interlog delay if there are unsent logs to transmit, // otherwise reset back to default. if (unsent_logs()) { DCHECK(state_ < SENDING_CURRENT_LOGS); interlog_duration_ = TimeDelta::FromSeconds(kUnsentLogDelay); } else { interlog_duration_ = TimeDelta::FromSeconds(kMinSecondsPerLog); } } StartLogTransmissionTimer(); } void MetricsService::HandleBadResponseCode() { VLOG(1) << "Verify your metrics logs are formatted correctly. Verify server " "is active at " << server_url_; if (!pending_log()) { VLOG(1) << "METRICS: Recorder shutdown during log transmission."; } else { // Send progressively less frequently. DCHECK(kBackoff > 1.0); interlog_duration_ = TimeDelta::FromMicroseconds( static_cast<int64>(kBackoff * interlog_duration_.InMicroseconds())); if (kMaxBackoff * TimeDelta::FromSeconds(kMinSecondsPerLog) < interlog_duration_) { interlog_duration_ = kMaxBackoff * TimeDelta::FromSeconds(kMinSecondsPerLog); } VLOG(1) << "METRICS: transmission retry being scheduled in " << interlog_duration_.InSeconds() << " seconds for " << compressed_log_; } } void MetricsService::LogWindowChange(NotificationType type, const NotificationSource& source, const NotificationDetails& details) { int controller_id = -1; uintptr_t window_or_tab = source.map_key(); MetricsLog::WindowEventType window_type; // Note: since we stop all logging when a single OTR session is active, it is // possible that we start getting notifications about a window that we don't // know about. if (window_map_.find(window_or_tab) == window_map_.end()) { controller_id = next_window_id_++; window_map_[window_or_tab] = controller_id; } else { controller_id = window_map_[window_or_tab]; } DCHECK_NE(controller_id, -1); switch (type.value) { case NotificationType::TAB_PARENTED: case NotificationType::BROWSER_OPENED: window_type = MetricsLog::WINDOW_CREATE; break; case NotificationType::TAB_CLOSING: case NotificationType::BROWSER_CLOSED: window_map_.erase(window_map_.find(window_or_tab)); window_type = MetricsLog::WINDOW_DESTROY; break; default: NOTREACHED(); return; } // TODO(brettw) we should have some kind of ID for the parent. current_log_->RecordWindowEvent(window_type, controller_id, 0); } void MetricsService::LogLoadComplete(NotificationType type, const NotificationSource& source, const NotificationDetails& details) { if (details == NotificationService::NoDetails()) return; // TODO(jar): There is a bug causing this to be called too many times, and // the log overflows. For now, we won't record these events. UMA_HISTOGRAM_COUNTS("UMA.LogLoadComplete called", 1); return; const Details<LoadNotificationDetails> load_details(details); int controller_id = window_map_[details.map_key()]; current_log_->RecordLoadEvent(controller_id, load_details->url(), load_details->origin(), load_details->session_index(), load_details->load_time()); } void MetricsService::IncrementPrefValue(const char* path) { PrefService* pref = g_browser_process->local_state(); DCHECK(pref); int value = pref->GetInteger(path); pref->SetInteger(path, value + 1); } void MetricsService::IncrementLongPrefsValue(const char* path) { PrefService* pref = g_browser_process->local_state(); DCHECK(pref); int64 value = pref->GetInt64(path); pref->SetInt64(path, value + 1); } void MetricsService::LogLoadStarted() { IncrementPrefValue(prefs::kStabilityPageLoadCount); IncrementLongPrefsValue(prefs::kUninstallMetricsPageLoadCount); // We need to save the prefs, as page load count is a critical stat, and it // might be lost due to a crash :-(. } void MetricsService::LogRendererCrash() { IncrementPrefValue(prefs::kStabilityRendererCrashCount); } void MetricsService::LogExtensionRendererCrash() { IncrementPrefValue(prefs::kStabilityExtensionRendererCrashCount); } void MetricsService::LogRendererHang() { IncrementPrefValue(prefs::kStabilityRendererHangCount); } void MetricsService::LogCleanShutdown() { RecordBooleanPrefValue(prefs::kStabilityExitedCleanly, true); } #if defined(OS_CHROMEOS) void MetricsService::LogChromeOSCrash(const std::string &crash_type) { if (crash_type == "user") IncrementPrefValue(prefs::kStabilityOtherUserCrashCount); else if (crash_type == "kernel") IncrementPrefValue(prefs::kStabilityKernelCrashCount); else if (crash_type == "uncleanshutdown") IncrementPrefValue(prefs::kStabilitySystemUncleanShutdownCount); else NOTREACHED() << "Unexpected Chrome OS crash type " << crash_type; // Wake up metrics logs sending if necessary now that new // log data is available. HandleIdleSinceLastTransmission(false); } #endif // OS_CHROMEOS void MetricsService::LogChildProcessChange( NotificationType type, const NotificationSource& source, const NotificationDetails& details) { Details<ChildProcessInfo> child_details(details); const std::wstring& child_name = child_details->name(); if (child_process_stats_buffer_.find(child_name) == child_process_stats_buffer_.end()) { child_process_stats_buffer_[child_name] = ChildProcessStats(child_details->type()); } ChildProcessStats& stats = child_process_stats_buffer_[child_name]; switch (type.value) { case NotificationType::CHILD_PROCESS_HOST_CONNECTED: stats.process_launches++; break; case NotificationType::CHILD_INSTANCE_CREATED: stats.instances++; break; case NotificationType::CHILD_PROCESS_CRASHED: stats.process_crashes++; // Exclude plugin crashes from the count below because we report them via // a separate UMA metric. if (child_details->type() != ChildProcessInfo::PLUGIN_PROCESS) { IncrementPrefValue(prefs::kStabilityChildProcessCrashCount); } break; default: NOTREACHED() << "Unexpected notification type " << type.value; return; } } // Recursively counts the number of bookmarks and folders in node. static void CountBookmarks(const BookmarkNode* node, int* bookmarks, int* folders) { if (node->type() == BookmarkNode::URL) (*bookmarks)++; else (*folders)++; for (int i = 0; i < node->child_count(); ++i) CountBookmarks(node->GetChild(i), bookmarks, folders); } void MetricsService::LogBookmarks(const BookmarkNode* node, const char* num_bookmarks_key, const char* num_folders_key) { DCHECK(node); int num_bookmarks = 0; int num_folders = 0; CountBookmarks(node, &num_bookmarks, &num_folders); num_folders--; // Don't include the root folder in the count. PrefService* pref = g_browser_process->local_state(); DCHECK(pref); pref->SetInteger(num_bookmarks_key, num_bookmarks); pref->SetInteger(num_folders_key, num_folders); } void MetricsService::LogBookmarks(BookmarkModel* model) { DCHECK(model); LogBookmarks(model->GetBookmarkBarNode(), prefs::kNumBookmarksOnBookmarkBar, prefs::kNumFoldersOnBookmarkBar); LogBookmarks(model->other_node(), prefs::kNumBookmarksInOtherBookmarkFolder, prefs::kNumFoldersInOtherBookmarkFolder); ScheduleNextStateSave(); } void MetricsService::LogKeywords(const TemplateURLModel* url_model) { DCHECK(url_model); PrefService* pref = g_browser_process->local_state(); DCHECK(pref); pref->SetInteger(prefs::kNumKeywords, static_cast<int>(url_model->GetTemplateURLs().size())); ScheduleNextStateSave(); } void MetricsService::RecordPluginChanges(PrefService* pref) { ListPrefUpdate update(pref, prefs::kStabilityPluginStats); ListValue* plugins = update.Get(); DCHECK(plugins); for (ListValue::iterator value_iter = plugins->begin(); value_iter != plugins->end(); ++value_iter) { if (!(*value_iter)->IsType(Value::TYPE_DICTIONARY)) { NOTREACHED(); continue; } DictionaryValue* plugin_dict = static_cast<DictionaryValue*>(*value_iter); std::string plugin_name; plugin_dict->GetString(prefs::kStabilityPluginName, &plugin_name); if (plugin_name.empty()) { NOTREACHED(); continue; } // TODO(viettrungluu): remove conversions if (child_process_stats_buffer_.find(UTF8ToWide(plugin_name)) == child_process_stats_buffer_.end()) continue; ChildProcessStats stats = child_process_stats_buffer_[UTF8ToWide(plugin_name)]; if (stats.process_launches) { int launches = 0; plugin_dict->GetInteger(prefs::kStabilityPluginLaunches, &launches); launches += stats.process_launches; plugin_dict->SetInteger(prefs::kStabilityPluginLaunches, launches); } if (stats.process_crashes) { int crashes = 0; plugin_dict->GetInteger(prefs::kStabilityPluginCrashes, &crashes); crashes += stats.process_crashes; plugin_dict->SetInteger(prefs::kStabilityPluginCrashes, crashes); } if (stats.instances) { int instances = 0; plugin_dict->GetInteger(prefs::kStabilityPluginInstances, &instances); instances += stats.instances; plugin_dict->SetInteger(prefs::kStabilityPluginInstances, instances); } child_process_stats_buffer_.erase(UTF8ToWide(plugin_name)); } // Now go through and add dictionaries for plugins that didn't already have // reports in Local State. for (std::map<std::wstring, ChildProcessStats>::iterator cache_iter = child_process_stats_buffer_.begin(); cache_iter != child_process_stats_buffer_.end(); ++cache_iter) { ChildProcessStats stats = cache_iter->second; // Insert only plugins information into the plugins list. if (ChildProcessInfo::PLUGIN_PROCESS != stats.process_type) continue; // TODO(viettrungluu): remove conversion std::string plugin_name = WideToUTF8(cache_iter->first); DictionaryValue* plugin_dict = new DictionaryValue; plugin_dict->SetString(prefs::kStabilityPluginName, plugin_name); plugin_dict->SetInteger(prefs::kStabilityPluginLaunches, stats.process_launches); plugin_dict->SetInteger(prefs::kStabilityPluginCrashes, stats.process_crashes); plugin_dict->SetInteger(prefs::kStabilityPluginInstances, stats.instances); plugins->Append(plugin_dict); } child_process_stats_buffer_.clear(); } bool MetricsService::CanLogNotification(NotificationType type, const NotificationSource& source, const NotificationDetails& details) { // We simply don't log anything to UMA if there is a single incognito // session visible. The problem is that we always notify using the orginal // profile in order to simplify notification processing. return !BrowserList::IsOffTheRecordSessionActive(); } void MetricsService::RecordBooleanPrefValue(const char* path, bool value) { DCHECK(IsSingleThreaded()); PrefService* pref = g_browser_process->local_state(); DCHECK(pref); pref->SetBoolean(path, value); RecordCurrentState(pref); } void MetricsService::RecordCurrentState(PrefService* pref) { pref->SetInt64(prefs::kStabilityLastTimestampSec, Time::Now().ToTimeT()); RecordPluginChanges(pref); } static bool IsSingleThreaded() { static base::PlatformThreadId thread_id = 0; if (!thread_id) thread_id = base::PlatformThread::CurrentId(); return base::PlatformThread::CurrentId() == thread_id; } #if defined(OS_CHROMEOS) void MetricsService::StartExternalMetrics() { external_metrics_ = new chromeos::ExternalMetrics; external_metrics_->Start(); } #endif