| // Copyright 2014 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| //------------------------------------------------------------------------------ |
| // Description of the life cycle of a instance of MetricsService. |
| // |
| // OVERVIEW |
| // |
| // A MetricsService instance is typically created at application startup. It is |
| // the central controller for the acquisition of log data, and the automatic |
| // transmission of that log data to an external server. Its major job is to |
| // manage logs, grouping them for transmission, and transmitting them. As part |
| // of its grouping, MS finalizes logs by including some just-in-time gathered |
| // memory statistics, snapshotting the current stats of numerous histograms, |
| // closing the logs, translating to protocol buffer format, and compressing the |
| // results for transmission. Transmission includes submitting a compressed log |
| // as data in a URL-post, and retransmitting (or retaining at process |
| // termination) if the attempted transmission failed. Retention across process |
| // terminations is done using the the PrefServices facilities. The retained logs |
| // (the ones that never got transmitted) are compressed and base64-encoded |
| // before being persisted. |
| // |
| // Logs fall into one of two categories: "initial logs," and "ongoing logs." |
| // There is at most one initial log sent for each complete run of Chrome (from |
| // startup, to browser shutdown). An initial log is generally transmitted some |
| // short time (1 minute?) after startup, and includes stats such as recent crash |
| // info, the number and types of plugins, etc. The external server's response |
| // to the initial log conceptually tells this MS if it should continue |
| // transmitting logs (during this session). The server response can actually be |
| // much more detailed, and always includes (at a minimum) how often additional |
| // ongoing logs should be sent. |
| // |
| // After the above initial log, a series of ongoing logs will be transmitted. |
| // The first ongoing log actually begins to accumulate information stating when |
| // the MS was first constructed. Note that even though the initial log is |
| // commonly sent a full minute after startup, the initial log does not include |
| // much in the way of user stats. The most common interlog period (delay) |
| // is 30 minutes. That time period starts when the first user action causes a |
| // logging event. This means that if there is no user action, there may be long |
| // periods without any (ongoing) log transmissions. Ongoing logs typically |
| // contain very detailed records of user activities (ex: opened tab, closed |
| // tab, fetched URL, maximized window, etc.) In addition, just before an |
| // ongoing log is closed out, a call is made to gather memory statistics. Those |
| // memory statistics are deposited into a histogram, and the log finalization |
| // code is then called. In the finalization, a call to a Histogram server |
| // acquires a list of all local histograms that have been flagged for upload |
| // to the UMA server. The finalization also acquires the most recent number |
| // of page loads, along with any counts of renderer or plugin crashes. |
| // |
| // When the browser shuts down, there will typically be a fragment of an ongoing |
| // log that has not yet been transmitted. At shutdown time, that fragment is |
| // closed (including snapshotting histograms), and persisted, for potential |
| // transmission during a future run of the product. |
| // |
| // There are two slightly abnormal shutdown conditions. There is a |
| // "disconnected scenario," and a "really fast startup and shutdown" scenario. |
| // In the "never connected" situation, the user has (during the running of the |
| // process) never established an internet connection. As a result, attempts to |
| // transmit the initial log have failed, and a lot(?) of data has accumulated in |
| // the ongoing log (which didn't yet get closed, because there was never even a |
| // contemplation of sending it). There is also a kindred "lost connection" |
| // situation, where a loss of connection prevented an ongoing log from being |
| // transmitted, and a (still open) log was stuck accumulating a lot(?) of data, |
| // while the earlier log retried its transmission. In both of these |
| // disconnected situations, two logs need to be, and are, persistently stored |
| // for future transmission. |
| // |
| // The other unusual shutdown condition, termed "really fast startup and |
| // shutdown," involves the deliberate user termination of the process before |
| // the initial log is even formed or transmitted. In that situation, no logging |
| // is done, but the historical crash statistics remain (unlogged) for inclusion |
| // in a future run's initial log. (i.e., we don't lose crash stats). |
| // |
| // With the above overview, we can now describe the state machine's various |
| // states, based on the State enum specified in the state_ member. Those states |
| // are: |
| // |
| // INITIALIZED, // Constructor was called. |
| // INIT_TASK_SCHEDULED, // Waiting for deferred init tasks to finish. |
| // INIT_TASK_DONE, // Waiting for timer to send initial log. |
| // SENDING_LOGS, // Sending logs and creating new ones when we run out. |
| // |
| // In more detail, we have: |
| // |
| // INITIALIZED, // Constructor was called. |
| // The MS has been constructed, but has taken no actions to compose the |
| // initial log. |
| // |
| // INIT_TASK_SCHEDULED, // Waiting for deferred init tasks to finish. |
| // Typically about 30 seconds after startup, a task is sent to a second thread |
| // (the file thread) to perform deferred (lower priority and slower) |
| // initialization steps such as getting the list of plugins. That task will |
| // (when complete) make an async callback (via a Task) to indicate the |
| // completion. |
| // |
| // INIT_TASK_DONE, // Waiting for timer to send initial log. |
| // The callback has arrived, and it is now possible for an initial log to be |
| // created. This callback typically arrives back less than one second after |
| // the deferred init task is dispatched. |
| // |
| // SENDING_LOGS, // Sending logs an creating new ones when we run out. |
| // Logs from previous sessions have been loaded, and initial logs have been |
| // created (an optional stability log and the first metrics log). We will |
| // send all of these logs, and when run out, we will start cutting new logs |
| // to send. We will also cut a new log if we expect a shutdown. |
| // |
| // The progression through the above states is simple, and sequential. |
| // States proceed from INITIAL to SENDING_LOGS, and remain in the latter until |
| // shutdown. |
| // |
| // Also note that whenever we successfully send a log, we mirror the list |
| // of logs into the PrefService. This ensures that IF we crash, we won't start |
| // up and retransmit our old logs again. |
| // |
| // Due to race conditions, it is always possible that a log file could be sent |
| // twice. For example, if a log file is sent, but not yet acknowledged by |
| // the external server, and the user shuts down, then a copy of the log may be |
| // saved for re-transmission. These duplicates could be filtered out server |
| // side, but are not expected to be a significant problem. |
| // |
| // |
| //------------------------------------------------------------------------------ |
| |
| #include "components/metrics/metrics_service.h" |
| |
| #include <stddef.h> |
| |
| #include <algorithm> |
| #include <utility> |
| |
| #include "base/bind.h" |
| #include "base/callback.h" |
| #include "base/location.h" |
| #include "base/memory/ptr_util.h" |
| #include "base/metrics/histogram_base.h" |
| #include "base/metrics/histogram_macros.h" |
| #include "base/metrics/histogram_samples.h" |
| #include "base/metrics/persistent_histogram_allocator.h" |
| #include "base/metrics/sparse_histogram.h" |
| #include "base/metrics/statistics_recorder.h" |
| #include "base/single_thread_task_runner.h" |
| #include "base/strings/string_number_conversions.h" |
| #include "base/strings/utf_string_conversions.h" |
| #include "base/threading/platform_thread.h" |
| #include "base/threading/thread.h" |
| #include "base/threading/thread_restrictions.h" |
| #include "base/threading/thread_task_runner_handle.h" |
| #include "base/time/time.h" |
| #include "base/tracked_objects.h" |
| #include "build/build_config.h" |
| #include "components/metrics/data_use_tracker.h" |
| #include "components/metrics/metrics_log.h" |
| #include "components/metrics/metrics_log_manager.h" |
| #include "components/metrics/metrics_log_uploader.h" |
| #include "components/metrics/metrics_pref_names.h" |
| #include "components/metrics/metrics_reporting_scheduler.h" |
| #include "components/metrics/metrics_service_client.h" |
| #include "components/metrics/metrics_state_manager.h" |
| #include "components/prefs/pref_registry_simple.h" |
| #include "components/prefs/pref_service.h" |
| #include "components/variations/entropy_provider.h" |
| |
| namespace metrics { |
| |
| namespace { |
| |
| // Check to see that we're being called on only one thread. |
| bool IsSingleThreaded() { |
| static base::PlatformThreadId thread_id = 0; |
| if (!thread_id) |
| thread_id = base::PlatformThread::CurrentId(); |
| return base::PlatformThread::CurrentId() == thread_id; |
| } |
| |
| // The delay, in seconds, after starting recording before doing expensive |
| // initialization work. |
| #if defined(OS_ANDROID) || defined(OS_IOS) |
| // On mobile devices, a significant portion of sessions last less than a minute. |
| // Use a shorter timer on these platforms to avoid losing data. |
| // TODO(dfalcantara): To avoid delaying startup, tighten up initialization so |
| // that it occurs after the user gets their initial page. |
| const int kInitializationDelaySeconds = 5; |
| #else |
| const int kInitializationDelaySeconds = 30; |
| #endif |
| |
| // The maximum number of events in a log uploaded to the UMA server. |
| const int kEventLimit = 2400; |
| |
| // If an upload fails, and the transmission was over this byte count, then we |
| // will discard the log, and not try to retransmit it. We also don't persist |
| // the log to the prefs for transmission during the next chrome session if this |
| // limit is exceeded. |
| const size_t kUploadLogAvoidRetransmitSize = 100 * 1024; |
| |
| // Interval, in minutes, between state saves. |
| const int kSaveStateIntervalMinutes = 5; |
| |
| enum ResponseStatus { |
| UNKNOWN_FAILURE, |
| SUCCESS, |
| BAD_REQUEST, // Invalid syntax or log too large. |
| NO_RESPONSE, |
| NUM_RESPONSE_STATUSES |
| }; |
| |
| ResponseStatus ResponseCodeToStatus(int response_code) { |
| switch (response_code) { |
| case -1: |
| return NO_RESPONSE; |
| case 200: |
| return SUCCESS; |
| case 400: |
| return BAD_REQUEST; |
| default: |
| return UNKNOWN_FAILURE; |
| } |
| } |
| |
| #if defined(OS_ANDROID) || defined(OS_IOS) |
| void MarkAppCleanShutdownAndCommit(CleanExitBeacon* clean_exit_beacon, |
| PrefService* local_state) { |
| clean_exit_beacon->WriteBeaconValue(true); |
| local_state->SetInteger(prefs::kStabilityExecutionPhase, |
| MetricsService::SHUTDOWN_COMPLETE); |
| // Start writing right away (write happens on a different thread). |
| local_state->CommitPendingWrite(); |
| } |
| #endif // defined(OS_ANDROID) || defined(OS_IOS) |
| |
| } // namespace |
| |
| // static |
| MetricsService::ShutdownCleanliness MetricsService::clean_shutdown_status_ = |
| MetricsService::CLEANLY_SHUTDOWN; |
| |
| MetricsService::ExecutionPhase MetricsService::execution_phase_ = |
| MetricsService::UNINITIALIZED_PHASE; |
| |
| // static |
| void MetricsService::RegisterPrefs(PrefRegistrySimple* registry) { |
| DCHECK(IsSingleThreaded()); |
| MetricsStateManager::RegisterPrefs(registry); |
| MetricsLog::RegisterPrefs(registry); |
| DataUseTracker::RegisterPrefs(registry); |
| |
| registry->RegisterInt64Pref(prefs::kInstallDate, 0); |
| |
| registry->RegisterInt64Pref(prefs::kStabilityLaunchTimeSec, 0); |
| registry->RegisterInt64Pref(prefs::kStabilityLastTimestampSec, 0); |
| registry->RegisterStringPref(prefs::kStabilityStatsVersion, std::string()); |
| registry->RegisterInt64Pref(prefs::kStabilityStatsBuildTime, 0); |
| registry->RegisterBooleanPref(prefs::kStabilityExitedCleanly, true); |
| registry->RegisterIntegerPref(prefs::kStabilityExecutionPhase, |
| UNINITIALIZED_PHASE); |
| registry->RegisterBooleanPref(prefs::kStabilitySessionEndCompleted, true); |
| registry->RegisterIntegerPref(prefs::kMetricsSessionID, -1); |
| |
| registry->RegisterListPref(prefs::kMetricsInitialLogs); |
| registry->RegisterListPref(prefs::kMetricsOngoingLogs); |
| |
| registry->RegisterInt64Pref(prefs::kUninstallLaunchCount, 0); |
| registry->RegisterInt64Pref(prefs::kUninstallMetricsUptimeSec, 0); |
| } |
| |
| MetricsService::MetricsService(MetricsStateManager* state_manager, |
| MetricsServiceClient* client, |
| PrefService* local_state) |
| : log_manager_(local_state, kUploadLogAvoidRetransmitSize), |
| histogram_snapshot_manager_(this), |
| state_manager_(state_manager), |
| client_(client), |
| local_state_(local_state), |
| clean_exit_beacon_(client->GetRegistryBackupKey(), local_state), |
| recording_state_(UNSET), |
| reporting_active_(false), |
| test_mode_active_(false), |
| state_(INITIALIZED), |
| log_upload_in_progress_(false), |
| idle_since_last_transmission_(false), |
| session_id_(-1), |
| data_use_tracker_(DataUseTracker::Create(local_state_)), |
| self_ptr_factory_(this), |
| state_saver_factory_(this) { |
| DCHECK(IsSingleThreaded()); |
| DCHECK(state_manager_); |
| DCHECK(client_); |
| DCHECK(local_state_); |
| |
| // Set the install date if this is our first run. |
| int64_t install_date = local_state_->GetInt64(prefs::kInstallDate); |
| if (install_date == 0) |
| local_state_->SetInt64(prefs::kInstallDate, base::Time::Now().ToTimeT()); |
| } |
| |
| MetricsService::~MetricsService() { |
| DisableRecording(); |
| } |
| |
| void MetricsService::InitializeMetricsRecordingState() { |
| InitializeMetricsState(); |
| |
| base::Closure upload_callback = |
| base::Bind(&MetricsService::StartScheduledUpload, |
| self_ptr_factory_.GetWeakPtr()); |
| scheduler_.reset( |
| new MetricsReportingScheduler( |
| upload_callback, |
| // MetricsServiceClient outlives MetricsService, and |
| // MetricsReportingScheduler is tied to the lifetime of |this|. |
| base::Bind(&MetricsServiceClient::GetStandardUploadInterval, |
| base::Unretained(client_)))); |
| |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->Init(); |
| } |
| |
| void MetricsService::Start() { |
| HandleIdleSinceLastTransmission(false); |
| EnableRecording(); |
| EnableReporting(); |
| } |
| |
| void MetricsService::StartRecordingForTests() { |
| test_mode_active_ = true; |
| EnableRecording(); |
| DisableReporting(); |
| } |
| |
| void MetricsService::Stop() { |
| HandleIdleSinceLastTransmission(false); |
| DisableReporting(); |
| DisableRecording(); |
| } |
| |
| void MetricsService::EnableReporting() { |
| if (reporting_active_) |
| return; |
| reporting_active_ = true; |
| StartSchedulerIfNecessary(); |
| } |
| |
| void MetricsService::DisableReporting() { |
| reporting_active_ = false; |
| } |
| |
| std::string MetricsService::GetClientId() { |
| return state_manager_->client_id(); |
| } |
| |
| int64_t MetricsService::GetInstallDate() { |
| return local_state_->GetInt64(prefs::kInstallDate); |
| } |
| |
| int64_t MetricsService::GetMetricsReportingEnabledDate() { |
| return local_state_->GetInt64(prefs::kMetricsReportingEnabledTimestamp); |
| } |
| |
| bool MetricsService::WasLastShutdownClean() const { |
| return clean_exit_beacon_.exited_cleanly(); |
| } |
| |
| void MetricsService::EnableRecording() { |
| DCHECK(IsSingleThreaded()); |
| |
| if (recording_state_ == ACTIVE) |
| return; |
| recording_state_ = ACTIVE; |
| |
| state_manager_->ForceClientIdCreation(); |
| client_->SetMetricsClientId(state_manager_->client_id()); |
| if (!log_manager_.current_log()) |
| OpenNewLog(); |
| |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->OnRecordingEnabled(); |
| |
| base::RemoveActionCallback(action_callback_); |
| action_callback_ = base::Bind(&MetricsService::OnUserAction, |
| base::Unretained(this)); |
| base::AddActionCallback(action_callback_); |
| } |
| |
| void MetricsService::DisableRecording() { |
| DCHECK(IsSingleThreaded()); |
| |
| if (recording_state_ == INACTIVE) |
| return; |
| recording_state_ = INACTIVE; |
| |
| base::RemoveActionCallback(action_callback_); |
| |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->OnRecordingDisabled(); |
| |
| PushPendingLogsToPersistentStorage(); |
| } |
| |
| bool MetricsService::recording_active() const { |
| DCHECK(IsSingleThreaded()); |
| return recording_state_ == ACTIVE; |
| } |
| |
| bool MetricsService::reporting_active() const { |
| DCHECK(IsSingleThreaded()); |
| return reporting_active_; |
| } |
| |
| void MetricsService::RecordDelta(const base::HistogramBase& histogram, |
| const base::HistogramSamples& snapshot) { |
| log_manager_.current_log()->RecordHistogramDelta(histogram.histogram_name(), |
| snapshot); |
| } |
| |
| void MetricsService::InconsistencyDetected( |
| base::HistogramBase::Inconsistency problem) { |
| UMA_HISTOGRAM_ENUMERATION("Histogram.InconsistenciesBrowser", |
| problem, base::HistogramBase::NEVER_EXCEEDED_VALUE); |
| } |
| |
| void MetricsService::UniqueInconsistencyDetected( |
| base::HistogramBase::Inconsistency problem) { |
| UMA_HISTOGRAM_ENUMERATION("Histogram.InconsistenciesBrowserUnique", |
| problem, base::HistogramBase::NEVER_EXCEEDED_VALUE); |
| } |
| |
| void MetricsService::InconsistencyDetectedInLoggedCount(int amount) { |
| UMA_HISTOGRAM_COUNTS("Histogram.InconsistentSnapshotBrowser", |
| std::abs(amount)); |
| } |
| |
| void MetricsService::HandleIdleSinceLastTransmission(bool in_idle) { |
| // If there wasn't a lot of action, maybe the computer was asleep, in which |
| // case, the log transmissions should have stopped. Here we start them up |
| // again. |
| if (!in_idle && idle_since_last_transmission_) |
| StartSchedulerIfNecessary(); |
| idle_since_last_transmission_ = in_idle; |
| } |
| |
| void MetricsService::OnApplicationNotIdle() { |
| if (recording_state_ == ACTIVE) |
| HandleIdleSinceLastTransmission(false); |
| } |
| |
| void MetricsService::RecordStartOfSessionEnd() { |
| LogCleanShutdown(); |
| RecordBooleanPrefValue(prefs::kStabilitySessionEndCompleted, false); |
| } |
| |
| void MetricsService::RecordCompletedSessionEnd() { |
| LogCleanShutdown(); |
| RecordBooleanPrefValue(prefs::kStabilitySessionEndCompleted, true); |
| } |
| |
| #if defined(OS_ANDROID) || defined(OS_IOS) |
| void MetricsService::OnAppEnterBackground() { |
| scheduler_->Stop(); |
| |
| MarkAppCleanShutdownAndCommit(&clean_exit_beacon_, local_state_); |
| |
| // Give providers a chance to persist histograms as part of being |
| // backgrounded. |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->OnAppEnterBackground(); |
| |
| // At this point, there's no way of knowing when the process will be |
| // killed, so this has to be treated similar to a shutdown, closing and |
| // persisting all logs. Unlinke a shutdown, the state is primed to be ready |
| // to continue logging and uploading if the process does return. |
| if (recording_active() && state_ >= SENDING_LOGS) { |
| PushPendingLogsToPersistentStorage(); |
| // Persisting logs closes the current log, so start recording a new log |
| // immediately to capture any background work that might be done before the |
| // process is killed. |
| OpenNewLog(); |
| } |
| } |
| |
| void MetricsService::OnAppEnterForeground() { |
| clean_exit_beacon_.WriteBeaconValue(false); |
| StartSchedulerIfNecessary(); |
| } |
| #else |
| void MetricsService::LogNeedForCleanShutdown() { |
| clean_exit_beacon_.WriteBeaconValue(false); |
| // Redundant setting to be sure we call for a clean shutdown. |
| clean_shutdown_status_ = NEED_TO_SHUTDOWN; |
| } |
| #endif // defined(OS_ANDROID) || defined(OS_IOS) |
| |
| // static |
| void MetricsService::SetExecutionPhase(ExecutionPhase execution_phase, |
| PrefService* local_state) { |
| execution_phase_ = execution_phase; |
| local_state->SetInteger(prefs::kStabilityExecutionPhase, execution_phase_); |
| } |
| |
| void MetricsService::RecordBreakpadRegistration(bool success) { |
| if (!success) |
| IncrementPrefValue(prefs::kStabilityBreakpadRegistrationFail); |
| else |
| IncrementPrefValue(prefs::kStabilityBreakpadRegistrationSuccess); |
| } |
| |
| void MetricsService::RecordBreakpadHasDebugger(bool has_debugger) { |
| if (!has_debugger) |
| IncrementPrefValue(prefs::kStabilityDebuggerNotPresent); |
| else |
| IncrementPrefValue(prefs::kStabilityDebuggerPresent); |
| } |
| |
| void MetricsService::ClearSavedStabilityMetrics() { |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->ClearSavedStabilityMetrics(); |
| |
| // Reset the prefs that are managed by MetricsService/MetricsLog directly. |
| local_state_->SetInteger(prefs::kStabilityBreakpadRegistrationSuccess, 0); |
| local_state_->SetInteger(prefs::kStabilityBreakpadRegistrationFail, 0); |
| local_state_->SetInteger(prefs::kStabilityCrashCount, 0); |
| local_state_->SetInteger(prefs::kStabilityDebuggerPresent, 0); |
| local_state_->SetInteger(prefs::kStabilityDebuggerNotPresent, 0); |
| local_state_->SetInteger(prefs::kStabilityExecutionPhase, |
| UNINITIALIZED_PHASE); |
| local_state_->SetInteger(prefs::kStabilityIncompleteSessionEndCount, 0); |
| local_state_->SetInteger(prefs::kStabilityLaunchCount, 0); |
| local_state_->SetBoolean(prefs::kStabilitySessionEndCompleted, true); |
| local_state_->SetInteger(prefs::kStabilityDeferredCount, 0); |
| // Note: kStabilityDiscardCount is not cleared as its intent is to measure |
| // the number of times data is discarded, even across versions. |
| local_state_->SetInteger(prefs::kStabilityVersionMismatchCount, 0); |
| } |
| |
| void MetricsService::PushExternalLog(const std::string& log) { |
| log_manager_.StoreLog(log, MetricsLog::ONGOING_LOG); |
| } |
| |
| UpdateUsagePrefCallbackType MetricsService::GetDataUseForwardingCallback() { |
| DCHECK(IsSingleThreaded()); |
| |
| if (data_use_tracker_) { |
| return data_use_tracker_->GetDataUseForwardingCallback( |
| base::ThreadTaskRunnerHandle::Get()); |
| } |
| return UpdateUsagePrefCallbackType(); |
| } |
| |
| void MetricsService::MergeHistogramDeltas() { |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->MergeHistogramDeltas(); |
| } |
| |
| //------------------------------------------------------------------------------ |
| // private methods |
| //------------------------------------------------------------------------------ |
| |
| |
| //------------------------------------------------------------------------------ |
| // Initialization methods |
| |
| void MetricsService::InitializeMetricsState() { |
| const int64_t buildtime = MetricsLog::GetBuildTime(); |
| const std::string version = client_->GetVersionString(); |
| bool version_changed = false; |
| int64_t previous_buildtime = |
| local_state_->GetInt64(prefs::kStabilityStatsBuildTime); |
| std::string previous_version = |
| local_state_->GetString(prefs::kStabilityStatsVersion); |
| if (previous_buildtime != buildtime || previous_version != version) { |
| local_state_->SetString(prefs::kStabilityStatsVersion, version); |
| local_state_->SetInt64(prefs::kStabilityStatsBuildTime, buildtime); |
| version_changed = true; |
| } |
| |
| log_manager_.LoadPersistedUnsentLogs(); |
| |
| session_id_ = local_state_->GetInteger(prefs::kMetricsSessionID); |
| |
| if (!clean_exit_beacon_.exited_cleanly()) { |
| IncrementPrefValue(prefs::kStabilityCrashCount); |
| // Reset flag, and wait until we call LogNeedForCleanShutdown() before |
| // monitoring. |
| clean_exit_beacon_.WriteBeaconValue(true); |
| } |
| |
| bool has_initial_stability_log = false; |
| bool providers_have_initial_stability_metrics = |
| ProvidersHaveInitialStabilityMetrics(); |
| if (!clean_exit_beacon_.exited_cleanly() || |
| providers_have_initial_stability_metrics) { |
| // TODO(rtenneti): On windows, consider saving/getting execution_phase from |
| // the registry. |
| int execution_phase = |
| local_state_->GetInteger(prefs::kStabilityExecutionPhase); |
| UMA_HISTOGRAM_SPARSE_SLOWLY("Chrome.Browser.CrashedExecutionPhase", |
| execution_phase); |
| |
| // If the previous session didn't exit cleanly, or if any provider |
| // explicitly requests it, prepare an initial stability log - |
| // provided UMA is enabled. |
| if (state_manager_->IsMetricsReportingEnabled()) { |
| has_initial_stability_log = PrepareInitialStabilityLog(previous_version); |
| if (!has_initial_stability_log) |
| IncrementPrefValue(prefs::kStabilityDeferredCount); |
| } |
| } |
| |
| // If no initial stability log was generated and there was a version upgrade, |
| // clear the stability stats from the previous version (so that they don't get |
| // attributed to the current version). This could otherwise happen due to a |
| // number of different edge cases, such as if the last version crashed before |
| // it could save off a system profile or if UMA reporting is disabled (which |
| // normally results in stats being accumulated). |
| if (!has_initial_stability_log && version_changed) { |
| ClearSavedStabilityMetrics(); |
| IncrementPrefValue(prefs::kStabilityDiscardCount); |
| } |
| |
| // Update session ID. |
| ++session_id_; |
| local_state_->SetInteger(prefs::kMetricsSessionID, session_id_); |
| |
| // Stability bookkeeping |
| IncrementPrefValue(prefs::kStabilityLaunchCount); |
| |
| DCHECK_EQ(UNINITIALIZED_PHASE, execution_phase_); |
| SetExecutionPhase(START_METRICS_RECORDING, local_state_); |
| |
| if (!local_state_->GetBoolean(prefs::kStabilitySessionEndCompleted)) { |
| IncrementPrefValue(prefs::kStabilityIncompleteSessionEndCount); |
| // This is marked false when we get a WM_ENDSESSION. |
| local_state_->SetBoolean(prefs::kStabilitySessionEndCompleted, true); |
| } |
| |
| // Call GetUptimes() for the first time, thus allowing all later calls |
| // to record incremental uptimes accurately. |
| base::TimeDelta ignored_uptime_parameter; |
| base::TimeDelta startup_uptime; |
| GetUptimes(local_state_, &startup_uptime, &ignored_uptime_parameter); |
| DCHECK_EQ(0, startup_uptime.InMicroseconds()); |
| // For backwards compatibility, leave this intact in case Omaha is checking |
| // them. prefs::kStabilityLastTimestampSec may also be useless now. |
| // TODO(jar): Delete these if they have no uses. |
| local_state_->SetInt64(prefs::kStabilityLaunchTimeSec, |
| base::Time::Now().ToTimeT()); |
| |
| // Bookkeeping for the uninstall metrics. |
| IncrementLongPrefsValue(prefs::kUninstallLaunchCount); |
| |
| // Kick off the process of saving the state (so the uptime numbers keep |
| // getting updated) every n minutes. |
| ScheduleNextStateSave(); |
| } |
| |
| void MetricsService::OnUserAction(const std::string& action) { |
| log_manager_.current_log()->RecordUserAction(action); |
| HandleIdleSinceLastTransmission(false); |
| } |
| |
| void MetricsService::FinishedInitTask() { |
| DCHECK_EQ(INIT_TASK_SCHEDULED, state_); |
| state_ = INIT_TASK_DONE; |
| |
| // Create the initial log. |
| if (!initial_metrics_log_.get()) { |
| initial_metrics_log_ = CreateLog(MetricsLog::ONGOING_LOG); |
| NotifyOnDidCreateMetricsLog(); |
| } |
| |
| scheduler_->InitTaskComplete(); |
| } |
| |
| void MetricsService::GetUptimes(PrefService* pref, |
| base::TimeDelta* incremental_uptime, |
| base::TimeDelta* uptime) { |
| base::TimeTicks now = base::TimeTicks::Now(); |
| // If this is the first call, init |first_updated_time_| and |
| // |last_updated_time_|. |
| if (last_updated_time_.is_null()) { |
| first_updated_time_ = now; |
| last_updated_time_ = now; |
| } |
| *incremental_uptime = now - last_updated_time_; |
| *uptime = now - first_updated_time_; |
| last_updated_time_ = now; |
| |
| const int64_t incremental_time_secs = incremental_uptime->InSeconds(); |
| if (incremental_time_secs > 0) { |
| int64_t metrics_uptime = pref->GetInt64(prefs::kUninstallMetricsUptimeSec); |
| metrics_uptime += incremental_time_secs; |
| pref->SetInt64(prefs::kUninstallMetricsUptimeSec, metrics_uptime); |
| } |
| } |
| |
| void MetricsService::NotifyOnDidCreateMetricsLog() { |
| DCHECK(IsSingleThreaded()); |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->OnDidCreateMetricsLog(); |
| } |
| |
| //------------------------------------------------------------------------------ |
| // State save methods |
| |
| void MetricsService::ScheduleNextStateSave() { |
| state_saver_factory_.InvalidateWeakPtrs(); |
| |
| base::ThreadTaskRunnerHandle::Get()->PostDelayedTask( |
| FROM_HERE, base::Bind(&MetricsService::SaveLocalState, |
| state_saver_factory_.GetWeakPtr()), |
| base::TimeDelta::FromMinutes(kSaveStateIntervalMinutes)); |
| } |
| |
| void MetricsService::SaveLocalState() { |
| RecordCurrentState(local_state_); |
| |
| // TODO(jar):110021 Does this run down the batteries???? |
| ScheduleNextStateSave(); |
| } |
| |
| |
| //------------------------------------------------------------------------------ |
| // Recording control methods |
| |
| void MetricsService::OpenNewLog() { |
| DCHECK(!log_manager_.current_log()); |
| |
| log_manager_.BeginLoggingWithLog(CreateLog(MetricsLog::ONGOING_LOG)); |
| NotifyOnDidCreateMetricsLog(); |
| if (state_ == INITIALIZED) { |
| // We only need to schedule that run once. |
| state_ = INIT_TASK_SCHEDULED; |
| |
| base::ThreadTaskRunnerHandle::Get()->PostDelayedTask( |
| FROM_HERE, base::Bind(&MetricsService::StartInitTask, |
| self_ptr_factory_.GetWeakPtr()), |
| base::TimeDelta::FromSeconds(kInitializationDelaySeconds)); |
| } |
| } |
| |
| void MetricsService::StartInitTask() { |
| client_->InitializeSystemProfileMetrics( |
| base::Bind(&MetricsService::FinishedInitTask, |
| self_ptr_factory_.GetWeakPtr())); |
| } |
| |
| void MetricsService::CloseCurrentLog() { |
| if (!log_manager_.current_log()) |
| return; |
| |
| // TODO(jar): Integrate bounds on log recording more consistently, so that we |
| // can stop recording logs that are too big much sooner. |
| if (log_manager_.current_log()->num_events() > kEventLimit) { |
| UMA_HISTOGRAM_COUNTS("UMA.Discarded Log Events", |
| log_manager_.current_log()->num_events()); |
| log_manager_.DiscardCurrentLog(); |
| OpenNewLog(); // Start trivial log to hold our histograms. |
| } |
| |
| // If a persistent allocator is in use, update its internal histograms (such |
| // as how much memory is being used) before reporting. |
| base::PersistentHistogramAllocator* allocator = |
| base::GlobalHistogramAllocator::Get(); |
| if (allocator) |
| allocator->UpdateTrackingHistograms(); |
| |
| // Put incremental data (histogram deltas, and realtime stats deltas) at the |
| // end of all log transmissions (initial log handles this separately). |
| // RecordIncrementalStabilityElements only exists on the derived |
| // MetricsLog class. |
| MetricsLog* current_log = log_manager_.current_log(); |
| DCHECK(current_log); |
| RecordCurrentEnvironment(current_log); |
| base::TimeDelta incremental_uptime; |
| base::TimeDelta uptime; |
| GetUptimes(local_state_, &incremental_uptime, &uptime); |
| current_log->RecordStabilityMetrics(metrics_providers_.get(), |
| incremental_uptime, uptime); |
| |
| current_log->RecordGeneralMetrics(metrics_providers_.get()); |
| RecordCurrentHistograms(); |
| |
| log_manager_.FinishCurrentLog(); |
| } |
| |
| void MetricsService::PushPendingLogsToPersistentStorage() { |
| if (state_ < SENDING_LOGS) |
| return; // We didn't and still don't have time to get plugin list etc. |
| |
| CloseCurrentLog(); |
| log_manager_.PersistUnsentLogs(); |
| } |
| |
| //------------------------------------------------------------------------------ |
| // Transmission of logs methods |
| |
| void MetricsService::StartSchedulerIfNecessary() { |
| // Never schedule cutting or uploading of logs in test mode. |
| if (test_mode_active_) |
| return; |
| |
| // Even if reporting is disabled, the scheduler is needed to trigger the |
| // creation of the initial log, which must be done in order for any logs to be |
| // persisted on shutdown or backgrounding. |
| if (recording_active() && |
| (reporting_active() || state_ < SENDING_LOGS)) { |
| scheduler_->Start(); |
| } |
| } |
| |
| void MetricsService::StartScheduledUpload() { |
| DCHECK(state_ >= INIT_TASK_DONE); |
| // If we're getting no notifications, then the log won't have much in it, and |
| // it's possible the computer is about to go to sleep, so don't upload and |
| // stop the scheduler. |
| // If recording has been turned off, the scheduler doesn't need to run. |
| // If reporting is off, proceed if the initial log hasn't been created, since |
| // that has to happen in order for logs to be cut and stored when persisting. |
| // TODO(stuartmorgan): Call Stop() on the scheduler when reporting and/or |
| // recording are turned off instead of letting it fire and then aborting. |
| if (idle_since_last_transmission_ || |
| !recording_active() || |
| (!reporting_active() && state_ >= SENDING_LOGS)) { |
| scheduler_->Stop(); |
| scheduler_->UploadCancelled(); |
| return; |
| } |
| |
| // If there are unsent logs, send the next one. If not, start the asynchronous |
| // process of finalizing the current log for upload. |
| if (state_ == SENDING_LOGS && log_manager_.has_unsent_logs()) { |
| SendNextLog(); |
| } else { |
| // There are no logs left to send, so start creating a new one. |
| client_->CollectFinalMetricsForLog( |
| base::Bind(&MetricsService::OnFinalLogInfoCollectionDone, |
| self_ptr_factory_.GetWeakPtr())); |
| } |
| } |
| |
| void MetricsService::OnFinalLogInfoCollectionDone() { |
| // If somehow there is a log upload in progress, we return and hope things |
| // work out. The scheduler isn't informed since if this happens, the scheduler |
| // will get a response from the upload. |
| DCHECK(!log_upload_in_progress_); |
| if (log_upload_in_progress_) |
| return; |
| |
| // Abort if metrics were turned off during the final info gathering. |
| if (!recording_active()) { |
| scheduler_->Stop(); |
| scheduler_->UploadCancelled(); |
| return; |
| } |
| |
| if (state_ == INIT_TASK_DONE) { |
| PrepareInitialMetricsLog(); |
| } else { |
| DCHECK_EQ(SENDING_LOGS, state_); |
| CloseCurrentLog(); |
| OpenNewLog(); |
| } |
| SendNextLog(); |
| } |
| |
| void MetricsService::SendNextLog() { |
| DCHECK_EQ(SENDING_LOGS, state_); |
| if (!reporting_active()) { |
| scheduler_->Stop(); |
| scheduler_->UploadCancelled(); |
| return; |
| } |
| if (!log_manager_.has_unsent_logs()) { |
| // Should only get here if serializing the log failed somehow. |
| // Just tell the scheduler it was uploaded and wait for the next log |
| // interval. |
| scheduler_->UploadFinished(true, log_manager_.has_unsent_logs()); |
| return; |
| } |
| if (!log_manager_.has_staged_log()) |
| log_manager_.StageNextLogForUpload(); |
| |
| // Proceed to stage the log for upload if log size satisfies cellular log |
| // upload constrains. |
| bool upload_canceled = false; |
| bool is_cellular_logic = client_->IsUMACellularUploadLogicEnabled(); |
| if (is_cellular_logic && data_use_tracker_ && |
| !data_use_tracker_->ShouldUploadLogOnCellular( |
| log_manager_.staged_log_hash().size())) { |
| scheduler_->UploadCancelled(); |
| upload_canceled = true; |
| } else { |
| SendStagedLog(); |
| } |
| if (is_cellular_logic) { |
| UMA_HISTOGRAM_BOOLEAN("UMA.LogUpload.Canceled.CellularConstraint", |
| upload_canceled); |
| } |
| } |
| |
| bool MetricsService::ProvidersHaveInitialStabilityMetrics() { |
| // Check whether any metrics provider has initial stability metrics. |
| for (MetricsProvider* provider : metrics_providers_) { |
| if (provider->HasInitialStabilityMetrics()) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| bool MetricsService::PrepareInitialStabilityLog( |
| const std::string& prefs_previous_version) { |
| DCHECK_EQ(INITIALIZED, state_); |
| |
| std::unique_ptr<MetricsLog> initial_stability_log( |
| CreateLog(MetricsLog::INITIAL_STABILITY_LOG)); |
| |
| // Do not call NotifyOnDidCreateMetricsLog here because the stability |
| // log describes stats from the _previous_ session. |
| std::string system_profile_app_version; |
| if (!initial_stability_log->LoadSavedEnvironmentFromPrefs( |
| &system_profile_app_version)) { |
| return false; |
| } |
| if (system_profile_app_version != prefs_previous_version) |
| IncrementPrefValue(prefs::kStabilityVersionMismatchCount); |
| |
| log_manager_.PauseCurrentLog(); |
| log_manager_.BeginLoggingWithLog(std::move(initial_stability_log)); |
| |
| // Note: Some stability providers may record stability stats via histograms, |
| // so this call has to be after BeginLoggingWithLog(). |
| log_manager_.current_log()->RecordStabilityMetrics( |
| metrics_providers_.get(), base::TimeDelta(), base::TimeDelta()); |
| RecordCurrentStabilityHistograms(); |
| |
| // Note: RecordGeneralMetrics() intentionally not called since this log is for |
| // stability stats from a previous session only. |
| |
| log_manager_.FinishCurrentLog(); |
| log_manager_.ResumePausedLog(); |
| |
| // Store unsent logs, including the stability log that was just saved, so |
| // that they're not lost in case of a crash before upload time. |
| log_manager_.PersistUnsentLogs(); |
| |
| return true; |
| } |
| |
| void MetricsService::PrepareInitialMetricsLog() { |
| DCHECK_EQ(INIT_TASK_DONE, state_); |
| |
| RecordCurrentEnvironment(initial_metrics_log_.get()); |
| base::TimeDelta incremental_uptime; |
| base::TimeDelta uptime; |
| GetUptimes(local_state_, &incremental_uptime, &uptime); |
| |
| // Histograms only get written to the current log, so make the new log current |
| // before writing them. |
| log_manager_.PauseCurrentLog(); |
| log_manager_.BeginLoggingWithLog(std::move(initial_metrics_log_)); |
| |
| // Note: Some stability providers may record stability stats via histograms, |
| // so this call has to be after BeginLoggingWithLog(). |
| MetricsLog* current_log = log_manager_.current_log(); |
| current_log->RecordStabilityMetrics(metrics_providers_.get(), |
| base::TimeDelta(), base::TimeDelta()); |
| current_log->RecordGeneralMetrics(metrics_providers_.get()); |
| RecordCurrentHistograms(); |
| |
| log_manager_.FinishCurrentLog(); |
| log_manager_.ResumePausedLog(); |
| |
| // Store unsent logs, including the initial log that was just saved, so |
| // that they're not lost in case of a crash before upload time. |
| log_manager_.PersistUnsentLogs(); |
| |
| state_ = SENDING_LOGS; |
| } |
| |
| void MetricsService::SendStagedLog() { |
| DCHECK(log_manager_.has_staged_log()); |
| if (!log_manager_.has_staged_log()) |
| return; |
| |
| DCHECK(!log_upload_in_progress_); |
| log_upload_in_progress_ = true; |
| |
| if (!log_uploader_) { |
| log_uploader_ = client_->CreateUploader( |
| base::Bind(&MetricsService::OnLogUploadComplete, |
| self_ptr_factory_.GetWeakPtr())); |
| } |
| |
| const std::string hash = |
| base::HexEncode(log_manager_.staged_log_hash().data(), |
| log_manager_.staged_log_hash().size()); |
| log_uploader_->UploadLog(log_manager_.staged_log(), hash); |
| |
| HandleIdleSinceLastTransmission(true); |
| } |
| |
| |
| void MetricsService::OnLogUploadComplete(int response_code) { |
| DCHECK_EQ(SENDING_LOGS, state_); |
| DCHECK(log_upload_in_progress_); |
| log_upload_in_progress_ = false; |
| |
| // Log a histogram to track response success vs. failure rates. |
| UMA_HISTOGRAM_ENUMERATION("UMA.UploadResponseStatus.Protobuf", |
| ResponseCodeToStatus(response_code), |
| NUM_RESPONSE_STATUSES); |
| |
| bool upload_succeeded = response_code == 200; |
| |
| // Provide boolean for error recovery (allow us to ignore response_code). |
| bool discard_log = false; |
| const size_t log_size = log_manager_.staged_log().length(); |
| if (upload_succeeded) { |
| UMA_HISTOGRAM_COUNTS_10000("UMA.LogSize.OnSuccess", log_size / 1024); |
| } else if (log_size > kUploadLogAvoidRetransmitSize) { |
| UMA_HISTOGRAM_COUNTS("UMA.Large Rejected Log was Discarded", |
| static_cast<int>(log_size)); |
| discard_log = true; |
| } else if (response_code == 400) { |
| // Bad syntax. Retransmission won't work. |
| discard_log = true; |
| } |
| |
| if (upload_succeeded || discard_log) { |
| log_manager_.DiscardStagedLog(); |
| // Store the updated list to disk now that the removed log is uploaded. |
| log_manager_.PersistUnsentLogs(); |
| } |
| |
| // Error 400 indicates a problem with the log, not with the server, so |
| // don't consider that a sign that the server is in trouble. |
| bool server_is_healthy = upload_succeeded || response_code == 400; |
| scheduler_->UploadFinished(server_is_healthy, log_manager_.has_unsent_logs()); |
| |
| if (server_is_healthy) |
| client_->OnLogUploadComplete(); |
| } |
| |
| void MetricsService::IncrementPrefValue(const char* path) { |
| int value = local_state_->GetInteger(path); |
| local_state_->SetInteger(path, value + 1); |
| } |
| |
| void MetricsService::IncrementLongPrefsValue(const char* path) { |
| int64_t value = local_state_->GetInt64(path); |
| local_state_->SetInt64(path, value + 1); |
| } |
| |
| bool MetricsService::UmaMetricsProperlyShutdown() { |
| CHECK(clean_shutdown_status_ == CLEANLY_SHUTDOWN || |
| clean_shutdown_status_ == NEED_TO_SHUTDOWN); |
| return clean_shutdown_status_ == CLEANLY_SHUTDOWN; |
| } |
| |
| void MetricsService::AddSyntheticTrialObserver( |
| variations::SyntheticTrialObserver* observer) { |
| synthetic_trial_observer_list_.AddObserver(observer); |
| if (!synthetic_trial_groups_.empty()) |
| observer->OnSyntheticTrialsChanged(synthetic_trial_groups_); |
| } |
| |
| void MetricsService::RemoveSyntheticTrialObserver( |
| variations::SyntheticTrialObserver* observer) { |
| synthetic_trial_observer_list_.RemoveObserver(observer); |
| } |
| |
| void MetricsService::RegisterSyntheticFieldTrial( |
| const variations::SyntheticTrialGroup& trial) { |
| for (size_t i = 0; i < synthetic_trial_groups_.size(); ++i) { |
| if (synthetic_trial_groups_[i].id.name == trial.id.name) { |
| if (synthetic_trial_groups_[i].id.group != trial.id.group) { |
| synthetic_trial_groups_[i].id.group = trial.id.group; |
| synthetic_trial_groups_[i].start_time = base::TimeTicks::Now(); |
| NotifySyntheticTrialObservers(); |
| } |
| return; |
| } |
| } |
| |
| variations::SyntheticTrialGroup trial_group = trial; |
| trial_group.start_time = base::TimeTicks::Now(); |
| synthetic_trial_groups_.push_back(trial_group); |
| NotifySyntheticTrialObservers(); |
| } |
| |
| void MetricsService::RegisterSyntheticMultiGroupFieldTrial( |
| uint32_t trial_name_hash, |
| const std::vector<uint32_t>& group_name_hashes) { |
| auto has_same_trial_name = |
| [trial_name_hash](const variations::SyntheticTrialGroup& x) { |
| return x.id.name == trial_name_hash; |
| }; |
| synthetic_trial_groups_.erase( |
| std::remove_if(synthetic_trial_groups_.begin(), |
| synthetic_trial_groups_.end(), has_same_trial_name), |
| synthetic_trial_groups_.end()); |
| |
| if (group_name_hashes.empty()) |
| return; |
| |
| variations::SyntheticTrialGroup trial_group(trial_name_hash, |
| group_name_hashes[0]); |
| trial_group.start_time = base::TimeTicks::Now(); |
| for (uint32_t group_name_hash : group_name_hashes) { |
| // Note: Adding the trial group will copy it, so this re-uses the same |
| // |trial_group| struct for convenience (e.g. so start_time's all match). |
| trial_group.id.group = group_name_hash; |
| synthetic_trial_groups_.push_back(trial_group); |
| } |
| NotifySyntheticTrialObservers(); |
| } |
| |
| void MetricsService::GetCurrentSyntheticFieldTrialsForTesting( |
| std::vector<variations::ActiveGroupId>* synthetic_trials) { |
| GetSyntheticFieldTrialsOlderThan(base::TimeTicks::Now(), synthetic_trials); |
| } |
| |
| void MetricsService::RegisterMetricsProvider( |
| std::unique_ptr<MetricsProvider> provider) { |
| DCHECK_EQ(INITIALIZED, state_); |
| metrics_providers_.push_back(std::move(provider)); |
| } |
| |
| void MetricsService::CheckForClonedInstall( |
| scoped_refptr<base::SingleThreadTaskRunner> task_runner) { |
| state_manager_->CheckForClonedInstall(task_runner); |
| } |
| |
| void MetricsService::NotifySyntheticTrialObservers() { |
| FOR_EACH_OBSERVER(variations::SyntheticTrialObserver, |
| synthetic_trial_observer_list_, |
| OnSyntheticTrialsChanged(synthetic_trial_groups_)); |
| } |
| |
| void MetricsService::GetSyntheticFieldTrialsOlderThan( |
| base::TimeTicks time, |
| std::vector<variations::ActiveGroupId>* synthetic_trials) { |
| DCHECK(synthetic_trials); |
| synthetic_trials->clear(); |
| for (size_t i = 0; i < synthetic_trial_groups_.size(); ++i) { |
| if (synthetic_trial_groups_[i].start_time <= time) |
| synthetic_trials->push_back(synthetic_trial_groups_[i].id); |
| } |
| } |
| |
| std::unique_ptr<MetricsLog> MetricsService::CreateLog( |
| MetricsLog::LogType log_type) { |
| return base::MakeUnique<MetricsLog>(state_manager_->client_id(), session_id_, |
| log_type, client_, local_state_); |
| } |
| |
| void MetricsService::RecordCurrentEnvironment(MetricsLog* log) { |
| std::vector<variations::ActiveGroupId> synthetic_trials; |
| GetSyntheticFieldTrialsOlderThan(log->creation_time(), &synthetic_trials); |
| log->RecordEnvironment(metrics_providers_.get(), synthetic_trials, |
| GetInstallDate(), GetMetricsReportingEnabledDate()); |
| } |
| |
| void MetricsService::RecordCurrentHistograms() { |
| DCHECK(log_manager_.current_log()); |
| SCOPED_UMA_HISTOGRAM_TIMER("UMA.MetricsService.RecordCurrentHistograms.Time"); |
| |
| // "true" to the begin() call indicates that StatisticsRecorder should include |
| // histograms held in persistent storage. |
| histogram_snapshot_manager_.PrepareDeltas( |
| base::StatisticsRecorder::begin(true), base::StatisticsRecorder::end(), |
| base::Histogram::kNoFlags, base::Histogram::kUmaTargetedHistogramFlag); |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->RecordHistogramSnapshots(&histogram_snapshot_manager_); |
| } |
| |
| void MetricsService::RecordCurrentStabilityHistograms() { |
| DCHECK(log_manager_.current_log()); |
| // "true" indicates that StatisticsRecorder should include histograms in |
| // persistent storage. |
| histogram_snapshot_manager_.PrepareDeltas( |
| base::StatisticsRecorder::begin(true), base::StatisticsRecorder::end(), |
| base::Histogram::kNoFlags, base::Histogram::kUmaStabilityHistogramFlag); |
| for (MetricsProvider* provider : metrics_providers_) |
| provider->RecordInitialHistogramSnapshots(&histogram_snapshot_manager_); |
| } |
| |
| void MetricsService::LogCleanShutdown() { |
| // Redundant setting to assure that we always reset this value at shutdown |
| // (and that we don't use some alternate path, and not call LogCleanShutdown). |
| clean_shutdown_status_ = CLEANLY_SHUTDOWN; |
| |
| clean_exit_beacon_.WriteBeaconValue(true); |
| RecordCurrentState(local_state_); |
| local_state_->SetInteger(prefs::kStabilityExecutionPhase, |
| MetricsService::SHUTDOWN_COMPLETE); |
| } |
| |
| void MetricsService::RecordBooleanPrefValue(const char* path, bool value) { |
| DCHECK(IsSingleThreaded()); |
| local_state_->SetBoolean(path, value); |
| RecordCurrentState(local_state_); |
| } |
| |
| void MetricsService::RecordCurrentState(PrefService* pref) { |
| pref->SetInt64(prefs::kStabilityLastTimestampSec, |
| base::Time::Now().ToTimeT()); |
| } |
| |
| void MetricsService::SkipAndDiscardUpload() { |
| log_manager_.DiscardStagedLog(); |
| scheduler_->UploadCancelled(); |
| log_upload_in_progress_ = false; |
| } |
| |
| } // namespace metrics |