| // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "base/prefs/json_pref_store.h" |
| |
| #include <algorithm> |
| |
| #include "base/bind.h" |
| #include "base/callback.h" |
| #include "base/files/file_path.h" |
| #include "base/files/file_util.h" |
| #include "base/json/json_file_value_serializer.h" |
| #include "base/json/json_string_value_serializer.h" |
| #include "base/memory/ref_counted.h" |
| #include "base/metrics/histogram.h" |
| #include "base/prefs/pref_filter.h" |
| #include "base/sequenced_task_runner.h" |
| #include "base/strings/string_number_conversions.h" |
| #include "base/strings/string_util.h" |
| #include "base/task_runner_util.h" |
| #include "base/threading/sequenced_worker_pool.h" |
| #include "base/time/default_clock.h" |
| #include "base/values.h" |
| |
| // Result returned from internal read tasks. |
| struct JsonPrefStore::ReadResult { |
| public: |
| ReadResult(); |
| ~ReadResult(); |
| |
| scoped_ptr<base::Value> value; |
| PrefReadError error; |
| bool no_dir; |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(ReadResult); |
| }; |
| |
| JsonPrefStore::ReadResult::ReadResult() |
| : error(PersistentPrefStore::PREF_READ_ERROR_NONE), no_dir(false) { |
| } |
| |
| JsonPrefStore::ReadResult::~ReadResult() { |
| } |
| |
| namespace { |
| |
| // Some extensions we'll tack on to copies of the Preferences files. |
| const base::FilePath::CharType kBadExtension[] = FILE_PATH_LITERAL("bad"); |
| |
| PersistentPrefStore::PrefReadError HandleReadErrors( |
| const base::Value* value, |
| const base::FilePath& path, |
| int error_code, |
| const std::string& error_msg) { |
| if (!value) { |
| DVLOG(1) << "Error while loading JSON file: " << error_msg |
| << ", file: " << path.value(); |
| switch (error_code) { |
| case JSONFileValueDeserializer::JSON_ACCESS_DENIED: |
| return PersistentPrefStore::PREF_READ_ERROR_ACCESS_DENIED; |
| break; |
| case JSONFileValueDeserializer::JSON_CANNOT_READ_FILE: |
| return PersistentPrefStore::PREF_READ_ERROR_FILE_OTHER; |
| break; |
| case JSONFileValueDeserializer::JSON_FILE_LOCKED: |
| return PersistentPrefStore::PREF_READ_ERROR_FILE_LOCKED; |
| break; |
| case JSONFileValueDeserializer::JSON_NO_SUCH_FILE: |
| return PersistentPrefStore::PREF_READ_ERROR_NO_FILE; |
| break; |
| default: |
| // JSON errors indicate file corruption of some sort. |
| // Since the file is corrupt, move it to the side and continue with |
| // empty preferences. This will result in them losing their settings. |
| // We keep the old file for possible support and debugging assistance |
| // as well as to detect if they're seeing these errors repeatedly. |
| // TODO(erikkay) Instead, use the last known good file. |
| base::FilePath bad = path.ReplaceExtension(kBadExtension); |
| |
| // If they've ever had a parse error before, put them in another bucket. |
| // TODO(erikkay) if we keep this error checking for very long, we may |
| // want to differentiate between recent and long ago errors. |
| bool bad_existed = base::PathExists(bad); |
| base::Move(path, bad); |
| return bad_existed ? PersistentPrefStore::PREF_READ_ERROR_JSON_REPEAT |
| : PersistentPrefStore::PREF_READ_ERROR_JSON_PARSE; |
| } |
| } else if (!value->IsType(base::Value::TYPE_DICTIONARY)) { |
| return PersistentPrefStore::PREF_READ_ERROR_JSON_TYPE; |
| } |
| return PersistentPrefStore::PREF_READ_ERROR_NONE; |
| } |
| |
| // Records a sample for |size| in the Settings.JsonDataReadSizeKilobytes |
| // histogram suffixed with the base name of the JSON file under |path|. |
| void RecordJsonDataSizeHistogram(const base::FilePath& path, size_t size) { |
| std::string spaceless_basename; |
| base::ReplaceChars(path.BaseName().MaybeAsASCII(), " ", "_", |
| &spaceless_basename); |
| |
| // The histogram below is an expansion of the UMA_HISTOGRAM_CUSTOM_COUNTS |
| // macro adapted to allow for a dynamically suffixed histogram name. |
| // Note: The factory creates and owns the histogram. |
| base::HistogramBase* histogram = base::Histogram::FactoryGet( |
| "Settings.JsonDataReadSizeKilobytes." + spaceless_basename, 1, 10000, 50, |
| base::HistogramBase::kUmaTargetedHistogramFlag); |
| histogram->Add(static_cast<int>(size) / 1024); |
| } |
| |
| scoped_ptr<JsonPrefStore::ReadResult> ReadPrefsFromDisk( |
| const base::FilePath& path, |
| const base::FilePath& alternate_path) { |
| if (!base::PathExists(path) && !alternate_path.empty() && |
| base::PathExists(alternate_path)) { |
| base::Move(alternate_path, path); |
| } |
| |
| int error_code; |
| std::string error_msg; |
| scoped_ptr<JsonPrefStore::ReadResult> read_result( |
| new JsonPrefStore::ReadResult); |
| JSONFileValueDeserializer deserializer(path); |
| read_result->value.reset(deserializer.Deserialize(&error_code, &error_msg)); |
| read_result->error = |
| HandleReadErrors(read_result->value.get(), path, error_code, error_msg); |
| read_result->no_dir = !base::PathExists(path.DirName()); |
| |
| if (read_result->error == PersistentPrefStore::PREF_READ_ERROR_NONE) |
| RecordJsonDataSizeHistogram(path, deserializer.get_last_read_size()); |
| |
| return read_result.Pass(); |
| } |
| |
| } // namespace |
| |
| // static |
| scoped_refptr<base::SequencedTaskRunner> JsonPrefStore::GetTaskRunnerForFile( |
| const base::FilePath& filename, |
| base::SequencedWorkerPool* worker_pool) { |
| std::string token("json_pref_store-"); |
| token.append(filename.AsUTF8Unsafe()); |
| return worker_pool->GetSequencedTaskRunnerWithShutdownBehavior( |
| worker_pool->GetNamedSequenceToken(token), |
| base::SequencedWorkerPool::BLOCK_SHUTDOWN); |
| } |
| |
| JsonPrefStore::JsonPrefStore( |
| const base::FilePath& filename, |
| const scoped_refptr<base::SequencedTaskRunner>& sequenced_task_runner, |
| scoped_ptr<PrefFilter> pref_filter) |
| : JsonPrefStore(filename, |
| base::FilePath(), |
| sequenced_task_runner, |
| pref_filter.Pass()) { |
| } |
| |
| JsonPrefStore::JsonPrefStore( |
| const base::FilePath& filename, |
| const base::FilePath& alternate_filename, |
| const scoped_refptr<base::SequencedTaskRunner>& sequenced_task_runner, |
| scoped_ptr<PrefFilter> pref_filter) |
| : path_(filename), |
| alternate_path_(alternate_filename), |
| sequenced_task_runner_(sequenced_task_runner), |
| prefs_(new base::DictionaryValue()), |
| read_only_(false), |
| writer_(filename, sequenced_task_runner), |
| pref_filter_(pref_filter.Pass()), |
| initialized_(false), |
| filtering_in_progress_(false), |
| pending_lossy_write_(false), |
| read_error_(PREF_READ_ERROR_NONE), |
| write_count_histogram_(writer_.commit_interval(), path_) { |
| DCHECK(!path_.empty()); |
| } |
| |
| bool JsonPrefStore::GetValue(const std::string& key, |
| const base::Value** result) const { |
| DCHECK(CalledOnValidThread()); |
| |
| base::Value* tmp = NULL; |
| if (!prefs_->Get(key, &tmp)) |
| return false; |
| |
| if (result) |
| *result = tmp; |
| return true; |
| } |
| |
| void JsonPrefStore::AddObserver(PrefStore::Observer* observer) { |
| DCHECK(CalledOnValidThread()); |
| |
| observers_.AddObserver(observer); |
| } |
| |
| void JsonPrefStore::RemoveObserver(PrefStore::Observer* observer) { |
| DCHECK(CalledOnValidThread()); |
| |
| observers_.RemoveObserver(observer); |
| } |
| |
| bool JsonPrefStore::HasObservers() const { |
| DCHECK(CalledOnValidThread()); |
| |
| return observers_.might_have_observers(); |
| } |
| |
| bool JsonPrefStore::IsInitializationComplete() const { |
| DCHECK(CalledOnValidThread()); |
| |
| return initialized_; |
| } |
| |
| bool JsonPrefStore::GetMutableValue(const std::string& key, |
| base::Value** result) { |
| DCHECK(CalledOnValidThread()); |
| |
| return prefs_->Get(key, result); |
| } |
| |
| void JsonPrefStore::SetValue(const std::string& key, |
| scoped_ptr<base::Value> value, |
| uint32 flags) { |
| DCHECK(CalledOnValidThread()); |
| |
| DCHECK(value); |
| base::Value* old_value = NULL; |
| prefs_->Get(key, &old_value); |
| if (!old_value || !value->Equals(old_value)) { |
| prefs_->Set(key, value.Pass()); |
| ReportValueChanged(key, flags); |
| } |
| } |
| |
| void JsonPrefStore::SetValueSilently(const std::string& key, |
| scoped_ptr<base::Value> value, |
| uint32 flags) { |
| DCHECK(CalledOnValidThread()); |
| |
| DCHECK(value); |
| base::Value* old_value = NULL; |
| prefs_->Get(key, &old_value); |
| if (!old_value || !value->Equals(old_value)) { |
| prefs_->Set(key, value.Pass()); |
| ScheduleWrite(flags); |
| } |
| } |
| |
| void JsonPrefStore::RemoveValue(const std::string& key, uint32 flags) { |
| DCHECK(CalledOnValidThread()); |
| |
| if (prefs_->RemovePath(key, NULL)) |
| ReportValueChanged(key, flags); |
| } |
| |
| void JsonPrefStore::RemoveValueSilently(const std::string& key, uint32 flags) { |
| DCHECK(CalledOnValidThread()); |
| |
| prefs_->RemovePath(key, NULL); |
| ScheduleWrite(flags); |
| } |
| |
| bool JsonPrefStore::ReadOnly() const { |
| DCHECK(CalledOnValidThread()); |
| |
| return read_only_; |
| } |
| |
| PersistentPrefStore::PrefReadError JsonPrefStore::GetReadError() const { |
| DCHECK(CalledOnValidThread()); |
| |
| return read_error_; |
| } |
| |
| PersistentPrefStore::PrefReadError JsonPrefStore::ReadPrefs() { |
| DCHECK(CalledOnValidThread()); |
| |
| OnFileRead(ReadPrefsFromDisk(path_, alternate_path_)); |
| return filtering_in_progress_ ? PREF_READ_ERROR_ASYNCHRONOUS_TASK_INCOMPLETE |
| : read_error_; |
| } |
| |
| void JsonPrefStore::ReadPrefsAsync(ReadErrorDelegate* error_delegate) { |
| DCHECK(CalledOnValidThread()); |
| |
| initialized_ = false; |
| error_delegate_.reset(error_delegate); |
| |
| // Weakly binds the read task so that it doesn't kick in during shutdown. |
| base::PostTaskAndReplyWithResult( |
| sequenced_task_runner_.get(), |
| FROM_HERE, |
| base::Bind(&ReadPrefsFromDisk, path_, alternate_path_), |
| base::Bind(&JsonPrefStore::OnFileRead, AsWeakPtr())); |
| } |
| |
| void JsonPrefStore::CommitPendingWrite() { |
| DCHECK(CalledOnValidThread()); |
| |
| // Schedule a write for any lossy writes that are outstanding to ensure that |
| // they get flushed when this function is called. |
| SchedulePendingLossyWrites(); |
| |
| if (writer_.HasPendingWrite() && !read_only_) |
| writer_.DoScheduledWrite(); |
| } |
| |
| void JsonPrefStore::SchedulePendingLossyWrites() { |
| if (pending_lossy_write_) |
| writer_.ScheduleWrite(this); |
| } |
| |
| void JsonPrefStore::ReportValueChanged(const std::string& key, uint32 flags) { |
| DCHECK(CalledOnValidThread()); |
| |
| if (pref_filter_) |
| pref_filter_->FilterUpdate(key); |
| |
| FOR_EACH_OBSERVER(PrefStore::Observer, observers_, OnPrefValueChanged(key)); |
| |
| ScheduleWrite(flags); |
| } |
| |
| void JsonPrefStore::RegisterOnNextSuccessfulWriteCallback( |
| const base::Closure& on_next_successful_write) { |
| DCHECK(CalledOnValidThread()); |
| |
| writer_.RegisterOnNextSuccessfulWriteCallback(on_next_successful_write); |
| } |
| |
| void JsonPrefStore::OnFileRead(scoped_ptr<ReadResult> read_result) { |
| DCHECK(CalledOnValidThread()); |
| |
| DCHECK(read_result); |
| |
| scoped_ptr<base::DictionaryValue> unfiltered_prefs(new base::DictionaryValue); |
| |
| read_error_ = read_result->error; |
| |
| bool initialization_successful = !read_result->no_dir; |
| |
| if (initialization_successful) { |
| switch (read_error_) { |
| case PREF_READ_ERROR_ACCESS_DENIED: |
| case PREF_READ_ERROR_FILE_OTHER: |
| case PREF_READ_ERROR_FILE_LOCKED: |
| case PREF_READ_ERROR_JSON_TYPE: |
| case PREF_READ_ERROR_FILE_NOT_SPECIFIED: |
| read_only_ = true; |
| break; |
| case PREF_READ_ERROR_NONE: |
| DCHECK(read_result->value.get()); |
| unfiltered_prefs.reset( |
| static_cast<base::DictionaryValue*>(read_result->value.release())); |
| break; |
| case PREF_READ_ERROR_NO_FILE: |
| // If the file just doesn't exist, maybe this is first run. In any case |
| // there's no harm in writing out default prefs in this case. |
| break; |
| case PREF_READ_ERROR_JSON_PARSE: |
| case PREF_READ_ERROR_JSON_REPEAT: |
| break; |
| case PREF_READ_ERROR_ASYNCHRONOUS_TASK_INCOMPLETE: |
| // This is a special error code to be returned by ReadPrefs when it |
| // can't complete synchronously, it should never be returned by the read |
| // operation itself. |
| NOTREACHED(); |
| break; |
| case PREF_READ_ERROR_LEVELDB_IO: |
| case PREF_READ_ERROR_LEVELDB_CORRUPTION_READ_ONLY: |
| case PREF_READ_ERROR_LEVELDB_CORRUPTION: |
| // These are specific to LevelDBPrefStore. |
| NOTREACHED(); |
| case PREF_READ_ERROR_MAX_ENUM: |
| NOTREACHED(); |
| break; |
| } |
| } |
| |
| if (pref_filter_) { |
| filtering_in_progress_ = true; |
| const PrefFilter::PostFilterOnLoadCallback post_filter_on_load_callback( |
| base::Bind( |
| &JsonPrefStore::FinalizeFileRead, AsWeakPtr(), |
| initialization_successful)); |
| pref_filter_->FilterOnLoad(post_filter_on_load_callback, |
| unfiltered_prefs.Pass()); |
| } else { |
| FinalizeFileRead(initialization_successful, unfiltered_prefs.Pass(), false); |
| } |
| } |
| |
| JsonPrefStore::~JsonPrefStore() { |
| CommitPendingWrite(); |
| } |
| |
| bool JsonPrefStore::SerializeData(std::string* output) { |
| DCHECK(CalledOnValidThread()); |
| |
| pending_lossy_write_ = false; |
| |
| write_count_histogram_.RecordWriteOccured(); |
| |
| if (pref_filter_) |
| pref_filter_->FilterSerializeData(prefs_.get()); |
| |
| JSONStringValueSerializer serializer(output); |
| // Not pretty-printing prefs shrinks pref file size by ~30%. To obtain |
| // readable prefs for debugging purposes, you can dump your prefs into any |
| // command-line or online JSON pretty printing tool. |
| serializer.set_pretty_print(false); |
| return serializer.Serialize(*prefs_); |
| } |
| |
| void JsonPrefStore::FinalizeFileRead(bool initialization_successful, |
| scoped_ptr<base::DictionaryValue> prefs, |
| bool schedule_write) { |
| DCHECK(CalledOnValidThread()); |
| |
| filtering_in_progress_ = false; |
| |
| if (!initialization_successful) { |
| FOR_EACH_OBSERVER(PrefStore::Observer, |
| observers_, |
| OnInitializationCompleted(false)); |
| return; |
| } |
| |
| prefs_ = prefs.Pass(); |
| |
| initialized_ = true; |
| |
| if (schedule_write) |
| ScheduleWrite(DEFAULT_PREF_WRITE_FLAGS); |
| |
| if (error_delegate_ && read_error_ != PREF_READ_ERROR_NONE) |
| error_delegate_->OnError(read_error_); |
| |
| FOR_EACH_OBSERVER(PrefStore::Observer, |
| observers_, |
| OnInitializationCompleted(true)); |
| |
| return; |
| } |
| |
| void JsonPrefStore::ScheduleWrite(uint32 flags) { |
| if (read_only_) |
| return; |
| |
| if (flags & LOSSY_PREF_WRITE_FLAG) |
| pending_lossy_write_ = true; |
| else |
| writer_.ScheduleWrite(this); |
| } |
| |
| // NOTE: This value should NOT be changed without renaming the histogram |
| // otherwise it will create incompatible buckets. |
| const int32_t |
| JsonPrefStore::WriteCountHistogram::kHistogramWriteReportIntervalMins = 5; |
| |
| JsonPrefStore::WriteCountHistogram::WriteCountHistogram( |
| const base::TimeDelta& commit_interval, |
| const base::FilePath& path) |
| : WriteCountHistogram(commit_interval, |
| path, |
| scoped_ptr<base::Clock>(new base::DefaultClock)) { |
| } |
| |
| JsonPrefStore::WriteCountHistogram::WriteCountHistogram( |
| const base::TimeDelta& commit_interval, |
| const base::FilePath& path, |
| scoped_ptr<base::Clock> clock) |
| : commit_interval_(commit_interval), |
| path_(path), |
| clock_(clock.release()), |
| report_interval_( |
| base::TimeDelta::FromMinutes(kHistogramWriteReportIntervalMins)), |
| last_report_time_(clock_->Now()), |
| writes_since_last_report_(0) { |
| } |
| |
| JsonPrefStore::WriteCountHistogram::~WriteCountHistogram() { |
| ReportOutstandingWrites(); |
| } |
| |
| void JsonPrefStore::WriteCountHistogram::RecordWriteOccured() { |
| ReportOutstandingWrites(); |
| |
| ++writes_since_last_report_; |
| } |
| |
| void JsonPrefStore::WriteCountHistogram::ReportOutstandingWrites() { |
| base::Time current_time = clock_->Now(); |
| base::TimeDelta time_since_last_report = current_time - last_report_time_; |
| |
| if (time_since_last_report <= report_interval_) |
| return; |
| |
| // If the time since the last report exceeds the report interval, report all |
| // the writes since the last report. They must have all occurred in the same |
| // report interval. |
| base::HistogramBase* histogram = GetHistogram(); |
| histogram->Add(writes_since_last_report_); |
| |
| // There may be several report intervals that elapsed that don't have any |
| // writes in them. Report these too. |
| int64 total_num_intervals_elapsed = |
| (time_since_last_report / report_interval_); |
| for (int64 i = 0; i < total_num_intervals_elapsed - 1; ++i) |
| histogram->Add(0); |
| |
| writes_since_last_report_ = 0; |
| last_report_time_ += total_num_intervals_elapsed * report_interval_; |
| } |
| |
| base::HistogramBase* JsonPrefStore::WriteCountHistogram::GetHistogram() { |
| std::string spaceless_basename; |
| base::ReplaceChars(path_.BaseName().MaybeAsASCII(), " ", "_", |
| &spaceless_basename); |
| std::string histogram_name = |
| "Settings.JsonDataWriteCount." + spaceless_basename; |
| |
| // The min value for a histogram is 1. The max value is the maximum number of |
| // writes that can occur in the window being recorded. The number of buckets |
| // used is the max value (plus the underflow/overflow buckets). |
| int32_t min_value = 1; |
| int32_t max_value = report_interval_ / commit_interval_; |
| int32_t num_buckets = max_value + 1; |
| |
| // NOTE: These values should NOT be changed without renaming the histogram |
| // otherwise it will create incompatible buckets. |
| DCHECK_EQ(30, max_value); |
| DCHECK_EQ(31, num_buckets); |
| |
| // The histogram below is an expansion of the UMA_HISTOGRAM_CUSTOM_COUNTS |
| // macro adapted to allow for a dynamically suffixed histogram name. |
| // Note: The factory creates and owns the histogram. |
| base::HistogramBase* histogram = base::Histogram::FactoryGet( |
| histogram_name, min_value, max_value, num_buckets, |
| base::HistogramBase::kUmaTargetedHistogramFlag); |
| return histogram; |
| } |