|  | // Copyright 2014 The Chromium Authors. All rights reserved. | 
|  | // Use of this source code is governed by a BSD-style license that can be | 
|  | // found in the LICENSE file. | 
|  |  | 
|  | #include "extensions/browser/computed_hashes.h" | 
|  |  | 
|  | #include <memory> | 
|  | #include <utility> | 
|  |  | 
|  | #include "base/base64.h" | 
|  | #include "base/files/file_path.h" | 
|  | #include "base/files/file_util.h" | 
|  | #include "base/json/json_reader.h" | 
|  | #include "base/json/json_writer.h" | 
|  | #include "base/metrics/histogram_macros.h" | 
|  | #include "base/stl_util.h" | 
|  | #include "base/timer/elapsed_timer.h" | 
|  | #include "base/values.h" | 
|  | #include "crypto/secure_hash.h" | 
|  | #include "crypto/sha2.h" | 
|  |  | 
|  | namespace extensions { | 
|  |  | 
|  | namespace computed_hashes { | 
|  | const char kBlockHashesKey[] = "block_hashes"; | 
|  | const char kBlockSizeKey[] = "block_size"; | 
|  | const char kFileHashesKey[] = "file_hashes"; | 
|  | const char kPathKey[] = "path"; | 
|  | const char kVersionKey[] = "version"; | 
|  | const int kVersion = 2; | 
|  | }  // namespace computed_hashes | 
|  |  | 
|  | namespace { | 
|  |  | 
|  | // Helper to record UMA for ComputedHashes::Reader::InitFromFile. | 
|  | // Records failure UMA if RecordSuccess() isn't explicitly called. | 
|  | class ScopedUMARecorder { | 
|  | public: | 
|  | ScopedUMARecorder() = default; | 
|  |  | 
|  | ~ScopedUMARecorder() { | 
|  | if (recorded_) | 
|  | return; | 
|  | RecordImpl(false); | 
|  | } | 
|  |  | 
|  | void RecordSuccess() { | 
|  | recorded_ = true; | 
|  | RecordImpl(true); | 
|  | } | 
|  |  | 
|  | private: | 
|  | void RecordImpl(bool succeeded) { | 
|  | UMA_HISTOGRAM_BOOLEAN( | 
|  | "Extensions.ContentVerification.ComputedHashesReadResult", succeeded); | 
|  | if (succeeded) { | 
|  | UMA_HISTOGRAM_TIMES( | 
|  | "Extensions.ContentVerification.ComputedHashesInitTime", | 
|  | timer_.Elapsed()); | 
|  | } | 
|  | } | 
|  |  | 
|  | bool recorded_ = false; | 
|  | base::ElapsedTimer timer_; | 
|  | DISALLOW_COPY_AND_ASSIGN(ScopedUMARecorder); | 
|  | }; | 
|  |  | 
|  | }  // namespace | 
|  |  | 
|  | ComputedHashes::Reader::Reader() { | 
|  | } | 
|  |  | 
|  | ComputedHashes::Reader::~Reader() { | 
|  | } | 
|  |  | 
|  | bool ComputedHashes::Reader::InitFromFile(const base::FilePath& path) { | 
|  | ScopedUMARecorder uma_recorder; | 
|  | std::string contents; | 
|  | if (!base::ReadFileToString(path, &contents)) | 
|  | return false; | 
|  |  | 
|  | base::DictionaryValue* top_dictionary = NULL; | 
|  | std::unique_ptr<base::Value> value(base::JSONReader::Read(contents)); | 
|  | if (!value.get() || !value->GetAsDictionary(&top_dictionary)) | 
|  | return false; | 
|  |  | 
|  | // For now we don't support forwards or backwards compatability in the | 
|  | // format, so we return false on version mismatch. | 
|  | int version = 0; | 
|  | if (!top_dictionary->GetInteger(computed_hashes::kVersionKey, &version) || | 
|  | version != computed_hashes::kVersion) | 
|  | return false; | 
|  |  | 
|  | base::ListValue* all_hashes = NULL; | 
|  | if (!top_dictionary->GetList(computed_hashes::kFileHashesKey, &all_hashes)) | 
|  | return false; | 
|  |  | 
|  | for (size_t i = 0; i < all_hashes->GetSize(); i++) { | 
|  | base::DictionaryValue* dictionary = NULL; | 
|  | if (!all_hashes->GetDictionary(i, &dictionary)) | 
|  | return false; | 
|  |  | 
|  | std::string relative_path_utf8; | 
|  | if (!dictionary->GetString(computed_hashes::kPathKey, &relative_path_utf8)) | 
|  | return false; | 
|  |  | 
|  | int block_size; | 
|  | if (!dictionary->GetInteger(computed_hashes::kBlockSizeKey, &block_size)) | 
|  | return false; | 
|  | if (block_size <= 0 || ((block_size % 1024) != 0)) { | 
|  | LOG(ERROR) << "Invalid block size: " << block_size; | 
|  | return false; | 
|  | } | 
|  |  | 
|  | base::ListValue* hashes_list = NULL; | 
|  | if (!dictionary->GetList(computed_hashes::kBlockHashesKey, &hashes_list)) | 
|  | return false; | 
|  |  | 
|  | base::FilePath relative_path = | 
|  | base::FilePath::FromUTF8Unsafe(relative_path_utf8); | 
|  | relative_path = relative_path.NormalizePathSeparatorsTo('/'); | 
|  |  | 
|  | data_[relative_path] = HashInfo(block_size, std::vector<std::string>()); | 
|  | std::vector<std::string>* hashes = &(data_[relative_path].second); | 
|  |  | 
|  | for (size_t j = 0; j < hashes_list->GetSize(); j++) { | 
|  | std::string encoded; | 
|  | if (!hashes_list->GetString(j, &encoded)) | 
|  | return false; | 
|  |  | 
|  | hashes->push_back(std::string()); | 
|  | std::string* decoded = &hashes->back(); | 
|  | if (!base::Base64Decode(encoded, decoded)) { | 
|  | hashes->clear(); | 
|  | return false; | 
|  | } | 
|  | } | 
|  | } | 
|  | uma_recorder.RecordSuccess(); | 
|  | return true; | 
|  | } | 
|  |  | 
|  | bool ComputedHashes::Reader::GetHashes(const base::FilePath& relative_path, | 
|  | int* block_size, | 
|  | std::vector<std::string>* hashes) const { | 
|  | base::FilePath path = relative_path.NormalizePathSeparatorsTo('/'); | 
|  | std::map<base::FilePath, HashInfo>::const_iterator i = data_.find(path); | 
|  | if (i == data_.end()) { | 
|  | // If we didn't find the entry using exact match, it's possible the | 
|  | // developer is using a path with some letters in the incorrect case, which | 
|  | // happens to work on windows/osx. So try doing a linear scan to look for a | 
|  | // case-insensitive match. In practice most extensions don't have that big | 
|  | // a list of files so the performance penalty is probably not too big | 
|  | // here. Also for crbug.com/29941 we plan to start warning developers when | 
|  | // they are making this mistake, since their extension will be broken on | 
|  | // linux/chromeos. | 
|  | for (i = data_.begin(); i != data_.end(); ++i) { | 
|  | const base::FilePath& entry = i->first; | 
|  | if (base::FilePath::CompareEqualIgnoreCase(entry.value(), path.value())) | 
|  | break; | 
|  | } | 
|  | if (i == data_.end()) | 
|  | return false; | 
|  | } | 
|  | const HashInfo& info = i->second; | 
|  | *block_size = info.first; | 
|  | *hashes = info.second; | 
|  | return true; | 
|  | } | 
|  |  | 
|  | ComputedHashes::Writer::Writer() : file_list_(new base::ListValue) { | 
|  | } | 
|  |  | 
|  | ComputedHashes::Writer::~Writer() { | 
|  | } | 
|  |  | 
|  | void ComputedHashes::Writer::AddHashes(const base::FilePath& relative_path, | 
|  | int block_size, | 
|  | const std::vector<std::string>& hashes) { | 
|  | auto block_hashes = std::make_unique<base::ListValue>(); | 
|  | block_hashes->GetList().reserve(hashes.size()); | 
|  | for (const auto& hash : hashes) { | 
|  | std::string encoded; | 
|  | base::Base64Encode(hash, &encoded); | 
|  | block_hashes->GetList().emplace_back(std::move(encoded)); | 
|  | } | 
|  |  | 
|  | auto dict = std::make_unique<base::DictionaryValue>(); | 
|  | dict->SetString(computed_hashes::kPathKey, | 
|  | relative_path.NormalizePathSeparatorsTo('/').AsUTF8Unsafe()); | 
|  | dict->SetInteger(computed_hashes::kBlockSizeKey, block_size); | 
|  | dict->Set(computed_hashes::kBlockHashesKey, std::move(block_hashes)); | 
|  | file_list_->Append(std::move(dict)); | 
|  | } | 
|  |  | 
|  | bool ComputedHashes::Writer::WriteToFile(const base::FilePath& path) { | 
|  | std::string json; | 
|  | base::DictionaryValue top_dictionary; | 
|  | top_dictionary.SetInteger(computed_hashes::kVersionKey, | 
|  | computed_hashes::kVersion); | 
|  | top_dictionary.Set(computed_hashes::kFileHashesKey, std::move(file_list_)); | 
|  |  | 
|  | if (!base::JSONWriter::Write(top_dictionary, &json)) | 
|  | return false; | 
|  | int written = base::WriteFile(path, json.data(), json.size()); | 
|  | if (static_cast<unsigned>(written) != json.size()) { | 
|  | LOG(ERROR) << "Error writing " << path.AsUTF8Unsafe() | 
|  | << " ; write result:" << written << " expected:" << json.size(); | 
|  | return false; | 
|  | } | 
|  | return true; | 
|  | } | 
|  |  | 
|  | void ComputedHashes::ComputeHashesForContent(const std::string& contents, | 
|  | size_t block_size, | 
|  | std::vector<std::string>* hashes) { | 
|  | size_t offset = 0; | 
|  | // Even when the contents is empty, we want to output at least one hash | 
|  | // block (the hash of the empty string). | 
|  | do { | 
|  | const char* block_start = contents.data() + offset; | 
|  | DCHECK(offset <= contents.size()); | 
|  | size_t bytes_to_read = std::min(contents.size() - offset, block_size); | 
|  | std::unique_ptr<crypto::SecureHash> hash( | 
|  | crypto::SecureHash::Create(crypto::SecureHash::SHA256)); | 
|  | hash->Update(block_start, bytes_to_read); | 
|  |  | 
|  | hashes->push_back(std::string()); | 
|  | std::string* buffer = &(hashes->back()); | 
|  | buffer->resize(crypto::kSHA256Length); | 
|  | hash->Finish(base::data(*buffer), buffer->size()); | 
|  |  | 
|  | // If |contents| is empty, then we want to just exit here. | 
|  | if (bytes_to_read == 0) | 
|  | break; | 
|  |  | 
|  | offset += bytes_to_read; | 
|  | } while (offset < contents.size()); | 
|  | } | 
|  |  | 
|  | }  // namespace extensions |