blob: 714b9f3be2965a3dbc3964385d49830d60021c2f [file] [log] [blame]
// Copyright 2014 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "extensions/browser/computed_hashes.h"
#include <memory>
#include <utility>
#include "base/base64.h"
#include "base/files/file_path.h"
#include "base/files/file_util.h"
#include "base/json/json_reader.h"
#include "base/json/json_writer.h"
#include "base/metrics/histogram_macros.h"
#include "base/stl_util.h"
#include "base/timer/elapsed_timer.h"
#include "base/values.h"
#include "crypto/secure_hash.h"
#include "crypto/sha2.h"
namespace extensions {
namespace computed_hashes {
const char kBlockHashesKey[] = "block_hashes";
const char kBlockSizeKey[] = "block_size";
const char kFileHashesKey[] = "file_hashes";
const char kPathKey[] = "path";
const char kVersionKey[] = "version";
const int kVersion = 2;
} // namespace computed_hashes
namespace {
// Helper to record UMA for ComputedHashes::Reader::InitFromFile.
// Records failure UMA if RecordSuccess() isn't explicitly called.
class ScopedUMARecorder {
public:
ScopedUMARecorder() = default;
~ScopedUMARecorder() {
if (recorded_)
return;
RecordImpl(false);
}
void RecordSuccess() {
recorded_ = true;
RecordImpl(true);
}
private:
void RecordImpl(bool succeeded) {
UMA_HISTOGRAM_BOOLEAN(
"Extensions.ContentVerification.ComputedHashesReadResult", succeeded);
if (succeeded) {
UMA_HISTOGRAM_TIMES(
"Extensions.ContentVerification.ComputedHashesInitTime",
timer_.Elapsed());
}
}
bool recorded_ = false;
base::ElapsedTimer timer_;
DISALLOW_COPY_AND_ASSIGN(ScopedUMARecorder);
};
} // namespace
ComputedHashes::Reader::Reader() {
}
ComputedHashes::Reader::~Reader() {
}
bool ComputedHashes::Reader::InitFromFile(const base::FilePath& path) {
ScopedUMARecorder uma_recorder;
std::string contents;
if (!base::ReadFileToString(path, &contents))
return false;
base::DictionaryValue* top_dictionary = NULL;
std::unique_ptr<base::Value> value(
base::JSONReader::ReadDeprecated(contents));
if (!value.get() || !value->GetAsDictionary(&top_dictionary))
return false;
// For now we don't support forwards or backwards compatability in the
// format, so we return false on version mismatch.
int version = 0;
if (!top_dictionary->GetInteger(computed_hashes::kVersionKey, &version) ||
version != computed_hashes::kVersion)
return false;
base::ListValue* all_hashes = NULL;
if (!top_dictionary->GetList(computed_hashes::kFileHashesKey, &all_hashes))
return false;
for (size_t i = 0; i < all_hashes->GetSize(); i++) {
base::DictionaryValue* dictionary = NULL;
if (!all_hashes->GetDictionary(i, &dictionary))
return false;
std::string relative_path_utf8;
if (!dictionary->GetString(computed_hashes::kPathKey, &relative_path_utf8))
return false;
int block_size;
if (!dictionary->GetInteger(computed_hashes::kBlockSizeKey, &block_size))
return false;
if (block_size <= 0 || ((block_size % 1024) != 0)) {
LOG(ERROR) << "Invalid block size: " << block_size;
return false;
}
base::ListValue* hashes_list = NULL;
if (!dictionary->GetList(computed_hashes::kBlockHashesKey, &hashes_list))
return false;
base::FilePath relative_path =
base::FilePath::FromUTF8Unsafe(relative_path_utf8);
relative_path = relative_path.NormalizePathSeparatorsTo('/');
data_[relative_path] = HashInfo(block_size, std::vector<std::string>());
std::vector<std::string>* hashes = &(data_[relative_path].second);
for (size_t j = 0; j < hashes_list->GetSize(); j++) {
std::string encoded;
if (!hashes_list->GetString(j, &encoded))
return false;
hashes->push_back(std::string());
std::string* decoded = &hashes->back();
if (!base::Base64Decode(encoded, decoded)) {
hashes->clear();
return false;
}
}
}
uma_recorder.RecordSuccess();
return true;
}
bool ComputedHashes::Reader::GetHashes(const base::FilePath& relative_path,
int* block_size,
std::vector<std::string>* hashes) const {
base::FilePath path = relative_path.NormalizePathSeparatorsTo('/');
auto i = data_.find(path);
if (i == data_.end()) {
// If we didn't find the entry using exact match, it's possible the
// developer is using a path with some letters in the incorrect case, which
// happens to work on windows/osx. So try doing a linear scan to look for a
// case-insensitive match. In practice most extensions don't have that big
// a list of files so the performance penalty is probably not too big
// here. Also for crbug.com/29941 we plan to start warning developers when
// they are making this mistake, since their extension will be broken on
// linux/chromeos.
for (i = data_.begin(); i != data_.end(); ++i) {
const base::FilePath& entry = i->first;
if (base::FilePath::CompareEqualIgnoreCase(entry.value(), path.value()))
break;
}
if (i == data_.end())
return false;
}
const HashInfo& info = i->second;
*block_size = info.first;
*hashes = info.second;
return true;
}
ComputedHashes::Writer::Writer() : file_list_(new base::ListValue) {
}
ComputedHashes::Writer::~Writer() {
}
void ComputedHashes::Writer::AddHashes(const base::FilePath& relative_path,
int block_size,
const std::vector<std::string>& hashes) {
auto block_hashes = std::make_unique<base::ListValue>();
block_hashes->GetList().reserve(hashes.size());
for (const auto& hash : hashes) {
std::string encoded;
base::Base64Encode(hash, &encoded);
block_hashes->GetList().emplace_back(std::move(encoded));
}
auto dict = std::make_unique<base::DictionaryValue>();
dict->SetString(computed_hashes::kPathKey,
relative_path.NormalizePathSeparatorsTo('/').AsUTF8Unsafe());
dict->SetInteger(computed_hashes::kBlockSizeKey, block_size);
dict->Set(computed_hashes::kBlockHashesKey, std::move(block_hashes));
file_list_->Append(std::move(dict));
}
bool ComputedHashes::Writer::WriteToFile(const base::FilePath& path) {
std::string json;
base::DictionaryValue top_dictionary;
top_dictionary.SetInteger(computed_hashes::kVersionKey,
computed_hashes::kVersion);
top_dictionary.Set(computed_hashes::kFileHashesKey, std::move(file_list_));
if (!base::JSONWriter::Write(top_dictionary, &json))
return false;
int written = base::WriteFile(path, json.data(), json.size());
if (static_cast<unsigned>(written) != json.size()) {
LOG(ERROR) << "Error writing " << path.AsUTF8Unsafe()
<< " ; write result:" << written << " expected:" << json.size();
return false;
}
return true;
}
void ComputedHashes::ComputeHashesForContent(const std::string& contents,
size_t block_size,
std::vector<std::string>* hashes) {
size_t offset = 0;
// Even when the contents is empty, we want to output at least one hash
// block (the hash of the empty string).
do {
const char* block_start = contents.data() + offset;
DCHECK(offset <= contents.size());
size_t bytes_to_read = std::min(contents.size() - offset, block_size);
std::unique_ptr<crypto::SecureHash> hash(
crypto::SecureHash::Create(crypto::SecureHash::SHA256));
hash->Update(block_start, bytes_to_read);
hashes->push_back(std::string());
std::string* buffer = &(hashes->back());
buffer->resize(crypto::kSHA256Length);
hash->Finish(base::data(*buffer), buffer->size());
// If |contents| is empty, then we want to just exit here.
if (bytes_to_read == 0)
break;
offset += bytes_to_read;
} while (offset < contents.size());
}
} // namespace extensions