blob: af4e55128449818c72be04a0c734ed1d1eb5096a [file] [log] [blame]
// Copyright 2014 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
// The basic usage of the Filter interface is described in the comment at
// the beginning of filter.h. If Filter::Factory is passed a vector of
// size greater than 1, that interface is implemented by a series of filters
// connected in a chain. In such a case the first filter
// in the chain proxies calls to ReadData() so that its return values
// apply to the entire chain.
//
// In a filter chain, the data flows from first filter (held by the
// caller) down the chain. When ReadData() is called on any filter
// except for the last filter, it proxies the call down the chain,
// filling in the input buffers of subsequent filters if needed (==
// that filter's last_status() value is FILTER_NEED_MORE_DATA) and
// available (== the current filter has data it can output). The last
// Filter will then output data if possible, and return
// FILTER_NEED_MORE_DATA if not. Because the indirection pushes
// data along the filter chain at each level if it's available and the
// next filter needs it, a return value of FILTER_NEED_MORE_DATA from the
// final filter will apply to the entire chain.
#include "net/filter/filter.h"
#include "base/files/file_path.h"
#include "base/metrics/histogram_macros.h"
#include "base/strings/string_util.h"
#include "net/base/filename_util_unsafe.h"
#include "net/base/io_buffer.h"
#include "net/base/mime_util.h"
#include "net/base/sdch_net_log_params.h"
#include "net/filter/gzip_filter.h"
#include "net/filter/sdch_filter.h"
#include "net/url_request/url_request_context.h"
#include "url/gurl.h"
namespace net {
namespace {
// Filter types (using canonical lower case only):
const char kDeflate[] = "deflate";
const char kGZip[] = "gzip";
const char kXGZip[] = "x-gzip";
const char kSdch[] = "sdch";
// compress and x-compress are currently not supported. If we decide to support
// them, we'll need the same mime type compatibility hack we have for gzip. For
// more information, see Firefox's nsHttpChannel::ProcessNormal.
// Mime types:
const char kApplicationXGzip[] = "application/x-gzip";
const char kApplicationGzip[] = "application/gzip";
const char kApplicationXGunzip[] = "application/x-gunzip";
const char kTextHtml[] = "text/html";
// Buffer size allocated when de-compressing data.
const int kFilterBufSize = 32 * 1024;
void LogSdchProblem(const FilterContext& filter_context,
SdchProblemCode problem) {
SdchManager::SdchErrorRecovery(problem);
filter_context.GetNetLog().AddEvent(
NetLog::TYPE_SDCH_DECODING_ERROR,
base::Bind(&NetLogSdchResourceProblemCallback, problem));
}
std::string FilterTypeAsString(Filter::FilterType type_id) {
switch (type_id) {
case Filter::FILTER_TYPE_DEFLATE:
return "FILTER_TYPE_DEFLATE";
case Filter::FILTER_TYPE_GZIP:
return "FILTER_TYPE_GZIP";
case Filter::FILTER_TYPE_GZIP_HELPING_SDCH:
return "FILTER_TYPE_GZIP_HELPING_SDCH";
case Filter::FILTER_TYPE_SDCH:
return "FILTER_TYPE_SDCH";
case Filter::FILTER_TYPE_SDCH_POSSIBLE :
return "FILTER_TYPE_SDCH_POSSIBLE ";
case Filter::FILTER_TYPE_UNSUPPORTED:
return "FILTER_TYPE_UNSUPPORTED";
}
return "";
}
} // namespace
FilterContext::~FilterContext() {
}
Filter::~Filter() {}
// static
Filter* Filter::Factory(const std::vector<FilterType>& filter_types,
const FilterContext& filter_context) {
if (filter_types.empty())
return NULL;
Filter* filter_list = NULL; // Linked list of filters.
for (size_t i = 0; i < filter_types.size(); i++) {
filter_list = PrependNewFilter(filter_types[i], filter_context,
kFilterBufSize, filter_list);
if (!filter_list)
return NULL;
}
return filter_list;
}
// static
Filter* Filter::GZipFactory() {
return InitGZipFilter(FILTER_TYPE_GZIP, kFilterBufSize);
}
// static
Filter* Filter::FactoryForTests(const std::vector<FilterType>& filter_types,
const FilterContext& filter_context,
int buffer_size) {
if (filter_types.empty())
return NULL;
Filter* filter_list = NULL; // Linked list of filters.
for (size_t i = 0; i < filter_types.size(); i++) {
filter_list = PrependNewFilter(filter_types[i], filter_context,
buffer_size, filter_list);
if (!filter_list)
return NULL;
}
return filter_list;
}
Filter::FilterStatus Filter::ReadData(char* dest_buffer, int* dest_len) {
const int dest_buffer_capacity = *dest_len;
if (last_status_ == FILTER_ERROR)
return last_status_;
if (!next_filter_.get())
return last_status_ = ReadFilteredData(dest_buffer, dest_len);
// This filter needs more data, but it's not clear that the rest of
// the chain does; delegate the actual status return to the next filter.
if (last_status_ == FILTER_NEED_MORE_DATA && !stream_data_len())
return next_filter_->ReadData(dest_buffer, dest_len);
do {
if (next_filter_->last_status() == FILTER_NEED_MORE_DATA) {
PushDataIntoNextFilter();
if (FILTER_ERROR == last_status_)
return FILTER_ERROR;
}
*dest_len = dest_buffer_capacity; // Reset the input/output parameter.
next_filter_->ReadData(dest_buffer, dest_len);
if (FILTER_NEED_MORE_DATA == last_status_)
return next_filter_->last_status();
// In the case where this filter has data internally, and is indicating such
// with a last_status_ of FILTER_OK, but at the same time the next filter in
// the chain indicated it FILTER_NEED_MORE_DATA, we have to be cautious
// about confusing the caller. The API confusion can appear if we return
// FILTER_OK (suggesting we have more data in aggregate), but yet we don't
// populate our output buffer. When that is the case, we need to
// alternately call our filter element, and the next_filter element until we
// get out of this state (by pumping data into the next filter until it
// outputs data, or it runs out of data and reports that it NEED_MORE_DATA.)
} while (FILTER_OK == last_status_ &&
FILTER_NEED_MORE_DATA == next_filter_->last_status() &&
0 == *dest_len);
if (next_filter_->last_status() == FILTER_ERROR)
return FILTER_ERROR;
return FILTER_OK;
}
bool Filter::FlushStreamBuffer(int stream_data_len) {
DCHECK_LE(stream_data_len, stream_buffer_size_);
if (stream_data_len <= 0 || stream_data_len > stream_buffer_size_)
return false;
DCHECK(stream_buffer());
// Bail out if there is more data in the stream buffer to be filtered.
if (!stream_buffer() || stream_data_len_)
return false;
next_stream_data_ = stream_buffer()->data();
stream_data_len_ = stream_data_len;
last_status_ = FILTER_OK;
return true;
}
// static
Filter::FilterType Filter::ConvertEncodingToType(
const std::string& filter_type) {
FilterType type_id;
if (LowerCaseEqualsASCII(filter_type, kDeflate)) {
type_id = FILTER_TYPE_DEFLATE;
} else if (LowerCaseEqualsASCII(filter_type, kGZip) ||
LowerCaseEqualsASCII(filter_type, kXGZip)) {
type_id = FILTER_TYPE_GZIP;
} else if (LowerCaseEqualsASCII(filter_type, kSdch)) {
type_id = FILTER_TYPE_SDCH;
} else {
// Note we also consider "identity" and "uncompressed" UNSUPPORTED as
// filter should be disabled in such cases.
type_id = FILTER_TYPE_UNSUPPORTED;
}
return type_id;
}
namespace {
// Result of running FixupEncodingTypes with a Content-Encoding of gzip. This
// enum is used for UMA and is available in histograms.xml as
// GzipEncodingFixupResult.
enum GzipEncodingFixupResult {
GZIP_ENCODING_LEFT_AS_IS = 0,
// Cleared because a resource with a GZIP MIME type was being transferred with
// a Content-Encoding of gzip.
GZIP_ENCODING_CLEARED_DUE_TO_GZIP_MIME_TYPE = 1,
// Cleared because the resource is known to be a download and the resulting
// file had a filename that indicating that the contents are GZIP.
GZIP_ENCODING_CLEARED_DUE_TO_DOWNLOAD_OF_GZIP_FILE = 2,
// Cleared because the resource is not a handled MIME type (and hence will be
// downloaded), and the predicted filename is one that indicates the contents
// to be GZIP.
GZIP_ENCODING_CLEARED_DUE_TO_UNHANDLED_GZIP_NAME = 3,
GZIP_ENCODING_FIXUP_RESULT_MAX_ENTRIES
};
} // namespace
// static
void Filter::FixupEncodingTypes(
const FilterContext& filter_context,
std::vector<FilterType>* encoding_types) {
std::string mime_type;
bool success = filter_context.GetMimeType(&mime_type);
DCHECK(success || mime_type.empty());
if ((1 == encoding_types->size()) &&
(FILTER_TYPE_GZIP == encoding_types->front())) {
GzipEncodingFixupResult fixup_result = GZIP_ENCODING_LEFT_AS_IS;
if (LowerCaseEqualsASCII(mime_type, kApplicationXGzip) ||
LowerCaseEqualsASCII(mime_type, kApplicationGzip) ||
LowerCaseEqualsASCII(mime_type, kApplicationXGunzip)) {
// The server has told us that it sent us gziped content with a gzip
// content encoding. Sadly, Apache mistakenly sets these headers for all
// .gz files. We match Firefox's nsHttpChannel::ProcessNormal and ignore
// the Content-Encoding here.
encoding_types->clear();
fixup_result = GZIP_ENCODING_CLEARED_DUE_TO_GZIP_MIME_TYPE;
}
GURL url;
std::string disposition;
success = filter_context.GetURL(&url);
DCHECK(success);
filter_context.GetContentDisposition(&disposition);
// Don't supply a MIME type here, since that may cause disk IO.
base::FilePath::StringType extension =
GenerateFileExtensionUnsafe(url, disposition, "UTF-8", "", "", "");
if (filter_context.IsDownload()) {
// We don't want to decompress gzipped files when the user explicitly
// asks to download them.
// For the case of svgz files, we use the extension to distinguish
// between svgz files and svg files compressed with gzip by the server.
// When viewing a .svgz file, we need to uncompress it, but we don't
// want to do that when downloading.
// See Firefox's nonDecodableExtensions in nsExternalHelperAppService.cpp
if (EndsWith(extension, FILE_PATH_LITERAL(".gz"), false) ||
LowerCaseEqualsASCII(extension, ".tgz") ||
LowerCaseEqualsASCII(extension, ".svgz")) {
encoding_types->clear();
fixup_result = GZIP_ENCODING_CLEARED_DUE_TO_DOWNLOAD_OF_GZIP_FILE;
}
} else {
// When the user does not explicitly ask to download a file, if we get a
// supported mime type, then we attempt to decompress in order to view it.
// However, if it's not a supported mime type, then we will attempt to
// download it, and in that case, don't decompress .gz/.tgz files.
if ((EndsWith(extension, FILE_PATH_LITERAL(".gz"), false) ||
LowerCaseEqualsASCII(extension, ".tgz")) &&
!IsSupportedMimeType(mime_type)) {
encoding_types->clear();
fixup_result = GZIP_ENCODING_CLEARED_DUE_TO_UNHANDLED_GZIP_NAME;
}
}
UMA_HISTOGRAM_ENUMERATION("Net.GzipEncodingFixupResult", fixup_result,
GZIP_ENCODING_FIXUP_RESULT_MAX_ENTRIES);
}
// If the request was for SDCH content, then we might need additional fixups.
if (!filter_context.SdchDictionariesAdvertised()) {
// It was not an SDCH request, so we'll just record stats.
if (1 < encoding_types->size()) {
// Multiple filters were intended to only be used for SDCH (thus far!)
LogSdchProblem(filter_context, SDCH_MULTIENCODING_FOR_NON_SDCH_REQUEST);
}
if ((1 == encoding_types->size()) &&
(FILTER_TYPE_SDCH == encoding_types->front())) {
LogSdchProblem(filter_context,
SDCH_SDCH_CONTENT_ENCODE_FOR_NON_SDCH_REQUEST);
}
return;
}
// The request was tagged as an SDCH request, which means the server supplied
// a dictionary, and we advertised it in the request. Some proxies will do
// very strange things to the request, or the response, so we have to handle
// them gracefully.
// If content encoding included SDCH, then everything is "relatively" fine.
if (!encoding_types->empty() &&
(FILTER_TYPE_SDCH == encoding_types->front())) {
// Some proxies (found currently in Argentina) strip the Content-Encoding
// text from "sdch,gzip" to a mere "sdch" without modifying the compressed
// payload. To handle this gracefully, we simulate the "probably" deleted
// ",gzip" by appending a tentative gzip decode, which will default to a
// no-op pass through filter if it doesn't get gzip headers where expected.
if (1 == encoding_types->size()) {
encoding_types->push_back(FILTER_TYPE_GZIP_HELPING_SDCH);
LogSdchProblem(filter_context, SDCH_OPTIONAL_GUNZIP_ENCODING_ADDED);
}
return;
}
// There are now several cases to handle for an SDCH request. Foremost, if
// the outbound request was stripped so as not to advertise support for
// encodings, we might get back content with no encoding, or (for example)
// just gzip. We have to be sure that any changes we make allow for such
// minimal coding to work. That issue is why we use TENTATIVE filters if we
// add any, as those filters sniff the content, and act as pass-through
// filters if headers are not found.
// If the outbound GET is not modified, then the server will generally try to
// send us SDCH encoded content. As that content returns, there are several
// corruptions of the header "content-encoding" that proxies may perform (and
// have been detected in the wild). We already dealt with the a honest
// content encoding of "sdch,gzip" being corrupted into "sdch" with on change
// of the actual content. Another common corruption is to either disscard
// the accurate content encoding, or to replace it with gzip only (again, with
// no change in actual content). The last observed corruption it to actually
// change the content, such as by re-gzipping it, and that may happen along
// with corruption of the stated content encoding (wow!).
// The one unresolved failure mode comes when we advertise a dictionary, and
// the server tries to *send* a gzipped file (not gzip encode content), and
// then we could do a gzip decode :-(. Since SDCH is only (currently)
// supported server side on paths that only send HTML content, this mode has
// never surfaced in the wild (and is unlikely to).
// We will gather a lot of stats as we perform the fixups
if (StartsWithASCII(mime_type, kTextHtml, false)) {
// Suspicious case: Advertised dictionary, but server didn't use sdch, and
// we're HTML tagged.
if (encoding_types->empty()) {
LogSdchProblem(filter_context, SDCH_ADDED_CONTENT_ENCODING);
} else if (1 == encoding_types->size()) {
LogSdchProblem(filter_context, SDCH_FIXED_CONTENT_ENCODING);
} else {
LogSdchProblem(filter_context, SDCH_FIXED_CONTENT_ENCODINGS);
}
} else {
// Remarkable case!?! We advertised an SDCH dictionary, content-encoding
// was not marked for SDCH processing: Why did the server suggest an SDCH
// dictionary in the first place??. Also, the content isn't
// tagged as HTML, despite the fact that SDCH encoding is mostly likely for
// HTML: Did some anti-virus system strip this tag (sometimes they strip
// accept-encoding headers on the request)?? Does the content encoding not
// start with "text/html" for some other reason?? We'll report this as a
// fixup to a binary file, but it probably really is text/html (some how).
if (encoding_types->empty()) {
LogSdchProblem(filter_context, SDCH_BINARY_ADDED_CONTENT_ENCODING);
} else if (1 == encoding_types->size()) {
LogSdchProblem(filter_context, SDCH_BINARY_FIXED_CONTENT_ENCODING);
} else {
LogSdchProblem(filter_context, SDCH_BINARY_FIXED_CONTENT_ENCODINGS);
}
}
// Leave the existing encoding type to be processed first, and add our
// tentative decodings to be done afterwards. Vodaphone UK reportedyl will
// perform a second layer of gzip encoding atop the server's sdch,gzip
// encoding, and then claim that the content encoding is a mere gzip. As a
// result we'll need (in that case) to do the gunzip, plus our tentative
// gunzip and tentative SDCH decoding.
// This approach nicely handles the empty() list as well, and should work with
// other (as yet undiscovered) proxies the choose to re-compressed with some
// other encoding (such as bzip2, etc.).
encoding_types->insert(encoding_types->begin(),
FILTER_TYPE_GZIP_HELPING_SDCH);
encoding_types->insert(encoding_types->begin(), FILTER_TYPE_SDCH_POSSIBLE);
return;
}
std::string Filter::OrderedFilterList() const {
if (next_filter_) {
return FilterTypeAsString(type_id_) + "," +
next_filter_->OrderedFilterList();
} else {
return FilterTypeAsString(type_id_);
}
}
Filter::Filter(FilterType type_id)
: stream_buffer_(NULL),
stream_buffer_size_(0),
next_stream_data_(NULL),
stream_data_len_(0),
last_status_(FILTER_NEED_MORE_DATA),
type_id_(type_id) {}
Filter::FilterStatus Filter::CopyOut(char* dest_buffer, int* dest_len) {
int out_len;
int input_len = *dest_len;
*dest_len = 0;
if (0 == stream_data_len_)
return Filter::FILTER_NEED_MORE_DATA;
out_len = std::min(input_len, stream_data_len_);
memcpy(dest_buffer, next_stream_data_, out_len);
*dest_len += out_len;
stream_data_len_ -= out_len;
if (0 == stream_data_len_) {
next_stream_data_ = NULL;
return Filter::FILTER_NEED_MORE_DATA;
} else {
next_stream_data_ += out_len;
return Filter::FILTER_OK;
}
}
// static
Filter* Filter::InitGZipFilter(FilterType type_id, int buffer_size) {
scoped_ptr<GZipFilter> gz_filter(new GZipFilter(type_id));
gz_filter->InitBuffer(buffer_size);
return gz_filter->InitDecoding(type_id) ? gz_filter.release() : NULL;
}
// static
Filter* Filter::InitSdchFilter(FilterType type_id,
const FilterContext& filter_context,
int buffer_size) {
scoped_ptr<SdchFilter> sdch_filter(new SdchFilter(type_id, filter_context));
sdch_filter->InitBuffer(buffer_size);
return sdch_filter->InitDecoding(type_id) ? sdch_filter.release() : NULL;
}
// static
Filter* Filter::PrependNewFilter(FilterType type_id,
const FilterContext& filter_context,
int buffer_size,
Filter* filter_list) {
scoped_ptr<Filter> first_filter; // Soon to be start of chain.
switch (type_id) {
case FILTER_TYPE_GZIP_HELPING_SDCH:
case FILTER_TYPE_DEFLATE:
case FILTER_TYPE_GZIP:
first_filter.reset(InitGZipFilter(type_id, buffer_size));
break;
case FILTER_TYPE_SDCH:
case FILTER_TYPE_SDCH_POSSIBLE:
if (filter_context.GetURLRequestContext()->sdch_manager() &&
SdchManager::sdch_enabled()) {
first_filter.reset(
InitSdchFilter(type_id, filter_context, buffer_size));
}
break;
default:
break;
}
if (!first_filter.get())
return NULL;
first_filter->next_filter_.reset(filter_list);
return first_filter.release();
}
void Filter::InitBuffer(int buffer_size) {
DCHECK(!stream_buffer());
DCHECK_GT(buffer_size, 0);
stream_buffer_ = new IOBuffer(buffer_size);
stream_buffer_size_ = buffer_size;
}
void Filter::PushDataIntoNextFilter() {
IOBuffer* next_buffer = next_filter_->stream_buffer();
int next_size = next_filter_->stream_buffer_size();
last_status_ = ReadFilteredData(next_buffer->data(), &next_size);
if (FILTER_ERROR != last_status_)
next_filter_->FlushStreamBuffer(next_size);
}
} // namespace net