blob: 368b91ae3eff103afdb77ae5bb48c5e20bc9eeb9 [file] [log] [blame]
// Copyright (c) 2012 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "net/url_request/url_request.h"
#include <utility>
#include "base/bind.h"
#include "base/bind_helpers.h"
#include "base/callback.h"
#include "base/compiler_specific.h"
#include "base/lazy_instance.h"
#include "base/memory/singleton.h"
#include "base/profiler/scoped_tracker.h"
#include "base/rand_util.h"
#include "base/stl_util.h"
#include "base/strings/utf_string_conversions.h"
#include "base/synchronization/lock.h"
#include "base/threading/thread_task_runner_handle.h"
#include "base/values.h"
#include "net/base/auth.h"
#include "net/base/host_port_pair.h"
#include "net/base/load_flags.h"
#include "net/base/load_timing_info.h"
#include "net/base/net_errors.h"
#include "net/base/network_change_notifier.h"
#include "net/base/network_delegate.h"
#include "net/base/upload_data_stream.h"
#include "net/http/http_response_headers.h"
#include "net/http/http_util.h"
#include "net/log/net_log.h"
#include "net/log/net_log_event_type.h"
#include "net/log/net_log_source_type.h"
#include "net/ssl/ssl_cert_request_info.h"
#include "net/url_request/redirect_info.h"
#include "net/url_request/url_request_context.h"
#include "net/url_request/url_request_error_job.h"
#include "net/url_request/url_request_job.h"
#include "net/url_request/url_request_job_manager.h"
#include "net/url_request/url_request_netlog_params.h"
#include "net/url_request/url_request_redirect_job.h"
#include "url/gurl.h"
#include "url/origin.h"
using base::Time;
using std::string;
namespace net {
namespace {
// Max number of http redirects to follow. Same number as gecko.
const int kMaxRedirects = 20;
// TODO(battre): Delete this, see http://crbug.com/89321:
// This counter keeps track of the identifiers used for URL requests so far.
// 0 is reserved to represent an invalid ID.
uint64_t g_next_url_request_identifier = 1;
// This lock protects g_next_url_request_identifier.
base::LazyInstance<base::Lock>::Leaky
g_next_url_request_identifier_lock = LAZY_INSTANCE_INITIALIZER;
// Returns an prior unused identifier for URL requests.
uint64_t GenerateURLRequestIdentifier() {
base::AutoLock lock(g_next_url_request_identifier_lock.Get());
return g_next_url_request_identifier++;
}
// True once the first URLRequest was started.
bool g_url_requests_started = false;
// True if cookies are accepted by default.
bool g_default_can_use_cookies = true;
// When the URLRequest first assempts load timing information, it has the times
// at which each event occurred. The API requires the time which the request
// was blocked on each phase. This function handles the conversion.
//
// In the case of reusing a SPDY session, old proxy results may have been
// reused, so proxy resolution times may be before the request was started.
//
// Due to preconnect and late binding, it is also possible for the connection
// attempt to start before a request has been started, or proxy resolution
// completed.
//
// This functions fixes both those cases.
void ConvertRealLoadTimesToBlockingTimes(LoadTimingInfo* load_timing_info) {
DCHECK(!load_timing_info->request_start.is_null());
// Earliest time possible for the request to be blocking on connect events.
base::TimeTicks block_on_connect = load_timing_info->request_start;
if (!load_timing_info->proxy_resolve_start.is_null()) {
DCHECK(!load_timing_info->proxy_resolve_end.is_null());
// Make sure the proxy times are after request start.
if (load_timing_info->proxy_resolve_start < load_timing_info->request_start)
load_timing_info->proxy_resolve_start = load_timing_info->request_start;
if (load_timing_info->proxy_resolve_end < load_timing_info->request_start)
load_timing_info->proxy_resolve_end = load_timing_info->request_start;
// Connect times must also be after the proxy times.
block_on_connect = load_timing_info->proxy_resolve_end;
}
// Make sure connection times are after start and proxy times.
LoadTimingInfo::ConnectTiming* connect_timing =
&load_timing_info->connect_timing;
if (!connect_timing->dns_start.is_null()) {
DCHECK(!connect_timing->dns_end.is_null());
if (connect_timing->dns_start < block_on_connect)
connect_timing->dns_start = block_on_connect;
if (connect_timing->dns_end < block_on_connect)
connect_timing->dns_end = block_on_connect;
}
if (!connect_timing->connect_start.is_null()) {
DCHECK(!connect_timing->connect_end.is_null());
if (connect_timing->connect_start < block_on_connect)
connect_timing->connect_start = block_on_connect;
if (connect_timing->connect_end < block_on_connect)
connect_timing->connect_end = block_on_connect;
}
if (!connect_timing->ssl_start.is_null()) {
DCHECK(!connect_timing->ssl_end.is_null());
if (connect_timing->ssl_start < block_on_connect)
connect_timing->ssl_start = block_on_connect;
if (connect_timing->ssl_end < block_on_connect)
connect_timing->ssl_end = block_on_connect;
}
}
} // namespace
///////////////////////////////////////////////////////////////////////////////
// URLRequest::Delegate
void URLRequest::Delegate::OnReceivedRedirect(URLRequest* request,
const RedirectInfo& redirect_info,
bool* defer_redirect) {
}
void URLRequest::Delegate::OnAuthRequired(URLRequest* request,
AuthChallengeInfo* auth_info) {
request->CancelAuth();
}
void URLRequest::Delegate::OnCertificateRequested(
URLRequest* request,
SSLCertRequestInfo* cert_request_info) {
request->CancelWithError(ERR_SSL_CLIENT_AUTH_CERT_NEEDED);
}
void URLRequest::Delegate::OnSSLCertificateError(URLRequest* request,
const SSLInfo& ssl_info,
bool is_hsts_ok) {
request->Cancel();
}
void URLRequest::Delegate::OnResponseStarted(URLRequest* request,
int net_error) {
OnResponseStarted(request);
}
void URLRequest::Delegate::OnResponseStarted(URLRequest* request) {
NOTREACHED();
}
///////////////////////////////////////////////////////////////////////////////
// URLRequest
URLRequest::~URLRequest() {
Cancel();
if (network_delegate_) {
network_delegate_->NotifyURLRequestDestroyed(this);
if (job_.get())
job_->NotifyURLRequestDestroyed();
}
// Delete job before |this|, since subclasses may do weird things, like depend
// on UserData associated with |this| and poke at it during teardown.
job_.reset();
DCHECK_EQ(1u, context_->url_requests()->count(this));
context_->url_requests()->erase(this);
int net_error = OK;
// Log error only on failure, not cancellation, as even successful requests
// are "cancelled" on destruction.
if (status_.status() == URLRequestStatus::FAILED)
net_error = status_.error();
net_log_.EndEventWithNetErrorCode(NetLogEventType::REQUEST_ALIVE, net_error);
}
void URLRequest::set_upload(std::unique_ptr<UploadDataStream> upload) {
upload_data_stream_ = std::move(upload);
}
const UploadDataStream* URLRequest::get_upload() const {
return upload_data_stream_.get();
}
bool URLRequest::has_upload() const {
return upload_data_stream_.get() != NULL;
}
void URLRequest::SetExtraRequestHeaderByName(const string& name,
const string& value,
bool overwrite) {
DCHECK(!is_pending_ || is_redirecting_);
if (overwrite) {
extra_request_headers_.SetHeader(name, value);
} else {
extra_request_headers_.SetHeaderIfMissing(name, value);
}
}
void URLRequest::RemoveRequestHeaderByName(const string& name) {
DCHECK(!is_pending_ || is_redirecting_);
extra_request_headers_.RemoveHeader(name);
}
void URLRequest::SetExtraRequestHeaders(
const HttpRequestHeaders& headers) {
DCHECK(!is_pending_);
extra_request_headers_ = headers;
// NOTE: This method will likely become non-trivial once the other setters
// for request headers are implemented.
}
bool URLRequest::GetFullRequestHeaders(HttpRequestHeaders* headers) const {
if (!job_.get())
return false;
return job_->GetFullRequestHeaders(headers);
}
int64_t URLRequest::GetTotalReceivedBytes() const {
if (!job_.get())
return 0;
return job_->GetTotalReceivedBytes();
}
int64_t URLRequest::GetTotalSentBytes() const {
if (!job_.get())
return 0;
return job_->GetTotalSentBytes();
}
int64_t URLRequest::GetRawBodyBytes() const {
if (!job_.get())
return 0;
return job_->prefilter_bytes_read();
}
LoadStateWithParam URLRequest::GetLoadState() const {
// The !blocked_by_.empty() check allows |this| to report it's blocked on a
// delegate before it has been started.
if (calling_delegate_ || !blocked_by_.empty()) {
return LoadStateWithParam(
LOAD_STATE_WAITING_FOR_DELEGATE,
use_blocked_by_as_load_param_ ? base::UTF8ToUTF16(blocked_by_) :
base::string16());
}
return LoadStateWithParam(job_.get() ? job_->GetLoadState() : LOAD_STATE_IDLE,
base::string16());
}
std::unique_ptr<base::Value> URLRequest::GetStateAsValue() const {
std::unique_ptr<base::DictionaryValue> dict(new base::DictionaryValue());
dict->SetString("url", original_url().possibly_invalid_spec());
if (url_chain_.size() > 1) {
std::unique_ptr<base::ListValue> list(new base::ListValue());
for (const GURL& url : url_chain_) {
list->AppendString(url.possibly_invalid_spec());
}
dict->Set("url_chain", std::move(list));
}
dict->SetInteger("load_flags", load_flags_);
LoadStateWithParam load_state = GetLoadState();
dict->SetInteger("load_state", load_state.state);
if (!load_state.param.empty())
dict->SetString("load_state_param", load_state.param);
if (!blocked_by_.empty())
dict->SetString("delegate_info", blocked_by_);
dict->SetString("method", method_);
dict->SetBoolean("has_upload", has_upload());
dict->SetBoolean("is_pending", is_pending_);
// Add the status of the request. The status should always be IO_PENDING, and
// the error should always be OK, unless something is holding onto a request
// that has finished or a request was leaked. Neither of these should happen.
switch (status_.status()) {
case URLRequestStatus::SUCCESS:
dict->SetString("status", "SUCCESS");
break;
case URLRequestStatus::IO_PENDING:
dict->SetString("status", "IO_PENDING");
break;
case URLRequestStatus::CANCELED:
dict->SetString("status", "CANCELED");
break;
case URLRequestStatus::FAILED:
dict->SetString("status", "FAILED");
break;
}
if (status_.error() != OK)
dict->SetInteger("net_error", status_.error());
return std::move(dict);
}
void URLRequest::LogBlockedBy(const char* blocked_by) {
DCHECK(blocked_by);
DCHECK_GT(strlen(blocked_by), 0u);
// Only log information to NetLog during startup and certain deferring calls
// to delegates. For all reads but the first, do nothing.
if (!calling_delegate_ && !response_info_.request_time.is_null())
return;
LogUnblocked();
blocked_by_ = blocked_by;
use_blocked_by_as_load_param_ = false;
net_log_.BeginEvent(NetLogEventType::DELEGATE_INFO,
NetLog::StringCallback("delegate_info", &blocked_by_));
}
void URLRequest::LogAndReportBlockedBy(const char* source) {
LogBlockedBy(source);
use_blocked_by_as_load_param_ = true;
}
void URLRequest::LogUnblocked() {
if (blocked_by_.empty())
return;
net_log_.EndEvent(NetLogEventType::DELEGATE_INFO);
blocked_by_.clear();
}
UploadProgress URLRequest::GetUploadProgress() const {
if (!job_.get()) {
// We haven't started or the request was cancelled
return UploadProgress();
}
if (final_upload_progress_.position()) {
// The first job completed and none of the subsequent series of
// GETs when following redirects will upload anything, so we return the
// cached results from the initial job, the POST.
return final_upload_progress_;
}
if (upload_data_stream_)
return upload_data_stream_->GetUploadProgress();
return UploadProgress();
}
void URLRequest::GetResponseHeaderByName(const string& name,
string* value) const {
DCHECK(value);
if (response_info_.headers.get()) {
response_info_.headers->GetNormalizedHeader(name, value);
} else {
value->clear();
}
}
HostPortPair URLRequest::GetSocketAddress() const {
DCHECK(job_.get());
return job_->GetSocketAddress();
}
HttpResponseHeaders* URLRequest::response_headers() const {
return response_info_.headers.get();
}
void URLRequest::GetLoadTimingInfo(LoadTimingInfo* load_timing_info) const {
*load_timing_info = load_timing_info_;
}
void URLRequest::PopulateNetErrorDetails(NetErrorDetails* details) const {
if (!job_)
return;
return job_->PopulateNetErrorDetails(details);
}
bool URLRequest::GetRemoteEndpoint(IPEndPoint* endpoint) const {
if (!job_)
return false;
return job_->GetRemoteEndpoint(endpoint);
}
void URLRequest::GetMimeType(string* mime_type) const {
DCHECK(job_.get());
job_->GetMimeType(mime_type);
}
void URLRequest::GetCharset(string* charset) const {
DCHECK(job_.get());
job_->GetCharset(charset);
}
int URLRequest::GetResponseCode() const {
DCHECK(job_.get());
return job_->GetResponseCode();
}
void URLRequest::SetLoadFlags(int flags) {
if ((load_flags_ & LOAD_IGNORE_LIMITS) != (flags & LOAD_IGNORE_LIMITS)) {
DCHECK(!job_.get());
DCHECK(flags & LOAD_IGNORE_LIMITS);
DCHECK_EQ(priority_, MAXIMUM_PRIORITY);
}
load_flags_ = flags;
// This should be a no-op given the above DCHECKs, but do this
// anyway for release mode.
if ((load_flags_ & LOAD_IGNORE_LIMITS) != 0)
SetPriority(MAXIMUM_PRIORITY);
}
// static
void URLRequest::SetDefaultCookiePolicyToBlock() {
CHECK(!g_url_requests_started);
g_default_can_use_cookies = false;
}
// static
bool URLRequest::IsHandledProtocol(const std::string& scheme) {
return URLRequestJobManager::SupportsScheme(scheme);
}
// static
bool URLRequest::IsHandledURL(const GURL& url) {
if (!url.is_valid()) {
// We handle error cases.
return true;
}
return IsHandledProtocol(url.scheme());
}
void URLRequest::set_first_party_for_cookies(
const GURL& first_party_for_cookies) {
DCHECK(!is_pending_);
first_party_for_cookies_ = first_party_for_cookies;
}
void URLRequest::set_first_party_url_policy(
FirstPartyURLPolicy first_party_url_policy) {
DCHECK(!is_pending_);
first_party_url_policy_ = first_party_url_policy;
}
void URLRequest::set_initiator(const base::Optional<url::Origin>& initiator) {
DCHECK(!is_pending_);
DCHECK(!initiator.has_value() || initiator.value().unique() ||
initiator.value().GetURL().is_valid());
initiator_ = initiator;
}
void URLRequest::set_method(const std::string& method) {
DCHECK(!is_pending_);
method_ = method;
}
void URLRequest::SetReferrer(const std::string& referrer) {
DCHECK(!is_pending_);
GURL referrer_url(referrer);
if (referrer_url.is_valid()) {
referrer_ = referrer_url.GetAsReferrer().spec();
} else {
referrer_ = referrer;
}
}
void URLRequest::set_referrer_policy(ReferrerPolicy referrer_policy) {
DCHECK(!is_pending_);
referrer_policy_ = referrer_policy;
}
void URLRequest::set_delegate(Delegate* delegate) {
DCHECK(!delegate_);
DCHECK(delegate);
delegate_ = delegate;
}
void URLRequest::Start() {
DCHECK(delegate_);
if (!status_.is_success())
return;
// TODO(pkasting): Remove ScopedTracker below once crbug.com/456327 is fixed.
tracked_objects::ScopedTracker tracking_profile(
FROM_HERE_WITH_EXPLICIT_FUNCTION("456327 URLRequest::Start"));
// Some values can be NULL, but the job factory must not be.
DCHECK(context_->job_factory());
// Anything that sets |blocked_by_| before start should have cleaned up after
// itself.
DCHECK(blocked_by_.empty());
g_url_requests_started = true;
response_info_.request_time = base::Time::Now();
load_timing_info_ = LoadTimingInfo();
load_timing_info_.request_start_time = response_info_.request_time;
load_timing_info_.request_start = base::TimeTicks::Now();
if (network_delegate_) {
// TODO(mmenke): Remove ScopedTracker below once crbug.com/456327 is fixed.
tracked_objects::ScopedTracker tracking_profile25(
FROM_HERE_WITH_EXPLICIT_FUNCTION("456327 URLRequest::Start 2.5"));
OnCallToDelegate();
int error = network_delegate_->NotifyBeforeURLRequest(
this, before_request_callback_, &delegate_redirect_url_);
// If ERR_IO_PENDING is returned, the delegate will invoke
// |before_request_callback_| later.
if (error != ERR_IO_PENDING)
BeforeRequestComplete(error);
return;
}
// TODO(mmenke): Remove ScopedTracker below once crbug.com/456327 is fixed.
tracked_objects::ScopedTracker tracking_profile2(
FROM_HERE_WITH_EXPLICIT_FUNCTION("456327 URLRequest::Start 2"));
StartJob(URLRequestJobManager::GetInstance()->CreateJob(
this, network_delegate_));
}
///////////////////////////////////////////////////////////////////////////////
URLRequest::URLRequest(const GURL& url,
RequestPriority priority,
Delegate* delegate,
const URLRequestContext* context,
NetworkDelegate* network_delegate)
: context_(context),
network_delegate_(network_delegate ? network_delegate
: context->network_delegate()),
net_log_(NetLogWithSource::Make(context->net_log(),
NetLogSourceType::URL_REQUEST)),
url_chain_(1, url),
method_("GET"),
referrer_policy_(CLEAR_REFERRER_ON_TRANSITION_FROM_SECURE_TO_INSECURE),
first_party_url_policy_(NEVER_CHANGE_FIRST_PARTY_URL),
load_flags_(LOAD_NORMAL),
delegate_(delegate),
status_(URLRequestStatus::FromError(OK)),
is_pending_(false),
is_redirecting_(false),
redirect_limit_(kMaxRedirects),
priority_(priority),
identifier_(GenerateURLRequestIdentifier()),
calling_delegate_(false),
use_blocked_by_as_load_param_(false),
before_request_callback_(base::Bind(&URLRequest::BeforeRequestComplete,
base::Unretained(this))),
has_notified_completion_(false),
received_response_content_length_(0),
creation_time_(base::TimeTicks::Now()) {
// Sanity check out environment.
DCHECK(base::ThreadTaskRunnerHandle::IsSet());
context->url_requests()->insert(this);
net_log_.BeginEvent(NetLogEventType::REQUEST_ALIVE);
}
void URLRequest::BeforeRequestComplete(int error) {
DCHECK(!job_.get());
DCHECK_NE(ERR_IO_PENDING, error);
// Check that there are no callbacks to already canceled requests.
DCHECK_NE(URLRequestStatus::CANCELED, status_.status());
OnCallToDelegateComplete();
if (error != OK) {
std::string source("delegate");
net_log_.AddEvent(NetLogEventType::CANCELLED,
NetLog::StringCallback("source", &source));
StartJob(new URLRequestErrorJob(this, network_delegate_, error));
} else if (!delegate_redirect_url_.is_empty()) {
GURL new_url;
new_url.Swap(&delegate_redirect_url_);
URLRequestRedirectJob* job = new URLRequestRedirectJob(
this, network_delegate_, new_url,
// Use status code 307 to preserve the method, so POST requests work.
URLRequestRedirectJob::REDIRECT_307_TEMPORARY_REDIRECT, "Delegate");
StartJob(job);
} else {
StartJob(URLRequestJobManager::GetInstance()->CreateJob(
this, network_delegate_));
}
}
void URLRequest::StartJob(URLRequestJob* job) {
// TODO(mmenke): Remove ScopedTracker below once crbug.com/456327 is fixed.
tracked_objects::ScopedTracker tracking_profile(
FROM_HERE_WITH_EXPLICIT_FUNCTION("456327 URLRequest::StartJob"));
DCHECK(!is_pending_);
DCHECK(!job_.get());
net_log_.BeginEvent(
NetLogEventType::URL_REQUEST_START_JOB,
base::Bind(&NetLogURLRequestStartCallback, &url(), &method_, load_flags_,
priority_,
upload_data_stream_ ? upload_data_stream_->identifier() : -1));
job_.reset(job);
job_->SetExtraRequestHeaders(extra_request_headers_);
job_->SetPriority(priority_);
if (upload_data_stream_.get())
job_->SetUpload(upload_data_stream_.get());
is_pending_ = true;
is_redirecting_ = false;
response_info_.was_cached = false;
if (GURL(referrer_) != URLRequestJob::ComputeReferrerForRedirect(
referrer_policy_, referrer_, url())) {
if (!network_delegate_ ||
!network_delegate_->CancelURLRequestWithPolicyViolatingReferrerHeader(
*this, url(), GURL(referrer_))) {
referrer_.clear();
} else {
// We need to clear the referrer anyway to avoid an infinite recursion
// when starting the error job.
referrer_.clear();
std::string source("delegate");
net_log_.AddEvent(NetLogEventType::CANCELLED,
NetLog::StringCallback("source", &source));
RestartWithJob(new URLRequestErrorJob(
this, network_delegate_, ERR_BLOCKED_BY_CLIENT));
return;
}
}
// Start() always completes asynchronously.
//
// Status is generally set by URLRequestJob itself, but Start() calls
// directly into the URLRequestJob subclass, so URLRequestJob can't set it
// here.
// TODO(mmenke): Make the URLRequest manage its own status.
status_ = URLRequestStatus::FromError(ERR_IO_PENDING);
job_->Start();
}
void URLRequest::Restart() {
// Should only be called if the original job didn't make any progress.
DCHECK(job_.get() && !job_->has_response_started());
RestartWithJob(
URLRequestJobManager::GetInstance()->CreateJob(this, network_delegate_));
}
void URLRequest::RestartWithJob(URLRequestJob *job) {
DCHECK(job->request() == this);
PrepareToRestart();
StartJob(job);
}
int URLRequest::Cancel() {
return DoCancel(ERR_ABORTED, SSLInfo());
}
int URLRequest::CancelWithError(int error) {
return DoCancel(error, SSLInfo());
}
void URLRequest::CancelWithSSLError(int error, const SSLInfo& ssl_info) {
// This should only be called on a started request.
if (!is_pending_ || !job_.get() || job_->has_response_started()) {
NOTREACHED();
return;
}
DoCancel(error, ssl_info);
}
int URLRequest::DoCancel(int error, const SSLInfo& ssl_info) {
DCHECK(error < 0);
// If cancelled while calling a delegate, clear delegate info.
if (calling_delegate_) {
LogUnblocked();
OnCallToDelegateComplete();
}
// If the URL request already has an error status, then canceling is a no-op.
// Plus, we don't want to change the error status once it has been set.
if (status_.is_success()) {
status_ = URLRequestStatus(URLRequestStatus::CANCELED, error);
response_info_.ssl_info = ssl_info;
// If the request hasn't already been completed, log a cancellation event.
if (!has_notified_completion_) {
// Don't log an error code on ERR_ABORTED, since that's redundant.
net_log_.AddEventWithNetErrorCode(NetLogEventType::CANCELLED,
error == ERR_ABORTED ? OK : error);
}
}
if (is_pending_ && job_.get())
job_->Kill();
// We need to notify about the end of this job here synchronously. The
// Job sends an asynchronous notification but by the time this is processed,
// our |context_| is NULL.
NotifyRequestCompleted();
// The Job will call our NotifyDone method asynchronously. This is done so
// that the Delegate implementation can call Cancel without having to worry
// about being called recursively.
return status_.error();
}
int URLRequest::Read(IOBuffer* dest, int dest_size) {
DCHECK(job_.get());
// If this is the first read, end the delegate call that may have started in
// OnResponseStarted.
OnCallToDelegateComplete();
// If the request has failed, Read() will return actual network error code.
if (!status_.is_success())
return status_.error();
// This handles reads after the request already completed successfully.
// TODO(ahendrickson): DCHECK() that it is not done after
// http://crbug.com/115705 is fixed.
if (job_->is_done())
return status_.error();
if (dest_size == 0) {
// Caller is not too bright. I guess we've done what they asked.
return OK;
}
int rv = job_->Read(dest, dest_size);
if (rv == ERR_IO_PENDING)
set_status(URLRequestStatus::FromError(ERR_IO_PENDING));
// If rv is not 0 or actual bytes read, the status cannot be success.
DCHECK(rv >= 0 || status_.status() != URLRequestStatus::SUCCESS);
if (rv == 0 && status_.is_success())
NotifyRequestCompleted();
return rv;
}
// Deprecated.
bool URLRequest::Read(IOBuffer* dest, int dest_size, int* bytes_read) {
int result = Read(dest, dest_size);
if (result >= 0) {
*bytes_read = result;
return true;
}
if (result == ERR_IO_PENDING) {
*bytes_read = 0;
} else {
*bytes_read = -1;
}
return false;
}
void URLRequest::StopCaching() {
DCHECK(job_.get());
job_->StopCaching();
}
void URLRequest::NotifyReceivedRedirect(const RedirectInfo& redirect_info,
bool* defer_redirect) {
is_redirecting_ = true;
// TODO(davidben): Pass the full RedirectInfo down to MaybeInterceptRedirect?
URLRequestJob* job =
URLRequestJobManager::GetInstance()->MaybeInterceptRedirect(
this, network_delegate_, redirect_info.new_url);
if (job) {
RestartWithJob(job);
} else {
OnCallToDelegate();
delegate_->OnReceivedRedirect(this, redirect_info, defer_redirect);
// |this| may be have been destroyed here.
}
}
void URLRequest::NotifyResponseStarted(const URLRequestStatus& status) {
// Change status if there was an error.
if (status.status() != URLRequestStatus::SUCCESS)
set_status(status);
int net_error = OK;
if (!status_.is_success())
net_error = status_.error();
net_log_.EndEventWithNetErrorCode(NetLogEventType::URL_REQUEST_START_JOB,
net_error);
URLRequestJob* job =
URLRequestJobManager::GetInstance()->MaybeInterceptResponse(
this, network_delegate_);
if (job) {
RestartWithJob(job);
} else {
// In some cases (e.g. an event was canceled), we might have sent the
// completion event and receive a NotifyResponseStarted() later.
if (!has_notified_completion_ && status_.is_success()) {
if (network_delegate_)
network_delegate_->NotifyResponseStarted(this, net_error);
}
// Notify in case the entire URL Request has been finished.
if (!has_notified_completion_ && !status_.is_success())
NotifyRequestCompleted();
OnCallToDelegate();
delegate_->OnResponseStarted(this, net_error);
// Nothing may appear below this line as OnResponseStarted may delete
// |this|.
}
}
void URLRequest::FollowDeferredRedirect() {
DCHECK(job_.get());
DCHECK(status_.is_success());
status_ = URLRequestStatus::FromError(ERR_IO_PENDING);
job_->FollowDeferredRedirect();
}
void URLRequest::SetAuth(const AuthCredentials& credentials) {
DCHECK(job_.get());
DCHECK(job_->NeedsAuth());
status_ = URLRequestStatus::FromError(ERR_IO_PENDING);
job_->SetAuth(credentials);
}
void URLRequest::CancelAuth() {
DCHECK(job_.get());
DCHECK(job_->NeedsAuth());
status_ = URLRequestStatus::FromError(ERR_IO_PENDING);
job_->CancelAuth();
}
void URLRequest::ContinueWithCertificate(X509Certificate* client_cert,
SSLPrivateKey* client_private_key) {
DCHECK(job_.get());
// Matches the call in NotifyCertificateRequested.
OnCallToDelegateComplete();
status_ = URLRequestStatus::FromError(ERR_IO_PENDING);
job_->ContinueWithCertificate(client_cert, client_private_key);
}
void URLRequest::ContinueDespiteLastError() {
DCHECK(job_.get());
// Matches the call in NotifySSLCertificateError.
OnCallToDelegateComplete();
status_ = URLRequestStatus::FromError(ERR_IO_PENDING);
job_->ContinueDespiteLastError();
}
void URLRequest::PrepareToRestart() {
DCHECK(job_.get());
// Close the current URL_REQUEST_START_JOB, since we will be starting a new
// one.
net_log_.EndEvent(NetLogEventType::URL_REQUEST_START_JOB);
job_.reset();
response_info_ = HttpResponseInfo();
response_info_.request_time = base::Time::Now();
load_timing_info_ = LoadTimingInfo();
load_timing_info_.request_start_time = response_info_.request_time;
load_timing_info_.request_start = base::TimeTicks::Now();
status_ = URLRequestStatus();
is_pending_ = false;
proxy_server_ = ProxyServer();
}
int URLRequest::Redirect(const RedirectInfo& redirect_info) {
// Matches call in NotifyReceivedRedirect.
OnCallToDelegateComplete();
if (net_log_.IsCapturing()) {
net_log_.AddEvent(
NetLogEventType::URL_REQUEST_REDIRECTED,
NetLog::StringCallback("location",
&redirect_info.new_url.possibly_invalid_spec()));
}
// TODO(davidben): Pass the full RedirectInfo to the NetworkDelegate.
if (network_delegate_)
network_delegate_->NotifyBeforeRedirect(this, redirect_info.new_url);
if (redirect_limit_ <= 0) {
DVLOG(1) << "disallowing redirect: exceeds limit";
return ERR_TOO_MANY_REDIRECTS;
}
if (!redirect_info.new_url.is_valid())
return ERR_INVALID_URL;
if (!job_->IsSafeRedirect(redirect_info.new_url)) {
DVLOG(1) << "disallowing redirect: unsafe protocol";
return ERR_UNSAFE_REDIRECT;
}
if (!final_upload_progress_.position() && upload_data_stream_)
final_upload_progress_ = upload_data_stream_->GetUploadProgress();
PrepareToRestart();
if (redirect_info.new_method != method_) {
// TODO(davidben): This logic still needs to be replicated at the consumers.
if (method_ == "POST") {
// If being switched from POST, must remove Origin header.
// TODO(jww): This is Origin header removal is probably layering violation
// and
// should be refactored into //content. See https://crbug.com/471397.
extra_request_headers_.RemoveHeader(HttpRequestHeaders::kOrigin);
}
// The inclusion of a multipart Content-Type header can cause problems with
// some
// servers:
// http://code.google.com/p/chromium/issues/detail?id=843
extra_request_headers_.RemoveHeader(HttpRequestHeaders::kContentLength);
extra_request_headers_.RemoveHeader(HttpRequestHeaders::kContentType);
upload_data_stream_.reset();
method_ = redirect_info.new_method;
}
// Cross-origin redirects should not result in an Origin header value that is
// equal to the original request's Origin header. This is necessary to prevent
// a reflection of POST requests to bypass CSRF protections. If the header was
// not set to "null", a POST request from origin A to a malicious origin M
// could be redirected by M back to A.
//
// This behavior is specified in step 1 of step 10 of the 301, 302, 303, 307,
// 308 block of step 5 of Section 4.2 of Fetch[1] (which supercedes the
// behavior outlined in RFC 6454[2].
//
// [1]: https://fetch.spec.whatwg.org/#concept-http-fetch
// [2]: https://tools.ietf.org/html/rfc6454#section-7
//
// TODO(jww): This is a layering violation and should be refactored somewhere
// up into //net's embedder. https://crbug.com/471397
if (!url::Origin(redirect_info.new_url)
.IsSameOriginWith(url::Origin(url())) &&
extra_request_headers_.HasHeader(HttpRequestHeaders::kOrigin)) {
extra_request_headers_.SetHeader(HttpRequestHeaders::kOrigin,
url::Origin().Serialize());
}
referrer_ = redirect_info.new_referrer;
referrer_policy_ = redirect_info.new_referrer_policy;
first_party_for_cookies_ = redirect_info.new_first_party_for_cookies;
token_binding_referrer_ = redirect_info.referred_token_binding_host;
url_chain_.push_back(redirect_info.new_url);
--redirect_limit_;
Start();
return OK;
}
const URLRequestContext* URLRequest::context() const {
return context_;
}
int64_t URLRequest::GetExpectedContentSize() const {
int64_t expected_content_size = -1;
if (job_.get())
expected_content_size = job_->expected_content_size();
return expected_content_size;
}
void URLRequest::SetPriority(RequestPriority priority) {
DCHECK_GE(priority, MINIMUM_PRIORITY);
DCHECK_LE(priority, MAXIMUM_PRIORITY);
if ((load_flags_ & LOAD_IGNORE_LIMITS) && (priority != MAXIMUM_PRIORITY)) {
NOTREACHED();
// Maintain the invariant that requests with IGNORE_LIMITS set
// have MAXIMUM_PRIORITY for release mode.
return;
}
if (priority_ == priority)
return;
priority_ = priority;
if (job_.get()) {
net_log_.AddEvent(
NetLogEventType::URL_REQUEST_SET_PRIORITY,
NetLog::StringCallback("priority", RequestPriorityToString(priority_)));
job_->SetPriority(priority_);
}
}
void URLRequest::NotifyAuthRequired(AuthChallengeInfo* auth_info) {
NetworkDelegate::AuthRequiredResponse rv =
NetworkDelegate::AUTH_REQUIRED_RESPONSE_NO_ACTION;
auth_info_ = auth_info;
if (network_delegate_) {
OnCallToDelegate();
rv = network_delegate_->NotifyAuthRequired(
this,
*auth_info,
base::Bind(&URLRequest::NotifyAuthRequiredComplete,
base::Unretained(this)),
&auth_credentials_);
if (rv == NetworkDelegate::AUTH_REQUIRED_RESPONSE_IO_PENDING)
return;
}
NotifyAuthRequiredComplete(rv);
}
void URLRequest::NotifyAuthRequiredComplete(
NetworkDelegate::AuthRequiredResponse result) {
OnCallToDelegateComplete();
// Check that there are no callbacks to already canceled requests.
DCHECK_NE(URLRequestStatus::CANCELED, status_.status());
// NotifyAuthRequired may be called multiple times, such as
// when an authentication attempt fails. Clear out the data
// so it can be reset on another round.
AuthCredentials credentials = auth_credentials_;
auth_credentials_ = AuthCredentials();
scoped_refptr<AuthChallengeInfo> auth_info;
auth_info.swap(auth_info_);
switch (result) {
case NetworkDelegate::AUTH_REQUIRED_RESPONSE_NO_ACTION:
// Defer to the URLRequest::Delegate, since the NetworkDelegate
// didn't take an action.
delegate_->OnAuthRequired(this, auth_info.get());
break;
case NetworkDelegate::AUTH_REQUIRED_RESPONSE_SET_AUTH:
SetAuth(credentials);
break;
case NetworkDelegate::AUTH_REQUIRED_RESPONSE_CANCEL_AUTH:
CancelAuth();
break;
case NetworkDelegate::AUTH_REQUIRED_RESPONSE_IO_PENDING:
NOTREACHED();
break;
}
}
void URLRequest::NotifyCertificateRequested(
SSLCertRequestInfo* cert_request_info) {
status_ = URLRequestStatus();
OnCallToDelegate();
delegate_->OnCertificateRequested(this, cert_request_info);
}
void URLRequest::NotifySSLCertificateError(const SSLInfo& ssl_info,
bool fatal) {
status_ = URLRequestStatus();
OnCallToDelegate();
delegate_->OnSSLCertificateError(this, ssl_info, fatal);
}
bool URLRequest::CanGetCookies(const CookieList& cookie_list) const {
DCHECK(!(load_flags_ & LOAD_DO_NOT_SEND_COOKIES));
if (network_delegate_) {
return network_delegate_->CanGetCookies(*this, cookie_list);
}
return g_default_can_use_cookies;
}
bool URLRequest::CanSetCookie(const std::string& cookie_line,
CookieOptions* options) const {
DCHECK(!(load_flags_ & LOAD_DO_NOT_SAVE_COOKIES));
if (network_delegate_) {
return network_delegate_->CanSetCookie(*this, cookie_line, options);
}
return g_default_can_use_cookies;
}
bool URLRequest::CanEnablePrivacyMode() const {
if (network_delegate_) {
return network_delegate_->CanEnablePrivacyMode(url(),
first_party_for_cookies_);
}
return !g_default_can_use_cookies;
}
void URLRequest::NotifyReadCompleted(int bytes_read) {
if (bytes_read > 0)
set_status(URLRequestStatus());
// Notify in case the entire URL Request has been finished.
if (bytes_read <= 0)
NotifyRequestCompleted();
// When URLRequestJob notices there was an error in URLRequest's |status_|,
// it calls this method with |bytes_read| set to -1. Set it to a real error
// here.
// TODO(maksims): NotifyReadCompleted take the error code as an argument on
// failure, rather than -1.
if (bytes_read == -1)
bytes_read = status_.error();
// Notify NetworkChangeNotifier that we just received network data.
// This is to identify cases where the NetworkChangeNotifier thinks we
// are off-line but we are still receiving network data (crbug.com/124069),
// and to get rough network connection measurements.
if (bytes_read > 0 && !was_cached())
NetworkChangeNotifier::NotifyDataReceived(*this, bytes_read);
delegate_->OnReadCompleted(this, bytes_read);
// Nothing below this line as OnReadCompleted may delete |this|.
}
void URLRequest::OnHeadersComplete() {
set_status(URLRequestStatus());
// Cache load timing information now, as information will be lost once the
// socket is closed and the ClientSocketHandle is Reset, which will happen
// once the body is complete. The start times should already be populated.
if (job_.get()) {
// Keep a copy of the two times the URLRequest sets.
base::TimeTicks request_start = load_timing_info_.request_start;
base::Time request_start_time = load_timing_info_.request_start_time;
// Clear load times. Shouldn't be neded, but gives the GetLoadTimingInfo a
// consistent place to start from.
load_timing_info_ = LoadTimingInfo();
job_->GetLoadTimingInfo(&load_timing_info_);
load_timing_info_.request_start = request_start;
load_timing_info_.request_start_time = request_start_time;
raw_header_size_ = GetTotalReceivedBytes();
ConvertRealLoadTimesToBlockingTimes(&load_timing_info_);
}
}
void URLRequest::NotifyRequestCompleted() {
// TODO(battre): Get rid of this check, according to willchan it should
// not be needed.
if (has_notified_completion_)
return;
is_pending_ = false;
is_redirecting_ = false;
has_notified_completion_ = true;
if (network_delegate_)
network_delegate_->NotifyCompleted(this, job_.get() != NULL,
status_.error());
}
void URLRequest::OnCallToDelegate() {
DCHECK(!calling_delegate_);
DCHECK(blocked_by_.empty());
calling_delegate_ = true;
net_log_.BeginEvent(NetLogEventType::URL_REQUEST_DELEGATE);
}
void URLRequest::OnCallToDelegateComplete() {
// This should have been cleared before resuming the request.
DCHECK(blocked_by_.empty());
if (!calling_delegate_)
return;
calling_delegate_ = false;
net_log_.EndEvent(NetLogEventType::URL_REQUEST_DELEGATE);
}
void URLRequest::GetConnectionAttempts(ConnectionAttempts* out) const {
if (job_)
job_->GetConnectionAttempts(out);
else
out->clear();
}
void URLRequest::set_status(URLRequestStatus status) {
DCHECK(status_.is_io_pending() || status_.is_success() ||
(!status.is_success() && !status.is_io_pending()));
status_ = status;
}
} // namespace net