| // Copyright (c) 2012 The Chromium Authors. All rights reserved. | 
 | // Use of this source code is governed by a BSD-style license that can be | 
 | // found in the LICENSE file. | 
 |  | 
 | #include "net/socket/transport_client_socket_pool.h" | 
 |  | 
 | #include <algorithm> | 
 | #include <utility> | 
 |  | 
 | #include "base/compiler_specific.h" | 
 | #include "base/lazy_instance.h" | 
 | #include "base/logging.h" | 
 | #include "base/message_loop/message_loop.h" | 
 | #include "base/metrics/histogram_macros.h" | 
 | #include "base/profiler/scoped_tracker.h" | 
 | #include "base/strings/string_util.h" | 
 | #include "base/synchronization/lock.h" | 
 | #include "base/time/time.h" | 
 | #include "base/trace_event/trace_event.h" | 
 | #include "base/values.h" | 
 | #include "net/base/ip_endpoint.h" | 
 | #include "net/base/net_errors.h" | 
 | #include "net/log/net_log.h" | 
 | #include "net/socket/client_socket_factory.h" | 
 | #include "net/socket/client_socket_handle.h" | 
 | #include "net/socket/client_socket_pool_base.h" | 
 | #include "net/socket/socket_net_log_params.h" | 
 | #include "net/socket/socket_performance_watcher.h" | 
 | #include "net/socket/socket_performance_watcher_factory.h" | 
 | #include "net/socket/tcp_client_socket.h" | 
 |  | 
 | using base::TimeDelta; | 
 |  | 
 | namespace net { | 
 |  | 
 | namespace { | 
 |  | 
 | // Returns true iff all addresses in |list| are in the IPv6 family. | 
 | bool AddressListOnlyContainsIPv6(const AddressList& list) { | 
 |   DCHECK(!list.empty()); | 
 |   for (AddressList::const_iterator iter = list.begin(); iter != list.end(); | 
 |        ++iter) { | 
 |     if (iter->GetFamily() != ADDRESS_FAMILY_IPV6) | 
 |       return false; | 
 |   } | 
 |   return true; | 
 | } | 
 |  | 
 | }  // namespace | 
 |  | 
 | // This lock protects |g_last_connect_time|. | 
 | static base::LazyInstance<base::Lock>::Leaky | 
 |     g_last_connect_time_lock = LAZY_INSTANCE_INITIALIZER; | 
 |  | 
 | // |g_last_connect_time| has the last time a connect() call is made. | 
 | static base::LazyInstance<base::TimeTicks>::Leaky | 
 |     g_last_connect_time = LAZY_INSTANCE_INITIALIZER; | 
 |  | 
 | TransportSocketParams::TransportSocketParams( | 
 |     const HostPortPair& host_port_pair, | 
 |     bool disable_resolver_cache, | 
 |     const OnHostResolutionCallback& host_resolution_callback, | 
 |     CombineConnectAndWritePolicy combine_connect_and_write_if_supported) | 
 |     : destination_(host_port_pair), | 
 |       host_resolution_callback_(host_resolution_callback), | 
 |       combine_connect_and_write_(combine_connect_and_write_if_supported) { | 
 |   if (disable_resolver_cache) | 
 |     destination_.set_allow_cached_response(false); | 
 |   // combine_connect_and_write currently translates to TCP FastOpen. | 
 |   // Enable TCP FastOpen if user wants it. | 
 |   if (combine_connect_and_write_ == COMBINE_CONNECT_AND_WRITE_DEFAULT) { | 
 |     IsTCPFastOpenUserEnabled() ? combine_connect_and_write_ = | 
 |         COMBINE_CONNECT_AND_WRITE_DESIRED : | 
 |         COMBINE_CONNECT_AND_WRITE_PROHIBITED; | 
 |   } | 
 | } | 
 |  | 
 | TransportSocketParams::~TransportSocketParams() {} | 
 |  | 
 | // TODO(eroman): The use of this constant needs to be re-evaluated. The time | 
 | // needed for TCPClientSocketXXX::Connect() can be arbitrarily long, since | 
 | // the address list may contain many alternatives, and most of those may | 
 | // timeout. Even worse, the per-connect timeout threshold varies greatly | 
 | // between systems (anywhere from 20 seconds to 190 seconds). | 
 | // See comment #12 at http://crbug.com/23364 for specifics. | 
 | const int TransportConnectJob::kTimeoutInSeconds = 240;  // 4 minutes. | 
 |  | 
 | // TODO(willchan): Base this off RTT instead of statically setting it. Note we | 
 | // choose a timeout that is different from the backup connect job timer so they | 
 | // don't synchronize. | 
 | const int TransportConnectJob::kIPv6FallbackTimerInMs = 300; | 
 |  | 
 | TransportConnectJob::TransportConnectJob( | 
 |     const std::string& group_name, | 
 |     RequestPriority priority, | 
 |     ClientSocketPool::RespectLimits respect_limits, | 
 |     const scoped_refptr<TransportSocketParams>& params, | 
 |     base::TimeDelta timeout_duration, | 
 |     ClientSocketFactory* client_socket_factory, | 
 |     SocketPerformanceWatcherFactory* socket_performance_watcher_factory, | 
 |     HostResolver* host_resolver, | 
 |     Delegate* delegate, | 
 |     NetLog* net_log) | 
 |     : ConnectJob(group_name, | 
 |                  timeout_duration, | 
 |                  priority, | 
 |                  respect_limits, | 
 |                  delegate, | 
 |                  BoundNetLog::Make(net_log, NetLog::SOURCE_CONNECT_JOB)), | 
 |       params_(params), | 
 |       resolver_(host_resolver), | 
 |       client_socket_factory_(client_socket_factory), | 
 |       next_state_(STATE_NONE), | 
 |       socket_performance_watcher_factory_(socket_performance_watcher_factory), | 
 |       interval_between_connects_(CONNECT_INTERVAL_GT_20MS), | 
 |       resolve_result_(OK) {} | 
 |  | 
 | TransportConnectJob::~TransportConnectJob() { | 
 |   // We don't worry about cancelling the host resolution and TCP connect, since | 
 |   // ~SingleRequestHostResolver and ~StreamSocket will take care of it. | 
 | } | 
 |  | 
 | LoadState TransportConnectJob::GetLoadState() const { | 
 |   switch (next_state_) { | 
 |     case STATE_RESOLVE_HOST: | 
 |     case STATE_RESOLVE_HOST_COMPLETE: | 
 |       return LOAD_STATE_RESOLVING_HOST; | 
 |     case STATE_TRANSPORT_CONNECT: | 
 |     case STATE_TRANSPORT_CONNECT_COMPLETE: | 
 |       return LOAD_STATE_CONNECTING; | 
 |     case STATE_NONE: | 
 |       return LOAD_STATE_IDLE; | 
 |   } | 
 |   NOTREACHED(); | 
 |   return LOAD_STATE_IDLE; | 
 | } | 
 |  | 
 | void TransportConnectJob::GetAdditionalErrorState(ClientSocketHandle* handle) { | 
 |   // If hostname resolution failed, record an empty endpoint and the result. | 
 |   // Also record any attempts made on either of the sockets. | 
 |   ConnectionAttempts attempts; | 
 |   if (resolve_result_ != OK) { | 
 |     DCHECK_EQ(0u, addresses_.size()); | 
 |     attempts.push_back(ConnectionAttempt(IPEndPoint(), resolve_result_)); | 
 |   } | 
 |   attempts.insert(attempts.begin(), connection_attempts_.begin(), | 
 |                   connection_attempts_.end()); | 
 |   attempts.insert(attempts.begin(), fallback_connection_attempts_.begin(), | 
 |                   fallback_connection_attempts_.end()); | 
 |   handle->set_connection_attempts(attempts); | 
 | } | 
 |  | 
 | // static | 
 | void TransportConnectJob::MakeAddressListStartWithIPv4(AddressList* list) { | 
 |   for (AddressList::iterator i = list->begin(); i != list->end(); ++i) { | 
 |     if (i->GetFamily() == ADDRESS_FAMILY_IPV4) { | 
 |       std::rotate(list->begin(), i, list->end()); | 
 |       break; | 
 |     } | 
 |   } | 
 | } | 
 |  | 
 | // static | 
 | base::TimeDelta TransportConnectJob::HistogramDuration( | 
 |     const LoadTimingInfo::ConnectTiming& connect_timing, | 
 |     RaceResult race_result) { | 
 |   DCHECK(!connect_timing.connect_start.is_null()); | 
 |   DCHECK(!connect_timing.dns_start.is_null()); | 
 |   base::TimeTicks now = base::TimeTicks::Now(); | 
 |   base::TimeDelta total_duration = now - connect_timing.dns_start; | 
 |   UMA_HISTOGRAM_CUSTOM_TIMES("Net.DNS_Resolution_And_TCP_Connection_Latency2", | 
 |                              total_duration, | 
 |                              base::TimeDelta::FromMilliseconds(1), | 
 |                              base::TimeDelta::FromMinutes(10), | 
 |                              100); | 
 |  | 
 |   base::TimeDelta connect_duration = now - connect_timing.connect_start; | 
 |   UMA_HISTOGRAM_CUSTOM_TIMES("Net.TCP_Connection_Latency", | 
 |                              connect_duration, | 
 |                              base::TimeDelta::FromMilliseconds(1), | 
 |                              base::TimeDelta::FromMinutes(10), | 
 |                              100); | 
 |  | 
 |   switch (race_result) { | 
 |     case RACE_IPV4_WINS: | 
 |       UMA_HISTOGRAM_CUSTOM_TIMES("Net.TCP_Connection_Latency_IPv4_Wins_Race", | 
 |                                  connect_duration, | 
 |                                  base::TimeDelta::FromMilliseconds(1), | 
 |                                  base::TimeDelta::FromMinutes(10), | 
 |                                  100); | 
 |       break; | 
 |  | 
 |     case RACE_IPV4_SOLO: | 
 |       UMA_HISTOGRAM_CUSTOM_TIMES("Net.TCP_Connection_Latency_IPv4_No_Race", | 
 |                                  connect_duration, | 
 |                                  base::TimeDelta::FromMilliseconds(1), | 
 |                                  base::TimeDelta::FromMinutes(10), | 
 |                                  100); | 
 |       break; | 
 |  | 
 |     case RACE_IPV6_WINS: | 
 |       UMA_HISTOGRAM_CUSTOM_TIMES("Net.TCP_Connection_Latency_IPv6_Raceable", | 
 |                                  connect_duration, | 
 |                                  base::TimeDelta::FromMilliseconds(1), | 
 |                                  base::TimeDelta::FromMinutes(10), | 
 |                                  100); | 
 |       break; | 
 |  | 
 |     case RACE_IPV6_SOLO: | 
 |       UMA_HISTOGRAM_CUSTOM_TIMES("Net.TCP_Connection_Latency_IPv6_Solo", | 
 |                                  connect_duration, | 
 |                                  base::TimeDelta::FromMilliseconds(1), | 
 |                                  base::TimeDelta::FromMinutes(10), | 
 |                                  100); | 
 |       break; | 
 |  | 
 |     default: | 
 |       NOTREACHED(); | 
 |       break; | 
 |   } | 
 |  | 
 |   return connect_duration; | 
 | } | 
 |  | 
 | void TransportConnectJob::OnIOComplete(int result) { | 
 |   result = DoLoop(result); | 
 |   if (result != ERR_IO_PENDING) | 
 |     NotifyDelegateOfCompletion(result);  // Deletes |this| | 
 | } | 
 |  | 
 | int TransportConnectJob::DoLoop(int result) { | 
 |   DCHECK_NE(next_state_, STATE_NONE); | 
 |  | 
 |   int rv = result; | 
 |   do { | 
 |     State state = next_state_; | 
 |     next_state_ = STATE_NONE; | 
 |     switch (state) { | 
 |       case STATE_RESOLVE_HOST: | 
 |         DCHECK_EQ(OK, rv); | 
 |         rv = DoResolveHost(); | 
 |         break; | 
 |       case STATE_RESOLVE_HOST_COMPLETE: | 
 |         rv = DoResolveHostComplete(rv); | 
 |         break; | 
 |       case STATE_TRANSPORT_CONNECT: | 
 |         DCHECK_EQ(OK, rv); | 
 |         rv = DoTransportConnect(); | 
 |         break; | 
 |       case STATE_TRANSPORT_CONNECT_COMPLETE: | 
 |         rv = DoTransportConnectComplete(rv); | 
 |         break; | 
 |       default: | 
 |         NOTREACHED(); | 
 |         rv = ERR_FAILED; | 
 |         break; | 
 |     } | 
 |   } while (rv != ERR_IO_PENDING && next_state_ != STATE_NONE); | 
 |  | 
 |   return rv; | 
 | } | 
 | int TransportConnectJob::DoResolveHost() { | 
 |   // TODO(ricea): Remove ScopedTracker below once crbug.com/436634 is fixed. | 
 |   tracked_objects::ScopedTracker tracking_profile( | 
 |       FROM_HERE_WITH_EXPLICIT_FUNCTION( | 
 |           "436634 TransportConnectJob::DoResolveHost")); | 
 |  | 
 |   next_state_ = STATE_RESOLVE_HOST_COMPLETE; | 
 |   connect_timing_.dns_start = base::TimeTicks::Now(); | 
 |  | 
 |   return resolver_.Resolve( | 
 |       params_->destination(), priority(), &addresses_, | 
 |       base::Bind(&TransportConnectJob::OnIOComplete, base::Unretained(this)), | 
 |       net_log()); | 
 | } | 
 |  | 
 | int TransportConnectJob::DoResolveHostComplete(int result) { | 
 |   TRACE_EVENT0("net", "TransportConnectJob::DoResolveHostComplete"); | 
 |   connect_timing_.dns_end = base::TimeTicks::Now(); | 
 |   // Overwrite connection start time, since for connections that do not go | 
 |   // through proxies, |connect_start| should not include dns lookup time. | 
 |   connect_timing_.connect_start = connect_timing_.dns_end; | 
 |   resolve_result_ = result; | 
 |  | 
 |   if (result != OK) | 
 |     return result; | 
 |  | 
 |   // Invoke callback, and abort if it fails. | 
 |   if (!params_->host_resolution_callback().is_null()) { | 
 |     result = params_->host_resolution_callback().Run(addresses_, net_log()); | 
 |     if (result != OK) | 
 |       return result; | 
 |   } | 
 |  | 
 |   next_state_ = STATE_TRANSPORT_CONNECT; | 
 |   return result; | 
 | } | 
 |  | 
 | int TransportConnectJob::DoTransportConnect() { | 
 |   base::TimeTicks now = base::TimeTicks::Now(); | 
 |   base::TimeTicks last_connect_time; | 
 |   { | 
 |     base::AutoLock lock(g_last_connect_time_lock.Get()); | 
 |     last_connect_time = g_last_connect_time.Get(); | 
 |     *g_last_connect_time.Pointer() = now; | 
 |   } | 
 |   if (last_connect_time.is_null()) { | 
 |     interval_between_connects_ = CONNECT_INTERVAL_GT_20MS; | 
 |   } else { | 
 |     int64_t interval = (now - last_connect_time).InMilliseconds(); | 
 |     if (interval <= 10) | 
 |       interval_between_connects_ = CONNECT_INTERVAL_LE_10MS; | 
 |     else if (interval <= 20) | 
 |       interval_between_connects_ = CONNECT_INTERVAL_LE_20MS; | 
 |     else | 
 |       interval_between_connects_ = CONNECT_INTERVAL_GT_20MS; | 
 |   } | 
 |  | 
 |   next_state_ = STATE_TRANSPORT_CONNECT_COMPLETE; | 
 |   // Create a |SocketPerformanceWatcher|, and pass the ownership. | 
 |   std::unique_ptr<SocketPerformanceWatcher> socket_performance_watcher; | 
 |   if (socket_performance_watcher_factory_) { | 
 |     socket_performance_watcher = | 
 |         socket_performance_watcher_factory_->CreateSocketPerformanceWatcher( | 
 |             SocketPerformanceWatcherFactory::PROTOCOL_TCP); | 
 |   } | 
 |   transport_socket_ = client_socket_factory_->CreateTransportClientSocket( | 
 |       addresses_, std::move(socket_performance_watcher), net_log().net_log(), | 
 |       net_log().source()); | 
 |  | 
 |   // If the list contains IPv6 and IPv4 addresses, the first address will | 
 |   // be IPv6, and the IPv4 addresses will be tried as fallback addresses, | 
 |   // per "Happy Eyeballs" (RFC 6555). | 
 |   bool try_ipv6_connect_with_ipv4_fallback = | 
 |       addresses_.front().GetFamily() == ADDRESS_FAMILY_IPV6 && | 
 |       !AddressListOnlyContainsIPv6(addresses_); | 
 |  | 
 |   // Enable TCP FastOpen if indicated by transport socket params. | 
 |   // Note: We currently do not turn on TCP FastOpen for destinations where | 
 |   // we try a TCP connect over IPv6 with fallback to IPv4. | 
 |   if (!try_ipv6_connect_with_ipv4_fallback && | 
 |       params_->combine_connect_and_write() == | 
 |           TransportSocketParams::COMBINE_CONNECT_AND_WRITE_DESIRED) { | 
 |     transport_socket_->EnableTCPFastOpenIfSupported(); | 
 |   } | 
 |  | 
 |   int rv = transport_socket_->Connect( | 
 |       base::Bind(&TransportConnectJob::OnIOComplete, base::Unretained(this))); | 
 |   if (rv == ERR_IO_PENDING && try_ipv6_connect_with_ipv4_fallback) { | 
 |     fallback_timer_.Start( | 
 |         FROM_HERE, base::TimeDelta::FromMilliseconds(kIPv6FallbackTimerInMs), | 
 |         this, &TransportConnectJob::DoIPv6FallbackTransportConnect); | 
 |   } | 
 |   return rv; | 
 | } | 
 |  | 
 | int TransportConnectJob::DoTransportConnectComplete(int result) { | 
 |   if (result == OK) { | 
 |     // Success will be returned via the main socket, so also include connection | 
 |     // attempts made on the fallback socket up to this point. (Unfortunately, | 
 |     // the only simple way to return information in the success case is through | 
 |     // the successfully-connected socket.) | 
 |     if (fallback_transport_socket_) { | 
 |       ConnectionAttempts fallback_attempts; | 
 |       fallback_transport_socket_->GetConnectionAttempts(&fallback_attempts); | 
 |       transport_socket_->AddConnectionAttempts(fallback_attempts); | 
 |     } | 
 |  | 
 |     bool is_ipv4 = addresses_.front().GetFamily() == ADDRESS_FAMILY_IPV4; | 
 |     RaceResult race_result = RACE_UNKNOWN; | 
 |     if (is_ipv4) | 
 |       race_result = RACE_IPV4_SOLO; | 
 |     else if (AddressListOnlyContainsIPv6(addresses_)) | 
 |       race_result = RACE_IPV6_SOLO; | 
 |     else | 
 |       race_result = RACE_IPV6_WINS; | 
 |     base::TimeDelta connect_duration = | 
 |         HistogramDuration(connect_timing_, race_result); | 
 |     switch (interval_between_connects_) { | 
 |       case CONNECT_INTERVAL_LE_10MS: | 
 |         UMA_HISTOGRAM_CUSTOM_TIMES( | 
 |             "Net.TCP_Connection_Latency_Interval_LessThanOrEqual_10ms", | 
 |             connect_duration, | 
 |             base::TimeDelta::FromMilliseconds(1), | 
 |             base::TimeDelta::FromMinutes(10), | 
 |             100); | 
 |         break; | 
 |       case CONNECT_INTERVAL_LE_20MS: | 
 |         UMA_HISTOGRAM_CUSTOM_TIMES( | 
 |             "Net.TCP_Connection_Latency_Interval_LessThanOrEqual_20ms", | 
 |             connect_duration, | 
 |             base::TimeDelta::FromMilliseconds(1), | 
 |             base::TimeDelta::FromMinutes(10), | 
 |             100); | 
 |         break; | 
 |       case CONNECT_INTERVAL_GT_20MS: | 
 |         UMA_HISTOGRAM_CUSTOM_TIMES( | 
 |             "Net.TCP_Connection_Latency_Interval_GreaterThan_20ms", | 
 |             connect_duration, | 
 |             base::TimeDelta::FromMilliseconds(1), | 
 |             base::TimeDelta::FromMinutes(10), | 
 |             100); | 
 |         break; | 
 |       default: | 
 |         NOTREACHED(); | 
 |         break; | 
 |     } | 
 |  | 
 |     SetSocket(std::move(transport_socket_)); | 
 |   } else { | 
 |     // Failure will be returned via |GetAdditionalErrorState|, so save | 
 |     // connection attempts from both sockets for use there. | 
 |     CopyConnectionAttemptsFromSockets(); | 
 |  | 
 |     transport_socket_.reset(); | 
 |   } | 
 |  | 
 |   fallback_timer_.Stop(); | 
 |   fallback_transport_socket_.reset(); | 
 |   fallback_addresses_.reset(); | 
 |  | 
 |   return result; | 
 | } | 
 |  | 
 | void TransportConnectJob::DoIPv6FallbackTransportConnect() { | 
 |   // The timer should only fire while we're waiting for the main connect to | 
 |   // succeed. | 
 |   if (next_state_ != STATE_TRANSPORT_CONNECT_COMPLETE) { | 
 |     NOTREACHED(); | 
 |     return; | 
 |   } | 
 |  | 
 |   DCHECK(!fallback_transport_socket_.get()); | 
 |   DCHECK(!fallback_addresses_.get()); | 
 |  | 
 |   // Create a |SocketPerformanceWatcher|, and pass the ownership. | 
 |   std::unique_ptr<SocketPerformanceWatcher> socket_performance_watcher; | 
 |   if (socket_performance_watcher_factory_) { | 
 |     socket_performance_watcher = | 
 |         socket_performance_watcher_factory_->CreateSocketPerformanceWatcher( | 
 |             SocketPerformanceWatcherFactory::PROTOCOL_TCP); | 
 |   } | 
 |  | 
 |   fallback_addresses_.reset(new AddressList(addresses_)); | 
 |   MakeAddressListStartWithIPv4(fallback_addresses_.get()); | 
 |   fallback_transport_socket_ = | 
 |       client_socket_factory_->CreateTransportClientSocket( | 
 |           *fallback_addresses_, std::move(socket_performance_watcher), | 
 |           net_log().net_log(), net_log().source()); | 
 |   fallback_connect_start_time_ = base::TimeTicks::Now(); | 
 |   int rv = fallback_transport_socket_->Connect( | 
 |       base::Bind( | 
 |           &TransportConnectJob::DoIPv6FallbackTransportConnectComplete, | 
 |           base::Unretained(this))); | 
 |   if (rv != ERR_IO_PENDING) | 
 |     DoIPv6FallbackTransportConnectComplete(rv); | 
 | } | 
 |  | 
 | void TransportConnectJob::DoIPv6FallbackTransportConnectComplete(int result) { | 
 |   // This should only happen when we're waiting for the main connect to succeed. | 
 |   if (next_state_ != STATE_TRANSPORT_CONNECT_COMPLETE) { | 
 |     NOTREACHED(); | 
 |     return; | 
 |   } | 
 |  | 
 |   DCHECK_NE(ERR_IO_PENDING, result); | 
 |   DCHECK(fallback_transport_socket_.get()); | 
 |   DCHECK(fallback_addresses_.get()); | 
 |  | 
 |   if (result == OK) { | 
 |     DCHECK(!fallback_connect_start_time_.is_null()); | 
 |  | 
 |     // Success will be returned via the fallback socket, so also include | 
 |     // connection attempts made on the main socket up to this point. | 
 |     // (Unfortunately, the only simple way to return information in the success | 
 |     // case is through the successfully-connected socket.) | 
 |     if (transport_socket_) { | 
 |       ConnectionAttempts attempts; | 
 |       transport_socket_->GetConnectionAttempts(&attempts); | 
 |       fallback_transport_socket_->AddConnectionAttempts(attempts); | 
 |     } | 
 |  | 
 |     connect_timing_.connect_start = fallback_connect_start_time_; | 
 |     HistogramDuration(connect_timing_, RACE_IPV4_WINS); | 
 |     SetSocket(std::move(fallback_transport_socket_)); | 
 |     next_state_ = STATE_NONE; | 
 |   } else { | 
 |     // Failure will be returned via |GetAdditionalErrorState|, so save | 
 |     // connection attempts from both sockets for use there. | 
 |     CopyConnectionAttemptsFromSockets(); | 
 |  | 
 |     fallback_transport_socket_.reset(); | 
 |     fallback_addresses_.reset(); | 
 |   } | 
 |  | 
 |   transport_socket_.reset(); | 
 |  | 
 |   NotifyDelegateOfCompletion(result);  // Deletes |this| | 
 | } | 
 |  | 
 | int TransportConnectJob::ConnectInternal() { | 
 |   next_state_ = STATE_RESOLVE_HOST; | 
 |   return DoLoop(OK); | 
 | } | 
 |  | 
 | void TransportConnectJob::CopyConnectionAttemptsFromSockets() { | 
 |   if (transport_socket_) | 
 |     transport_socket_->GetConnectionAttempts(&connection_attempts_); | 
 |   if (fallback_transport_socket_) { | 
 |     fallback_transport_socket_->GetConnectionAttempts( | 
 |         &fallback_connection_attempts_); | 
 |   } | 
 | } | 
 |  | 
 | std::unique_ptr<ConnectJob> | 
 | TransportClientSocketPool::TransportConnectJobFactory::NewConnectJob( | 
 |     const std::string& group_name, | 
 |     const PoolBase::Request& request, | 
 |     ConnectJob::Delegate* delegate) const { | 
 |   return std::unique_ptr<ConnectJob>(new TransportConnectJob( | 
 |       group_name, request.priority(), request.respect_limits(), | 
 |       request.params(), ConnectionTimeout(), client_socket_factory_, | 
 |       socket_performance_watcher_factory_, host_resolver_, delegate, net_log_)); | 
 | } | 
 |  | 
 | base::TimeDelta | 
 |     TransportClientSocketPool::TransportConnectJobFactory::ConnectionTimeout() | 
 |     const { | 
 |   return base::TimeDelta::FromSeconds(TransportConnectJob::kTimeoutInSeconds); | 
 | } | 
 |  | 
 | TransportClientSocketPool::TransportClientSocketPool( | 
 |     int max_sockets, | 
 |     int max_sockets_per_group, | 
 |     HostResolver* host_resolver, | 
 |     ClientSocketFactory* client_socket_factory, | 
 |     SocketPerformanceWatcherFactory* socket_performance_watcher_factory, | 
 |     NetLog* net_log) | 
 |     : base_(NULL, | 
 |             max_sockets, | 
 |             max_sockets_per_group, | 
 |             ClientSocketPool::unused_idle_socket_timeout(), | 
 |             ClientSocketPool::used_idle_socket_timeout(), | 
 |             new TransportConnectJobFactory(client_socket_factory, | 
 |                                            host_resolver, | 
 |                                            socket_performance_watcher_factory, | 
 |                                            net_log)) { | 
 |   base_.EnableConnectBackupJobs(); | 
 | } | 
 |  | 
 | TransportClientSocketPool::~TransportClientSocketPool() {} | 
 |  | 
 | int TransportClientSocketPool::RequestSocket(const std::string& group_name, | 
 |                                              const void* params, | 
 |                                              RequestPriority priority, | 
 |                                              RespectLimits respect_limits, | 
 |                                              ClientSocketHandle* handle, | 
 |                                              const CompletionCallback& callback, | 
 |                                              const BoundNetLog& net_log) { | 
 |   const scoped_refptr<TransportSocketParams>* casted_params = | 
 |       static_cast<const scoped_refptr<TransportSocketParams>*>(params); | 
 |  | 
 |   NetLogTcpClientSocketPoolRequestedSocket(net_log, casted_params); | 
 |  | 
 |   return base_.RequestSocket(group_name, *casted_params, priority, | 
 |                              respect_limits, handle, callback, net_log); | 
 | } | 
 |  | 
 | void TransportClientSocketPool::NetLogTcpClientSocketPoolRequestedSocket( | 
 |     const BoundNetLog& net_log, | 
 |     const scoped_refptr<TransportSocketParams>* casted_params) { | 
 |   if (net_log.IsCapturing()) { | 
 |     // TODO(eroman): Split out the host and port parameters. | 
 |     net_log.AddEvent( | 
 |         NetLog::TYPE_TCP_CLIENT_SOCKET_POOL_REQUESTED_SOCKET, | 
 |         CreateNetLogHostPortPairCallback( | 
 |             &casted_params->get()->destination().host_port_pair())); | 
 |   } | 
 | } | 
 |  | 
 | void TransportClientSocketPool::RequestSockets( | 
 |     const std::string& group_name, | 
 |     const void* params, | 
 |     int num_sockets, | 
 |     const BoundNetLog& net_log) { | 
 |   const scoped_refptr<TransportSocketParams>* casted_params = | 
 |       static_cast<const scoped_refptr<TransportSocketParams>*>(params); | 
 |  | 
 |   if (net_log.IsCapturing()) { | 
 |     // TODO(eroman): Split out the host and port parameters. | 
 |     net_log.AddEvent( | 
 |         NetLog::TYPE_TCP_CLIENT_SOCKET_POOL_REQUESTED_SOCKETS, | 
 |         CreateNetLogHostPortPairCallback( | 
 |             &casted_params->get()->destination().host_port_pair())); | 
 |   } | 
 |  | 
 |   base_.RequestSockets(group_name, *casted_params, num_sockets, net_log); | 
 | } | 
 |  | 
 | void TransportClientSocketPool::CancelRequest( | 
 |     const std::string& group_name, | 
 |     ClientSocketHandle* handle) { | 
 |   base_.CancelRequest(group_name, handle); | 
 | } | 
 |  | 
 | void TransportClientSocketPool::ReleaseSocket( | 
 |     const std::string& group_name, | 
 |     std::unique_ptr<StreamSocket> socket, | 
 |     int id) { | 
 |   base_.ReleaseSocket(group_name, std::move(socket), id); | 
 | } | 
 |  | 
 | void TransportClientSocketPool::FlushWithError(int error) { | 
 |   base_.FlushWithError(error); | 
 | } | 
 |  | 
 | void TransportClientSocketPool::CloseIdleSockets() { | 
 |   base_.CloseIdleSockets(); | 
 | } | 
 |  | 
 | int TransportClientSocketPool::IdleSocketCount() const { | 
 |   return base_.idle_socket_count(); | 
 | } | 
 |  | 
 | int TransportClientSocketPool::IdleSocketCountInGroup( | 
 |     const std::string& group_name) const { | 
 |   return base_.IdleSocketCountInGroup(group_name); | 
 | } | 
 |  | 
 | LoadState TransportClientSocketPool::GetLoadState( | 
 |     const std::string& group_name, const ClientSocketHandle* handle) const { | 
 |   return base_.GetLoadState(group_name, handle); | 
 | } | 
 |  | 
 | std::unique_ptr<base::DictionaryValue> | 
 | TransportClientSocketPool::GetInfoAsValue(const std::string& name, | 
 |                                           const std::string& type, | 
 |                                           bool include_nested_pools) const { | 
 |   return base_.GetInfoAsValue(name, type); | 
 | } | 
 |  | 
 | base::TimeDelta TransportClientSocketPool::ConnectionTimeout() const { | 
 |   return base_.ConnectionTimeout(); | 
 | } | 
 |  | 
 | bool TransportClientSocketPool::IsStalled() const { | 
 |   return base_.IsStalled(); | 
 | } | 
 |  | 
 | void TransportClientSocketPool::AddHigherLayeredPool( | 
 |     HigherLayeredPool* higher_pool) { | 
 |   base_.AddHigherLayeredPool(higher_pool); | 
 | } | 
 |  | 
 | void TransportClientSocketPool::RemoveHigherLayeredPool( | 
 |     HigherLayeredPool* higher_pool) { | 
 |   base_.RemoveHigherLayeredPool(higher_pool); | 
 | } | 
 |  | 
 | }  // namespace net |