// Copyright (c) 2012 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "net/url_request/url_request.h" #include "base/bind.h" #include "base/bind_helpers.h" #include "base/callback.h" #include "base/compiler_specific.h" #include "base/debug/stack_trace.h" #include "base/lazy_instance.h" #include "base/memory/singleton.h" #include "base/message_loop/message_loop.h" #include "base/metrics/stats_counters.h" #include "base/stl_util.h" #include "base/strings/utf_string_conversions.h" #include "base/synchronization/lock.h" #include "base/values.h" #include "net/base/auth.h" #include "net/base/host_port_pair.h" #include "net/base/load_flags.h" #include "net/base/load_timing_info.h" #include "net/base/net_errors.h" #include "net/base/net_log.h" #include "net/base/network_change_notifier.h" #include "net/base/network_delegate.h" #include "net/base/upload_data_stream.h" #include "net/http/http_response_headers.h" #include "net/http/http_util.h" #include "net/ssl/ssl_cert_request_info.h" #include "net/url_request/redirect_info.h" #include "net/url_request/url_request_context.h" #include "net/url_request/url_request_error_job.h" #include "net/url_request/url_request_job.h" #include "net/url_request/url_request_job_manager.h" #include "net/url_request/url_request_netlog_params.h" #include "net/url_request/url_request_redirect_job.h" using base::Time; using std::string; namespace net { namespace { // Max number of http redirects to follow. Same number as gecko. const int kMaxRedirects = 20; // Discard headers which have meaning in POST (Content-Length, Content-Type, // Origin). void StripPostSpecificHeaders(HttpRequestHeaders* headers) { // These are headers that may be attached to a POST. headers->RemoveHeader(HttpRequestHeaders::kContentLength); headers->RemoveHeader(HttpRequestHeaders::kContentType); headers->RemoveHeader(HttpRequestHeaders::kOrigin); } // TODO(battre): Delete this, see http://crbug.com/89321: // This counter keeps track of the identifiers used for URL requests so far. // 0 is reserved to represent an invalid ID. uint64 g_next_url_request_identifier = 1; // This lock protects g_next_url_request_identifier. base::LazyInstance<base::Lock>::Leaky g_next_url_request_identifier_lock = LAZY_INSTANCE_INITIALIZER; // Returns an prior unused identifier for URL requests. uint64 GenerateURLRequestIdentifier() { base::AutoLock lock(g_next_url_request_identifier_lock.Get()); return g_next_url_request_identifier++; } // True once the first URLRequest was started. bool g_url_requests_started = false; // True if cookies are accepted by default. bool g_default_can_use_cookies = true; // When the URLRequest first assempts load timing information, it has the times // at which each event occurred. The API requires the time which the request // was blocked on each phase. This function handles the conversion. // // In the case of reusing a SPDY session, old proxy results may have been // reused, so proxy resolution times may be before the request was started. // // Due to preconnect and late binding, it is also possible for the connection // attempt to start before a request has been started, or proxy resolution // completed. // // This functions fixes both those cases. void ConvertRealLoadTimesToBlockingTimes( net::LoadTimingInfo* load_timing_info) { DCHECK(!load_timing_info->request_start.is_null()); // Earliest time possible for the request to be blocking on connect events. base::TimeTicks block_on_connect = load_timing_info->request_start; if (!load_timing_info->proxy_resolve_start.is_null()) { DCHECK(!load_timing_info->proxy_resolve_end.is_null()); // Make sure the proxy times are after request start. if (load_timing_info->proxy_resolve_start < load_timing_info->request_start) load_timing_info->proxy_resolve_start = load_timing_info->request_start; if (load_timing_info->proxy_resolve_end < load_timing_info->request_start) load_timing_info->proxy_resolve_end = load_timing_info->request_start; // Connect times must also be after the proxy times. block_on_connect = load_timing_info->proxy_resolve_end; } // Make sure connection times are after start and proxy times. net::LoadTimingInfo::ConnectTiming* connect_timing = &load_timing_info->connect_timing; if (!connect_timing->dns_start.is_null()) { DCHECK(!connect_timing->dns_end.is_null()); if (connect_timing->dns_start < block_on_connect) connect_timing->dns_start = block_on_connect; if (connect_timing->dns_end < block_on_connect) connect_timing->dns_end = block_on_connect; } if (!connect_timing->connect_start.is_null()) { DCHECK(!connect_timing->connect_end.is_null()); if (connect_timing->connect_start < block_on_connect) connect_timing->connect_start = block_on_connect; if (connect_timing->connect_end < block_on_connect) connect_timing->connect_end = block_on_connect; } if (!connect_timing->ssl_start.is_null()) { DCHECK(!connect_timing->ssl_end.is_null()); if (connect_timing->ssl_start < block_on_connect) connect_timing->ssl_start = block_on_connect; if (connect_timing->ssl_end < block_on_connect) connect_timing->ssl_end = block_on_connect; } } } // namespace void URLRequest::Deprecated::RegisterRequestInterceptor( Interceptor* interceptor) { URLRequest::RegisterRequestInterceptor(interceptor); } void URLRequest::Deprecated::UnregisterRequestInterceptor( Interceptor* interceptor) { URLRequest::UnregisterRequestInterceptor(interceptor); } /////////////////////////////////////////////////////////////////////////////// // URLRequest::Interceptor URLRequestJob* URLRequest::Interceptor::MaybeInterceptRedirect( URLRequest* request, NetworkDelegate* network_delegate, const GURL& location) { return NULL; } URLRequestJob* URLRequest::Interceptor::MaybeInterceptResponse( URLRequest* request, NetworkDelegate* network_delegate) { return NULL; } /////////////////////////////////////////////////////////////////////////////// // URLRequest::Delegate void URLRequest::Delegate::OnReceivedRedirect(URLRequest* request, const RedirectInfo& redirect_info, bool* defer_redirect) { } void URLRequest::Delegate::OnAuthRequired(URLRequest* request, AuthChallengeInfo* auth_info) { request->CancelAuth(); } void URLRequest::Delegate::OnCertificateRequested( URLRequest* request, SSLCertRequestInfo* cert_request_info) { request->Cancel(); } void URLRequest::Delegate::OnSSLCertificateError(URLRequest* request, const SSLInfo& ssl_info, bool is_hsts_ok) { request->Cancel(); } void URLRequest::Delegate::OnBeforeNetworkStart(URLRequest* request, bool* defer) { } /////////////////////////////////////////////////////////////////////////////// // URLRequest URLRequest::~URLRequest() { Cancel(); if (network_delegate_) { network_delegate_->NotifyURLRequestDestroyed(this); if (job_.get()) job_->NotifyURLRequestDestroyed(); } if (job_.get()) OrphanJob(); int deleted = context_->url_requests()->erase(this); CHECK_EQ(1, deleted); int net_error = OK; // Log error only on failure, not cancellation, as even successful requests // are "cancelled" on destruction. if (status_.status() == URLRequestStatus::FAILED) net_error = status_.error(); net_log_.EndEventWithNetErrorCode(NetLog::TYPE_REQUEST_ALIVE, net_error); } void URLRequest::EnableChunkedUpload() { DCHECK(!upload_data_stream_ || upload_data_stream_->is_chunked()); if (!upload_data_stream_) { upload_data_stream_.reset( new UploadDataStream(UploadDataStream::CHUNKED, 0)); } } void URLRequest::AppendChunkToUpload(const char* bytes, int bytes_len, bool is_last_chunk) { DCHECK(upload_data_stream_); DCHECK(upload_data_stream_->is_chunked()); DCHECK_GT(bytes_len, 0); upload_data_stream_->AppendChunk(bytes, bytes_len, is_last_chunk); } void URLRequest::set_upload(scoped_ptr<UploadDataStream> upload) { DCHECK(!upload->is_chunked()); upload_data_stream_ = upload.Pass(); } const UploadDataStream* URLRequest::get_upload() const { return upload_data_stream_.get(); } bool URLRequest::has_upload() const { return upload_data_stream_.get() != NULL; } void URLRequest::SetExtraRequestHeaderById(int id, const string& value, bool overwrite) { DCHECK(!is_pending_ || is_redirecting_); NOTREACHED() << "implement me!"; } void URLRequest::SetExtraRequestHeaderByName(const string& name, const string& value, bool overwrite) { DCHECK(!is_pending_ || is_redirecting_); if (overwrite) { extra_request_headers_.SetHeader(name, value); } else { extra_request_headers_.SetHeaderIfMissing(name, value); } } void URLRequest::RemoveRequestHeaderByName(const string& name) { DCHECK(!is_pending_ || is_redirecting_); extra_request_headers_.RemoveHeader(name); } void URLRequest::SetExtraRequestHeaders( const HttpRequestHeaders& headers) { DCHECK(!is_pending_); extra_request_headers_ = headers; // NOTE: This method will likely become non-trivial once the other setters // for request headers are implemented. } bool URLRequest::GetFullRequestHeaders(HttpRequestHeaders* headers) const { if (!job_.get()) return false; return job_->GetFullRequestHeaders(headers); } int64 URLRequest::GetTotalReceivedBytes() const { if (!job_.get()) return 0; return job_->GetTotalReceivedBytes(); } LoadStateWithParam URLRequest::GetLoadState() const { // The !blocked_by_.empty() check allows |this| to report it's blocked on a // delegate before it has been started. if (calling_delegate_ || !blocked_by_.empty()) { return LoadStateWithParam( LOAD_STATE_WAITING_FOR_DELEGATE, use_blocked_by_as_load_param_ ? base::UTF8ToUTF16(blocked_by_) : base::string16()); } return LoadStateWithParam(job_.get() ? job_->GetLoadState() : LOAD_STATE_IDLE, base::string16()); } base::Value* URLRequest::GetStateAsValue() const { base::DictionaryValue* dict = new base::DictionaryValue(); dict->SetString("url", original_url().possibly_invalid_spec()); if (url_chain_.size() > 1) { base::ListValue* list = new base::ListValue(); for (std::vector<GURL>::const_iterator url = url_chain_.begin(); url != url_chain_.end(); ++url) { list->AppendString(url->possibly_invalid_spec()); } dict->Set("url_chain", list); } dict->SetInteger("load_flags", load_flags_); LoadStateWithParam load_state = GetLoadState(); dict->SetInteger("load_state", load_state.state); if (!load_state.param.empty()) dict->SetString("load_state_param", load_state.param); if (!blocked_by_.empty()) dict->SetString("delegate_info", blocked_by_); dict->SetString("method", method_); dict->SetBoolean("has_upload", has_upload()); dict->SetBoolean("is_pending", is_pending_); // Add the status of the request. The status should always be IO_PENDING, and // the error should always be OK, unless something is holding onto a request // that has finished or a request was leaked. Neither of these should happen. switch (status_.status()) { case URLRequestStatus::SUCCESS: dict->SetString("status", "SUCCESS"); break; case URLRequestStatus::IO_PENDING: dict->SetString("status", "IO_PENDING"); break; case URLRequestStatus::CANCELED: dict->SetString("status", "CANCELED"); break; case URLRequestStatus::FAILED: dict->SetString("status", "FAILED"); break; } if (status_.error() != OK) dict->SetInteger("net_error", status_.error()); return dict; } void URLRequest::LogBlockedBy(const char* blocked_by) { DCHECK(blocked_by); DCHECK_GT(strlen(blocked_by), 0u); // Only log information to NetLog during startup and certain deferring calls // to delegates. For all reads but the first, do nothing. if (!calling_delegate_ && !response_info_.request_time.is_null()) return; LogUnblocked(); blocked_by_ = blocked_by; use_blocked_by_as_load_param_ = false; net_log_.BeginEvent( NetLog::TYPE_DELEGATE_INFO, NetLog::StringCallback("delegate_info", &blocked_by_)); } void URLRequest::LogAndReportBlockedBy(const char* source) { LogBlockedBy(source); use_blocked_by_as_load_param_ = true; } void URLRequest::LogUnblocked() { if (blocked_by_.empty()) return; net_log_.EndEvent(NetLog::TYPE_DELEGATE_INFO); blocked_by_.clear(); } UploadProgress URLRequest::GetUploadProgress() const { if (!job_.get()) { // We haven't started or the request was cancelled return UploadProgress(); } if (final_upload_progress_.position()) { // The first job completed and none of the subsequent series of // GETs when following redirects will upload anything, so we return the // cached results from the initial job, the POST. return final_upload_progress_; } return job_->GetUploadProgress(); } void URLRequest::GetResponseHeaderById(int id, string* value) { DCHECK(job_.get()); NOTREACHED() << "implement me!"; } void URLRequest::GetResponseHeaderByName(const string& name, string* value) { DCHECK(value); if (response_info_.headers.get()) { response_info_.headers->GetNormalizedHeader(name, value); } else { value->clear(); } } void URLRequest::GetAllResponseHeaders(string* headers) { DCHECK(headers); if (response_info_.headers.get()) { response_info_.headers->GetNormalizedHeaders(headers); } else { headers->clear(); } } HostPortPair URLRequest::GetSocketAddress() const { DCHECK(job_.get()); return job_->GetSocketAddress(); } HttpResponseHeaders* URLRequest::response_headers() const { return response_info_.headers.get(); } void URLRequest::GetLoadTimingInfo(LoadTimingInfo* load_timing_info) const { *load_timing_info = load_timing_info_; } bool URLRequest::GetResponseCookies(ResponseCookies* cookies) { DCHECK(job_.get()); return job_->GetResponseCookies(cookies); } void URLRequest::GetMimeType(string* mime_type) const { DCHECK(job_.get()); job_->GetMimeType(mime_type); } void URLRequest::GetCharset(string* charset) const { DCHECK(job_.get()); job_->GetCharset(charset); } int URLRequest::GetResponseCode() const { DCHECK(job_.get()); return job_->GetResponseCode(); } void URLRequest::SetLoadFlags(int flags) { if ((load_flags_ & LOAD_IGNORE_LIMITS) != (flags & LOAD_IGNORE_LIMITS)) { DCHECK(!job_.get()); DCHECK(flags & LOAD_IGNORE_LIMITS); DCHECK_EQ(priority_, MAXIMUM_PRIORITY); } load_flags_ = flags; // This should be a no-op given the above DCHECKs, but do this // anyway for release mode. if ((load_flags_ & LOAD_IGNORE_LIMITS) != 0) SetPriority(MAXIMUM_PRIORITY); } // static void URLRequest::SetDefaultCookiePolicyToBlock() { CHECK(!g_url_requests_started); g_default_can_use_cookies = false; } // static bool URLRequest::IsHandledProtocol(const std::string& scheme) { return URLRequestJobManager::SupportsScheme(scheme); } // static bool URLRequest::IsHandledURL(const GURL& url) { if (!url.is_valid()) { // We handle error cases. return true; } return IsHandledProtocol(url.scheme()); } void URLRequest::set_first_party_for_cookies( const GURL& first_party_for_cookies) { DCHECK(!is_pending_); first_party_for_cookies_ = first_party_for_cookies; } void URLRequest::set_first_party_url_policy( FirstPartyURLPolicy first_party_url_policy) { DCHECK(!is_pending_); first_party_url_policy_ = first_party_url_policy; } void URLRequest::set_method(const std::string& method) { DCHECK(!is_pending_); method_ = method; } // static std::string URLRequest::ComputeMethodForRedirect( const std::string& method, int http_status_code) { // For 303 redirects, all request methods except HEAD are converted to GET, // as per the latest httpbis draft. The draft also allows POST requests to // be converted to GETs when following 301/302 redirects, for historical // reasons. Most major browsers do this and so shall we. Both RFC 2616 and // the httpbis draft say to prompt the user to confirm the generation of new // requests, other than GET and HEAD requests, but IE omits these prompts and // so shall we. // See: https://tools.ietf.org/html/draft-ietf-httpbis-p2-semantics-17#section-7.3 if ((http_status_code == 303 && method != "HEAD") || ((http_status_code == 301 || http_status_code == 302) && method == "POST")) { return "GET"; } return method; } void URLRequest::SetReferrer(const std::string& referrer) { DCHECK(!is_pending_); GURL referrer_url(referrer); if (referrer_url.is_valid()) { referrer_ = referrer_url.GetAsReferrer().spec(); } else { referrer_ = referrer; } } void URLRequest::set_referrer_policy(ReferrerPolicy referrer_policy) { DCHECK(!is_pending_); referrer_policy_ = referrer_policy; } void URLRequest::set_delegate(Delegate* delegate) { delegate_ = delegate; } void URLRequest::Start() { // Some values can be NULL, but the job factory must not be. DCHECK(context_->job_factory()); // Anything that sets |blocked_by_| before start should have cleaned up after // itself. DCHECK(blocked_by_.empty()); g_url_requests_started = true; response_info_.request_time = base::Time::Now(); load_timing_info_ = LoadTimingInfo(); load_timing_info_.request_start_time = response_info_.request_time; load_timing_info_.request_start = base::TimeTicks::Now(); // Only notify the delegate for the initial request. if (network_delegate_) { OnCallToDelegate(); int error = network_delegate_->NotifyBeforeURLRequest( this, before_request_callback_, &delegate_redirect_url_); // If ERR_IO_PENDING is returned, the delegate will invoke // |before_request_callback_| later. if (error != ERR_IO_PENDING) BeforeRequestComplete(error); return; } StartJob(URLRequestJobManager::GetInstance()->CreateJob( this, network_delegate_)); } /////////////////////////////////////////////////////////////////////////////// URLRequest::URLRequest(const GURL& url, RequestPriority priority, Delegate* delegate, const URLRequestContext* context, CookieStore* cookie_store, NetworkDelegate* network_delegate) : context_(context), network_delegate_(network_delegate ? network_delegate : context->network_delegate()), net_log_(BoundNetLog::Make(context->net_log(), NetLog::SOURCE_URL_REQUEST)), url_chain_(1, url), method_("GET"), referrer_policy_(CLEAR_REFERRER_ON_TRANSITION_FROM_SECURE_TO_INSECURE), first_party_url_policy_(NEVER_CHANGE_FIRST_PARTY_URL), load_flags_(LOAD_NORMAL), delegate_(delegate), is_pending_(false), is_redirecting_(false), redirect_limit_(kMaxRedirects), priority_(priority), identifier_(GenerateURLRequestIdentifier()), calling_delegate_(false), use_blocked_by_as_load_param_(false), before_request_callback_(base::Bind(&URLRequest::BeforeRequestComplete, base::Unretained(this))), has_notified_completion_(false), received_response_content_length_(0), creation_time_(base::TimeTicks::Now()), notified_before_network_start_(false), cookie_store_(cookie_store ? cookie_store : context->cookie_store()) { SIMPLE_STATS_COUNTER("URLRequestCount"); // Sanity check out environment. DCHECK(base::MessageLoop::current()) << "The current base::MessageLoop must exist"; context->url_requests()->insert(this); net_log_.BeginEvent(NetLog::TYPE_REQUEST_ALIVE); } // static void URLRequest::RegisterRequestInterceptor(Interceptor* interceptor) { URLRequestJobManager::GetInstance()->RegisterRequestInterceptor(interceptor); } // static void URLRequest::UnregisterRequestInterceptor(Interceptor* interceptor) { URLRequestJobManager::GetInstance()->UnregisterRequestInterceptor( interceptor); } void URLRequest::BeforeRequestComplete(int error) { DCHECK(!job_.get()); DCHECK_NE(ERR_IO_PENDING, error); // Check that there are no callbacks to already canceled requests. DCHECK_NE(URLRequestStatus::CANCELED, status_.status()); OnCallToDelegateComplete(); if (error != OK) { std::string source("delegate"); net_log_.AddEvent(NetLog::TYPE_CANCELLED, NetLog::StringCallback("source", &source)); StartJob(new URLRequestErrorJob(this, network_delegate_, error)); } else if (!delegate_redirect_url_.is_empty()) { GURL new_url; new_url.Swap(&delegate_redirect_url_); URLRequestRedirectJob* job = new URLRequestRedirectJob( this, network_delegate_, new_url, // Use status code 307 to preserve the method, so POST requests work. URLRequestRedirectJob::REDIRECT_307_TEMPORARY_REDIRECT, "Delegate"); StartJob(job); } else { StartJob(URLRequestJobManager::GetInstance()->CreateJob( this, network_delegate_)); } } void URLRequest::StartJob(URLRequestJob* job) { DCHECK(!is_pending_); DCHECK(!job_.get()); net_log_.BeginEvent( NetLog::TYPE_URL_REQUEST_START_JOB, base::Bind(&NetLogURLRequestStartCallback, &url(), &method_, load_flags_, priority_, upload_data_stream_ ? upload_data_stream_->identifier() : -1)); job_ = job; job_->SetExtraRequestHeaders(extra_request_headers_); job_->SetPriority(priority_); if (upload_data_stream_.get()) job_->SetUpload(upload_data_stream_.get()); is_pending_ = true; is_redirecting_ = false; response_info_.was_cached = false; // If the referrer is secure, but the requested URL is not, the referrer // policy should be something non-default. If you hit this, please file a // bug. if (referrer_policy_ == CLEAR_REFERRER_ON_TRANSITION_FROM_SECURE_TO_INSECURE && GURL(referrer_).SchemeIsSecure() && !url().SchemeIsSecure()) { if (!network_delegate_ || !network_delegate_->CancelURLRequestWithPolicyViolatingReferrerHeader( *this, url(), GURL(referrer_))) { referrer_.clear(); } else { // We need to clear the referrer anyway to avoid an infinite recursion // when starting the error job. referrer_.clear(); std::string source("delegate"); net_log_.AddEvent(NetLog::TYPE_CANCELLED, NetLog::StringCallback("source", &source)); RestartWithJob(new URLRequestErrorJob( this, network_delegate_, ERR_BLOCKED_BY_CLIENT)); return; } } // Don't allow errors to be sent from within Start(). // TODO(brettw) this may cause NotifyDone to be sent synchronously, // we probably don't want this: they should be sent asynchronously so // the caller does not get reentered. job_->Start(); } void URLRequest::Restart() { // Should only be called if the original job didn't make any progress. DCHECK(job_.get() && !job_->has_response_started()); RestartWithJob( URLRequestJobManager::GetInstance()->CreateJob(this, network_delegate_)); } void URLRequest::RestartWithJob(URLRequestJob *job) { DCHECK(job->request() == this); PrepareToRestart(); StartJob(job); } void URLRequest::Cancel() { DoCancel(ERR_ABORTED, SSLInfo()); } void URLRequest::CancelWithError(int error) { DoCancel(error, SSLInfo()); } void URLRequest::CancelWithSSLError(int error, const SSLInfo& ssl_info) { // This should only be called on a started request. if (!is_pending_ || !job_.get() || job_->has_response_started()) { NOTREACHED(); return; } DoCancel(error, ssl_info); } void URLRequest::DoCancel(int error, const SSLInfo& ssl_info) { DCHECK(error < 0); // If cancelled while calling a delegate, clear delegate info. if (calling_delegate_) { LogUnblocked(); OnCallToDelegateComplete(); } // If the URL request already has an error status, then canceling is a no-op. // Plus, we don't want to change the error status once it has been set. if (status_.is_success()) { status_.set_status(URLRequestStatus::CANCELED); status_.set_error(error); response_info_.ssl_info = ssl_info; // If the request hasn't already been completed, log a cancellation event. if (!has_notified_completion_) { // Don't log an error code on ERR_ABORTED, since that's redundant. net_log_.AddEventWithNetErrorCode(NetLog::TYPE_CANCELLED, error == ERR_ABORTED ? OK : error); } } if (is_pending_ && job_.get()) job_->Kill(); // We need to notify about the end of this job here synchronously. The // Job sends an asynchronous notification but by the time this is processed, // our |context_| is NULL. NotifyRequestCompleted(); // The Job will call our NotifyDone method asynchronously. This is done so // that the Delegate implementation can call Cancel without having to worry // about being called recursively. } bool URLRequest::Read(IOBuffer* dest, int dest_size, int* bytes_read) { DCHECK(job_.get()); DCHECK(bytes_read); *bytes_read = 0; // If this is the first read, end the delegate call that may have started in // OnResponseStarted. OnCallToDelegateComplete(); // This handles a cancel that happens while paused. // TODO(ahendrickson): DCHECK() that it is not done after // http://crbug.com/115705 is fixed. if (job_->is_done()) return false; if (dest_size == 0) { // Caller is not too bright. I guess we've done what they asked. return true; } // Once the request fails or is cancelled, read will just return 0 bytes // to indicate end of stream. if (!status_.is_success()) { return true; } bool rv = job_->Read(dest, dest_size, bytes_read); // If rv is false, the status cannot be success. DCHECK(rv || status_.status() != URLRequestStatus::SUCCESS); if (rv && *bytes_read <= 0 && status_.is_success()) NotifyRequestCompleted(); return rv; } void URLRequest::StopCaching() { DCHECK(job_.get()); job_->StopCaching(); } void URLRequest::NotifyReceivedRedirect(const RedirectInfo& redirect_info, bool* defer_redirect) { is_redirecting_ = true; // TODO(davidben): Pass the full RedirectInfo down to MaybeInterceptRedirect? URLRequestJob* job = URLRequestJobManager::GetInstance()->MaybeInterceptRedirect( this, network_delegate_, redirect_info.new_url); if (job) { RestartWithJob(job); } else if (delegate_) { OnCallToDelegate(); delegate_->OnReceivedRedirect(this, redirect_info, defer_redirect); // |this| may be have been destroyed here. } } void URLRequest::NotifyBeforeNetworkStart(bool* defer) { if (delegate_ && !notified_before_network_start_) { OnCallToDelegate(); delegate_->OnBeforeNetworkStart(this, defer); if (!*defer) OnCallToDelegateComplete(); notified_before_network_start_ = true; } } void URLRequest::ResumeNetworkStart() { DCHECK(job_.get()); DCHECK(notified_before_network_start_); OnCallToDelegateComplete(); job_->ResumeNetworkStart(); } void URLRequest::NotifyResponseStarted() { int net_error = OK; if (!status_.is_success()) net_error = status_.error(); net_log_.EndEventWithNetErrorCode(NetLog::TYPE_URL_REQUEST_START_JOB, net_error); URLRequestJob* job = URLRequestJobManager::GetInstance()->MaybeInterceptResponse( this, network_delegate_); if (job) { RestartWithJob(job); } else { if (delegate_) { // In some cases (e.g. an event was canceled), we might have sent the // completion event and receive a NotifyResponseStarted() later. if (!has_notified_completion_ && status_.is_success()) { if (network_delegate_) network_delegate_->NotifyResponseStarted(this); } // Notify in case the entire URL Request has been finished. if (!has_notified_completion_ && !status_.is_success()) NotifyRequestCompleted(); OnCallToDelegate(); delegate_->OnResponseStarted(this); // Nothing may appear below this line as OnResponseStarted may delete // |this|. } } } void URLRequest::FollowDeferredRedirect() { CHECK(job_.get()); CHECK(status_.is_success()); job_->FollowDeferredRedirect(); } void URLRequest::SetAuth(const AuthCredentials& credentials) { DCHECK(job_.get()); DCHECK(job_->NeedsAuth()); job_->SetAuth(credentials); } void URLRequest::CancelAuth() { DCHECK(job_.get()); DCHECK(job_->NeedsAuth()); job_->CancelAuth(); } void URLRequest::ContinueWithCertificate(X509Certificate* client_cert) { DCHECK(job_.get()); job_->ContinueWithCertificate(client_cert); } void URLRequest::ContinueDespiteLastError() { DCHECK(job_.get()); job_->ContinueDespiteLastError(); } void URLRequest::PrepareToRestart() { DCHECK(job_.get()); // Close the current URL_REQUEST_START_JOB, since we will be starting a new // one. net_log_.EndEvent(NetLog::TYPE_URL_REQUEST_START_JOB); OrphanJob(); response_info_ = HttpResponseInfo(); response_info_.request_time = base::Time::Now(); load_timing_info_ = LoadTimingInfo(); load_timing_info_.request_start_time = response_info_.request_time; load_timing_info_.request_start = base::TimeTicks::Now(); status_ = URLRequestStatus(); is_pending_ = false; } void URLRequest::OrphanJob() { // When calling this function, please check that URLRequestHttpJob is // not in between calling NetworkDelegate::NotifyHeadersReceived receiving // the call back. This is currently guaranteed by the following strategies: // - OrphanJob is called on JobRestart, in this case the URLRequestJob cannot // be receiving any headers at that time. // - OrphanJob is called in ~URLRequest, in this case // NetworkDelegate::NotifyURLRequestDestroyed notifies the NetworkDelegate // that the callback becomes invalid. job_->Kill(); job_->DetachRequest(); // ensures that the job will not call us again job_ = NULL; } int URLRequest::Redirect(const RedirectInfo& redirect_info) { // Matches call in NotifyReceivedRedirect. OnCallToDelegateComplete(); if (net_log_.IsLogging()) { net_log_.AddEvent( NetLog::TYPE_URL_REQUEST_REDIRECTED, NetLog::StringCallback("location", &redirect_info.new_url.possibly_invalid_spec())); } // TODO(davidben): Pass the full RedirectInfo to the NetworkDelegate. if (network_delegate_) network_delegate_->NotifyBeforeRedirect(this, redirect_info.new_url); if (redirect_limit_ <= 0) { DVLOG(1) << "disallowing redirect: exceeds limit"; return ERR_TOO_MANY_REDIRECTS; } if (!redirect_info.new_url.is_valid()) return ERR_INVALID_URL; if (!job_->IsSafeRedirect(redirect_info.new_url)) { DVLOG(1) << "disallowing redirect: unsafe protocol"; return ERR_UNSAFE_REDIRECT; } if (!final_upload_progress_.position()) final_upload_progress_ = job_->GetUploadProgress(); PrepareToRestart(); if (redirect_info.new_method != method_) { // TODO(davidben): This logic still needs to be replicated at the consumers. if (method_ == "POST") { // If being switched from POST, must remove headers that were specific to // the POST and don't have meaning in other methods. For example the // inclusion of a multipart Content-Type header in GET can cause problems // with some servers: // http://code.google.com/p/chromium/issues/detail?id=843 StripPostSpecificHeaders(&extra_request_headers_); } upload_data_stream_.reset(); method_ = redirect_info.new_method; } referrer_ = redirect_info.new_referrer; first_party_for_cookies_ = redirect_info.new_first_party_for_cookies; url_chain_.push_back(redirect_info.new_url); --redirect_limit_; Start(); return OK; } const URLRequestContext* URLRequest::context() const { return context_; } int64 URLRequest::GetExpectedContentSize() const { int64 expected_content_size = -1; if (job_.get()) expected_content_size = job_->expected_content_size(); return expected_content_size; } void URLRequest::SetPriority(RequestPriority priority) { DCHECK_GE(priority, MINIMUM_PRIORITY); DCHECK_LE(priority, MAXIMUM_PRIORITY); if ((load_flags_ & LOAD_IGNORE_LIMITS) && (priority != MAXIMUM_PRIORITY)) { NOTREACHED(); // Maintain the invariant that requests with IGNORE_LIMITS set // have MAXIMUM_PRIORITY for release mode. return; } if (priority_ == priority) return; priority_ = priority; if (job_.get()) { net_log_.AddEvent(NetLog::TYPE_URL_REQUEST_SET_PRIORITY, NetLog::IntegerCallback("priority", priority_)); job_->SetPriority(priority_); } } bool URLRequest::GetHSTSRedirect(GURL* redirect_url) const { const GURL& url = this->url(); if (!url.SchemeIs("http")) return false; TransportSecurityState* state = context()->transport_security_state(); if (state && state->ShouldUpgradeToSSL(url.host())) { url::Replacements<char> replacements; const char kNewScheme[] = "https"; replacements.SetScheme(kNewScheme, url::Component(0, strlen(kNewScheme))); *redirect_url = url.ReplaceComponents(replacements); return true; } return false; } void URLRequest::NotifyAuthRequired(AuthChallengeInfo* auth_info) { NetworkDelegate::AuthRequiredResponse rv = NetworkDelegate::AUTH_REQUIRED_RESPONSE_NO_ACTION; auth_info_ = auth_info; if (network_delegate_) { OnCallToDelegate(); rv = network_delegate_->NotifyAuthRequired( this, *auth_info, base::Bind(&URLRequest::NotifyAuthRequiredComplete, base::Unretained(this)), &auth_credentials_); if (rv == NetworkDelegate::AUTH_REQUIRED_RESPONSE_IO_PENDING) return; } NotifyAuthRequiredComplete(rv); } void URLRequest::NotifyAuthRequiredComplete( NetworkDelegate::AuthRequiredResponse result) { OnCallToDelegateComplete(); // Check that there are no callbacks to already canceled requests. DCHECK_NE(URLRequestStatus::CANCELED, status_.status()); // NotifyAuthRequired may be called multiple times, such as // when an authentication attempt fails. Clear out the data // so it can be reset on another round. AuthCredentials credentials = auth_credentials_; auth_credentials_ = AuthCredentials(); scoped_refptr<AuthChallengeInfo> auth_info; auth_info.swap(auth_info_); switch (result) { case NetworkDelegate::AUTH_REQUIRED_RESPONSE_NO_ACTION: // Defer to the URLRequest::Delegate, since the NetworkDelegate // didn't take an action. if (delegate_) delegate_->OnAuthRequired(this, auth_info.get()); break; case NetworkDelegate::AUTH_REQUIRED_RESPONSE_SET_AUTH: SetAuth(credentials); break; case NetworkDelegate::AUTH_REQUIRED_RESPONSE_CANCEL_AUTH: CancelAuth(); break; case NetworkDelegate::AUTH_REQUIRED_RESPONSE_IO_PENDING: NOTREACHED(); break; } } void URLRequest::NotifyCertificateRequested( SSLCertRequestInfo* cert_request_info) { if (delegate_) delegate_->OnCertificateRequested(this, cert_request_info); } void URLRequest::NotifySSLCertificateError(const SSLInfo& ssl_info, bool fatal) { if (delegate_) delegate_->OnSSLCertificateError(this, ssl_info, fatal); } bool URLRequest::CanGetCookies(const CookieList& cookie_list) const { DCHECK(!(load_flags_ & LOAD_DO_NOT_SEND_COOKIES)); if (network_delegate_) { return network_delegate_->CanGetCookies(*this, cookie_list); } return g_default_can_use_cookies; } bool URLRequest::CanSetCookie(const std::string& cookie_line, CookieOptions* options) const { DCHECK(!(load_flags_ & LOAD_DO_NOT_SAVE_COOKIES)); if (network_delegate_) { return network_delegate_->CanSetCookie(*this, cookie_line, options); } return g_default_can_use_cookies; } bool URLRequest::CanEnablePrivacyMode() const { if (network_delegate_) { return network_delegate_->CanEnablePrivacyMode(url(), first_party_for_cookies_); } return !g_default_can_use_cookies; } void URLRequest::NotifyReadCompleted(int bytes_read) { // Notify in case the entire URL Request has been finished. if (bytes_read <= 0) NotifyRequestCompleted(); // Notify NetworkChangeNotifier that we just received network data. // This is to identify cases where the NetworkChangeNotifier thinks we // are off-line but we are still receiving network data (crbug.com/124069), // and to get rough network connection measurements. if (bytes_read > 0 && !was_cached()) NetworkChangeNotifier::NotifyDataReceived(*this, bytes_read); if (delegate_) delegate_->OnReadCompleted(this, bytes_read); // Nothing below this line as OnReadCompleted may delete |this|. } void URLRequest::OnHeadersComplete() { // Cache load timing information now, as information will be lost once the // socket is closed and the ClientSocketHandle is Reset, which will happen // once the body is complete. The start times should already be populated. if (job_.get()) { // Keep a copy of the two times the URLRequest sets. base::TimeTicks request_start = load_timing_info_.request_start; base::Time request_start_time = load_timing_info_.request_start_time; // Clear load times. Shouldn't be neded, but gives the GetLoadTimingInfo a // consistent place to start from. load_timing_info_ = LoadTimingInfo(); job_->GetLoadTimingInfo(&load_timing_info_); load_timing_info_.request_start = request_start; load_timing_info_.request_start_time = request_start_time; ConvertRealLoadTimesToBlockingTimes(&load_timing_info_); } } void URLRequest::NotifyRequestCompleted() { // TODO(battre): Get rid of this check, according to willchan it should // not be needed. if (has_notified_completion_) return; is_pending_ = false; is_redirecting_ = false; has_notified_completion_ = true; if (network_delegate_) network_delegate_->NotifyCompleted(this, job_.get() != NULL); } void URLRequest::OnCallToDelegate() { DCHECK(!calling_delegate_); DCHECK(blocked_by_.empty()); calling_delegate_ = true; net_log_.BeginEvent(NetLog::TYPE_URL_REQUEST_DELEGATE); } void URLRequest::OnCallToDelegateComplete() { // This should have been cleared before resuming the request. DCHECK(blocked_by_.empty()); if (!calling_delegate_) return; calling_delegate_ = false; net_log_.EndEvent(NetLog::TYPE_URL_REQUEST_DELEGATE); } void URLRequest::set_stack_trace(const base::debug::StackTrace& stack_trace) { base::debug::StackTrace* stack_trace_copy = new base::debug::StackTrace(NULL, 0); *stack_trace_copy = stack_trace; stack_trace_.reset(stack_trace_copy); } const base::debug::StackTrace* URLRequest::stack_trace() const { return stack_trace_.get(); } } // namespace net