diff options
Diffstat (limited to 'chromium/net/http/http_cache_transaction.cc')
-rw-r--r-- | chromium/net/http/http_cache_transaction.cc | 2612 |
1 files changed, 2612 insertions, 0 deletions
diff --git a/chromium/net/http/http_cache_transaction.cc b/chromium/net/http/http_cache_transaction.cc new file mode 100644 index 00000000000..8d4a45573b8 --- /dev/null +++ b/chromium/net/http/http_cache_transaction.cc @@ -0,0 +1,2612 @@ +// Copyright (c) 2012 The Chromium Authors. All rights reserved. +// Use of this source code is governed by a BSD-style license that can be +// found in the LICENSE file. + +#include "net/http/http_cache_transaction.h" + +#include "build/build_config.h" + +#if defined(OS_POSIX) +#include <unistd.h> +#endif + +#include <algorithm> +#include <string> + +#include "base/bind.h" +#include "base/compiler_specific.h" +#include "base/memory/ref_counted.h" +#include "base/metrics/field_trial.h" +#include "base/metrics/histogram.h" +#include "base/metrics/sparse_histogram.h" +#include "base/rand_util.h" +#include "base/strings/string_number_conversions.h" +#include "base/strings/string_util.h" +#include "base/time/time.h" +#include "net/base/completion_callback.h" +#include "net/base/io_buffer.h" +#include "net/base/load_flags.h" +#include "net/base/load_timing_info.h" +#include "net/base/net_errors.h" +#include "net/base/net_log.h" +#include "net/base/upload_data_stream.h" +#include "net/cert/cert_status_flags.h" +#include "net/disk_cache/disk_cache.h" +#include "net/http/http_network_session.h" +#include "net/http/http_request_info.h" +#include "net/http/http_response_headers.h" +#include "net/http/http_transaction.h" +#include "net/http/http_transaction_delegate.h" +#include "net/http/http_util.h" +#include "net/http/partial_data.h" +#include "net/ssl/ssl_cert_request_info.h" +#include "net/ssl/ssl_config_service.h" + +using base::Time; +using base::TimeDelta; +using base::TimeTicks; + +namespace { + +// From http://tools.ietf.org/html/draft-ietf-httpbis-p6-cache-21#section-6 +// a "non-error response" is one with a 2xx (Successful) or 3xx +// (Redirection) status code. +bool NonErrorResponse(int status_code) { + int status_code_range = status_code / 100; + return status_code_range == 2 || status_code_range == 3; +} + +// Error codes that will be considered indicative of a page being offline/ +// unreachable for LOAD_FROM_CACHE_IF_OFFLINE. +bool IsOfflineError(int error) { + return (error == net::ERR_NAME_NOT_RESOLVED || + error == net::ERR_INTERNET_DISCONNECTED || + error == net::ERR_ADDRESS_UNREACHABLE || + error == net::ERR_CONNECTION_TIMED_OUT); +} + +// Enum for UMA, indicating the status (with regard to offline mode) of +// a particular request. +enum RequestOfflineStatus { + // A cache transaction hit in cache (data was present and not stale) + // and returned it. + OFFLINE_STATUS_FRESH_CACHE, + + // A network request was required for a cache entry, and it succeeded. + OFFLINE_STATUS_NETWORK_SUCCEEDED, + + // A network request was required for a cache entry, and it failed with + // a non-offline error. + OFFLINE_STATUS_NETWORK_FAILED, + + // A network request was required for a cache entry, it failed with an + // offline error, and we could serve stale data if + // LOAD_FROM_CACHE_IF_OFFLINE was set. + OFFLINE_STATUS_DATA_AVAILABLE_OFFLINE, + + // A network request was required for a cache entry, it failed with + // an offline error, and there was no servable data in cache (even + // stale data). + OFFLINE_STATUS_DATA_UNAVAILABLE_OFFLINE, + + OFFLINE_STATUS_MAX_ENTRIES +}; + +void RecordOfflineStatus(int load_flags, RequestOfflineStatus status) { + // Restrict to main frame to keep statistics close to + // "would have shown them something useful if offline mode was enabled". + if (load_flags & net::LOAD_MAIN_FRAME) { + UMA_HISTOGRAM_ENUMERATION("HttpCache.OfflineStatus", status, + OFFLINE_STATUS_MAX_ENTRIES); + } +} + +} // namespace + +namespace net { + +struct HeaderNameAndValue { + const char* name; + const char* value; +}; + +// If the request includes one of these request headers, then avoid caching +// to avoid getting confused. +static const HeaderNameAndValue kPassThroughHeaders[] = { + { "if-unmodified-since", NULL }, // causes unexpected 412s + { "if-match", NULL }, // causes unexpected 412s + { "if-range", NULL }, + { NULL, NULL } +}; + +struct ValidationHeaderInfo { + const char* request_header_name; + const char* related_response_header_name; +}; + +static const ValidationHeaderInfo kValidationHeaders[] = { + { "if-modified-since", "last-modified" }, + { "if-none-match", "etag" }, +}; + +// If the request includes one of these request headers, then avoid reusing +// our cached copy if any. +static const HeaderNameAndValue kForceFetchHeaders[] = { + { "cache-control", "no-cache" }, + { "pragma", "no-cache" }, + { NULL, NULL } +}; + +// If the request includes one of these request headers, then force our +// cached copy (if any) to be revalidated before reusing it. +static const HeaderNameAndValue kForceValidateHeaders[] = { + { "cache-control", "max-age=0" }, + { NULL, NULL } +}; + +static bool HeaderMatches(const HttpRequestHeaders& headers, + const HeaderNameAndValue* search) { + for (; search->name; ++search) { + std::string header_value; + if (!headers.GetHeader(search->name, &header_value)) + continue; + + if (!search->value) + return true; + + HttpUtil::ValuesIterator v(header_value.begin(), header_value.end(), ','); + while (v.GetNext()) { + if (LowerCaseEqualsASCII(v.value_begin(), v.value_end(), search->value)) + return true; + } + } + return false; +} + +//----------------------------------------------------------------------------- + +HttpCache::Transaction::Transaction( + RequestPriority priority, + HttpCache* cache, + HttpTransactionDelegate* transaction_delegate) + : next_state_(STATE_NONE), + request_(NULL), + priority_(priority), + cache_(cache->AsWeakPtr()), + entry_(NULL), + new_entry_(NULL), + new_response_(NULL), + mode_(NONE), + target_state_(STATE_NONE), + reading_(false), + invalid_range_(false), + truncated_(false), + is_sparse_(false), + range_requested_(false), + handling_206_(false), + cache_pending_(false), + done_reading_(false), + vary_mismatch_(false), + couldnt_conditionalize_request_(false), + io_buf_len_(0), + read_offset_(0), + effective_load_flags_(0), + write_len_(0), + weak_factory_(this), + io_callback_(base::Bind(&Transaction::OnIOComplete, + weak_factory_.GetWeakPtr())), + transaction_pattern_(PATTERN_UNDEFINED), + defer_cache_sensitivity_delay_(false), + transaction_delegate_(transaction_delegate) { + COMPILE_ASSERT(HttpCache::Transaction::kNumValidationHeaders == + arraysize(kValidationHeaders), + Invalid_number_of_validation_headers); + base::StringToInt( + base::FieldTrialList::FindFullName("CacheSensitivityAnalysis"), + &sensitivity_analysis_percent_increase_); +} + +HttpCache::Transaction::~Transaction() { + // We may have to issue another IO, but we should never invoke the callback_ + // after this point. + callback_.Reset(); + + transaction_delegate_ = NULL; + cache_io_start_ = base::TimeTicks(); + deferred_cache_sensitivity_delay_ = base::TimeDelta(); + + if (cache_.get()) { + if (entry_) { + bool cancel_request = reading_; + if (cancel_request) { + if (partial_.get()) { + entry_->disk_entry->CancelSparseIO(); + } else { + cancel_request &= (response_.headers->response_code() == 200); + } + } + + cache_->DoneWithEntry(entry_, this, cancel_request); + } else if (cache_pending_) { + cache_->RemovePendingTransaction(this); + } + } + + // Cancel any outstanding callbacks before we drop our reference to the + // HttpCache. This probably isn't strictly necessary, but might as well. + weak_factory_.InvalidateWeakPtrs(); + + // We could still have a cache read or write in progress, so we just null the + // cache_ pointer to signal that we are dead. See DoCacheReadCompleted. + cache_.reset(); +} + +int HttpCache::Transaction::WriteMetadata(IOBuffer* buf, int buf_len, + const CompletionCallback& callback) { + DCHECK(buf); + DCHECK_GT(buf_len, 0); + DCHECK(!callback.is_null()); + if (!cache_.get() || !entry_) + return ERR_UNEXPECTED; + + // We don't need to track this operation for anything. + // It could be possible to check if there is something already written and + // avoid writing again (it should be the same, right?), but let's allow the + // caller to "update" the contents with something new. + return entry_->disk_entry->WriteData(kMetadataIndex, 0, buf, buf_len, + callback, true); +} + +bool HttpCache::Transaction::AddTruncatedFlag() { + DCHECK(mode_ & WRITE || mode_ == NONE); + + // Don't set the flag for sparse entries. + if (partial_.get() && !truncated_) + return true; + + if (!CanResume(true)) + return false; + + // We may have received the whole resource already. + if (done_reading_) + return true; + + truncated_ = true; + target_state_ = STATE_NONE; + next_state_ = STATE_CACHE_WRITE_TRUNCATED_RESPONSE; + DoLoop(OK); + return true; +} + +LoadState HttpCache::Transaction::GetWriterLoadState() const { + if (network_trans_.get()) + return network_trans_->GetLoadState(); + if (entry_ || !request_) + return LOAD_STATE_IDLE; + return LOAD_STATE_WAITING_FOR_CACHE; +} + +const BoundNetLog& HttpCache::Transaction::net_log() const { + return net_log_; +} + +int HttpCache::Transaction::Start(const HttpRequestInfo* request, + const CompletionCallback& callback, + const BoundNetLog& net_log) { + DCHECK(request); + DCHECK(!callback.is_null()); + + // Ensure that we only have one asynchronous call at a time. + DCHECK(callback_.is_null()); + DCHECK(!reading_); + DCHECK(!network_trans_.get()); + DCHECK(!entry_); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + SetRequest(net_log, request); + + // We have to wait until the backend is initialized so we start the SM. + next_state_ = STATE_GET_BACKEND; + int rv = DoLoop(OK); + + // Setting this here allows us to check for the existence of a callback_ to + // determine if we are still inside Start. + if (rv == ERR_IO_PENDING) + callback_ = callback; + + return rv; +} + +int HttpCache::Transaction::RestartIgnoringLastError( + const CompletionCallback& callback) { + DCHECK(!callback.is_null()); + + // Ensure that we only have one asynchronous call at a time. + DCHECK(callback_.is_null()); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + int rv = RestartNetworkRequest(); + + if (rv == ERR_IO_PENDING) + callback_ = callback; + + return rv; +} + +int HttpCache::Transaction::RestartWithCertificate( + X509Certificate* client_cert, + const CompletionCallback& callback) { + DCHECK(!callback.is_null()); + + // Ensure that we only have one asynchronous call at a time. + DCHECK(callback_.is_null()); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + int rv = RestartNetworkRequestWithCertificate(client_cert); + + if (rv == ERR_IO_PENDING) + callback_ = callback; + + return rv; +} + +int HttpCache::Transaction::RestartWithAuth( + const AuthCredentials& credentials, + const CompletionCallback& callback) { + DCHECK(auth_response_.headers.get()); + DCHECK(!callback.is_null()); + + // Ensure that we only have one asynchronous call at a time. + DCHECK(callback_.is_null()); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + // Clear the intermediate response since we are going to start over. + auth_response_ = HttpResponseInfo(); + + int rv = RestartNetworkRequestWithAuth(credentials); + + if (rv == ERR_IO_PENDING) + callback_ = callback; + + return rv; +} + +bool HttpCache::Transaction::IsReadyToRestartForAuth() { + if (!network_trans_.get()) + return false; + return network_trans_->IsReadyToRestartForAuth(); +} + +int HttpCache::Transaction::Read(IOBuffer* buf, int buf_len, + const CompletionCallback& callback) { + DCHECK(buf); + DCHECK_GT(buf_len, 0); + DCHECK(!callback.is_null()); + + DCHECK(callback_.is_null()); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + // If we have an intermediate auth response at this point, then it means the + // user wishes to read the network response (the error page). If there is a + // previous response in the cache then we should leave it intact. + if (auth_response_.headers.get() && mode_ != NONE) { + UpdateTransactionPattern(PATTERN_NOT_COVERED); + DCHECK(mode_ & WRITE); + DoneWritingToEntry(mode_ == READ_WRITE); + mode_ = NONE; + } + + reading_ = true; + int rv; + + switch (mode_) { + case READ_WRITE: + DCHECK(partial_.get()); + if (!network_trans_.get()) { + // We are just reading from the cache, but we may be writing later. + rv = ReadFromEntry(buf, buf_len); + break; + } + case NONE: + case WRITE: + DCHECK(network_trans_.get()); + rv = ReadFromNetwork(buf, buf_len); + break; + case READ: + rv = ReadFromEntry(buf, buf_len); + break; + default: + NOTREACHED(); + rv = ERR_FAILED; + } + + if (rv == ERR_IO_PENDING) { + DCHECK(callback_.is_null()); + callback_ = callback; + } + return rv; +} + +void HttpCache::Transaction::StopCaching() { + // We really don't know where we are now. Hopefully there is no operation in + // progress, but nothing really prevents this method to be called after we + // returned ERR_IO_PENDING. We cannot attempt to truncate the entry at this + // point because we need the state machine for that (and even if we are really + // free, that would be an asynchronous operation). In other words, keep the + // entry how it is (it will be marked as truncated at destruction), and let + // the next piece of code that executes know that we are now reading directly + // from the net. + if (cache_.get() && entry_ && (mode_ & WRITE) && network_trans_.get() && + !is_sparse_ && !range_requested_) { + mode_ = NONE; + } +} + +bool HttpCache::Transaction::GetFullRequestHeaders( + HttpRequestHeaders* headers) const { + if (network_trans_) + return network_trans_->GetFullRequestHeaders(headers); + + // TODO(ttuttle): Read headers from cache. + return false; +} + +void HttpCache::Transaction::DoneReading() { + if (cache_.get() && entry_) { + DCHECK(reading_); + DCHECK_NE(mode_, UPDATE); + if (mode_ & WRITE) + DoneWritingToEntry(true); + } +} + +const HttpResponseInfo* HttpCache::Transaction::GetResponseInfo() const { + // Null headers means we encountered an error or haven't a response yet + if (auth_response_.headers.get()) + return &auth_response_; + return (response_.headers.get() || response_.ssl_info.cert.get() || + response_.cert_request_info.get()) + ? &response_ + : NULL; +} + +LoadState HttpCache::Transaction::GetLoadState() const { + LoadState state = GetWriterLoadState(); + if (state != LOAD_STATE_WAITING_FOR_CACHE) + return state; + + if (cache_.get()) + return cache_->GetLoadStateForPendingTransaction(this); + + return LOAD_STATE_IDLE; +} + +UploadProgress HttpCache::Transaction::GetUploadProgress() const { + if (network_trans_.get()) + return network_trans_->GetUploadProgress(); + return final_upload_progress_; +} + +bool HttpCache::Transaction::GetLoadTimingInfo( + LoadTimingInfo* load_timing_info) const { + if (network_trans_) + return network_trans_->GetLoadTimingInfo(load_timing_info); + + if (old_network_trans_load_timing_) { + *load_timing_info = *old_network_trans_load_timing_; + return true; + } + + if (first_cache_access_since_.is_null()) + return false; + + // If the cache entry was opened, return that time. + load_timing_info->send_start = first_cache_access_since_; + // This time doesn't make much sense when reading from the cache, so just use + // the same time as send_start. + load_timing_info->send_end = first_cache_access_since_; + return true; +} + +void HttpCache::Transaction::SetPriority(RequestPriority priority) { + priority_ = priority; + if (network_trans_) + network_trans_->SetPriority(priority_); +} + +//----------------------------------------------------------------------------- + +void HttpCache::Transaction::DoCallback(int rv) { + DCHECK(rv != ERR_IO_PENDING); + DCHECK(!callback_.is_null()); + + read_buf_ = NULL; // Release the buffer before invoking the callback. + + // Since Run may result in Read being called, clear callback_ up front. + CompletionCallback c = callback_; + callback_.Reset(); + c.Run(rv); +} + +int HttpCache::Transaction::HandleResult(int rv) { + DCHECK(rv != ERR_IO_PENDING); + if (!callback_.is_null()) + DoCallback(rv); + + return rv; +} + +// A few common patterns: (Foo* means Foo -> FooComplete) +// +// Not-cached entry: +// Start(): +// GetBackend* -> InitEntry -> OpenEntry* -> CreateEntry* -> AddToEntry* -> +// SendRequest* -> SuccessfulSendRequest -> OverwriteCachedResponse -> +// CacheWriteResponse* -> TruncateCachedData* -> TruncateCachedMetadata* -> +// PartialHeadersReceived +// +// Read(): +// NetworkRead* -> CacheWriteData* +// +// Cached entry, no validation: +// Start(): +// GetBackend* -> InitEntry -> OpenEntry* -> AddToEntry* -> CacheReadResponse* +// -> BeginPartialCacheValidation() -> BeginCacheValidation() +// +// Read(): +// CacheReadData* +// +// Cached entry, validation (304): +// Start(): +// GetBackend* -> InitEntry -> OpenEntry* -> AddToEntry* -> CacheReadResponse* +// -> BeginPartialCacheValidation() -> BeginCacheValidation() -> +// SendRequest* -> SuccessfulSendRequest -> UpdateCachedResponse -> +// CacheWriteResponse* -> UpdateCachedResponseComplete -> +// OverwriteCachedResponse -> PartialHeadersReceived +// +// Read(): +// CacheReadData* +// +// Cached entry, validation and replace (200): +// Start(): +// GetBackend* -> InitEntry -> OpenEntry* -> AddToEntry* -> CacheReadResponse* +// -> BeginPartialCacheValidation() -> BeginCacheValidation() -> +// SendRequest* -> SuccessfulSendRequest -> OverwriteCachedResponse -> +// CacheWriteResponse* -> DoTruncateCachedData* -> TruncateCachedMetadata* -> +// PartialHeadersReceived +// +// Read(): +// NetworkRead* -> CacheWriteData* +// +// Sparse entry, partially cached, byte range request: +// Start(): +// GetBackend* -> InitEntry -> OpenEntry* -> AddToEntry* -> CacheReadResponse* +// -> BeginPartialCacheValidation() -> CacheQueryData* -> +// ValidateEntryHeadersAndContinue() -> StartPartialCacheValidation -> +// CompletePartialCacheValidation -> BeginCacheValidation() -> SendRequest* -> +// SuccessfulSendRequest -> UpdateCachedResponse -> CacheWriteResponse* -> +// UpdateCachedResponseComplete -> OverwriteCachedResponse -> +// PartialHeadersReceived +// +// Read() 1: +// NetworkRead* -> CacheWriteData* +// +// Read() 2: +// NetworkRead* -> CacheWriteData* -> StartPartialCacheValidation -> +// CompletePartialCacheValidation -> CacheReadData* -> +// +// Read() 3: +// CacheReadData* -> StartPartialCacheValidation -> +// CompletePartialCacheValidation -> BeginCacheValidation() -> SendRequest* -> +// SuccessfulSendRequest -> UpdateCachedResponse* -> OverwriteCachedResponse +// -> PartialHeadersReceived -> NetworkRead* -> CacheWriteData* +// +int HttpCache::Transaction::DoLoop(int result) { + DCHECK(next_state_ != STATE_NONE); + + int rv = result; + do { + State state = next_state_; + next_state_ = STATE_NONE; + switch (state) { + case STATE_GET_BACKEND: + DCHECK_EQ(OK, rv); + rv = DoGetBackend(); + break; + case STATE_GET_BACKEND_COMPLETE: + rv = DoGetBackendComplete(rv); + break; + case STATE_SEND_REQUEST: + DCHECK_EQ(OK, rv); + rv = DoSendRequest(); + break; + case STATE_SEND_REQUEST_COMPLETE: + rv = DoSendRequestComplete(rv); + break; + case STATE_SUCCESSFUL_SEND_REQUEST: + DCHECK_EQ(OK, rv); + rv = DoSuccessfulSendRequest(); + break; + case STATE_NETWORK_READ: + DCHECK_EQ(OK, rv); + rv = DoNetworkRead(); + break; + case STATE_NETWORK_READ_COMPLETE: + rv = DoNetworkReadComplete(rv); + break; + case STATE_INIT_ENTRY: + DCHECK_EQ(OK, rv); + rv = DoInitEntry(); + break; + case STATE_OPEN_ENTRY: + DCHECK_EQ(OK, rv); + rv = DoOpenEntry(); + break; + case STATE_OPEN_ENTRY_COMPLETE: + rv = DoOpenEntryComplete(rv); + break; + case STATE_CREATE_ENTRY: + DCHECK_EQ(OK, rv); + rv = DoCreateEntry(); + break; + case STATE_CREATE_ENTRY_COMPLETE: + rv = DoCreateEntryComplete(rv); + break; + case STATE_DOOM_ENTRY: + DCHECK_EQ(OK, rv); + rv = DoDoomEntry(); + break; + case STATE_DOOM_ENTRY_COMPLETE: + rv = DoDoomEntryComplete(rv); + break; + case STATE_ADD_TO_ENTRY: + DCHECK_EQ(OK, rv); + rv = DoAddToEntry(); + break; + case STATE_ADD_TO_ENTRY_COMPLETE: + rv = DoAddToEntryComplete(rv); + break; + case STATE_ADD_TO_ENTRY_COMPLETE_AFTER_DELAY: + rv = DoAddToEntryCompleteAfterDelay(rv); + break; + case STATE_START_PARTIAL_CACHE_VALIDATION: + DCHECK_EQ(OK, rv); + rv = DoStartPartialCacheValidation(); + break; + case STATE_COMPLETE_PARTIAL_CACHE_VALIDATION: + rv = DoCompletePartialCacheValidation(rv); + break; + case STATE_UPDATE_CACHED_RESPONSE: + DCHECK_EQ(OK, rv); + rv = DoUpdateCachedResponse(); + break; + case STATE_UPDATE_CACHED_RESPONSE_COMPLETE: + rv = DoUpdateCachedResponseComplete(rv); + break; + case STATE_OVERWRITE_CACHED_RESPONSE: + DCHECK_EQ(OK, rv); + rv = DoOverwriteCachedResponse(); + break; + case STATE_TRUNCATE_CACHED_DATA: + DCHECK_EQ(OK, rv); + rv = DoTruncateCachedData(); + break; + case STATE_TRUNCATE_CACHED_DATA_COMPLETE: + rv = DoTruncateCachedDataComplete(rv); + break; + case STATE_TRUNCATE_CACHED_METADATA: + DCHECK_EQ(OK, rv); + rv = DoTruncateCachedMetadata(); + break; + case STATE_TRUNCATE_CACHED_METADATA_COMPLETE: + rv = DoTruncateCachedMetadataComplete(rv); + break; + case STATE_PARTIAL_HEADERS_RECEIVED: + DCHECK_EQ(OK, rv); + rv = DoPartialHeadersReceived(); + break; + case STATE_CACHE_READ_RESPONSE: + DCHECK_EQ(OK, rv); + rv = DoCacheReadResponse(); + break; + case STATE_CACHE_READ_RESPONSE_COMPLETE: + rv = DoCacheReadResponseComplete(rv); + break; + case STATE_CACHE_WRITE_RESPONSE: + DCHECK_EQ(OK, rv); + rv = DoCacheWriteResponse(); + break; + case STATE_CACHE_WRITE_TRUNCATED_RESPONSE: + DCHECK_EQ(OK, rv); + rv = DoCacheWriteTruncatedResponse(); + break; + case STATE_CACHE_WRITE_RESPONSE_COMPLETE: + rv = DoCacheWriteResponseComplete(rv); + break; + case STATE_CACHE_READ_METADATA: + DCHECK_EQ(OK, rv); + rv = DoCacheReadMetadata(); + break; + case STATE_CACHE_READ_METADATA_COMPLETE: + rv = DoCacheReadMetadataComplete(rv); + break; + case STATE_CACHE_QUERY_DATA: + DCHECK_EQ(OK, rv); + rv = DoCacheQueryData(); + break; + case STATE_CACHE_QUERY_DATA_COMPLETE: + rv = DoCacheQueryDataComplete(rv); + break; + case STATE_CACHE_READ_DATA: + DCHECK_EQ(OK, rv); + rv = DoCacheReadData(); + break; + case STATE_CACHE_READ_DATA_COMPLETE: + rv = DoCacheReadDataComplete(rv); + break; + case STATE_CACHE_WRITE_DATA: + rv = DoCacheWriteData(rv); + break; + case STATE_CACHE_WRITE_DATA_COMPLETE: + rv = DoCacheWriteDataComplete(rv); + break; + default: + NOTREACHED() << "bad state"; + rv = ERR_FAILED; + break; + } + } while (rv != ERR_IO_PENDING && next_state_ != STATE_NONE); + + if (rv != ERR_IO_PENDING) + HandleResult(rv); + + return rv; +} + +int HttpCache::Transaction::DoGetBackend() { + cache_pending_ = true; + next_state_ = STATE_GET_BACKEND_COMPLETE; + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_GET_BACKEND); + ReportCacheActionStart(); + return cache_->GetBackendForTransaction(this); +} + +int HttpCache::Transaction::DoGetBackendComplete(int result) { + DCHECK(result == OK || result == ERR_FAILED); + ReportCacheActionFinish(); + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_GET_BACKEND, + result); + cache_pending_ = false; + + if (!ShouldPassThrough()) { + cache_key_ = cache_->GenerateCacheKey(request_); + + // Requested cache access mode. + if (effective_load_flags_ & LOAD_ONLY_FROM_CACHE) { + mode_ = READ; + } else if (effective_load_flags_ & LOAD_BYPASS_CACHE) { + mode_ = WRITE; + } else { + mode_ = READ_WRITE; + } + + // Downgrade to UPDATE if the request has been externally conditionalized. + if (external_validation_.initialized) { + if (mode_ & WRITE) { + // Strip off the READ_DATA bit (and maybe add back a READ_META bit + // in case READ was off). + mode_ = UPDATE; + } else { + mode_ = NONE; + } + } + } + + // Use PUT and DELETE only to invalidate existing stored entries. + if ((request_->method == "PUT" || request_->method == "DELETE") && + mode_ != READ_WRITE && mode_ != WRITE) { + mode_ = NONE; + } + + // If must use cache, then we must fail. This can happen for back/forward + // navigations to a page generated via a form post. + if (!(mode_ & READ) && effective_load_flags_ & LOAD_ONLY_FROM_CACHE) + return ERR_CACHE_MISS; + + if (mode_ == NONE) { + if (partial_.get()) { + partial_->RestoreHeaders(&custom_request_->extra_headers); + partial_.reset(); + } + next_state_ = STATE_SEND_REQUEST; + } else { + next_state_ = STATE_INIT_ENTRY; + } + + // This is only set if we have something to do with the response. + range_requested_ = (partial_.get() != NULL); + + return OK; +} + +int HttpCache::Transaction::DoSendRequest() { + DCHECK(mode_ & WRITE || mode_ == NONE); + DCHECK(!network_trans_.get()); + + send_request_since_ = TimeTicks::Now(); + + // Create a network transaction. + int rv = cache_->network_layer_->CreateTransaction( + priority_, &network_trans_, NULL); + if (rv != OK) + return rv; + + // Old load timing information, if any, is now obsolete. + old_network_trans_load_timing_.reset(); + + ReportNetworkActionStart(); + next_state_ = STATE_SEND_REQUEST_COMPLETE; + rv = network_trans_->Start(request_, io_callback_, net_log_); + return rv; +} + +int HttpCache::Transaction::DoSendRequestComplete(int result) { + ReportNetworkActionFinish(); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + // If requested, and we have a readable cache entry, and we have + // an error indicating that we're offline as opposed to in contact + // with a bad server, read from cache anyway. + if (IsOfflineError(result)) { + if (mode_ == READ_WRITE && entry_ && !partial_) { + RecordOfflineStatus(effective_load_flags_, + OFFLINE_STATUS_DATA_AVAILABLE_OFFLINE); + if (effective_load_flags_ & LOAD_FROM_CACHE_IF_OFFLINE) { + UpdateTransactionPattern(PATTERN_NOT_COVERED); + response_.server_data_unavailable = true; + return SetupEntryForRead(); + } + } else { + RecordOfflineStatus(effective_load_flags_, + OFFLINE_STATUS_DATA_UNAVAILABLE_OFFLINE); + } + } else { + RecordOfflineStatus(effective_load_flags_, + (result == OK ? OFFLINE_STATUS_NETWORK_SUCCEEDED : + OFFLINE_STATUS_NETWORK_FAILED)); + } + + // If we tried to conditionalize the request and failed, we know + // we won't be reading from the cache after this point. + if (couldnt_conditionalize_request_) + mode_ = WRITE; + + if (result == OK) { + next_state_ = STATE_SUCCESSFUL_SEND_REQUEST; + return OK; + } + + // Do not record requests that have network errors or restarts. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + if (IsCertificateError(result)) { + const HttpResponseInfo* response = network_trans_->GetResponseInfo(); + // If we get a certificate error, then there is a certificate in ssl_info, + // so GetResponseInfo() should never return NULL here. + DCHECK(response); + response_.ssl_info = response->ssl_info; + } else if (result == ERR_SSL_CLIENT_AUTH_CERT_NEEDED) { + const HttpResponseInfo* response = network_trans_->GetResponseInfo(); + DCHECK(response); + response_.cert_request_info = response->cert_request_info; + } else if (response_.was_cached) { + DoneWritingToEntry(true); + } + return result; +} + +// We received the response headers and there is no error. +int HttpCache::Transaction::DoSuccessfulSendRequest() { + DCHECK(!new_response_); + const HttpResponseInfo* new_response = network_trans_->GetResponseInfo(); + + if (new_response->headers->response_code() == 401 || + new_response->headers->response_code() == 407) { + auth_response_ = *new_response; + return OK; + } + + new_response_ = new_response; + if (!ValidatePartialResponse() && !auth_response_.headers.get()) { + // Something went wrong with this request and we have to restart it. + // If we have an authentication response, we are exposed to weird things + // hapenning if the user cancels the authentication before we receive + // the new response. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + response_ = HttpResponseInfo(); + ResetNetworkTransaction(); + new_response_ = NULL; + next_state_ = STATE_SEND_REQUEST; + return OK; + } + if (handling_206_ && mode_ == READ_WRITE && !truncated_ && !is_sparse_) { + // We have stored the full entry, but it changed and the server is + // sending a range. We have to delete the old entry. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + DoneWritingToEntry(false); + } + if (new_response_->headers->response_code() == 416 && + (request_->method == "GET" || request_->method == "POST")) { + DCHECK_EQ(NONE, mode_); + response_ = *new_response_; + return OK; + } + + if (mode_ == WRITE && + transaction_pattern_ != PATTERN_ENTRY_CANT_CONDITIONALIZE) { + UpdateTransactionPattern(PATTERN_ENTRY_NOT_CACHED); + } + + if (mode_ == WRITE && + (request_->method == "PUT" || request_->method == "DELETE")) { + if (NonErrorResponse(new_response->headers->response_code())) { + int ret = cache_->DoomEntry(cache_key_, NULL); + DCHECK_EQ(OK, ret); + } + cache_->DoneWritingToEntry(entry_, true); + entry_ = NULL; + mode_ = NONE; + } + + if (mode_ != NONE && request_->method == "POST" && + NonErrorResponse(new_response->headers->response_code())) { + cache_->DoomMainEntryForUrl(request_->url); + } + + // Are we expecting a response to a conditional query? + if (mode_ == READ_WRITE || mode_ == UPDATE) { + if (new_response->headers->response_code() == 304 || handling_206_) { + UpdateTransactionPattern(PATTERN_ENTRY_VALIDATED); + next_state_ = STATE_UPDATE_CACHED_RESPONSE; + return OK; + } + UpdateTransactionPattern(PATTERN_ENTRY_UPDATED); + mode_ = WRITE; + } + + next_state_ = STATE_OVERWRITE_CACHED_RESPONSE; + return OK; +} + +int HttpCache::Transaction::DoNetworkRead() { + ReportNetworkActionStart(); + next_state_ = STATE_NETWORK_READ_COMPLETE; + return network_trans_->Read(read_buf_.get(), io_buf_len_, io_callback_); +} + +int HttpCache::Transaction::DoNetworkReadComplete(int result) { + DCHECK(mode_ & WRITE || mode_ == NONE); + + ReportNetworkActionFinish(); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + // If there is an error or we aren't saving the data, we are done; just wait + // until the destructor runs to see if we can keep the data. + if (mode_ == NONE || result < 0) + return result; + + next_state_ = STATE_CACHE_WRITE_DATA; + return result; +} + +int HttpCache::Transaction::DoInitEntry() { + DCHECK(!new_entry_); + + if (!cache_.get()) + return ERR_UNEXPECTED; + + if (mode_ == WRITE) { + next_state_ = STATE_DOOM_ENTRY; + return OK; + } + + next_state_ = STATE_OPEN_ENTRY; + return OK; +} + +int HttpCache::Transaction::DoOpenEntry() { + DCHECK(!new_entry_); + next_state_ = STATE_OPEN_ENTRY_COMPLETE; + cache_pending_ = true; + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_OPEN_ENTRY); + first_cache_access_since_ = TimeTicks::Now(); + ReportCacheActionStart(); + defer_cache_sensitivity_delay_ = true; + return ResetCacheIOStart(cache_->OpenEntry(cache_key_, &new_entry_, this)); +} + +int HttpCache::Transaction::DoOpenEntryComplete(int result) { + // It is important that we go to STATE_ADD_TO_ENTRY whenever the result is + // OK, otherwise the cache will end up with an active entry without any + // transaction attached. + ReportCacheActionFinish(); + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_OPEN_ENTRY, result); + cache_pending_ = false; + if (result == OK) { + next_state_ = STATE_ADD_TO_ENTRY; + return OK; + } + + if (result == ERR_CACHE_RACE) { + next_state_ = STATE_INIT_ENTRY; + return OK; + } + + if (request_->method == "PUT" || request_->method == "DELETE") { + DCHECK(mode_ == READ_WRITE || mode_ == WRITE); + mode_ = NONE; + next_state_ = STATE_SEND_REQUEST; + return OK; + } + + if (mode_ == READ_WRITE) { + mode_ = WRITE; + next_state_ = STATE_CREATE_ENTRY; + return OK; + } + if (mode_ == UPDATE) { + // There is no cache entry to update; proceed without caching. + mode_ = NONE; + next_state_ = STATE_SEND_REQUEST; + return OK; + } + if (cache_->mode() == PLAYBACK) + DVLOG(1) << "Playback Cache Miss: " << request_->url; + + // The entry does not exist, and we are not permitted to create a new entry, + // so we must fail. + return ERR_CACHE_MISS; +} + +int HttpCache::Transaction::DoCreateEntry() { + DCHECK(!new_entry_); + next_state_ = STATE_CREATE_ENTRY_COMPLETE; + cache_pending_ = true; + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_CREATE_ENTRY); + ReportCacheActionStart(); + defer_cache_sensitivity_delay_ = true; + return ResetCacheIOStart(cache_->CreateEntry(cache_key_, &new_entry_, this)); +} + +int HttpCache::Transaction::DoCreateEntryComplete(int result) { + // It is important that we go to STATE_ADD_TO_ENTRY whenever the result is + // OK, otherwise the cache will end up with an active entry without any + // transaction attached. + ReportCacheActionFinish(); + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_CREATE_ENTRY, + result); + cache_pending_ = false; + next_state_ = STATE_ADD_TO_ENTRY; + + if (result == ERR_CACHE_RACE) { + next_state_ = STATE_INIT_ENTRY; + return OK; + } + + if (result == OK) { + UMA_HISTOGRAM_BOOLEAN("HttpCache.OpenToCreateRace", false); + } else { + UMA_HISTOGRAM_BOOLEAN("HttpCache.OpenToCreateRace", true); + // We have a race here: Maybe we failed to open the entry and decided to + // create one, but by the time we called create, another transaction already + // created the entry. If we want to eliminate this issue, we need an atomic + // OpenOrCreate() method exposed by the disk cache. + DLOG(WARNING) << "Unable to create cache entry"; + mode_ = NONE; + if (partial_.get()) + partial_->RestoreHeaders(&custom_request_->extra_headers); + next_state_ = STATE_SEND_REQUEST; + } + return OK; +} + +int HttpCache::Transaction::DoDoomEntry() { + next_state_ = STATE_DOOM_ENTRY_COMPLETE; + cache_pending_ = true; + if (first_cache_access_since_.is_null()) + first_cache_access_since_ = TimeTicks::Now(); + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_DOOM_ENTRY); + ReportCacheActionStart(); + return ResetCacheIOStart(cache_->DoomEntry(cache_key_, this)); +} + +int HttpCache::Transaction::DoDoomEntryComplete(int result) { + ReportCacheActionFinish(); + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_DOOM_ENTRY, result); + next_state_ = STATE_CREATE_ENTRY; + cache_pending_ = false; + if (result == ERR_CACHE_RACE) + next_state_ = STATE_INIT_ENTRY; + return OK; +} + +int HttpCache::Transaction::DoAddToEntry() { + DCHECK(new_entry_); + cache_pending_ = true; + next_state_ = STATE_ADD_TO_ENTRY_COMPLETE; + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_ADD_TO_ENTRY); + DCHECK(entry_lock_waiting_since_.is_null()); + entry_lock_waiting_since_ = TimeTicks::Now(); + return cache_->AddTransactionToEntry(new_entry_, this); +} + +int HttpCache::Transaction::DoAddToEntryComplete(int result) { + DCHECK(defer_cache_sensitivity_delay_); + defer_cache_sensitivity_delay_ = false; + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_ADD_TO_ENTRY, + result); + const TimeDelta entry_lock_wait = + TimeTicks::Now() - entry_lock_waiting_since_; + UMA_HISTOGRAM_TIMES("HttpCache.EntryLockWait", entry_lock_wait); + + entry_lock_waiting_since_ = TimeTicks(); + DCHECK(new_entry_); + cache_pending_ = false; + + if (result == OK) + entry_ = new_entry_; + + // If there is a failure, the cache should have taken care of new_entry_. + new_entry_ = NULL; + + next_state_ = STATE_ADD_TO_ENTRY_COMPLETE_AFTER_DELAY; + + if (deferred_cache_sensitivity_delay_ == base::TimeDelta()) + return result; + + base::TimeDelta delay = deferred_cache_sensitivity_delay_; + deferred_cache_sensitivity_delay_ = base::TimeDelta(); + ScheduleDelayedLoop(delay, result); + return ERR_IO_PENDING; +} + +int HttpCache::Transaction::DoAddToEntryCompleteAfterDelay(int result) { + if (result == ERR_CACHE_RACE) { + next_state_ = STATE_INIT_ENTRY; + return OK; + } + + if (result != OK) { + NOTREACHED(); + return result; + } + + if (mode_ == WRITE) { + if (partial_.get()) + partial_->RestoreHeaders(&custom_request_->extra_headers); + next_state_ = STATE_SEND_REQUEST; + } else { + // We have to read the headers from the cached entry. + DCHECK(mode_ & READ_META); + next_state_ = STATE_CACHE_READ_RESPONSE; + } + return OK; +} + +// We may end up here multiple times for a given request. +int HttpCache::Transaction::DoStartPartialCacheValidation() { + if (mode_ == NONE) + return OK; + + next_state_ = STATE_COMPLETE_PARTIAL_CACHE_VALIDATION; + return ResetCacheIOStart( + partial_->ShouldValidateCache(entry_->disk_entry, io_callback_)); +} + +int HttpCache::Transaction::DoCompletePartialCacheValidation(int result) { + if (!result) { + // This is the end of the request. + if (mode_ & WRITE) { + DoneWritingToEntry(true); + } else { + cache_->DoneReadingFromEntry(entry_, this); + entry_ = NULL; + } + return result; + } + + if (result < 0) + return result; + + partial_->PrepareCacheValidation(entry_->disk_entry, + &custom_request_->extra_headers); + + if (reading_ && partial_->IsCurrentRangeCached()) { + next_state_ = STATE_CACHE_READ_DATA; + return OK; + } + + return BeginCacheValidation(); +} + +// We received 304 or 206 and we want to update the cached response headers. +int HttpCache::Transaction::DoUpdateCachedResponse() { + next_state_ = STATE_UPDATE_CACHED_RESPONSE_COMPLETE; + int rv = OK; + // Update cached response based on headers in new_response. + // TODO(wtc): should we update cached certificate (response_.ssl_info), too? + response_.headers->Update(*new_response_->headers.get()); + response_.response_time = new_response_->response_time; + response_.request_time = new_response_->request_time; + response_.network_accessed = new_response_->network_accessed; + + if (response_.headers->HasHeaderValue("cache-control", "no-store")) { + if (!entry_->doomed) { + int ret = cache_->DoomEntry(cache_key_, NULL); + DCHECK_EQ(OK, ret); + } + } else { + // If we are already reading, we already updated the headers for this + // request; doing it again will change Content-Length. + if (!reading_) { + target_state_ = STATE_UPDATE_CACHED_RESPONSE_COMPLETE; + next_state_ = STATE_CACHE_WRITE_RESPONSE; + rv = OK; + } + } + return rv; +} + +int HttpCache::Transaction::DoUpdateCachedResponseComplete(int result) { + if (mode_ == UPDATE) { + DCHECK(!handling_206_); + // We got a "not modified" response and already updated the corresponding + // cache entry above. + // + // By closing the cached entry now, we make sure that the 304 rather than + // the cached 200 response, is what will be returned to the user. + DoneWritingToEntry(true); + } else if (entry_ && !handling_206_) { + DCHECK_EQ(READ_WRITE, mode_); + if (!partial_.get() || partial_->IsLastRange()) { + cache_->ConvertWriterToReader(entry_); + mode_ = READ; + } + // We no longer need the network transaction, so destroy it. + final_upload_progress_ = network_trans_->GetUploadProgress(); + ResetNetworkTransaction(); + } else if (entry_ && handling_206_ && truncated_ && + partial_->initial_validation()) { + // We just finished the validation of a truncated entry, and the server + // is willing to resume the operation. Now we go back and start serving + // the first part to the user. + ResetNetworkTransaction(); + new_response_ = NULL; + next_state_ = STATE_START_PARTIAL_CACHE_VALIDATION; + partial_->SetRangeToStartDownload(); + return OK; + } + next_state_ = STATE_OVERWRITE_CACHED_RESPONSE; + return OK; +} + +int HttpCache::Transaction::DoOverwriteCachedResponse() { + if (mode_ & READ) { + next_state_ = STATE_PARTIAL_HEADERS_RECEIVED; + return OK; + } + + // We change the value of Content-Length for partial content. + if (handling_206_ && partial_.get()) + partial_->FixContentLength(new_response_->headers.get()); + + response_ = *new_response_; + + if (handling_206_ && !CanResume(false)) { + // There is no point in storing this resource because it will never be used. + DoneWritingToEntry(false); + if (partial_.get()) + partial_->FixResponseHeaders(response_.headers.get(), true); + next_state_ = STATE_PARTIAL_HEADERS_RECEIVED; + return OK; + } + + target_state_ = STATE_TRUNCATE_CACHED_DATA; + next_state_ = truncated_ ? STATE_CACHE_WRITE_TRUNCATED_RESPONSE : + STATE_CACHE_WRITE_RESPONSE; + return OK; +} + +int HttpCache::Transaction::DoTruncateCachedData() { + next_state_ = STATE_TRUNCATE_CACHED_DATA_COMPLETE; + if (!entry_) + return OK; + if (net_log_.IsLoggingAllEvents()) + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_WRITE_DATA); + ReportCacheActionStart(); + // Truncate the stream. + return ResetCacheIOStart( + WriteToEntry(kResponseContentIndex, 0, NULL, 0, io_callback_)); +} + +int HttpCache::Transaction::DoTruncateCachedDataComplete(int result) { + if (entry_) { + ReportCacheActionFinish(); + if (net_log_.IsLoggingAllEvents()) { + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_WRITE_DATA, + result); + } + } + + next_state_ = STATE_TRUNCATE_CACHED_METADATA; + return OK; +} + +int HttpCache::Transaction::DoTruncateCachedMetadata() { + next_state_ = STATE_TRUNCATE_CACHED_METADATA_COMPLETE; + if (!entry_) + return OK; + + if (net_log_.IsLoggingAllEvents()) + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_WRITE_INFO); + ReportCacheActionStart(); + return ResetCacheIOStart( + WriteToEntry(kMetadataIndex, 0, NULL, 0, io_callback_)); +} + +int HttpCache::Transaction::DoTruncateCachedMetadataComplete(int result) { + if (entry_) { + ReportCacheActionFinish(); + if (net_log_.IsLoggingAllEvents()) { + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_WRITE_INFO, + result); + } + } + + // If this response is a redirect, then we can stop writing now. (We don't + // need to cache the response body of a redirect.) + if (response_.headers->IsRedirect(NULL)) + DoneWritingToEntry(true); + next_state_ = STATE_PARTIAL_HEADERS_RECEIVED; + return OK; +} + +int HttpCache::Transaction::DoPartialHeadersReceived() { + new_response_ = NULL; + if (entry_ && !partial_.get() && + entry_->disk_entry->GetDataSize(kMetadataIndex)) + next_state_ = STATE_CACHE_READ_METADATA; + + if (!partial_.get()) + return OK; + + if (reading_) { + if (network_trans_.get()) { + next_state_ = STATE_NETWORK_READ; + } else { + next_state_ = STATE_CACHE_READ_DATA; + } + } else if (mode_ != NONE) { + // We are about to return the headers for a byte-range request to the user, + // so let's fix them. + partial_->FixResponseHeaders(response_.headers.get(), true); + } + return OK; +} + +int HttpCache::Transaction::DoCacheReadResponse() { + DCHECK(entry_); + next_state_ = STATE_CACHE_READ_RESPONSE_COMPLETE; + + io_buf_len_ = entry_->disk_entry->GetDataSize(kResponseInfoIndex); + read_buf_ = new IOBuffer(io_buf_len_); + + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_READ_INFO); + ReportCacheActionStart(); + return ResetCacheIOStart(entry_->disk_entry->ReadData( + kResponseInfoIndex, 0, read_buf_.get(), io_buf_len_, io_callback_)); +} + +int HttpCache::Transaction::DoCacheReadResponseComplete(int result) { + ReportCacheActionFinish(); + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_READ_INFO, result); + if (result != io_buf_len_ || + !HttpCache::ParseResponseInfo(read_buf_->data(), io_buf_len_, + &response_, &truncated_)) { + return OnCacheReadError(result, true); + } + + // Some resources may have slipped in as truncated when they're not. + int current_size = entry_->disk_entry->GetDataSize(kResponseContentIndex); + if (response_.headers->GetContentLength() == current_size) + truncated_ = false; + + // We now have access to the cache entry. + // + // o if we are a reader for the transaction, then we can start reading the + // cache entry. + // + // o if we can read or write, then we should check if the cache entry needs + // to be validated and then issue a network request if needed or just read + // from the cache if the cache entry is already valid. + // + // o if we are set to UPDATE, then we are handling an externally + // conditionalized request (if-modified-since / if-none-match). We check + // if the request headers define a validation request. + // + switch (mode_) { + case READ: + UpdateTransactionPattern(PATTERN_ENTRY_USED); + result = BeginCacheRead(); + break; + case READ_WRITE: + result = BeginPartialCacheValidation(); + break; + case UPDATE: + result = BeginExternallyConditionalizedRequest(); + break; + case WRITE: + default: + NOTREACHED(); + result = ERR_FAILED; + } + return result; +} + +int HttpCache::Transaction::DoCacheWriteResponse() { + if (entry_) { + if (net_log_.IsLoggingAllEvents()) + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_WRITE_INFO); + ReportCacheActionStart(); + } + return WriteResponseInfoToEntry(false); +} + +int HttpCache::Transaction::DoCacheWriteTruncatedResponse() { + if (entry_) { + if (net_log_.IsLoggingAllEvents()) + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_WRITE_INFO); + ReportCacheActionStart(); + } + return WriteResponseInfoToEntry(true); +} + +int HttpCache::Transaction::DoCacheWriteResponseComplete(int result) { + next_state_ = target_state_; + target_state_ = STATE_NONE; + if (!entry_) + return OK; + ReportCacheActionFinish(); + if (net_log_.IsLoggingAllEvents()) { + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_WRITE_INFO, + result); + } + + // Balance the AddRef from WriteResponseInfoToEntry. + if (result != io_buf_len_) { + DLOG(ERROR) << "failed to write response info to cache"; + DoneWritingToEntry(false); + } + return OK; +} + +int HttpCache::Transaction::DoCacheReadMetadata() { + DCHECK(entry_); + DCHECK(!response_.metadata.get()); + next_state_ = STATE_CACHE_READ_METADATA_COMPLETE; + + response_.metadata = + new IOBufferWithSize(entry_->disk_entry->GetDataSize(kMetadataIndex)); + + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_READ_INFO); + ReportCacheActionStart(); + return ResetCacheIOStart( + entry_->disk_entry->ReadData(kMetadataIndex, + 0, + response_.metadata.get(), + response_.metadata->size(), + io_callback_)); +} + +int HttpCache::Transaction::DoCacheReadMetadataComplete(int result) { + ReportCacheActionFinish(); + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_READ_INFO, result); + if (result != response_.metadata->size()) + return OnCacheReadError(result, false); + return OK; +} + +int HttpCache::Transaction::DoCacheQueryData() { + next_state_ = STATE_CACHE_QUERY_DATA_COMPLETE; + + // Balanced in DoCacheQueryDataComplete. + return ResetCacheIOStart( + entry_->disk_entry->ReadyForSparseIO(io_callback_)); +} + +int HttpCache::Transaction::DoCacheQueryDataComplete(int result) { + DCHECK_EQ(OK, result); + if (!cache_.get()) + return ERR_UNEXPECTED; + + return ValidateEntryHeadersAndContinue(); +} + +int HttpCache::Transaction::DoCacheReadData() { + DCHECK(entry_); + next_state_ = STATE_CACHE_READ_DATA_COMPLETE; + + if (net_log_.IsLoggingAllEvents()) + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_READ_DATA); + ReportCacheActionStart(); + if (partial_.get()) { + return ResetCacheIOStart(partial_->CacheRead( + entry_->disk_entry, read_buf_.get(), io_buf_len_, io_callback_)); + } + + return ResetCacheIOStart(entry_->disk_entry->ReadData(kResponseContentIndex, + read_offset_, + read_buf_.get(), + io_buf_len_, + io_callback_)); +} + +int HttpCache::Transaction::DoCacheReadDataComplete(int result) { + ReportCacheActionFinish(); + if (net_log_.IsLoggingAllEvents()) { + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_READ_DATA, + result); + } + + if (!cache_.get()) + return ERR_UNEXPECTED; + + if (partial_.get()) { + // Partial requests are confusing to report in histograms because they may + // have multiple underlying requests. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + return DoPartialCacheReadCompleted(result); + } + + if (result > 0) { + read_offset_ += result; + } else if (result == 0) { // End of file. + RecordHistograms(); + cache_->DoneReadingFromEntry(entry_, this); + entry_ = NULL; + } else { + return OnCacheReadError(result, false); + } + return result; +} + +int HttpCache::Transaction::DoCacheWriteData(int num_bytes) { + next_state_ = STATE_CACHE_WRITE_DATA_COMPLETE; + write_len_ = num_bytes; + if (entry_) { + if (net_log_.IsLoggingAllEvents()) + net_log_.BeginEvent(NetLog::TYPE_HTTP_CACHE_WRITE_DATA); + ReportCacheActionStart(); + } + + return ResetCacheIOStart( + AppendResponseDataToEntry(read_buf_.get(), num_bytes, io_callback_)); +} + +int HttpCache::Transaction::DoCacheWriteDataComplete(int result) { + if (entry_) { + ReportCacheActionFinish(); + if (net_log_.IsLoggingAllEvents()) { + net_log_.EndEventWithNetErrorCode(NetLog::TYPE_HTTP_CACHE_WRITE_DATA, + result); + } + } + // Balance the AddRef from DoCacheWriteData. + if (!cache_.get()) + return ERR_UNEXPECTED; + + if (result != write_len_) { + DLOG(ERROR) << "failed to write response data to cache"; + DoneWritingToEntry(false); + + // We want to ignore errors writing to disk and just keep reading from + // the network. + result = write_len_; + } else if (!done_reading_ && entry_) { + int current_size = entry_->disk_entry->GetDataSize(kResponseContentIndex); + int64 body_size = response_.headers->GetContentLength(); + if (body_size >= 0 && body_size <= current_size) + done_reading_ = true; + } + + if (partial_.get()) { + // This may be the last request. + if (!(result == 0 && !truncated_ && + (partial_->IsLastRange() || mode_ == WRITE))) + return DoPartialNetworkReadCompleted(result); + } + + if (result == 0) { + // End of file. This may be the result of a connection problem so see if we + // have to keep the entry around to be flagged as truncated later on. + if (done_reading_ || !entry_ || partial_.get() || + response_.headers->GetContentLength() <= 0) + DoneWritingToEntry(true); + } + + return result; +} + +//----------------------------------------------------------------------------- + +void HttpCache::Transaction::SetRequest(const BoundNetLog& net_log, + const HttpRequestInfo* request) { + net_log_ = net_log; + request_ = request; + effective_load_flags_ = request_->load_flags; + + switch (cache_->mode()) { + case NORMAL: + break; + case RECORD: + // When in record mode, we want to NEVER load from the cache. + // The reason for this is beacuse we save the Set-Cookie headers + // (intentionally). If we read from the cache, we replay them + // prematurely. + effective_load_flags_ |= LOAD_BYPASS_CACHE; + break; + case PLAYBACK: + // When in playback mode, we want to load exclusively from the cache. + effective_load_flags_ |= LOAD_ONLY_FROM_CACHE; + break; + case DISABLE: + effective_load_flags_ |= LOAD_DISABLE_CACHE; + break; + } + + // Some headers imply load flags. The order here is significant. + // + // LOAD_DISABLE_CACHE : no cache read or write + // LOAD_BYPASS_CACHE : no cache read + // LOAD_VALIDATE_CACHE : no cache read unless validation + // + // The former modes trump latter modes, so if we find a matching header we + // can stop iterating kSpecialHeaders. + // + static const struct { + const HeaderNameAndValue* search; + int load_flag; + } kSpecialHeaders[] = { + { kPassThroughHeaders, LOAD_DISABLE_CACHE }, + { kForceFetchHeaders, LOAD_BYPASS_CACHE }, + { kForceValidateHeaders, LOAD_VALIDATE_CACHE }, + }; + + bool range_found = false; + bool external_validation_error = false; + + if (request_->extra_headers.HasHeader(HttpRequestHeaders::kRange)) + range_found = true; + + for (size_t i = 0; i < ARRAYSIZE_UNSAFE(kSpecialHeaders); ++i) { + if (HeaderMatches(request_->extra_headers, kSpecialHeaders[i].search)) { + effective_load_flags_ |= kSpecialHeaders[i].load_flag; + break; + } + } + + // Check for conditionalization headers which may correspond with a + // cache validation request. + for (size_t i = 0; i < arraysize(kValidationHeaders); ++i) { + const ValidationHeaderInfo& info = kValidationHeaders[i]; + std::string validation_value; + if (request_->extra_headers.GetHeader( + info.request_header_name, &validation_value)) { + if (!external_validation_.values[i].empty() || + validation_value.empty()) { + external_validation_error = true; + } + external_validation_.values[i] = validation_value; + external_validation_.initialized = true; + } + } + + // We don't support ranges and validation headers. + if (range_found && external_validation_.initialized) { + LOG(WARNING) << "Byte ranges AND validation headers found."; + effective_load_flags_ |= LOAD_DISABLE_CACHE; + } + + // If there is more than one validation header, we can't treat this request as + // a cache validation, since we don't know for sure which header the server + // will give us a response for (and they could be contradictory). + if (external_validation_error) { + LOG(WARNING) << "Multiple or malformed validation headers found."; + effective_load_flags_ |= LOAD_DISABLE_CACHE; + } + + if (range_found && !(effective_load_flags_ & LOAD_DISABLE_CACHE)) { + UpdateTransactionPattern(PATTERN_NOT_COVERED); + partial_.reset(new PartialData); + if (request_->method == "GET" && partial_->Init(request_->extra_headers)) { + // We will be modifying the actual range requested to the server, so + // let's remove the header here. + custom_request_.reset(new HttpRequestInfo(*request_)); + custom_request_->extra_headers.RemoveHeader(HttpRequestHeaders::kRange); + request_ = custom_request_.get(); + partial_->SetHeaders(custom_request_->extra_headers); + } else { + // The range is invalid or we cannot handle it properly. + VLOG(1) << "Invalid byte range found."; + effective_load_flags_ |= LOAD_DISABLE_CACHE; + partial_.reset(NULL); + } + } +} + +bool HttpCache::Transaction::ShouldPassThrough() { + // We may have a null disk_cache if there is an error we cannot recover from, + // like not enough disk space, or sharing violations. + if (!cache_->disk_cache_.get()) + return true; + + // When using the record/playback modes, we always use the cache + // and we never pass through. + if (cache_->mode() == RECORD || cache_->mode() == PLAYBACK) + return false; + + if (effective_load_flags_ & LOAD_DISABLE_CACHE) + return true; + + if (request_->method == "GET") + return false; + + if (request_->method == "POST" && request_->upload_data_stream && + request_->upload_data_stream->identifier()) { + return false; + } + + if (request_->method == "PUT" && request_->upload_data_stream) + return false; + + if (request_->method == "DELETE") + return false; + + // TODO(darin): add support for caching HEAD responses + return true; +} + +int HttpCache::Transaction::BeginCacheRead() { + // We don't support any combination of LOAD_ONLY_FROM_CACHE and byte ranges. + if (response_.headers->response_code() == 206 || partial_.get()) { + NOTREACHED(); + return ERR_CACHE_MISS; + } + + // We don't have the whole resource. + if (truncated_) + return ERR_CACHE_MISS; + + if (entry_->disk_entry->GetDataSize(kMetadataIndex)) + next_state_ = STATE_CACHE_READ_METADATA; + + return OK; +} + +int HttpCache::Transaction::BeginCacheValidation() { + DCHECK(mode_ == READ_WRITE); + + bool skip_validation = !RequiresValidation(); + + if (truncated_) { + // Truncated entries can cause partial gets, so we shouldn't record this + // load in histograms. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + skip_validation = !partial_->initial_validation(); + } + + if (partial_.get() && (is_sparse_ || truncated_) && + (!partial_->IsCurrentRangeCached() || invalid_range_)) { + // Force revalidation for sparse or truncated entries. Note that we don't + // want to ignore the regular validation logic just because a byte range was + // part of the request. + skip_validation = false; + } + + if (skip_validation) { + UpdateTransactionPattern(PATTERN_ENTRY_USED); + RecordOfflineStatus(effective_load_flags_, OFFLINE_STATUS_FRESH_CACHE); + return SetupEntryForRead(); + } else { + // Make the network request conditional, to see if we may reuse our cached + // response. If we cannot do so, then we just resort to a normal fetch. + // Our mode remains READ_WRITE for a conditional request. Even if the + // conditionalization fails, we don't switch to WRITE mode until we + // know we won't be falling back to using the cache entry in the + // LOAD_FROM_CACHE_IF_OFFLINE case. + if (!ConditionalizeRequest()) { + couldnt_conditionalize_request_ = true; + UpdateTransactionPattern(PATTERN_ENTRY_CANT_CONDITIONALIZE); + if (partial_.get()) + return DoRestartPartialRequest(); + + DCHECK_NE(206, response_.headers->response_code()); + } + next_state_ = STATE_SEND_REQUEST; + } + return OK; +} + +int HttpCache::Transaction::BeginPartialCacheValidation() { + DCHECK(mode_ == READ_WRITE); + + if (response_.headers->response_code() != 206 && !partial_.get() && + !truncated_) { + return BeginCacheValidation(); + } + + // Partial requests should not be recorded in histograms. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + if (range_requested_) { + next_state_ = STATE_CACHE_QUERY_DATA; + return OK; + } + // The request is not for a range, but we have stored just ranges. + partial_.reset(new PartialData()); + partial_->SetHeaders(request_->extra_headers); + if (!custom_request_.get()) { + custom_request_.reset(new HttpRequestInfo(*request_)); + request_ = custom_request_.get(); + } + + return ValidateEntryHeadersAndContinue(); +} + +// This should only be called once per request. +int HttpCache::Transaction::ValidateEntryHeadersAndContinue() { + DCHECK(mode_ == READ_WRITE); + + if (!partial_->UpdateFromStoredHeaders( + response_.headers.get(), entry_->disk_entry, truncated_)) { + return DoRestartPartialRequest(); + } + + if (response_.headers->response_code() == 206) + is_sparse_ = true; + + if (!partial_->IsRequestedRangeOK()) { + // The stored data is fine, but the request may be invalid. + invalid_range_ = true; + } + + next_state_ = STATE_START_PARTIAL_CACHE_VALIDATION; + return OK; +} + +int HttpCache::Transaction::BeginExternallyConditionalizedRequest() { + DCHECK_EQ(UPDATE, mode_); + DCHECK(external_validation_.initialized); + + for (size_t i = 0; i < arraysize(kValidationHeaders); i++) { + if (external_validation_.values[i].empty()) + continue; + // Retrieve either the cached response's "etag" or "last-modified" header. + std::string validator; + response_.headers->EnumerateHeader( + NULL, + kValidationHeaders[i].related_response_header_name, + &validator); + + if (response_.headers->response_code() != 200 || truncated_ || + validator.empty() || validator != external_validation_.values[i]) { + // The externally conditionalized request is not a validation request + // for our existing cache entry. Proceed with caching disabled. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + DoneWritingToEntry(true); + } + } + + next_state_ = STATE_SEND_REQUEST; + return OK; +} + +int HttpCache::Transaction::RestartNetworkRequest() { + DCHECK(mode_ & WRITE || mode_ == NONE); + DCHECK(network_trans_.get()); + DCHECK_EQ(STATE_NONE, next_state_); + + ReportNetworkActionStart(); + next_state_ = STATE_SEND_REQUEST_COMPLETE; + int rv = network_trans_->RestartIgnoringLastError(io_callback_); + if (rv != ERR_IO_PENDING) + return DoLoop(rv); + return rv; +} + +int HttpCache::Transaction::RestartNetworkRequestWithCertificate( + X509Certificate* client_cert) { + DCHECK(mode_ & WRITE || mode_ == NONE); + DCHECK(network_trans_.get()); + DCHECK_EQ(STATE_NONE, next_state_); + + ReportNetworkActionStart(); + next_state_ = STATE_SEND_REQUEST_COMPLETE; + int rv = network_trans_->RestartWithCertificate(client_cert, io_callback_); + if (rv != ERR_IO_PENDING) + return DoLoop(rv); + return rv; +} + +int HttpCache::Transaction::RestartNetworkRequestWithAuth( + const AuthCredentials& credentials) { + DCHECK(mode_ & WRITE || mode_ == NONE); + DCHECK(network_trans_.get()); + DCHECK_EQ(STATE_NONE, next_state_); + + ReportNetworkActionStart(); + next_state_ = STATE_SEND_REQUEST_COMPLETE; + int rv = network_trans_->RestartWithAuth(credentials, io_callback_); + if (rv != ERR_IO_PENDING) + return DoLoop(rv); + return rv; +} + +bool HttpCache::Transaction::RequiresValidation() { + // TODO(darin): need to do more work here: + // - make sure we have a matching request method + // - watch out for cached responses that depend on authentication + + // In playback mode, nothing requires validation. + if (cache_->mode() == net::HttpCache::PLAYBACK) + return false; + + if (response_.vary_data.is_valid() && + !response_.vary_data.MatchesRequest(*request_, + *response_.headers.get())) { + vary_mismatch_ = true; + return true; + } + + if (effective_load_flags_ & LOAD_PREFERRING_CACHE) + return false; + + if (effective_load_flags_ & LOAD_VALIDATE_CACHE) + return true; + + if (request_->method == "PUT" || request_->method == "DELETE") + return true; + + if (response_.headers->RequiresValidation( + response_.request_time, response_.response_time, Time::Now())) { + return true; + } + + return false; +} + +bool HttpCache::Transaction::ConditionalizeRequest() { + DCHECK(response_.headers.get()); + + if (request_->method == "PUT" || request_->method == "DELETE") + return false; + + // This only makes sense for cached 200 or 206 responses. + if (response_.headers->response_code() != 200 && + response_.headers->response_code() != 206) { + return false; + } + + // We should have handled this case before. + DCHECK(response_.headers->response_code() != 206 || + response_.headers->HasStrongValidators()); + + // Just use the first available ETag and/or Last-Modified header value. + // TODO(darin): Or should we use the last? + + std::string etag_value; + if (response_.headers->GetHttpVersion() >= HttpVersion(1, 1)) + response_.headers->EnumerateHeader(NULL, "etag", &etag_value); + + std::string last_modified_value; + if (!vary_mismatch_) { + response_.headers->EnumerateHeader(NULL, "last-modified", + &last_modified_value); + } + + if (etag_value.empty() && last_modified_value.empty()) + return false; + + if (!partial_.get()) { + // Need to customize the request, so this forces us to allocate :( + custom_request_.reset(new HttpRequestInfo(*request_)); + request_ = custom_request_.get(); + } + DCHECK(custom_request_.get()); + + bool use_if_range = partial_.get() && !partial_->IsCurrentRangeCached() && + !invalid_range_; + + if (!etag_value.empty()) { + if (use_if_range) { + // We don't want to switch to WRITE mode if we don't have this block of a + // byte-range request because we may have other parts cached. + custom_request_->extra_headers.SetHeader( + HttpRequestHeaders::kIfRange, etag_value); + } else { + custom_request_->extra_headers.SetHeader( + HttpRequestHeaders::kIfNoneMatch, etag_value); + } + // For byte-range requests, make sure that we use only one way to validate + // the request. + if (partial_.get() && !partial_->IsCurrentRangeCached()) + return true; + } + + if (!last_modified_value.empty()) { + if (use_if_range) { + custom_request_->extra_headers.SetHeader( + HttpRequestHeaders::kIfRange, last_modified_value); + } else { + custom_request_->extra_headers.SetHeader( + HttpRequestHeaders::kIfModifiedSince, last_modified_value); + } + } + + return true; +} + +// We just received some headers from the server. We may have asked for a range, +// in which case partial_ has an object. This could be the first network request +// we make to fulfill the original request, or we may be already reading (from +// the net and / or the cache). If we are not expecting a certain response, we +// just bypass the cache for this request (but again, maybe we are reading), and +// delete partial_ (so we are not able to "fix" the headers that we return to +// the user). This results in either a weird response for the caller (we don't +// expect it after all), or maybe a range that was not exactly what it was asked +// for. +// +// If the server is simply telling us that the resource has changed, we delete +// the cached entry and restart the request as the caller intended (by returning +// false from this method). However, we may not be able to do that at any point, +// for instance if we already returned the headers to the user. +// +// WARNING: Whenever this code returns false, it has to make sure that the next +// time it is called it will return true so that we don't keep retrying the +// request. +bool HttpCache::Transaction::ValidatePartialResponse() { + const HttpResponseHeaders* headers = new_response_->headers.get(); + int response_code = headers->response_code(); + bool partial_response = (response_code == 206); + handling_206_ = false; + + if (!entry_ || request_->method != "GET") + return true; + + if (invalid_range_) { + // We gave up trying to match this request with the stored data. If the + // server is ok with the request, delete the entry, otherwise just ignore + // this request + DCHECK(!reading_); + if (partial_response || response_code == 200) { + DoomPartialEntry(true); + mode_ = NONE; + } else { + if (response_code == 304) + FailRangeRequest(); + IgnoreRangeRequest(); + } + return true; + } + + if (!partial_.get()) { + // We are not expecting 206 but we may have one. + if (partial_response) + IgnoreRangeRequest(); + + return true; + } + + // TODO(rvargas): Do we need to consider other results here?. + bool failure = response_code == 200 || response_code == 416; + + if (partial_->IsCurrentRangeCached()) { + // We asked for "If-None-Match: " so a 206 means a new object. + if (partial_response) + failure = true; + + if (response_code == 304 && partial_->ResponseHeadersOK(headers)) + return true; + } else { + // We asked for "If-Range: " so a 206 means just another range. + if (partial_response && partial_->ResponseHeadersOK(headers)) { + handling_206_ = true; + return true; + } + + if (!reading_ && !is_sparse_ && !partial_response) { + // See if we can ignore the fact that we issued a byte range request. + // If the server sends 200, just store it. If it sends an error, redirect + // or something else, we may store the response as long as we didn't have + // anything already stored. + if (response_code == 200 || + (!truncated_ && response_code != 304 && response_code != 416)) { + // The server is sending something else, and we can save it. + DCHECK((truncated_ && !partial_->IsLastRange()) || range_requested_); + partial_.reset(); + truncated_ = false; + return true; + } + } + + // 304 is not expected here, but we'll spare the entry (unless it was + // truncated). + if (truncated_) + failure = true; + } + + if (failure) { + // We cannot truncate this entry, it has to be deleted. + UpdateTransactionPattern(PATTERN_NOT_COVERED); + DoomPartialEntry(false); + mode_ = NONE; + if (!reading_ && !partial_->IsLastRange()) { + // We'll attempt to issue another network request, this time without us + // messing up the headers. + partial_->RestoreHeaders(&custom_request_->extra_headers); + partial_.reset(); + truncated_ = false; + return false; + } + LOG(WARNING) << "Failed to revalidate partial entry"; + partial_.reset(); + return true; + } + + IgnoreRangeRequest(); + return true; +} + +void HttpCache::Transaction::IgnoreRangeRequest() { + // We have a problem. We may or may not be reading already (in which case we + // returned the headers), but we'll just pretend that this request is not + // using the cache and see what happens. Most likely this is the first + // response from the server (it's not changing its mind midway, right?). + UpdateTransactionPattern(PATTERN_NOT_COVERED); + if (mode_ & WRITE) + DoneWritingToEntry(mode_ != WRITE); + else if (mode_ & READ && entry_) + cache_->DoneReadingFromEntry(entry_, this); + + partial_.reset(NULL); + entry_ = NULL; + mode_ = NONE; +} + +void HttpCache::Transaction::FailRangeRequest() { + response_ = *new_response_; + partial_->FixResponseHeaders(response_.headers.get(), false); +} + +int HttpCache::Transaction::SetupEntryForRead() { + if (network_trans_) + ResetNetworkTransaction(); + if (partial_.get()) { + if (truncated_ || is_sparse_ || !invalid_range_) { + // We are going to return the saved response headers to the caller, so + // we may need to adjust them first. + next_state_ = STATE_PARTIAL_HEADERS_RECEIVED; + return OK; + } else { + partial_.reset(); + } + } + cache_->ConvertWriterToReader(entry_); + mode_ = READ; + + if (entry_->disk_entry->GetDataSize(kMetadataIndex)) + next_state_ = STATE_CACHE_READ_METADATA; + return OK; +} + + +int HttpCache::Transaction::ReadFromNetwork(IOBuffer* data, int data_len) { + read_buf_ = data; + io_buf_len_ = data_len; + next_state_ = STATE_NETWORK_READ; + return DoLoop(OK); +} + +int HttpCache::Transaction::ReadFromEntry(IOBuffer* data, int data_len) { + read_buf_ = data; + io_buf_len_ = data_len; + next_state_ = STATE_CACHE_READ_DATA; + return DoLoop(OK); +} + +int HttpCache::Transaction::WriteToEntry(int index, int offset, + IOBuffer* data, int data_len, + const CompletionCallback& callback) { + if (!entry_) + return data_len; + + int rv = 0; + if (!partial_.get() || !data_len) { + rv = entry_->disk_entry->WriteData(index, offset, data, data_len, callback, + true); + } else { + rv = partial_->CacheWrite(entry_->disk_entry, data, data_len, callback); + } + return rv; +} + +int HttpCache::Transaction::WriteResponseInfoToEntry(bool truncated) { + next_state_ = STATE_CACHE_WRITE_RESPONSE_COMPLETE; + if (!entry_) + return OK; + + // Do not cache no-store content (unless we are record mode). Do not cache + // content with cert errors either. This is to prevent not reporting net + // errors when loading a resource from the cache. When we load a page over + // HTTPS with a cert error we show an SSL blocking page. If the user clicks + // proceed we reload the resource ignoring the errors. The loaded resource + // is then cached. If that resource is subsequently loaded from the cache, + // no net error is reported (even though the cert status contains the actual + // errors) and no SSL blocking page is shown. An alternative would be to + // reverse-map the cert status to a net error and replay the net error. + if ((cache_->mode() != RECORD && + response_.headers->HasHeaderValue("cache-control", "no-store")) || + net::IsCertStatusError(response_.ssl_info.cert_status)) { + DoneWritingToEntry(false); + ReportCacheActionFinish(); + if (net_log_.IsLoggingAllEvents()) + net_log_.EndEvent(NetLog::TYPE_HTTP_CACHE_WRITE_INFO); + return OK; + } + + // When writing headers, we normally only write the non-transient + // headers; when in record mode, record everything. + bool skip_transient_headers = (cache_->mode() != RECORD); + + if (truncated) + DCHECK_EQ(200, response_.headers->response_code()); + + scoped_refptr<PickledIOBuffer> data(new PickledIOBuffer()); + response_.Persist(data->pickle(), skip_transient_headers, truncated); + data->Done(); + + io_buf_len_ = data->pickle()->size(); + return ResetCacheIOStart(entry_->disk_entry->WriteData( + kResponseInfoIndex, 0, data.get(), io_buf_len_, io_callback_, true)); +} + +int HttpCache::Transaction::AppendResponseDataToEntry( + IOBuffer* data, int data_len, const CompletionCallback& callback) { + if (!entry_ || !data_len) + return data_len; + + int current_size = entry_->disk_entry->GetDataSize(kResponseContentIndex); + return WriteToEntry(kResponseContentIndex, current_size, data, data_len, + callback); +} + +void HttpCache::Transaction::DoneWritingToEntry(bool success) { + if (!entry_) + return; + + RecordHistograms(); + + cache_->DoneWritingToEntry(entry_, success); + entry_ = NULL; + mode_ = NONE; // switch to 'pass through' mode +} + +int HttpCache::Transaction::OnCacheReadError(int result, bool restart) { + DLOG(ERROR) << "ReadData failed: " << result; + const int result_for_histogram = std::max(0, -result); + if (restart) { + UMA_HISTOGRAM_SPARSE_SLOWLY("HttpCache.ReadErrorRestartable", + result_for_histogram); + } else { + UMA_HISTOGRAM_SPARSE_SLOWLY("HttpCache.ReadErrorNonRestartable", + result_for_histogram); + } + + // Avoid using this entry in the future. + if (cache_.get()) + cache_->DoomActiveEntry(cache_key_); + + if (restart) { + DCHECK(!reading_); + DCHECK(!network_trans_.get()); + cache_->DoneWithEntry(entry_, this, false); + entry_ = NULL; + is_sparse_ = false; + partial_.reset(); + next_state_ = STATE_GET_BACKEND; + return OK; + } + + return ERR_CACHE_READ_FAILURE; +} + +void HttpCache::Transaction::DoomPartialEntry(bool delete_object) { + DVLOG(2) << "DoomPartialEntry"; + int rv = cache_->DoomEntry(cache_key_, NULL); + DCHECK_EQ(OK, rv); + cache_->DoneWithEntry(entry_, this, false); + entry_ = NULL; + is_sparse_ = false; + if (delete_object) + partial_.reset(NULL); +} + +int HttpCache::Transaction::DoPartialNetworkReadCompleted(int result) { + partial_->OnNetworkReadCompleted(result); + + if (result == 0) { + // We need to move on to the next range. + ResetNetworkTransaction(); + next_state_ = STATE_START_PARTIAL_CACHE_VALIDATION; + } + return result; +} + +int HttpCache::Transaction::DoPartialCacheReadCompleted(int result) { + partial_->OnCacheReadCompleted(result); + + if (result == 0 && mode_ == READ_WRITE) { + // We need to move on to the next range. + next_state_ = STATE_START_PARTIAL_CACHE_VALIDATION; + } else if (result < 0) { + return OnCacheReadError(result, false); + } + return result; +} + +int HttpCache::Transaction::DoRestartPartialRequest() { + // The stored data cannot be used. Get rid of it and restart this request. + // We need to also reset the |truncated_| flag as a new entry is created. + DoomPartialEntry(!range_requested_); + mode_ = WRITE; + truncated_ = false; + next_state_ = STATE_INIT_ENTRY; + return OK; +} + +// Histogram data from the end of 2010 show the following distribution of +// response headers: +// +// Content-Length............... 87% +// Date......................... 98% +// Last-Modified................ 49% +// Etag......................... 19% +// Accept-Ranges: bytes......... 25% +// Accept-Ranges: none.......... 0.4% +// Strong Validator............. 50% +// Strong Validator + ranges.... 24% +// Strong Validator + CL........ 49% +// +bool HttpCache::Transaction::CanResume(bool has_data) { + // Double check that there is something worth keeping. + if (has_data && !entry_->disk_entry->GetDataSize(kResponseContentIndex)) + return false; + + if (request_->method != "GET") + return false; + + if (response_.headers->GetContentLength() <= 0 || + response_.headers->HasHeaderValue("Accept-Ranges", "none") || + !response_.headers->HasStrongValidators()) { + return false; + } + + return true; +} + +void HttpCache::Transaction::OnIOComplete(int result) { + if (!cache_io_start_.is_null()) { + base::TimeDelta cache_time = base::TimeTicks::Now() - cache_io_start_; + cache_io_start_ = base::TimeTicks(); + if (sensitivity_analysis_percent_increase_ > 0) { + cache_time *= sensitivity_analysis_percent_increase_; + cache_time /= 100; + if (!defer_cache_sensitivity_delay_) { + ScheduleDelayedLoop(cache_time, result); + return; + } else { + deferred_cache_sensitivity_delay_ += cache_time; + } + } + } + DCHECK(cache_io_start_.is_null()); + DoLoop(result); +} + +void HttpCache::Transaction::ScheduleDelayedLoop(base::TimeDelta delay, + int result) { + base::MessageLoop::current()->PostDelayedTask( + FROM_HERE, + base::Bind(&HttpCache::Transaction::RunDelayedLoop, + weak_factory_.GetWeakPtr(), + base::TimeTicks::Now(), + delay, + result), + delay); +} + +void HttpCache::Transaction::RunDelayedLoop(base::TimeTicks delay_start_time, + base::TimeDelta intended_delay, + int result) { + base::TimeDelta actual_delay = base::TimeTicks::Now() - delay_start_time; + int64 ratio; + int64 inverse_ratio; + if (intended_delay.InMicroseconds() > 0) { + ratio = + 100 * actual_delay.InMicroseconds() / intended_delay.InMicroseconds(); + } else { + ratio = 0; + } + if (actual_delay.InMicroseconds() > 0) { + inverse_ratio = + 100 * intended_delay.InMicroseconds() / actual_delay.InMicroseconds(); + } else { + inverse_ratio = 0; + } + bool ratio_sample = base::RandInt(0, 99) < ratio; + bool inverse_ratio_sample = base::RandInt(0, 99) < inverse_ratio; + int intended_delay_ms = intended_delay.InMilliseconds(); + UMA_HISTOGRAM_COUNTS_10000( + "HttpCache.CacheSensitivityAnalysis_IntendedDelayMs", + intended_delay_ms); + if (ratio_sample) { + UMA_HISTOGRAM_COUNTS_10000( + "HttpCache.CacheSensitivityAnalysis_RatioByIntendedDelayMs", + intended_delay_ms); + } + if (inverse_ratio_sample) { + UMA_HISTOGRAM_COUNTS_10000( + "HttpCache.CacheSensitivityAnalysis_InverseRatioByIntendedDelayMs", + intended_delay_ms); + } + + DCHECK(cache_io_start_.is_null()); + DCHECK(deferred_cache_sensitivity_delay_ == base::TimeDelta()); + DoLoop(result); +} + +void HttpCache::Transaction::ReportCacheActionStart() { + if (transaction_delegate_) + transaction_delegate_->OnCacheActionStart(); +} + +void HttpCache::Transaction::ReportCacheActionFinish() { + if (transaction_delegate_) + transaction_delegate_->OnCacheActionFinish(); +} + +void HttpCache::Transaction::ReportNetworkActionStart() { + if (transaction_delegate_) + transaction_delegate_->OnNetworkActionStart(); +} + +void HttpCache::Transaction::ReportNetworkActionFinish() { + if (transaction_delegate_) + transaction_delegate_->OnNetworkActionFinish(); +} + +void HttpCache::Transaction::UpdateTransactionPattern( + TransactionPattern new_transaction_pattern) { + if (transaction_pattern_ == PATTERN_NOT_COVERED) + return; + DCHECK(transaction_pattern_ == PATTERN_UNDEFINED || + new_transaction_pattern == PATTERN_NOT_COVERED); + transaction_pattern_ = new_transaction_pattern; +} + +void HttpCache::Transaction::RecordHistograms() { + DCHECK_NE(PATTERN_UNDEFINED, transaction_pattern_); + if (!cache_.get() || !cache_->GetCurrentBackend() || + cache_->GetCurrentBackend()->GetCacheType() != DISK_CACHE || + cache_->mode() != NORMAL || request_->method != "GET") { + return; + } + UMA_HISTOGRAM_ENUMERATION( + "HttpCache.Pattern", transaction_pattern_, PATTERN_MAX); + if (transaction_pattern_ == PATTERN_NOT_COVERED) + return; + DCHECK(!range_requested_); + DCHECK(!first_cache_access_since_.is_null()); + + TimeDelta total_time = base::TimeTicks::Now() - first_cache_access_since_; + + UMA_HISTOGRAM_TIMES("HttpCache.AccessToDone", total_time); + + bool did_send_request = !send_request_since_.is_null(); + DCHECK( + (did_send_request && + (transaction_pattern_ == PATTERN_ENTRY_NOT_CACHED || + transaction_pattern_ == PATTERN_ENTRY_VALIDATED || + transaction_pattern_ == PATTERN_ENTRY_UPDATED || + transaction_pattern_ == PATTERN_ENTRY_CANT_CONDITIONALIZE)) || + (!did_send_request && transaction_pattern_ == PATTERN_ENTRY_USED)); + + if (!did_send_request) { + DCHECK(transaction_pattern_ == PATTERN_ENTRY_USED); + UMA_HISTOGRAM_TIMES("HttpCache.AccessToDone.Used", total_time); + return; + } + + TimeDelta before_send_time = send_request_since_ - first_cache_access_since_; + int before_send_percent = + total_time.ToInternalValue() == 0 ? 0 + : before_send_time * 100 / total_time; + DCHECK_LE(0, before_send_percent); + DCHECK_GE(100, before_send_percent); + + UMA_HISTOGRAM_TIMES("HttpCache.AccessToDone.SentRequest", total_time); + UMA_HISTOGRAM_TIMES("HttpCache.BeforeSend", before_send_time); + UMA_HISTOGRAM_PERCENTAGE("HttpCache.PercentBeforeSend", before_send_percent); + + // TODO(gavinp): Remove or minimize these histograms, particularly the ones + // below this comment after we have received initial data. + switch (transaction_pattern_) { + case PATTERN_ENTRY_CANT_CONDITIONALIZE: { + UMA_HISTOGRAM_TIMES("HttpCache.BeforeSend.CantConditionalize", + before_send_time); + UMA_HISTOGRAM_PERCENTAGE("HttpCache.PercentBeforeSend.CantConditionalize", + before_send_percent); + break; + } + case PATTERN_ENTRY_NOT_CACHED: { + UMA_HISTOGRAM_TIMES("HttpCache.BeforeSend.NotCached", before_send_time); + UMA_HISTOGRAM_PERCENTAGE("HttpCache.PercentBeforeSend.NotCached", + before_send_percent); + break; + } + case PATTERN_ENTRY_VALIDATED: { + UMA_HISTOGRAM_TIMES("HttpCache.BeforeSend.Validated", before_send_time); + UMA_HISTOGRAM_PERCENTAGE("HttpCache.PercentBeforeSend.Validated", + before_send_percent); + break; + } + case PATTERN_ENTRY_UPDATED: { + UMA_HISTOGRAM_TIMES("HttpCache.BeforeSend.Updated", before_send_time); + UMA_HISTOGRAM_PERCENTAGE("HttpCache.PercentBeforeSend.Updated", + before_send_percent); + break; + } + default: + NOTREACHED(); + } +} + +int HttpCache::Transaction::ResetCacheIOStart(int return_value) { + DCHECK(cache_io_start_.is_null()); + if (return_value == ERR_IO_PENDING) + cache_io_start_ = base::TimeTicks::Now(); + return return_value; +} + +void HttpCache::Transaction::ResetNetworkTransaction() { + DCHECK(!old_network_trans_load_timing_); + DCHECK(network_trans_); + LoadTimingInfo load_timing; + if (network_trans_->GetLoadTimingInfo(&load_timing)) + old_network_trans_load_timing_.reset(new LoadTimingInfo(load_timing)); + network_trans_.reset(); +} + +} // namespace net |