diff options
Diffstat (limited to 'chromium/net/http/http_stream_parser_unittest.cc')
-rw-r--r-- | chromium/net/http/http_stream_parser_unittest.cc | 416 |
1 files changed, 416 insertions, 0 deletions
diff --git a/chromium/net/http/http_stream_parser_unittest.cc b/chromium/net/http/http_stream_parser_unittest.cc new file mode 100644 index 00000000000..84775945ec1 --- /dev/null +++ b/chromium/net/http/http_stream_parser_unittest.cc @@ -0,0 +1,416 @@ +// Copyright (c) 2012 The Chromium Authors. All rights reserved. +// Use of this source code is governed by a BSD-style license that can be +// found in the LICENSE file. + +#include "net/http/http_stream_parser.h" + +#include "base/file_util.h" +#include "base/files/file_path.h" +#include "base/files/scoped_temp_dir.h" +#include "base/memory/ref_counted.h" +#include "base/strings/string_piece.h" +#include "base/strings/stringprintf.h" +#include "net/base/io_buffer.h" +#include "net/base/net_errors.h" +#include "net/base/test_completion_callback.h" +#include "net/base/upload_bytes_element_reader.h" +#include "net/base/upload_data_stream.h" +#include "net/base/upload_file_element_reader.h" +#include "net/http/http_request_headers.h" +#include "net/http/http_request_info.h" +#include "net/http/http_response_info.h" +#include "net/socket/client_socket_handle.h" +#include "net/socket/socket_test_util.h" +#include "testing/gtest/include/gtest/gtest.h" +#include "url/gurl.h" + +namespace net { + +const size_t kOutputSize = 1024; // Just large enough for this test. +// The number of bytes that can fit in a buffer of kOutputSize. +const size_t kMaxPayloadSize = + kOutputSize - HttpStreamParser::kChunkHeaderFooterSize; + +// The empty payload is how the last chunk is encoded. +TEST(HttpStreamParser, EncodeChunk_EmptyPayload) { + char output[kOutputSize]; + + const base::StringPiece kPayload = ""; + const base::StringPiece kExpected = "0\r\n\r\n"; + const int num_bytes_written = + HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output)); + ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written)); + EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written)); +} + +TEST(HttpStreamParser, EncodeChunk_ShortPayload) { + char output[kOutputSize]; + + const std::string kPayload("foo\x00\x11\x22", 6); + // 11 = payload size + sizeof("6") + CRLF x 2. + const std::string kExpected("6\r\nfoo\x00\x11\x22\r\n", 11); + const int num_bytes_written = + HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output)); + ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written)); + EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written)); +} + +TEST(HttpStreamParser, EncodeChunk_LargePayload) { + char output[kOutputSize]; + + const std::string kPayload(1000, '\xff'); // '\xff' x 1000. + // 3E8 = 1000 in hex. + const std::string kExpected = "3E8\r\n" + kPayload + "\r\n"; + const int num_bytes_written = + HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output)); + ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written)); + EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written)); +} + +TEST(HttpStreamParser, EncodeChunk_FullPayload) { + char output[kOutputSize]; + + const std::string kPayload(kMaxPayloadSize, '\xff'); + // 3F4 = 1012 in hex. + const std::string kExpected = "3F4\r\n" + kPayload + "\r\n"; + const int num_bytes_written = + HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output)); + ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written)); + EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written)); +} + +TEST(HttpStreamParser, EncodeChunk_TooLargePayload) { + char output[kOutputSize]; + + // The payload is one byte larger the output buffer size. + const std::string kPayload(kMaxPayloadSize + 1, '\xff'); + const int num_bytes_written = + HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output)); + ASSERT_EQ(ERR_INVALID_ARGUMENT, num_bytes_written); +} + +TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_NoBody) { + // Shouldn't be merged if upload data is non-existent. + ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody( + "some header", NULL)); +} + +TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_EmptyBody) { + ScopedVector<UploadElementReader> element_readers; + scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0)); + ASSERT_EQ(OK, body->Init(CompletionCallback())); + // Shouldn't be merged if upload data is empty. + ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody( + "some header", body.get())); +} + +TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_ChunkedBody) { + const std::string payload = "123"; + scoped_ptr<UploadDataStream> body( + new UploadDataStream(UploadDataStream::CHUNKED, 0)); + body->AppendChunk(payload.data(), payload.size(), true); + ASSERT_EQ(OK, body->Init(CompletionCallback())); + // Shouldn't be merged if upload data carries chunked data. + ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody( + "some header", body.get())); +} + +TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_FileBody) { + ScopedVector<UploadElementReader> element_readers; + + // Create an empty temporary file. + base::ScopedTempDir temp_dir; + ASSERT_TRUE(temp_dir.CreateUniqueTempDir()); + base::FilePath temp_file_path; + ASSERT_TRUE(file_util::CreateTemporaryFileInDir(temp_dir.path(), + &temp_file_path)); + + element_readers.push_back( + new UploadFileElementReader(base::MessageLoopProxy::current().get(), + temp_file_path, + 0, + 0, + base::Time())); + + scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0)); + TestCompletionCallback callback; + ASSERT_EQ(ERR_IO_PENDING, body->Init(callback.callback())); + ASSERT_EQ(OK, callback.WaitForResult()); + // Shouldn't be merged if upload data carries a file, as it's not in-memory. + ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody( + "some header", body.get())); +} + +TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_SmallBodyInMemory) { + ScopedVector<UploadElementReader> element_readers; + const std::string payload = "123"; + element_readers.push_back(new UploadBytesElementReader( + payload.data(), payload.size())); + + scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0)); + ASSERT_EQ(OK, body->Init(CompletionCallback())); + // Yes, should be merged if the in-memory body is small here. + ASSERT_TRUE(HttpStreamParser::ShouldMergeRequestHeadersAndBody( + "some header", body.get())); +} + +TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_LargeBodyInMemory) { + ScopedVector<UploadElementReader> element_readers; + const std::string payload(10000, 'a'); // 'a' x 10000. + element_readers.push_back(new UploadBytesElementReader( + payload.data(), payload.size())); + + scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0)); + ASSERT_EQ(OK, body->Init(CompletionCallback())); + // Shouldn't be merged if the in-memory body is large here. + ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody( + "some header", body.get())); +} + +// Test to ensure the HttpStreamParser state machine does not get confused +// when sending a request with a chunked body, where chunks become available +// asynchronously, over a socket where writes may also complete +// asynchronously. +// This is a regression test for http://crbug.com/132243 +TEST(HttpStreamParser, AsyncChunkAndAsyncSocket) { + // The chunks that will be written in the request, as reflected in the + // MockWrites below. + static const char kChunk1[] = "Chunk 1"; + static const char kChunk2[] = "Chunky 2"; + static const char kChunk3[] = "Test 3"; + + MockWrite writes[] = { + MockWrite(ASYNC, 0, + "GET /one.html HTTP/1.1\r\n" + "Host: localhost\r\n" + "Transfer-Encoding: chunked\r\n" + "Connection: keep-alive\r\n\r\n"), + MockWrite(ASYNC, 1, "7\r\nChunk 1\r\n"), + MockWrite(ASYNC, 2, "8\r\nChunky 2\r\n"), + MockWrite(ASYNC, 3, "6\r\nTest 3\r\n"), + MockWrite(ASYNC, 4, "0\r\n\r\n"), + }; + + // The size of the response body, as reflected in the Content-Length of the + // MockRead below. + static const int kBodySize = 8; + + MockRead reads[] = { + MockRead(ASYNC, 5, "HTTP/1.1 200 OK\r\n"), + MockRead(ASYNC, 6, "Content-Length: 8\r\n\r\n"), + MockRead(ASYNC, 7, "one.html"), + MockRead(SYNCHRONOUS, 0, 8), // EOF + }; + + UploadDataStream upload_stream(UploadDataStream::CHUNKED, 0); + upload_stream.AppendChunk(kChunk1, arraysize(kChunk1) - 1, false); + ASSERT_EQ(OK, upload_stream.Init(CompletionCallback())); + + DeterministicSocketData data(reads, arraysize(reads), + writes, arraysize(writes)); + data.set_connect_data(MockConnect(SYNCHRONOUS, OK)); + + scoped_ptr<DeterministicMockTCPClientSocket> transport( + new DeterministicMockTCPClientSocket(NULL, &data)); + data.set_delegate(transport->AsWeakPtr()); + + TestCompletionCallback callback; + int rv = transport->Connect(callback.callback()); + rv = callback.GetResult(rv); + ASSERT_EQ(OK, rv); + + scoped_ptr<ClientSocketHandle> socket_handle(new ClientSocketHandle); + socket_handle->SetSocket(transport.PassAs<StreamSocket>()); + + HttpRequestInfo request_info; + request_info.method = "GET"; + request_info.url = GURL("http://localhost"); + request_info.load_flags = LOAD_NORMAL; + request_info.upload_data_stream = &upload_stream; + + scoped_refptr<GrowableIOBuffer> read_buffer(new GrowableIOBuffer); + HttpStreamParser parser( + socket_handle.get(), &request_info, read_buffer.get(), BoundNetLog()); + + HttpRequestHeaders request_headers; + request_headers.SetHeader("Host", "localhost"); + request_headers.SetHeader("Transfer-Encoding", "chunked"); + request_headers.SetHeader("Connection", "keep-alive"); + + HttpResponseInfo response_info; + // This will attempt to Write() the initial request and headers, which will + // complete asynchronously. + rv = parser.SendRequest("GET /one.html HTTP/1.1\r\n", request_headers, + &response_info, callback.callback()); + ASSERT_EQ(ERR_IO_PENDING, rv); + + // Complete the initial request write. Additionally, this should enqueue the + // first chunk. + data.RunFor(1); + ASSERT_FALSE(callback.have_result()); + + // Now append another chunk (while the first write is still pending), which + // should not confuse the state machine. + upload_stream.AppendChunk(kChunk2, arraysize(kChunk2) - 1, false); + ASSERT_FALSE(callback.have_result()); + + // Complete writing the first chunk, which should then enqueue the second + // chunk for writing and return, because it is set to complete + // asynchronously. + data.RunFor(1); + ASSERT_FALSE(callback.have_result()); + + // Complete writing the second chunk. However, because no chunks are + // available yet, no further writes should be called until a new chunk is + // added. + data.RunFor(1); + ASSERT_FALSE(callback.have_result()); + + // Add the final chunk. This will enqueue another write, but it will not + // complete due to the async nature. + upload_stream.AppendChunk(kChunk3, arraysize(kChunk3) - 1, true); + ASSERT_FALSE(callback.have_result()); + + // Finalize writing the last chunk, which will enqueue the trailer. + data.RunFor(1); + ASSERT_FALSE(callback.have_result()); + + // Finalize writing the trailer. + data.RunFor(1); + ASSERT_TRUE(callback.have_result()); + + // Warning: This will hang if the callback doesn't already have a result, + // due to the deterministic socket provider. Do not remove the above + // ASSERT_TRUE, which will avoid this hang. + rv = callback.WaitForResult(); + ASSERT_EQ(OK, rv); + + // Attempt to read the response status and the response headers. + rv = parser.ReadResponseHeaders(callback.callback()); + ASSERT_EQ(ERR_IO_PENDING, rv); + data.RunFor(2); + + ASSERT_TRUE(callback.have_result()); + rv = callback.WaitForResult(); + ASSERT_GT(rv, 0); + + // Finally, attempt to read the response body. + scoped_refptr<IOBuffer> body_buffer(new IOBuffer(kBodySize)); + rv = parser.ReadResponseBody( + body_buffer.get(), kBodySize, callback.callback()); + ASSERT_EQ(ERR_IO_PENDING, rv); + data.RunFor(1); + + ASSERT_TRUE(callback.have_result()); + rv = callback.WaitForResult(); + ASSERT_EQ(kBodySize, rv); +} + +TEST(HttpStreamParser, TruncatedHeaders) { + MockRead truncated_status_reads[] = { + MockRead(SYNCHRONOUS, 1, "HTTP/1.1 20"), + MockRead(SYNCHRONOUS, 0, 2), // EOF + }; + + MockRead truncated_after_status_reads[] = { + MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\n"), + MockRead(SYNCHRONOUS, 0, 2), // EOF + }; + + MockRead truncated_in_header_reads[] = { + MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHead"), + MockRead(SYNCHRONOUS, 0, 2), // EOF + }; + + MockRead truncated_after_header_reads[] = { + MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHeader: foo\r\n"), + MockRead(SYNCHRONOUS, 0, 2), // EOF + }; + + MockRead truncated_after_final_newline_reads[] = { + MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHeader: foo\r\n\r"), + MockRead(SYNCHRONOUS, 0, 2), // EOF + }; + + MockRead not_truncated_reads[] = { + MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHeader: foo\r\n\r\n"), + MockRead(SYNCHRONOUS, 0, 2), // EOF + }; + + MockRead* reads[] = { + truncated_status_reads, + truncated_after_status_reads, + truncated_in_header_reads, + truncated_after_header_reads, + truncated_after_final_newline_reads, + not_truncated_reads, + }; + + MockWrite writes[] = { + MockWrite(SYNCHRONOUS, 0, "GET / HTTP/1.1\r\n\r\n"), + }; + + enum { + HTTP = 0, + HTTPS, + NUM_PROTOCOLS, + }; + + for (size_t protocol = 0; protocol < NUM_PROTOCOLS; protocol++) { + SCOPED_TRACE(protocol); + + for (size_t i = 0; i < arraysize(reads); i++) { + SCOPED_TRACE(i); + DeterministicSocketData data(reads[i], 2, writes, arraysize(writes)); + data.set_connect_data(MockConnect(SYNCHRONOUS, OK)); + data.SetStop(3); + + scoped_ptr<DeterministicMockTCPClientSocket> transport( + new DeterministicMockTCPClientSocket(NULL, &data)); + data.set_delegate(transport->AsWeakPtr()); + + TestCompletionCallback callback; + int rv = transport->Connect(callback.callback()); + rv = callback.GetResult(rv); + ASSERT_EQ(OK, rv); + + scoped_ptr<ClientSocketHandle> socket_handle(new ClientSocketHandle); + socket_handle->SetSocket(transport.PassAs<StreamSocket>()); + + HttpRequestInfo request_info; + request_info.method = "GET"; + if (protocol == HTTP) { + request_info.url = GURL("http://localhost"); + } else { + request_info.url = GURL("https://localhost"); + } + request_info.load_flags = LOAD_NORMAL; + + scoped_refptr<GrowableIOBuffer> read_buffer(new GrowableIOBuffer); + HttpStreamParser parser( + socket_handle.get(), &request_info, read_buffer.get(), BoundNetLog()); + + HttpRequestHeaders request_headers; + HttpResponseInfo response_info; + rv = parser.SendRequest("GET / HTTP/1.1\r\n", request_headers, + &response_info, callback.callback()); + ASSERT_EQ(OK, rv); + + rv = parser.ReadResponseHeaders(callback.callback()); + if (i == arraysize(reads) - 1) { + EXPECT_EQ(OK, rv); + EXPECT_TRUE(response_info.headers.get()); + } else { + if (protocol == HTTP) { + EXPECT_EQ(ERR_CONNECTION_CLOSED, rv); + EXPECT_TRUE(response_info.headers.get()); + } else { + EXPECT_EQ(ERR_RESPONSE_HEADERS_TRUNCATED, rv); + EXPECT_FALSE(response_info.headers.get()); + } + } + } + } +} + +} // namespace net |