[go: nahoru, domu]

blob: cf5fa4140b098e2e4893884ef5bcf1ecf0dd036f [file] [log] [blame]
// Copyright (c) 2012 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "net/http/http_stream_parser.h"
#include "base/file_util.h"
#include "base/files/file_path.h"
#include "base/files/scoped_temp_dir.h"
#include "base/memory/ref_counted.h"
#include "base/strings/string_piece.h"
#include "base/strings/stringprintf.h"
#include "googleurl/src/gurl.h"
#include "net/base/io_buffer.h"
#include "net/base/net_errors.h"
#include "net/base/test_completion_callback.h"
#include "net/base/upload_bytes_element_reader.h"
#include "net/base/upload_data_stream.h"
#include "net/base/upload_file_element_reader.h"
#include "net/http/http_request_headers.h"
#include "net/http/http_request_info.h"
#include "net/http/http_response_info.h"
#include "net/socket/client_socket_handle.h"
#include "net/socket/socket_test_util.h"
#include "testing/gtest/include/gtest/gtest.h"
namespace net {
const size_t kOutputSize = 1024; // Just large enough for this test.
// The number of bytes that can fit in a buffer of kOutputSize.
const size_t kMaxPayloadSize =
kOutputSize - HttpStreamParser::kChunkHeaderFooterSize;
// The empty payload is how the last chunk is encoded.
TEST(HttpStreamParser, EncodeChunk_EmptyPayload) {
char output[kOutputSize];
const base::StringPiece kPayload = "";
const base::StringPiece kExpected = "0\r\n\r\n";
const int num_bytes_written =
HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output));
ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written));
EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written));
}
TEST(HttpStreamParser, EncodeChunk_ShortPayload) {
char output[kOutputSize];
const std::string kPayload("foo\x00\x11\x22", 6);
// 11 = payload size + sizeof("6") + CRLF x 2.
const std::string kExpected("6\r\nfoo\x00\x11\x22\r\n", 11);
const int num_bytes_written =
HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output));
ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written));
EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written));
}
TEST(HttpStreamParser, EncodeChunk_LargePayload) {
char output[kOutputSize];
const std::string kPayload(1000, '\xff'); // '\xff' x 1000.
// 3E8 = 1000 in hex.
const std::string kExpected = "3E8\r\n" + kPayload + "\r\n";
const int num_bytes_written =
HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output));
ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written));
EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written));
}
TEST(HttpStreamParser, EncodeChunk_FullPayload) {
char output[kOutputSize];
const std::string kPayload(kMaxPayloadSize, '\xff');
// 3F4 = 1012 in hex.
const std::string kExpected = "3F4\r\n" + kPayload + "\r\n";
const int num_bytes_written =
HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output));
ASSERT_EQ(kExpected.size(), static_cast<size_t>(num_bytes_written));
EXPECT_EQ(kExpected, base::StringPiece(output, num_bytes_written));
}
TEST(HttpStreamParser, EncodeChunk_TooLargePayload) {
char output[kOutputSize];
// The payload is one byte larger the output buffer size.
const std::string kPayload(kMaxPayloadSize + 1, '\xff');
const int num_bytes_written =
HttpStreamParser::EncodeChunk(kPayload, output, sizeof(output));
ASSERT_EQ(ERR_INVALID_ARGUMENT, num_bytes_written);
}
TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_NoBody) {
// Shouldn't be merged if upload data is non-existent.
ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody(
"some header", NULL));
}
TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_EmptyBody) {
ScopedVector<UploadElementReader> element_readers;
scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0));
ASSERT_EQ(OK, body->Init(CompletionCallback()));
// Shouldn't be merged if upload data is empty.
ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody(
"some header", body.get()));
}
TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_ChunkedBody) {
const std::string payload = "123";
scoped_ptr<UploadDataStream> body(
new UploadDataStream(UploadDataStream::CHUNKED, 0));
body->AppendChunk(payload.data(), payload.size(), true);
ASSERT_EQ(OK, body->Init(CompletionCallback()));
// Shouldn't be merged if upload data carries chunked data.
ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody(
"some header", body.get()));
}
TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_FileBody) {
ScopedVector<UploadElementReader> element_readers;
// Create an empty temporary file.
base::ScopedTempDir temp_dir;
ASSERT_TRUE(temp_dir.CreateUniqueTempDir());
base::FilePath temp_file_path;
ASSERT_TRUE(file_util::CreateTemporaryFileInDir(temp_dir.path(),
&temp_file_path));
element_readers.push_back(
new UploadFileElementReader(base::MessageLoopProxy::current().get(),
temp_file_path,
0,
0,
base::Time()));
scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0));
TestCompletionCallback callback;
ASSERT_EQ(ERR_IO_PENDING, body->Init(callback.callback()));
ASSERT_EQ(OK, callback.WaitForResult());
// Shouldn't be merged if upload data carries a file, as it's not in-memory.
ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody(
"some header", body.get()));
}
TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_SmallBodyInMemory) {
ScopedVector<UploadElementReader> element_readers;
const std::string payload = "123";
element_readers.push_back(new UploadBytesElementReader(
payload.data(), payload.size()));
scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0));
ASSERT_EQ(OK, body->Init(CompletionCallback()));
// Yes, should be merged if the in-memory body is small here.
ASSERT_TRUE(HttpStreamParser::ShouldMergeRequestHeadersAndBody(
"some header", body.get()));
}
TEST(HttpStreamParser, ShouldMergeRequestHeadersAndBody_LargeBodyInMemory) {
ScopedVector<UploadElementReader> element_readers;
const std::string payload(10000, 'a'); // 'a' x 10000.
element_readers.push_back(new UploadBytesElementReader(
payload.data(), payload.size()));
scoped_ptr<UploadDataStream> body(new UploadDataStream(&element_readers, 0));
ASSERT_EQ(OK, body->Init(CompletionCallback()));
// Shouldn't be merged if the in-memory body is large here.
ASSERT_FALSE(HttpStreamParser::ShouldMergeRequestHeadersAndBody(
"some header", body.get()));
}
// Test to ensure the HttpStreamParser state machine does not get confused
// when sending a request with a chunked body, where chunks become available
// asynchronously, over a socket where writes may also complete
// asynchronously.
// This is a regression test for http://crbug.com/132243
TEST(HttpStreamParser, AsyncChunkAndAsyncSocket) {
// The chunks that will be written in the request, as reflected in the
// MockWrites below.
static const char kChunk1[] = "Chunk 1";
static const char kChunk2[] = "Chunky 2";
static const char kChunk3[] = "Test 3";
MockWrite writes[] = {
MockWrite(ASYNC, 0,
"GET /one.html HTTP/1.1\r\n"
"Host: localhost\r\n"
"Transfer-Encoding: chunked\r\n"
"Connection: keep-alive\r\n\r\n"),
MockWrite(ASYNC, 1, "7\r\nChunk 1\r\n"),
MockWrite(ASYNC, 2, "8\r\nChunky 2\r\n"),
MockWrite(ASYNC, 3, "6\r\nTest 3\r\n"),
MockWrite(ASYNC, 4, "0\r\n\r\n"),
};
// The size of the response body, as reflected in the Content-Length of the
// MockRead below.
static const int kBodySize = 8;
MockRead reads[] = {
MockRead(ASYNC, 5, "HTTP/1.1 200 OK\r\n"),
MockRead(ASYNC, 6, "Content-Length: 8\r\n\r\n"),
MockRead(ASYNC, 7, "one.html"),
MockRead(SYNCHRONOUS, 0, 8), // EOF
};
UploadDataStream upload_stream(UploadDataStream::CHUNKED, 0);
upload_stream.AppendChunk(kChunk1, arraysize(kChunk1) - 1, false);
ASSERT_EQ(OK, upload_stream.Init(CompletionCallback()));
DeterministicSocketData data(reads, arraysize(reads),
writes, arraysize(writes));
data.set_connect_data(MockConnect(SYNCHRONOUS, OK));
scoped_ptr<DeterministicMockTCPClientSocket> transport(
new DeterministicMockTCPClientSocket(NULL, &data));
data.set_delegate(transport->AsWeakPtr());
TestCompletionCallback callback;
int rv = transport->Connect(callback.callback());
rv = callback.GetResult(rv);
ASSERT_EQ(OK, rv);
scoped_ptr<ClientSocketHandle> socket_handle(new ClientSocketHandle);
socket_handle->set_socket(transport.release());
HttpRequestInfo request_info;
request_info.method = "GET";
request_info.url = GURL("http://localhost");
request_info.load_flags = LOAD_NORMAL;
request_info.upload_data_stream = &upload_stream;
scoped_refptr<GrowableIOBuffer> read_buffer(new GrowableIOBuffer);
HttpStreamParser parser(
socket_handle.get(), &request_info, read_buffer.get(), BoundNetLog());
HttpRequestHeaders request_headers;
request_headers.SetHeader("Host", "localhost");
request_headers.SetHeader("Transfer-Encoding", "chunked");
request_headers.SetHeader("Connection", "keep-alive");
HttpResponseInfo response_info;
// This will attempt to Write() the initial request and headers, which will
// complete asynchronously.
rv = parser.SendRequest("GET /one.html HTTP/1.1\r\n", request_headers,
&response_info, callback.callback());
ASSERT_EQ(ERR_IO_PENDING, rv);
// Complete the initial request write. Additionally, this should enqueue the
// first chunk.
data.RunFor(1);
ASSERT_FALSE(callback.have_result());
// Now append another chunk (while the first write is still pending), which
// should not confuse the state machine.
upload_stream.AppendChunk(kChunk2, arraysize(kChunk2) - 1, false);
ASSERT_FALSE(callback.have_result());
// Complete writing the first chunk, which should then enqueue the second
// chunk for writing and return, because it is set to complete
// asynchronously.
data.RunFor(1);
ASSERT_FALSE(callback.have_result());
// Complete writing the second chunk. However, because no chunks are
// available yet, no further writes should be called until a new chunk is
// added.
data.RunFor(1);
ASSERT_FALSE(callback.have_result());
// Add the final chunk. This will enqueue another write, but it will not
// complete due to the async nature.
upload_stream.AppendChunk(kChunk3, arraysize(kChunk3) - 1, true);
ASSERT_FALSE(callback.have_result());
// Finalize writing the last chunk, which will enqueue the trailer.
data.RunFor(1);
ASSERT_FALSE(callback.have_result());
// Finalize writing the trailer.
data.RunFor(1);
ASSERT_TRUE(callback.have_result());
// Warning: This will hang if the callback doesn't already have a result,
// due to the deterministic socket provider. Do not remove the above
// ASSERT_TRUE, which will avoid this hang.
rv = callback.WaitForResult();
ASSERT_EQ(OK, rv);
// Attempt to read the response status and the response headers.
rv = parser.ReadResponseHeaders(callback.callback());
ASSERT_EQ(ERR_IO_PENDING, rv);
data.RunFor(2);
ASSERT_TRUE(callback.have_result());
rv = callback.WaitForResult();
ASSERT_GT(rv, 0);
// Finally, attempt to read the response body.
scoped_refptr<IOBuffer> body_buffer(new IOBuffer(kBodySize));
rv = parser.ReadResponseBody(
body_buffer.get(), kBodySize, callback.callback());
ASSERT_EQ(ERR_IO_PENDING, rv);
data.RunFor(1);
ASSERT_TRUE(callback.have_result());
rv = callback.WaitForResult();
ASSERT_EQ(kBodySize, rv);
}
TEST(HttpStreamParser, TruncatedHeaders) {
MockRead truncated_status_reads[] = {
MockRead(SYNCHRONOUS, 1, "HTTP/1.1 20"),
MockRead(SYNCHRONOUS, 0, 2), // EOF
};
MockRead truncated_after_status_reads[] = {
MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\n"),
MockRead(SYNCHRONOUS, 0, 2), // EOF
};
MockRead truncated_in_header_reads[] = {
MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHead"),
MockRead(SYNCHRONOUS, 0, 2), // EOF
};
MockRead truncated_after_header_reads[] = {
MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHeader: foo\r\n"),
MockRead(SYNCHRONOUS, 0, 2), // EOF
};
MockRead truncated_after_final_newline_reads[] = {
MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHeader: foo\r\n\r"),
MockRead(SYNCHRONOUS, 0, 2), // EOF
};
MockRead not_truncated_reads[] = {
MockRead(SYNCHRONOUS, 1, "HTTP/1.1 200 Ok\r\nHeader: foo\r\n\r\n"),
MockRead(SYNCHRONOUS, 0, 2), // EOF
};
MockRead* reads[] = {
truncated_status_reads,
truncated_after_status_reads,
truncated_in_header_reads,
truncated_after_header_reads,
truncated_after_final_newline_reads,
not_truncated_reads,
};
MockWrite writes[] = {
MockWrite(SYNCHRONOUS, 0, "GET / HTTP/1.1\r\n\r\n"),
};
enum {
HTTP = 0,
HTTPS,
NUM_PROTOCOLS,
};
for (size_t protocol = 0; protocol < NUM_PROTOCOLS; protocol++) {
SCOPED_TRACE(protocol);
for (size_t i = 0; i < arraysize(reads); i++) {
SCOPED_TRACE(i);
DeterministicSocketData data(reads[i], 2, writes, arraysize(writes));
data.set_connect_data(MockConnect(SYNCHRONOUS, OK));
data.SetStop(3);
scoped_ptr<DeterministicMockTCPClientSocket> transport(
new DeterministicMockTCPClientSocket(NULL, &data));
data.set_delegate(transport->AsWeakPtr());
TestCompletionCallback callback;
int rv = transport->Connect(callback.callback());
rv = callback.GetResult(rv);
ASSERT_EQ(OK, rv);
scoped_ptr<ClientSocketHandle> socket_handle(new ClientSocketHandle);
socket_handle->set_socket(transport.release());
HttpRequestInfo request_info;
request_info.method = "GET";
if (protocol == HTTP) {
request_info.url = GURL("http://localhost");
} else {
request_info.url = GURL("https://localhost");
}
request_info.load_flags = LOAD_NORMAL;
scoped_refptr<GrowableIOBuffer> read_buffer(new GrowableIOBuffer);
HttpStreamParser parser(
socket_handle.get(), &request_info, read_buffer.get(), BoundNetLog());
HttpRequestHeaders request_headers;
HttpResponseInfo response_info;
rv = parser.SendRequest("GET / HTTP/1.1\r\n", request_headers,
&response_info, callback.callback());
ASSERT_EQ(OK, rv);
rv = parser.ReadResponseHeaders(callback.callback());
if (i == arraysize(reads) - 1) {
EXPECT_EQ(OK, rv);
EXPECT_TRUE(response_info.headers.get());
} else {
if (protocol == HTTP) {
EXPECT_EQ(ERR_CONNECTION_CLOSED, rv);
EXPECT_TRUE(response_info.headers.get());
} else {
EXPECT_EQ(ERR_HEADERS_TRUNCATED, rv);
EXPECT_FALSE(response_info.headers.get());
}
}
}
}
}
} // namespace net