|  | // Copyright (c) 2013 The Chromium Authors. All rights reserved. | 
|  | // Use of this source code is governed by a BSD-style license that can be | 
|  | // found in the LICENSE file. | 
|  |  | 
|  | #include <stddef.h> | 
|  |  | 
|  | #include "tools/gn/input_file.h" | 
|  | #include "tools/gn/token.h" | 
|  | #include "tools/gn/tokenizer.h" | 
|  | #include "util/test/test.h" | 
|  |  | 
|  | namespace { | 
|  |  | 
|  | struct TokenExpectation { | 
|  | Token::Type type; | 
|  | const char* value; | 
|  | }; | 
|  |  | 
|  | template <size_t len> | 
|  | bool CheckTokenizer(const char* input, const TokenExpectation (&expect)[len]) { | 
|  | InputFile input_file(SourceFile("/test")); | 
|  | input_file.SetContents(input); | 
|  |  | 
|  | Err err; | 
|  | std::vector<Token> results = Tokenizer::Tokenize(&input_file, &err); | 
|  |  | 
|  | if (results.size() != len) | 
|  | return false; | 
|  | for (size_t i = 0; i < len; i++) { | 
|  | if (expect[i].type != results[i].type()) | 
|  | return false; | 
|  | if (expect[i].value != results[i].value()) | 
|  | return false; | 
|  | } | 
|  | return true; | 
|  | } | 
|  |  | 
|  | }  // namespace | 
|  |  | 
|  | TEST(Tokenizer, Empty) { | 
|  | InputFile empty_string_input(SourceFile("/test")); | 
|  | empty_string_input.SetContents(""); | 
|  |  | 
|  | Err err; | 
|  | std::vector<Token> results = Tokenizer::Tokenize(&empty_string_input, &err); | 
|  | EXPECT_TRUE(results.empty()); | 
|  |  | 
|  | InputFile whitespace_input(SourceFile("/test")); | 
|  | whitespace_input.SetContents("  \r \n \r\n"); | 
|  |  | 
|  | results = Tokenizer::Tokenize(&whitespace_input, &err); | 
|  | EXPECT_TRUE(results.empty()); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, Identifier) { | 
|  | TokenExpectation one_ident[] = {{Token::IDENTIFIER, "foo"}}; | 
|  | EXPECT_TRUE(CheckTokenizer("  foo ", one_ident)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, Integer) { | 
|  | TokenExpectation integers[] = {{Token::INTEGER, "123"}, | 
|  | {Token::INTEGER, "-123"}}; | 
|  | EXPECT_TRUE(CheckTokenizer("  123 -123 ", integers)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, IntegerNoSpace) { | 
|  | TokenExpectation integers[] = {{Token::INTEGER, "123"}, | 
|  | {Token::INTEGER, "-123"}}; | 
|  | EXPECT_TRUE(CheckTokenizer("  123-123 ", integers)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, String) { | 
|  | TokenExpectation strings[] = {{Token::STRING, "\"foo\""}, | 
|  | {Token::STRING, "\"bar\\\"baz\""}, | 
|  | {Token::STRING, "\"asdf\\\\\""}}; | 
|  | EXPECT_TRUE( | 
|  | CheckTokenizer("  \"foo\" \"bar\\\"baz\" \"asdf\\\\\" ", strings)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, Operator) { | 
|  | TokenExpectation operators[] = { | 
|  | {Token::MINUS, "-"}, | 
|  | {Token::PLUS, "+"}, | 
|  | {Token::EQUAL, "="}, | 
|  | {Token::PLUS_EQUALS, "+="}, | 
|  | {Token::MINUS_EQUALS, "-="}, | 
|  | {Token::NOT_EQUAL, "!="}, | 
|  | {Token::EQUAL_EQUAL, "=="}, | 
|  | {Token::LESS_THAN, "<"}, | 
|  | {Token::GREATER_THAN, ">"}, | 
|  | {Token::LESS_EQUAL, "<="}, | 
|  | {Token::GREATER_EQUAL, ">="}, | 
|  | {Token::BANG, "!"}, | 
|  | {Token::BOOLEAN_OR, "||"}, | 
|  | {Token::BOOLEAN_AND, "&&"}, | 
|  | {Token::DOT, "."}, | 
|  | {Token::COMMA, ","}, | 
|  | }; | 
|  | EXPECT_TRUE( | 
|  | CheckTokenizer("- + = += -= != ==  < > <= >= ! || && . ,", operators)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, Scoper) { | 
|  | TokenExpectation scopers[] = { | 
|  | {Token::LEFT_BRACE, "{"},    {Token::LEFT_BRACKET, "["}, | 
|  | {Token::RIGHT_BRACKET, "]"}, {Token::RIGHT_BRACE, "}"}, | 
|  | {Token::LEFT_PAREN, "("},    {Token::RIGHT_PAREN, ")"}, | 
|  | }; | 
|  | EXPECT_TRUE(CheckTokenizer("{[ ]} ()", scopers)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, FunctionCall) { | 
|  | TokenExpectation fn[] = { | 
|  | {Token::IDENTIFIER, "fun"}, {Token::LEFT_PAREN, "("}, | 
|  | {Token::STRING, "\"foo\""}, {Token::RIGHT_PAREN, ")"}, | 
|  | {Token::LEFT_BRACE, "{"},   {Token::IDENTIFIER, "foo"}, | 
|  | {Token::EQUAL, "="},        {Token::INTEGER, "12"}, | 
|  | {Token::RIGHT_BRACE, "}"}, | 
|  | }; | 
|  | EXPECT_TRUE(CheckTokenizer("fun(\"foo\") {\nfoo = 12}", fn)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, Locations) { | 
|  | InputFile input(SourceFile("/test")); | 
|  | input.SetContents("1 2 \"three\"\n  4"); | 
|  | Err err; | 
|  | std::vector<Token> results = Tokenizer::Tokenize(&input, &err); | 
|  |  | 
|  | ASSERT_EQ(4u, results.size()); | 
|  | ASSERT_TRUE(results[0].location() == Location(&input, 1, 1, 1)); | 
|  | ASSERT_TRUE(results[1].location() == Location(&input, 1, 3, 3)); | 
|  | ASSERT_TRUE(results[2].location() == Location(&input, 1, 5, 5)); | 
|  | ASSERT_TRUE(results[3].location() == Location(&input, 2, 3, 8)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, ByteOffsetOfNthLine) { | 
|  | EXPECT_EQ(0u, Tokenizer::ByteOffsetOfNthLine("foo", 1)); | 
|  |  | 
|  | // Windows and Posix have different line endings, so check the byte at the | 
|  | // location rather than the offset. | 
|  | char input1[] = "aaa\nxaa\n\nya"; | 
|  | EXPECT_EQ('x', input1[Tokenizer::ByteOffsetOfNthLine(input1, 2)]); | 
|  | EXPECT_EQ('y', input1[Tokenizer::ByteOffsetOfNthLine(input1, 4)]); | 
|  |  | 
|  | char input2[3]; | 
|  | input2[0] = 'a'; | 
|  | input2[1] = '\n';  // Manually set to avoid Windows double-byte endings. | 
|  | input2[2] = 0; | 
|  | EXPECT_EQ(0u, Tokenizer::ByteOffsetOfNthLine(input2, 1)); | 
|  | EXPECT_EQ(2u, Tokenizer::ByteOffsetOfNthLine(input2, 2)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, Comments) { | 
|  | TokenExpectation fn[] = { | 
|  | {Token::LINE_COMMENT, "# Stuff"}, | 
|  | {Token::IDENTIFIER, "fun"}, | 
|  | {Token::LEFT_PAREN, "("}, | 
|  | {Token::STRING, "\"foo\""}, | 
|  | {Token::RIGHT_PAREN, ")"}, | 
|  | {Token::LEFT_BRACE, "{"}, | 
|  | {Token::SUFFIX_COMMENT, "# Things"}, | 
|  | {Token::LINE_COMMENT, "#Wee"}, | 
|  | {Token::IDENTIFIER, "foo"}, | 
|  | {Token::EQUAL, "="}, | 
|  | {Token::INTEGER, "12"}, | 
|  | {Token::SUFFIX_COMMENT, "#Zip"}, | 
|  | {Token::RIGHT_BRACE, "}"}, | 
|  | }; | 
|  | EXPECT_TRUE( | 
|  | CheckTokenizer("# Stuff\n" | 
|  | "fun(\"foo\") {  # Things\n" | 
|  | "#Wee\n" | 
|  | "foo = 12 #Zip\n" | 
|  | "}", | 
|  | fn)); | 
|  | } | 
|  |  | 
|  | TEST(Tokenizer, CommentsContinued) { | 
|  | // In the first test, the comments aren't horizontally aligned, so they're | 
|  | // considered separate. In the second test, they are, so "B" is a | 
|  | // continuation of "A" (another SUFFIX comment). | 
|  | TokenExpectation fn1[] = { | 
|  | {Token::IDENTIFIER, "fun"},   {Token::LEFT_PAREN, "("}, | 
|  | {Token::STRING, "\"foo\""},   {Token::RIGHT_PAREN, ")"}, | 
|  | {Token::LEFT_BRACE, "{"},     {Token::SUFFIX_COMMENT, "# A"}, | 
|  | {Token::LINE_COMMENT, "# B"}, {Token::RIGHT_BRACE, "}"}, | 
|  | }; | 
|  | EXPECT_TRUE( | 
|  | CheckTokenizer("fun(\"foo\") {  # A\n" | 
|  | "  # B\n" | 
|  | "}", | 
|  | fn1)); | 
|  |  | 
|  | TokenExpectation fn2[] = { | 
|  | {Token::IDENTIFIER, "fun"},     {Token::LEFT_PAREN, "("}, | 
|  | {Token::STRING, "\"foo\""},     {Token::RIGHT_PAREN, ")"}, | 
|  | {Token::LEFT_BRACE, "{"},       {Token::SUFFIX_COMMENT, "# A"}, | 
|  | {Token::SUFFIX_COMMENT, "# B"}, {Token::RIGHT_BRACE, "}"}, | 
|  | }; | 
|  | EXPECT_TRUE(CheckTokenizer( | 
|  | "fun(\"foo\") {  # A\n" | 
|  | "              # B\n"  // Note that these are aligned, the \"s move A out. | 
|  | "}", | 
|  | fn2)); | 
|  | } |