// __ _____ _____ _____ // __| | __| | | | JSON for Modern C++ (supporting code) // | | |__ | | | | | | version 3.11.2 // |_____|_____|_____|_|___| https://github.com/nlohmann/json // // SPDX-FileCopyrightText: 2013-2022 Niels Lohmann // SPDX-License-Identifier: MIT #include "doctest_compatibility.h" #define JSON_TESTS_PRIVATE #include using nlohmann::json; namespace { // shortcut to scan a string literal json::lexer::token_type scan_string(const char* s, bool ignore_comments = false); json::lexer::token_type scan_string(const char* s, const bool ignore_comments) { auto ia = nlohmann::detail::input_adapter(s); return nlohmann::detail::lexer(std::move(ia), ignore_comments).scan(); // NOLINT(hicpp-move-const-arg,performance-move-const-arg) } } // namespace std::string get_error_message(const char* s, bool ignore_comments = false); std::string get_error_message(const char* s, const bool ignore_comments) { auto ia = nlohmann::detail::input_adapter(s); auto lexer = nlohmann::detail::lexer(std::move(ia), ignore_comments); // NOLINT(hicpp-move-const-arg,performance-move-const-arg) lexer.scan(); return lexer.get_error_message(); } TEST_CASE("lexer class") { SECTION("scan") { SECTION("structural characters") { CHECK((scan_string("[") == json::lexer::token_type::begin_array)); CHECK((scan_string("]") == json::lexer::token_type::end_array)); CHECK((scan_string("{") == json::lexer::token_type::begin_object)); CHECK((scan_string("}") == json::lexer::token_type::end_object)); CHECK((scan_string(",") == json::lexer::token_type::value_separator)); CHECK((scan_string(":") == json::lexer::token_type::name_separator)); } SECTION("literal names") { CHECK((scan_string("null") == json::lexer::token_type::literal_null)); CHECK((scan_string("true") == json::lexer::token_type::literal_true)); CHECK((scan_string("false") == json::lexer::token_type::literal_false)); } SECTION("numbers") { CHECK((scan_string("0") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("1") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("2") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("3") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("4") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("5") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("6") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("7") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("8") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("9") == json::lexer::token_type::value_unsigned)); CHECK((scan_string("-0") == json::lexer::token_type::value_integer)); CHECK((scan_string("-1") == json::lexer::token_type::value_integer)); CHECK((scan_string("1.1") == json::lexer::token_type::value_float)); CHECK((scan_string("-1.1") == json::lexer::token_type::value_float)); CHECK((scan_string("1E10") == json::lexer::token_type::value_float)); } SECTION("whitespace") { // result is end_of_input, because not token is following CHECK((scan_string(" ") == json::lexer::token_type::end_of_input)); CHECK((scan_string("\t") == json::lexer::token_type::end_of_input)); CHECK((scan_string("\n") == json::lexer::token_type::end_of_input)); CHECK((scan_string("\r") == json::lexer::token_type::end_of_input)); CHECK((scan_string(" \t\n\r\n\t ") == json::lexer::token_type::end_of_input)); } } SECTION("token_type_name") { CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::uninitialized)) == "")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::literal_true)) == "true literal")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::literal_false)) == "false literal")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::literal_null)) == "null literal")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::value_string)) == "string literal")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::value_unsigned)) == "number literal")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::value_integer)) == "number literal")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::value_float)) == "number literal")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::begin_array)) == "'['")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::begin_object)) == "'{'")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::end_array)) == "']'")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::end_object)) == "'}'")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::name_separator)) == "':'")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::value_separator)) == "','")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::parse_error)) == "")); CHECK((std::string(json::lexer::token_type_name(json::lexer::token_type::end_of_input)) == "end of input")); } SECTION("parse errors on first character") { for (int c = 1; c < 128; ++c) { // create string from the ASCII code const auto s = std::string(1, static_cast(c)); // store scan() result const auto res = scan_string(s.c_str()); CAPTURE(s) switch (c) { // single characters that are valid tokens case ('['): case (']'): case ('{'): case ('}'): case (','): case (':'): case ('0'): case ('1'): case ('2'): case ('3'): case ('4'): case ('5'): case ('6'): case ('7'): case ('8'): case ('9'): { CHECK((res != json::lexer::token_type::parse_error)); break; } // whitespace case (' '): case ('\t'): case ('\n'): case ('\r'): { CHECK((res == json::lexer::token_type::end_of_input)); break; } // anything else is not expected default: { CHECK((res == json::lexer::token_type::parse_error)); break; } } } } SECTION("very large string") { // strings larger than 1024 bytes yield a resize of the lexer's yytext buffer std::string s("\""); s += std::string(2048, 'x'); s += "\""; CHECK((scan_string(s.c_str()) == json::lexer::token_type::value_string)); } SECTION("fail on comments") { CHECK((scan_string("/", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/", false) == "invalid literal"); CHECK((scan_string("/!", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/!", false) == "invalid literal"); CHECK((scan_string("/*", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/*", false) == "invalid literal"); CHECK((scan_string("/**", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/**", false) == "invalid literal"); CHECK((scan_string("//", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("//", false) == "invalid literal"); CHECK((scan_string("/**/", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/**/", false) == "invalid literal"); CHECK((scan_string("/** /", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/** /", false) == "invalid literal"); CHECK((scan_string("/***/", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/***/", false) == "invalid literal"); CHECK((scan_string("/* true */", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/* true */", false) == "invalid literal"); CHECK((scan_string("/*/**/", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/*/**/", false) == "invalid literal"); CHECK((scan_string("/*/* */", false) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/*/* */", false) == "invalid literal"); } SECTION("ignore comments") { CHECK((scan_string("/", true) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/", true) == "invalid comment; expecting '/' or '*' after '/'"); CHECK((scan_string("/!", true) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/!", true) == "invalid comment; expecting '/' or '*' after '/'"); CHECK((scan_string("/*", true) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/*", true) == "invalid comment; missing closing '*/'"); CHECK((scan_string("/**", true) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/**", true) == "invalid comment; missing closing '*/'"); CHECK((scan_string("//", true) == json::lexer::token_type::end_of_input)); CHECK((scan_string("/**/", true) == json::lexer::token_type::end_of_input)); CHECK((scan_string("/** /", true) == json::lexer::token_type::parse_error)); CHECK(get_error_message("/** /", true) == "invalid comment; missing closing '*/'"); CHECK((scan_string("/***/", true) == json::lexer::token_type::end_of_input)); CHECK((scan_string("/* true */", true) == json::lexer::token_type::end_of_input)); CHECK((scan_string("/*/**/", true) == json::lexer::token_type::end_of_input)); CHECK((scan_string("/*/* */", true) == json::lexer::token_type::end_of_input)); CHECK((scan_string("//\n//\n", true) == json::lexer::token_type::end_of_input)); CHECK((scan_string("/**//**//**/", true) == json::lexer::token_type::end_of_input)); } }