|  | // Protocol Buffers - Google's data interchange format | 
|  | // Copyright 2008 Google Inc.  All rights reserved. | 
|  | // | 
|  | // Use of this source code is governed by a BSD-style | 
|  | // license that can be found in the LICENSE file or at | 
|  | // https://developers.google.com/open-source/licenses/bsd | 
|  |  | 
|  | // This file contains a program for running the test suite in a separate | 
|  | // process.  The other alternative is to run the suite in-process.  See | 
|  | // conformance.proto for pros/cons of these two options. | 
|  | // | 
|  | // This program will fork the process under test and communicate with it over | 
|  | // its stdin/stdout: | 
|  | // | 
|  | //     +--------+   pipe   +----------+ | 
|  | //     | tester | <------> | testee   | | 
|  | //     |        |          |          | | 
|  | //     |  C++   |          | any lang | | 
|  | //     +--------+          +----------+ | 
|  | // | 
|  | // The tester contains all of the test cases and their expected output. | 
|  | // The testee is a simple program written in the target language that reads | 
|  | // each test case and attempts to produce acceptable output for it. | 
|  | // | 
|  | // Every test consists of a ConformanceRequest/ConformanceResponse | 
|  | // request/reply pair.  The protocol on the pipe is simply: | 
|  | // | 
|  | //   1. tester sends 4-byte length N (little endian) | 
|  | //   2. tester sends N bytes representing a ConformanceRequest proto | 
|  | //   3. testee sends 4-byte length M (little endian) | 
|  | //   4. testee sends M bytes representing a ConformanceResponse proto | 
|  |  | 
|  | #include <signal.h> | 
|  | #include <stdio.h> | 
|  | #include <sys/types.h> | 
|  | #include <sys/wait.h> | 
|  | #include <unistd.h> | 
|  |  | 
|  | #include <algorithm> | 
|  | #include <cctype> | 
|  | #include <cstdio> | 
|  | #include <cstdlib> | 
|  | #include <cstring> | 
|  | #include <fstream> | 
|  | #include <string> | 
|  | #include <vector> | 
|  |  | 
|  | #include "absl/container/flat_hash_set.h" | 
|  | #include "absl/strings/ascii.h" | 
|  | #include "absl/strings/str_cat.h" | 
|  | #include "conformance/conformance.pb.h" | 
|  | #include "conformance_test.h" | 
|  | #include "fork_pipe_runner.h" | 
|  |  | 
|  | using google::protobuf::ConformanceTestSuite; | 
|  |  | 
|  | namespace google { | 
|  | namespace protobuf { | 
|  | namespace { | 
|  |  | 
|  | void ParseFailureList(const char *filename, | 
|  | conformance::FailureSet *failure_list) { | 
|  | std::ifstream infile(filename); | 
|  |  | 
|  | if (!infile.is_open()) { | 
|  | fprintf(stderr, "Couldn't open failure list file: %s\n", filename); | 
|  | exit(1); | 
|  | } | 
|  |  | 
|  | for (std::string line; std::getline(infile, line);) { | 
|  | // Remove comments. | 
|  | std::string test_name = line.substr(0, line.find('#')); | 
|  |  | 
|  | test_name.erase( | 
|  | std::remove_if(test_name.begin(), test_name.end(), ::isspace), | 
|  | test_name.end()); | 
|  |  | 
|  | if (test_name.empty()) {  // Skip empty lines. | 
|  | continue; | 
|  | } | 
|  |  | 
|  | // If we remove whitespace from the beginning of a line, and what we have | 
|  | // left at first is a '#', then we have a comment. | 
|  | if (test_name[0] != '#') { | 
|  | // Find our failure message if it exists. Will be set to an empty string | 
|  | // if no message is found. Empty failure messages also pass our tests. | 
|  | size_t check_message = line.find('#'); | 
|  | std::string message; | 
|  | if (check_message != std::string::npos) { | 
|  | message = line.substr(check_message + 1);  // +1 to skip the delimiter | 
|  | // If we had only whitespace after the delimiter, we will have an empty | 
|  | // failure message and the test will still pass. | 
|  | message = std::string(absl::StripAsciiWhitespace(message)); | 
|  | } | 
|  | conformance::TestStatus *test = failure_list->add_test(); | 
|  | test->set_name(test_name); | 
|  | test->set_failure_message(message); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void UsageError() { | 
|  | fprintf(stderr, "Usage: conformance-test-runner [options] <test-program>\n"); | 
|  | fprintf(stderr, "\n"); | 
|  | fprintf(stderr, "Options:\n"); | 
|  | fprintf(stderr, | 
|  | "  --failure_list <filename>   Use to specify list of tests\n"); | 
|  | fprintf(stderr, | 
|  | "                              that are expected to fail.  File\n"); | 
|  | fprintf(stderr, | 
|  | "                              should contain one test name per\n"); | 
|  | fprintf(stderr, | 
|  | "                              line.  Use '#' for comments.\n\n"); | 
|  | fprintf(stderr, | 
|  | "  --text_format_failure_list <filename>   Use to specify list \n"); | 
|  | fprintf(stderr, | 
|  | "                              of tests that are expected to \n"); | 
|  | fprintf(stderr, "                              fail in the \n"); | 
|  | fprintf(stderr, | 
|  | "                              text_format_conformance_suite.  \n"); | 
|  | fprintf(stderr, | 
|  | "                              File should contain one test name \n"); | 
|  | fprintf(stderr, | 
|  | "                              per line.  Use '#' for comments.\n\n"); | 
|  |  | 
|  | fprintf(stderr, | 
|  | "  --enforce_recommended       Enforce that recommended test\n"); | 
|  | fprintf(stderr, | 
|  | "                              cases are also passing. Specify\n"); | 
|  | fprintf(stderr, | 
|  | "                              this flag if you want to be\n"); | 
|  | fprintf(stderr, | 
|  | "                              strictly conforming to protobuf\n"); | 
|  | fprintf(stderr, "                              spec.\n\n"); | 
|  | fprintf(stderr, | 
|  | "  --maximum_edition <edition> Only run conformance tests up to\n"); | 
|  | fprintf(stderr, | 
|  | "                              and including the specified\n"); | 
|  | fprintf(stderr, "                              edition.\n\n"); | 
|  | fprintf(stderr, | 
|  | "  --output_dir                <dirname> Directory to write\n" | 
|  | "                              output files.\n\n"); | 
|  | fprintf(stderr, "  --test <test_name>          Only run\n"); | 
|  | fprintf(stderr, | 
|  | "                              the specified test. Multiple tests\n" | 
|  | "                              can be specified by repeating the \n" | 
|  | "                              flag.\n\n"); | 
|  | fprintf(stderr, | 
|  | "  --debug                     Enable debug mode\n" | 
|  | "                              to produce octal serialized\n" | 
|  | "                              ConformanceRequest for the tests\n" | 
|  | "                              passed to --test (required)\n\n"); | 
|  | fprintf(stderr, "  --performance               Boolean option\n"); | 
|  | fprintf(stderr, "                              for enabling run of\n"); | 
|  | fprintf(stderr, "                              performance tests.\n"); | 
|  | exit(1); | 
|  | } | 
|  |  | 
|  | }  // namespace | 
|  |  | 
|  | int RunConformanceTests(int argc, char *argv[], | 
|  | const std::vector<ConformanceTestSuite *> &suites) { | 
|  | if (suites.empty()) { | 
|  | fprintf(stderr, "No test suites found.\n"); | 
|  | return EXIT_FAILURE; | 
|  | } | 
|  |  | 
|  | std::string program; | 
|  | std::string testee; | 
|  | std::vector<std::string> program_args; | 
|  | bool performance = false; | 
|  | bool debug = false; | 
|  | absl::flat_hash_set<std::string> names_to_test; | 
|  | bool enforce_recommended = false; | 
|  | Edition maximum_edition = EDITION_UNKNOWN; | 
|  | std::string output_dir; | 
|  | bool verbose = false; | 
|  | bool isolated = false; | 
|  |  | 
|  | for (int arg = 1; arg < argc; ++arg) { | 
|  | if (strcmp(argv[arg], "--performance") == 0) { | 
|  | performance = true; | 
|  | } else if (strcmp(argv[arg], "--debug") == 0) { | 
|  | debug = true; | 
|  | } else if (strcmp(argv[arg], "--verbose") == 0) { | 
|  | verbose = true; | 
|  | } else if (strcmp(argv[arg], "--enforce_recommended") == 0) { | 
|  | enforce_recommended = true; | 
|  | } else if (strcmp(argv[arg], "--maximum_edition") == 0) { | 
|  | if (++arg == argc) UsageError(); | 
|  | Edition edition = EDITION_UNKNOWN; | 
|  | if (!Edition_Parse(absl::StrCat("EDITION_", argv[arg]), &edition)) { | 
|  | fprintf(stderr, "Unknown edition: %s\n", argv[arg]); | 
|  | UsageError(); | 
|  | } | 
|  | maximum_edition = edition; | 
|  | } else if (strcmp(argv[arg], "--output_dir") == 0) { | 
|  | if (++arg == argc) UsageError(); | 
|  | output_dir = argv[arg]; | 
|  |  | 
|  | } else if (strcmp(argv[arg], "--test") == 0) { | 
|  | if (++arg == argc) UsageError(); | 
|  | names_to_test.insert(argv[arg]); | 
|  |  | 
|  | } else if (argv[arg][0] == '-') { | 
|  | bool recognized_flag = false; | 
|  | for (ConformanceTestSuite *suite : suites) { | 
|  | if (strcmp(argv[arg], suite->GetFailureListFlagName().c_str()) == 0) { | 
|  | if (++arg == argc) UsageError(); | 
|  | recognized_flag = true; | 
|  | } | 
|  | } | 
|  | if (!recognized_flag) { | 
|  | fprintf(stderr, "Unknown option: %s\n", argv[arg]); | 
|  | UsageError(); | 
|  | } | 
|  | } else { | 
|  | program += argv[arg++]; | 
|  | while (arg < argc) { | 
|  | program_args.push_back(argv[arg]); | 
|  | arg++; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | if (debug && names_to_test.empty()) { | 
|  | UsageError(); | 
|  | } | 
|  |  | 
|  | if (!names_to_test.empty()) { | 
|  | isolated = true; | 
|  | } | 
|  |  | 
|  | bool all_ok = true; | 
|  | for (ConformanceTestSuite *suite : suites) { | 
|  | std::string failure_list_filename; | 
|  | conformance::FailureSet failure_list; | 
|  | for (int arg = 1; arg < argc; ++arg) { | 
|  | if (strcmp(argv[arg], suite->GetFailureListFlagName().c_str()) == 0) { | 
|  | if (++arg == argc) UsageError(); | 
|  | failure_list_filename = argv[arg]; | 
|  | ParseFailureList(argv[arg], &failure_list); | 
|  | } | 
|  | } | 
|  | suite->SetPerformance(performance); | 
|  | suite->SetVerbose(verbose); | 
|  | suite->SetEnforceRecommended(enforce_recommended); | 
|  | suite->SetMaximumEdition(maximum_edition); | 
|  | suite->SetOutputDir(output_dir); | 
|  | suite->SetDebug(debug); | 
|  | suite->SetNamesToTest(names_to_test); | 
|  | suite->SetTestee(program); | 
|  | suite->SetIsolated(isolated); | 
|  |  | 
|  | ForkPipeRunner runner(program, program_args); | 
|  |  | 
|  | std::string output; | 
|  | all_ok = all_ok && suite->RunSuite(&runner, &output, failure_list_filename, | 
|  | &failure_list); | 
|  |  | 
|  | names_to_test = suite->GetExpectedTestsNotRun(); | 
|  | fwrite(output.c_str(), 1, output.size(), stderr); | 
|  | } | 
|  |  | 
|  | if (!names_to_test.empty()) { | 
|  | fprintf(stderr, | 
|  | "These tests were requested to be ran isolated, but they do " | 
|  | "not exist. Revise the test names:\n\n"); | 
|  | for (const std::string &test_name : names_to_test) { | 
|  | fprintf(stderr, "  %s\n", test_name.c_str()); | 
|  | } | 
|  | fprintf(stderr, "\n\n"); | 
|  | } | 
|  | return all_ok ? EXIT_SUCCESS : EXIT_FAILURE; | 
|  | } | 
|  |  | 
|  | }  // namespace protobuf | 
|  | }  // namespace google |