blob: b1a1f9fefd8e2eea0e193f5b04449cc872a77013 [file] [log] [blame]
// Copyright 2013 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "base/test/launcher/unit_test_launcher.h"
#include <map>
#include <memory>
#include <utility>
#include "base/base_paths.h"
#include "base/base_switches.h"
#include "base/command_line.h"
#include "base/compiler_specific.h"
#include "base/debug/debugger.h"
#include "base/files/file_util.h"
#include "base/format_macros.h"
#include "base/functional/bind.h"
#include "base/functional/callback_helpers.h"
#include "base/i18n/icu_util.h"
#include "base/location.h"
#include "base/logging.h"
#include "base/message_loop/message_pump_type.h"
#include "base/sequence_checker.h"
#include "base/strings/strcat.h"
#include "base/strings/string_number_conversions.h"
#include "base/strings/string_util.h"
#include "base/system/sys_info.h"
#include "base/task/single_thread_task_executor.h"
#include "base/task/single_thread_task_runner.h"
#include "base/test/allow_check_is_test_for_testing.h"
#include "base/test/launcher/test_launcher.h"
#include "base/test/scoped_block_tests_writing_to_special_dirs.h"
#include "base/test/test_switches.h"
#include "base/test/test_timeouts.h"
#include "base/threading/thread_checker.h"
#include "build/build_config.h"
#include "testing/gtest/include/gtest/gtest.h"
#if BUILDFLAG(IS_POSIX)
#include "base/files/file_descriptor_watcher_posix.h"
#endif
#if BUILDFLAG(IS_IOS)
#include "base/test/test_support_ios.h"
#endif
namespace base {
namespace {
// This constant controls how many tests are run in a single batch by default.
const size_t kDefaultTestBatchLimit =
#if BUILDFLAG(IS_IOS)
100;
#else
10;
#endif
#if !BUILDFLAG(IS_ANDROID)
void PrintUsage() {
fprintf(
stdout,
"Runs tests using the gtest framework, each batch of tests being\n"
"run in their own process. Supported command-line flags:\n"
"\n"
" Common flags:\n"
" --gtest_filter=...\n"
" Runs a subset of tests (see --gtest_help for more info).\n"
"\n"
" --help\n"
" Shows this message.\n"
"\n"
" --gtest_help\n"
" Shows the gtest help message.\n"
"\n"
" --test-launcher-jobs=N\n"
" Sets the number of parallel test jobs to N.\n"
"\n"
" --single-process-tests\n"
" Runs the tests and the launcher in the same process. Useful\n"
" for debugging a specific test in a debugger.\n"
"\n"
" Other flags:\n"
" --test-launcher-filter-file=PATH\n"
" Like --gtest_filter, but read the test filter from PATH.\n"
" Supports multiple filter paths separated by ';'.\n"
" One pattern per line; lines starting with '-' are exclusions.\n"
" See also //testing/buildbot/filters/README.md file.\n"
"\n"
" --test-launcher-batch-limit=N\n"
" Sets the limit of test batch to run in a single process to N.\n"
"\n"
" --test-launcher-debug-launcher\n"
" Disables autodetection of debuggers and similar tools,\n"
" making it possible to use them to debug launcher itself.\n"
"\n"
" --test-launcher-retry-limit=N\n"
" Sets the limit of test retries on failures to N.\n"
" --gtest_repeat=N\n"
" Forces the launcher to run every test N times. -1 is a special"
" value, causing the infinite amount of iterations."
" Repeated tests are run in parallel, unless the number of"
" iterations is infinite or --gtest_break_on_failure is specified"
" (see below)."
" Consider using --test_launcher-jobs flag to speed up the"
" parallel execution."
"\n"
" --gtest_break_on_failure\n"
" Stop running repeated tests as soon as one repeat of the test fails."
" This flag forces sequential repeats and prevents parallelised"
" execution."
"\n"
" --test-launcher-summary-output=PATH\n"
" Saves a JSON machine-readable summary of the run.\n"
"\n"
" --test-launcher-print-test-stdio=auto|always|never\n"
" Controls when full test output is printed.\n"
" auto means to print it when the test failed.\n"
"\n"
" --test-launcher-test-part-results-limit=N\n"
" Sets the limit of failed EXPECT/ASSERT entries in the xml and\n"
" JSON outputs per test to N (default N=10). Negative value \n"
" will disable this limit.\n"
"\n"
" --test-launcher-total-shards=N\n"
" Sets the total number of shards to N.\n"
"\n"
" --test-launcher-shard-index=N\n"
" Sets the shard index to run to N (from 0 to TOTAL - 1).\n"
"\n"
" --test-launcher-print-temp-leaks\n"
" Prints information about leaked files and/or directories in\n"
" child process's temporary directories (Windows and macOS).\n");
fflush(stdout);
}
bool GetSwitchValueAsInt(const std::string& switch_name, int* result) {
if (!CommandLine::ForCurrentProcess()->HasSwitch(switch_name))
return true;
std::string switch_value =
CommandLine::ForCurrentProcess()->GetSwitchValueASCII(switch_name);
if (!StringToInt(switch_value, result) || *result < 0) {
LOG(ERROR) << "Invalid value for " << switch_name << ": " << switch_value;
return false;
}
return true;
}
int RunTestSuite(RunTestSuiteCallback run_test_suite,
size_t parallel_jobs,
int default_batch_limit,
size_t retry_limit,
bool use_job_objects,
RepeatingClosure timeout_callback,
OnceClosure gtest_init) {
bool force_single_process = false;
if (CommandLine::ForCurrentProcess()->HasSwitch(
switches::kTestLauncherDebugLauncher)) {
fprintf(stdout, "Forcing test launcher debugging mode.\n");
fflush(stdout);
} else {
if (base::debug::BeingDebugged()) {
fprintf(stdout,
"Debugger detected, switching to single process mode.\n"
"Pass --test-launcher-debug-launcher to debug the launcher "
"itself.\n");
fflush(stdout);
force_single_process = true;
}
}
if (CommandLine::ForCurrentProcess()->HasSwitch(kGTestHelpFlag) ||
CommandLine::ForCurrentProcess()->HasSwitch(kGTestListTestsFlag) ||
CommandLine::ForCurrentProcess()->HasSwitch(
switches::kSingleProcessTests) ||
CommandLine::ForCurrentProcess()->HasSwitch(
switches::kTestChildProcess) ||
force_single_process) {
return std::move(run_test_suite).Run();
}
// ICU must be initialized before any attempts to format times, e.g. for logs.
if (!base::i18n::InitializeICU()) {
return false;
}
if (CommandLine::ForCurrentProcess()->HasSwitch(switches::kHelpFlag)) {
PrintUsage();
return 0;
}
TimeTicks start_time(TimeTicks::Now());
std::move(gtest_init).Run();
TestTimeouts::Initialize();
int batch_limit = default_batch_limit;
if (!GetSwitchValueAsInt(switches::kTestLauncherBatchLimit, &batch_limit))
return 1;
fprintf(stdout,
"IMPORTANT DEBUGGING NOTE: batches of tests are run inside their\n"
"own process. For debugging a test inside a debugger, use the\n"
"--gtest_filter=<your_test_name> flag along with\n"
"--single-process-tests.\n");
fflush(stdout);
base::SingleThreadTaskExecutor executor(base::MessagePumpType::IO);
#if BUILDFLAG(IS_POSIX)
FileDescriptorWatcher file_descriptor_watcher(executor.task_runner());
#endif
DefaultUnitTestPlatformDelegate platform_delegate;
UnitTestLauncherDelegate delegate(&platform_delegate, batch_limit,
use_job_objects, timeout_callback);
TestLauncher launcher(&delegate, parallel_jobs, retry_limit);
bool success = launcher.Run();
fprintf(stdout, "Tests took %" PRId64 " seconds.\n",
(TimeTicks::Now() - start_time).InSeconds());
fflush(stdout);
return (success ? 0 : 1);
}
#endif
int LaunchUnitTestsInternal(RunTestSuiteCallback run_test_suite,
size_t parallel_jobs,
int default_batch_limit,
size_t retry_limit,
bool use_job_objects,
RepeatingClosure timeout_callback,
OnceClosure gtest_init) {
base::test::AllowCheckIsTestForTesting();
#if BUILDFLAG(IS_ANDROID)
// We can't easily fork on Android, just run the test suite directly.
return std::move(run_test_suite).Run();
#elif BUILDFLAG(IS_IOS)
InitIOSRunHook(base::BindOnce(&RunTestSuite, std::move(run_test_suite),
parallel_jobs, default_batch_limit, retry_limit,
use_job_objects, timeout_callback,
std::move(gtest_init)));
return RunTestsFromIOSApp();
#else
ScopedBlockTestsWritingToSpecialDirs scoped_blocker(
{
// Please keep these in alphabetic order within each platform type.
base::DIR_SRC_TEST_DATA_ROOT, base::DIR_USER_DESKTOP,
#if BUILDFLAG(IS_WIN)
base::DIR_COMMON_DESKTOP, base::DIR_START_MENU,
base::DIR_USER_STARTUP,
#endif // BUILDFLAG(IS_WIN)
},
([](const base::FilePath& path) {
ADD_FAILURE()
<< "Attempting to write file in dir " << path
<< " Use ScopedPathOverride or other mechanism to not write to this"
" directory.";
}));
return RunTestSuite(std::move(run_test_suite), parallel_jobs,
default_batch_limit, retry_limit, use_job_objects,
timeout_callback, std::move(gtest_init));
#endif
}
void InitGoogleTestChar(int* argc, char** argv) {
testing::InitGoogleTest(argc, argv);
}
#if BUILDFLAG(IS_WIN)
void InitGoogleTestWChar(int* argc, wchar_t** argv) {
testing::InitGoogleTest(argc, argv);
}
#endif // BUILDFLAG(IS_WIN)
} // namespace
MergeTestFilterSwitchHandler::~MergeTestFilterSwitchHandler() = default;
void MergeTestFilterSwitchHandler::ResolveDuplicate(
base::StringPiece key,
CommandLine::StringPieceType new_value,
CommandLine::StringType& out_value) {
if (key != switches::kTestLauncherFilterFile) {
out_value = CommandLine::StringType(new_value);
return;
}
if (!out_value.empty()) {
#if BUILDFLAG(IS_WIN)
StrAppend(&out_value, {L";"});
#else
StrAppend(&out_value, {";"});
#endif
}
StrAppend(&out_value, {new_value});
}
int LaunchUnitTests(int argc,
char** argv,
RunTestSuiteCallback run_test_suite,
size_t retry_limit) {
CommandLine::SetDuplicateSwitchHandler(
std::make_unique<MergeTestFilterSwitchHandler>());
CommandLine::Init(argc, argv);
size_t parallel_jobs = NumParallelJobs(/*cores_per_job=*/1);
if (parallel_jobs == 0U) {
return 1;
}
return LaunchUnitTestsInternal(std::move(run_test_suite), parallel_jobs,
kDefaultTestBatchLimit, retry_limit, true,
DoNothing(),
BindOnce(&InitGoogleTestChar, &argc, argv));
}
int LaunchUnitTestsSerially(int argc,
char** argv,
RunTestSuiteCallback run_test_suite) {
CommandLine::Init(argc, argv);
return LaunchUnitTestsInternal(std::move(run_test_suite), 1U,
kDefaultTestBatchLimit, 1U, true, DoNothing(),
BindOnce(&InitGoogleTestChar, &argc, argv));
}
int LaunchUnitTestsWithOptions(int argc,
char** argv,
size_t parallel_jobs,
int default_batch_limit,
bool use_job_objects,
RepeatingClosure timeout_callback,
RunTestSuiteCallback run_test_suite) {
CommandLine::Init(argc, argv);
return LaunchUnitTestsInternal(std::move(run_test_suite), parallel_jobs,
default_batch_limit, 1U, use_job_objects,
timeout_callback,
BindOnce(&InitGoogleTestChar, &argc, argv));
}
#if BUILDFLAG(IS_WIN)
int LaunchUnitTests(int argc,
wchar_t** argv,
bool use_job_objects,
RunTestSuiteCallback run_test_suite) {
// Windows CommandLine::Init ignores argv anyway.
CommandLine::Init(argc, NULL);
size_t parallel_jobs = NumParallelJobs(/*cores_per_job=*/1);
if (parallel_jobs == 0U) {
return 1;
}
return LaunchUnitTestsInternal(std::move(run_test_suite), parallel_jobs,
kDefaultTestBatchLimit, 1U, use_job_objects,
DoNothing(),
BindOnce(&InitGoogleTestWChar, &argc, argv));
}
#endif // BUILDFLAG(IS_WIN)
DefaultUnitTestPlatformDelegate::DefaultUnitTestPlatformDelegate() = default;
bool DefaultUnitTestPlatformDelegate::GetTests(
std::vector<TestIdentifier>* output) {
*output = GetCompiledInTests();
return true;
}
bool DefaultUnitTestPlatformDelegate::CreateResultsFile(
const base::FilePath& temp_dir,
base::FilePath* path) {
if (!CreateTemporaryDirInDir(temp_dir, FilePath::StringType(), path))
return false;
*path = path->AppendASCII("test_results.xml");
return true;
}
bool DefaultUnitTestPlatformDelegate::CreateTemporaryFile(
const base::FilePath& temp_dir,
base::FilePath* path) {
if (temp_dir.empty())
return false;
return CreateTemporaryFileInDir(temp_dir, path);
}
CommandLine DefaultUnitTestPlatformDelegate::GetCommandLineForChildGTestProcess(
const std::vector<std::string>& test_names,
const base::FilePath& output_file,
const base::FilePath& flag_file) {
CommandLine new_cmd_line(*CommandLine::ForCurrentProcess());
CHECK(base::PathExists(flag_file));
std::string long_flags(
StrCat({"--", kGTestFilterFlag, "=", JoinString(test_names, ":")}));
CHECK(WriteFile(flag_file, long_flags));
new_cmd_line.AppendSwitchPath(switches::kTestLauncherOutput, output_file);
new_cmd_line.AppendSwitchPath(kGTestFlagfileFlag, flag_file);
new_cmd_line.AppendSwitch(switches::kSingleProcessTests);
return new_cmd_line;
}
std::string DefaultUnitTestPlatformDelegate::GetWrapperForChildGTestProcess() {
return std::string();
}
UnitTestLauncherDelegate::UnitTestLauncherDelegate(
UnitTestPlatformDelegate* platform_delegate,
size_t batch_limit,
bool use_job_objects,
RepeatingClosure timeout_callback)
: platform_delegate_(platform_delegate),
batch_limit_(batch_limit),
use_job_objects_(use_job_objects),
timeout_callback_(timeout_callback) {}
UnitTestLauncherDelegate::~UnitTestLauncherDelegate() {
DCHECK(thread_checker_.CalledOnValidThread());
}
bool UnitTestLauncherDelegate::GetTests(std::vector<TestIdentifier>* output) {
DCHECK(thread_checker_.CalledOnValidThread());
return platform_delegate_->GetTests(output);
}
CommandLine UnitTestLauncherDelegate::GetCommandLine(
const std::vector<std::string>& test_names,
const FilePath& temp_dir,
FilePath* output_file) {
CHECK(!test_names.empty());
// Create a dedicated temporary directory to store the xml result data
// per run to ensure clean state and make it possible to launch multiple
// processes in parallel.
CHECK(platform_delegate_->CreateResultsFile(temp_dir, output_file));
FilePath flag_file;
platform_delegate_->CreateTemporaryFile(temp_dir, &flag_file);
return CommandLine(platform_delegate_->GetCommandLineForChildGTestProcess(
test_names, *output_file, flag_file));
}
std::string UnitTestLauncherDelegate::GetWrapper() {
return platform_delegate_->GetWrapperForChildGTestProcess();
}
int UnitTestLauncherDelegate::GetLaunchOptions() {
return use_job_objects_ ? TestLauncher::USE_JOB_OBJECTS : 0;
}
TimeDelta UnitTestLauncherDelegate::GetTimeout() {
return TestTimeouts::test_launcher_timeout();
}
size_t UnitTestLauncherDelegate::GetBatchSize() {
return batch_limit_;
}
void UnitTestLauncherDelegate::OnTestTimedOut(const CommandLine& cmd_line) {
timeout_callback_.Run();
}
} // namespace base