/* * Copyright 2011 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "CrashHandler.h" #include "GrContext.h" #include "GrContextFactory.h" #include "OverwriteLine.h" #include "PathOpsDebug.h" #include "Resources.h" #include "SkCommonFlags.h" #include "SkGraphics.h" #include "SkOSFile.h" #include "SkPathOpsDebug.h" #include "SkTArray.h" #include "SkTaskGroup.h" #include "SkTemplates.h" #include "SkTime.h" #include "Test.h" #include <atomic> using namespace skiatest; using namespace sk_gpu_test; DEFINE_bool2(dumpOp, d, false, "dump the pathOps to a file to recover mid-crash."); DEFINE_bool2(extendedTest, x, false, "run extended tests for pathOps."); DEFINE_bool2(runFail, f, false, "check for success on tests known to fail."); DEFINE_bool2(verifyOp, y, false, "compare the pathOps result against a region."); DEFINE_string2(json, J, "", "write json version of tests."); #if DEBUG_COIN DEFINE_bool2(coinTest, c, false, "detect unused coincidence algorithms."); #endif // need to explicitly declare this, or we get some weird infinite loop llist template TestRegistry* TestRegistry::gHead; void (*gVerboseFinalize)() = nullptr; // The threads report back to this object when they are done. class Status { public: explicit Status(int total) : fDone(0), fTestCount(0), fFailCount(0), fTotal(total) {} // Threadsafe. void endTest(const char* testName, bool success, SkMSec elapsed, int testCount) { const int done = ++fDone; fTestCount += testCount; if (!success) { SkDebugf("\n---- %s FAILED", testName); } SkString prefix(kSkOverwriteLine); SkString time; if (FLAGS_verbose) { prefix.printf("\n"); time.printf("%5dms ", elapsed); } SkDebugf("%s[%3d/%3d] %s%s", prefix.c_str(), done, fTotal, time.c_str(), testName); } void reportFailure() { fFailCount++; } int32_t testCount() { return fTestCount; } int32_t failCount() { return fFailCount; } private: std::atomic<int32_t> fDone; std::atomic<int32_t> fTestCount; std::atomic<int32_t> fFailCount; const int fTotal; }; class SkTestRunnable { public: SkTestRunnable(const Test& test, Status* status) : fTest(test), fStatus(status) {} void operator()() { struct TestReporter : public skiatest::Reporter { public: TestReporter() : fStats(nullptr), fError(false), fTestCount(0) {} void bumpTestCount() override { ++fTestCount; } bool allowExtendedTest() const override { return FLAGS_extendedTest; } bool verbose() const override { return FLAGS_veryVerbose; } void reportFailed(const skiatest::Failure& failure) override { SkDebugf("\nFAILED: %s", failure.toString().c_str()); fError = true; } void* stats() const override { return fStats; } void* fStats; bool fError; int fTestCount; } reporter; const Timer timer; fTest.proc(&reporter, GrContextOptions()); SkMSec elapsed = timer.elapsedMsInt(); if (reporter.fError) { fStatus->reportFailure(); } fStatus->endTest(fTest.name, !reporter.fError, elapsed, reporter.fTestCount); } private: Test fTest; Status* fStatus; }; static bool should_run(const char* testName, bool isGPUTest) { if (SkCommandLineFlags::ShouldSkip(FLAGS_match, testName)) { return false; } if (!FLAGS_cpu && !isGPUTest) { return false; } if (!FLAGS_gpu && isGPUTest) { return false; } return true; } int main(int argc, char** argv) { SkCommandLineFlags::Parse(argc, argv); #if DEBUG_DUMP_VERIFY SkPathOpsDebug::gDumpOp = FLAGS_dumpOp; SkPathOpsDebug::gVerifyOp = FLAGS_verifyOp; #endif SkPathOpsDebug::gRunFail = FLAGS_runFail; SkPathOpsDebug::gVeryVerbose = FLAGS_veryVerbose; PathOpsDebug::gOutFirst = true; PathOpsDebug::gCheckForDuplicateNames = false; PathOpsDebug::gOutputSVG = false; if ((PathOpsDebug::gJson = !FLAGS_json.isEmpty())) { PathOpsDebug::gOut = fopen(FLAGS_json[0], "wb"); fprintf(PathOpsDebug::gOut, "{\n"); FLAGS_threads = 0; PathOpsDebug::gMarkJsonFlaky = false; } SetupCrashHandler(); SkAutoGraphics ag; { SkString header("Skia UnitTests:"); if (!FLAGS_match.isEmpty()) { header.appendf(" --match"); for (int index = 0; index < FLAGS_match.count(); ++index) { header.appendf(" %s", FLAGS_match[index]); } } SkString tmpDir = skiatest::GetTmpDir(); if (!tmpDir.isEmpty()) { header.appendf(" --tmpDir %s", tmpDir.c_str()); } SkString resourcePath = GetResourcePath(); if (!resourcePath.isEmpty()) { header.appendf(" --resourcePath %s", resourcePath.c_str()); } #if DEBUG_COIN if (FLAGS_coinTest) { header.appendf(" -c"); } #endif if (FLAGS_dumpOp) { header.appendf(" -d"); } #ifdef SK_DEBUG if (FLAGS_runFail) { header.appendf(" -f"); } #endif if (FLAGS_verbose) { header.appendf(" -v"); } if (FLAGS_veryVerbose) { header.appendf(" -V"); } if (FLAGS_extendedTest) { header.appendf(" -x"); } if (FLAGS_verifyOp) { header.appendf(" -y"); } #ifdef SK_DEBUG header.append(" SK_DEBUG"); #else header.append(" SK_RELEASE"); #endif if (FLAGS_veryVerbose) { header.appendf("\n"); } SkDebugf("%s", header.c_str()); } // Count tests first. int total = 0; int toRun = 0; for (const Test& test : TestRegistry::Range()) { if (should_run(test.name, test.needsGpu)) { toRun++; } total++; } // Now run them. int skipCount = 0; SkTaskGroup::Enabler enabled(FLAGS_threads); SkTaskGroup cpuTests; SkTArray<const Test*> gpuTests; Status status(toRun); for (const Test& test : TestRegistry::Range()) { if (!should_run(test.name, test.needsGpu)) { ++skipCount; } else if (test.needsGpu) { gpuTests.push_back(&test); } else { cpuTests.add(SkTestRunnable(test, &status)); } } // Run GPU tests on this thread. for (int i = 0; i < gpuTests.count(); i++) { SkTestRunnable(*gpuTests[i], &status)(); } // Block until threaded tests finish. cpuTests.wait(); if (FLAGS_verbose) { SkDebugf( "\nFinished %d tests, %d failures, %d skipped. " "(%d internal tests)", toRun, status.failCount(), skipCount, status.testCount()); if (gVerboseFinalize) { (*gVerboseFinalize)(); } } SkDebugf("\n"); #if DEBUG_COIN if (FLAGS_coinTest) { SkPathOpsDebug::DumpCoinDict(); } #endif if (PathOpsDebug::gJson) { fprintf(PathOpsDebug::gOut, "\n}\n"); fclose(PathOpsDebug::gOut); } return (status.failCount() == 0) ? 0 : 1; }