#include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include using namespace nix; static Path gcRootsDir; struct MyArgs : MixEvalArgs, MixCommonArgs { Path releaseExpr; bool flake = false; bool dryRun = false; size_t nrWorkers = 1; size_t maxMemorySize = 4096; MyArgs() : MixCommonArgs("hydra-eval-jobs") { addFlag({ .longName = "help", .description = "show usage information", .handler = {[&]() { printf("USAGE: hydra-eval-jobs [options] expr\n\n"); for (const auto & [name, flag] : longFlags) { if (hiddenCategories.count(flag->category)) { continue; } printf(" --%-20s %s\n", name.c_str(), flag->description.c_str()); } }}, }); addFlag({ .longName = "gc-roots-dir", .description = "garbage collector roots directory", .labels = {"path"}, .handler = {&gcRootsDir} }); addFlag({ .longName = "workers", .description = "number of evaluate workers", .labels = {"workers"}, .handler = {[=](std::string s) { nrWorkers = std::stoi(s); }} }); addFlag({ .longName = "max-memory-size", .description = "maximum evaluation memory size", .labels = {"size"}, .handler = {[=](std::string s) { maxMemorySize = std::stoi(s); }} }); addFlag({ .longName = "dry-run", .description = "don't create store derivations", .handler = {&dryRun, true} }); addFlag({ .longName = "flake", .description = "build a flake", .handler = {&flake, true} }); expectArg("expr", &releaseExpr); } }; static MyArgs myArgs; static std::string queryMetaStrings(EvalState & state, DrvInfo & drv, const string & name, const string & subAttribute) { Strings res; std::function rec; rec = [&](Value & v) { state.forceValue(v); if (v.type() == nString) res.push_back(v.string.s); else if (v.isList()) for (unsigned int n = 0; n < v.listSize(); ++n) rec(*v.listElems()[n]); else if (v.type() == nAttrs) { auto a = v.attrs->find(state.symbols.create(subAttribute)); if (a != v.attrs->end()) res.push_back(state.forceString(*a->value)); } }; Value * v = drv.queryMeta(name); if (v) rec(*v); return concatStringsSep(", ", res); } static void worker( EvalState & state, Bindings & autoArgs, AutoCloseFD & to, AutoCloseFD & from) { Value vTop; if (myArgs.flake) { using namespace flake; auto flakeRef = parseFlakeRef(myArgs.releaseExpr); auto vFlake = state.allocValue(); auto lockedFlake = lockFlake(state, flakeRef, LockFlags { .updateLockFile = false, .useRegistries = false, .allowMutable = false, }); callFlake(state, lockedFlake, *vFlake); auto vOutputs = vFlake->attrs->get(state.symbols.create("outputs"))->value; state.forceValue(*vOutputs); auto aHydraJobs = vOutputs->attrs->get(state.symbols.create("hydraJobs")); if (!aHydraJobs) aHydraJobs = vOutputs->attrs->get(state.symbols.create("checks")); if (!aHydraJobs) throw Error("flake '%s' does not provide any Hydra jobs or checks", flakeRef); vTop = *aHydraJobs->value; } else { state.evalFile(lookupFileArg(state, myArgs.releaseExpr), vTop); } auto vRoot = state.allocValue(); state.autoCallFunction(autoArgs, vTop, *vRoot); while (true) { /* Wait for the master to send us a job name. */ writeLine(to.get(), "next"); auto s = readLine(from.get()); if (s == "exit") break; if (!hasPrefix(s, "do ")) abort(); std::string attrPath(s, 3); debug("worker process %d at '%s'", getpid(), attrPath); /* Evaluate it and send info back to the master. */ nlohmann::json reply; try { auto vTmp = findAlongAttrPath(state, attrPath, autoArgs, *vRoot).first; auto v = state.allocValue(); state.autoCallFunction(autoArgs, *vTmp, *v); if (auto drv = getDerivation(state, *v, false)) { DrvInfo::Outputs outputs = drv->queryOutputs(); if (drv->querySystem() == "unknown") throw EvalError("derivation must have a 'system' attribute"); auto drvPath = drv->queryDrvPath(); nlohmann::json job; job["nixName"] = drv->queryName(); job["system"] =drv->querySystem(); job["drvPath"] = drvPath; job["description"] = drv->queryMetaString("description"); job["license"] = queryMetaStrings(state, *drv, "license", "shortName"); job["homepage"] = drv->queryMetaString("homepage"); job["maintainers"] = queryMetaStrings(state, *drv, "maintainers", "email"); job["schedulingPriority"] = drv->queryMetaInt("schedulingPriority", 100); job["timeout"] = drv->queryMetaInt("timeout", 36000); job["maxSilent"] = drv->queryMetaInt("maxSilent", 7200); job["isChannel"] = drv->queryMetaBool("isHydraChannel", false); /* If this is an aggregate, then get its constituents. */ auto a = v->attrs->get(state.symbols.create("_hydraAggregate")); if (a && state.forceBool(*a->value, *a->pos)) { auto a = v->attrs->get(state.symbols.create("constituents")); if (!a) throw EvalError("derivation must have a ‘constituents’ attribute"); PathSet context; state.coerceToString(*a->pos, *a->value, context, true, false); for (auto & i : context) if (i.at(0) == '!') { size_t index = i.find("!", 1); job["constituents"].push_back(string(i, index + 1)); } state.forceList(*a->value, *a->pos); for (unsigned int n = 0; n < a->value->listSize(); ++n) { auto v = a->value->listElems()[n]; state.forceValue(*v); if (v->type() == nString) job["namedConstituents"].push_back(state.forceStringNoCtx(*v)); } } /* Register the derivation as a GC root. !!! This registers roots for jobs that we may have already done. */ auto localStore = state.store.dynamic_pointer_cast(); if (gcRootsDir != "" && localStore) { Path root = gcRootsDir + "/" + std::string(baseNameOf(drvPath)); if (!pathExists(root)) localStore->addPermRoot(localStore->parseStorePath(drvPath), root); } nlohmann::json out; for (auto & j : outputs) out[j.first] = j.second; job["outputs"] = std::move(out); reply["job"] = std::move(job); } else if (v->type() == nAttrs) { auto attrs = nlohmann::json::array(); StringSet ss; for (auto & i : v->attrs->lexicographicOrder()) { std::string name(i->name); if (name.find('.') != std::string::npos || name.find(' ') != std::string::npos) { printError("skipping job with illegal name '%s'", name); continue; } attrs.push_back(name); } reply["attrs"] = std::move(attrs); } else if (v->type() == nNull) ; else throw TypeError("attribute '%s' is %s, which is not supported", attrPath, showType(*v)); } catch (EvalError & e) { auto msg = e.msg(); // Transmits the error we got from the previous evaluation // in the JSON output. reply["error"] = filterANSIEscapes(msg, true); // Don't forget to print it into the STDERR log, this is // what's shown in the Hydra UI. printError(msg); } writeLine(to.get(), reply.dump()); /* If our RSS exceeds the maximum, exit. The master will start a new process. */ struct rusage r; getrusage(RUSAGE_SELF, &r); if ((size_t) r.ru_maxrss > myArgs.maxMemorySize * 1024) break; } writeLine(to.get(), "restart"); } int main(int argc, char * * argv) { /* Prevent undeclared dependencies in the evaluation via $NIX_PATH. */ unsetenv("NIX_PATH"); return handleExceptions(argv[0], [&]() { initNix(); initGC(); myArgs.parseCmdline(argvToStrings(argc, argv)); /* FIXME: The build hook in conjunction with import-from-derivation is causing "unexpected EOF" during eval */ settings.builders = ""; /* Prevent access to paths outside of the Nix search path and to the environment. */ evalSettings.restrictEval = true; /* When building a flake, use pure evaluation (no access to 'getEnv', 'currentSystem' etc. */ evalSettings.pureEval = myArgs.flake; if (myArgs.dryRun) settings.readOnlyMode = true; if (myArgs.releaseExpr == "") throw UsageError("no expression specified"); if (gcRootsDir == "") printMsg(lvlError, "warning: `--gc-roots-dir' not specified"); struct State { std::set todo{""}; std::set active; nlohmann::json jobs; std::exception_ptr exc; }; std::condition_variable wakeup; Sync state_; /* Start a handler thread per worker process. */ auto handler = [&]() { try { pid_t pid = -1; AutoCloseFD from, to; while (true) { /* Start a new worker process if necessary. */ if (pid == -1) { Pipe toPipe, fromPipe; toPipe.create(); fromPipe.create(); pid = startProcess( [&, to{std::make_shared(std::move(fromPipe.writeSide))}, from{std::make_shared(std::move(toPipe.readSide))} ]() { try { EvalState state(myArgs.searchPath, openStore()); Bindings & autoArgs = *myArgs.getAutoArgs(state); worker(state, autoArgs, *to, *from); } catch (Error & e) { nlohmann::json err; auto msg = e.msg(); err["error"] = filterANSIEscapes(msg, true); printError(msg); writeLine(to->get(), err.dump()); // Don't forget to print it into the STDERR log, this is // what's shown in the Hydra UI. writeLine(to->get(), "restart"); } }, ProcessOptions { .allowVfork = false }); from = std::move(fromPipe.readSide); to = std::move(toPipe.writeSide); debug("created worker process %d", pid); } /* Check whether the existing worker process is still there. */ auto s = readLine(from.get()); if (s == "restart") { pid = -1; continue; } else if (s != "next") { auto json = nlohmann::json::parse(s); throw Error("worker error: %s", (std::string) json["error"]); } /* Wait for a job name to become available. */ std::string attrPath; while (true) { checkInterrupt(); auto state(state_.lock()); if ((state->todo.empty() && state->active.empty()) || state->exc) { writeLine(to.get(), "exit"); return; } if (!state->todo.empty()) { attrPath = *state->todo.begin(); state->todo.erase(state->todo.begin()); state->active.insert(attrPath); break; } else state.wait(wakeup); } /* Tell the worker to evaluate it. */ writeLine(to.get(), "do " + attrPath); /* Wait for the response. */ auto response = nlohmann::json::parse(readLine(from.get())); /* Handle the response. */ StringSet newAttrs; if (response.find("job") != response.end()) { auto state(state_.lock()); state->jobs[attrPath] = response["job"]; } if (response.find("attrs") != response.end()) { for (auto & i : response["attrs"]) { auto s = (attrPath.empty() ? "" : attrPath + ".") + (std::string) i; newAttrs.insert(s); } } if (response.find("error") != response.end()) { auto state(state_.lock()); state->jobs[attrPath]["error"] = response["error"]; } /* Add newly discovered job names to the queue. */ { auto state(state_.lock()); state->active.erase(attrPath); for (auto & s : newAttrs) state->todo.insert(s); wakeup.notify_all(); } } } catch (...) { auto state(state_.lock()); state->exc = std::current_exception(); wakeup.notify_all(); } }; std::vector threads; for (size_t i = 0; i < myArgs.nrWorkers; i++) threads.emplace_back(std::thread(handler)); for (auto & thread : threads) thread.join(); auto state(state_.lock()); if (state->exc) std::rethrow_exception(state->exc); /* For aggregate jobs that have named consistuents (i.e. constituents that are a job name rather than a derivation), look up the referenced job and add it to the dependencies of the aggregate derivation. */ auto store = openStore(); for (auto i = state->jobs.begin(); i != state->jobs.end(); ++i) { auto jobName = i.key(); auto & job = i.value(); auto named = job.find("namedConstituents"); if (named == job.end()) continue; if (myArgs.dryRun) { for (std::string jobName2 : *named) { auto job2 = state->jobs.find(jobName2); if (job2 == state->jobs.end()) throw Error("aggregate job '%s' references non-existent job '%s'", jobName, jobName2); std::string drvPath2 = (*job2)["drvPath"]; job["constituents"].push_back(drvPath2); } } else { auto drvPath = store->parseStorePath((std::string) job["drvPath"]); auto drv = store->readDerivation(drvPath); for (std::string jobName2 : *named) { auto job2 = state->jobs.find(jobName2); if (job2 == state->jobs.end()) throw Error("aggregate job '%s' references non-existent job '%s'", jobName, jobName2); auto drvPath2 = store->parseStorePath((std::string) (*job2)["drvPath"]); auto drv2 = store->readDerivation(drvPath2); job["constituents"].push_back(store->printStorePath(drvPath2)); drv.inputDrvs[drvPath2] = {drv2.outputs.begin()->first}; } std::string drvName(drvPath.name()); assert(hasSuffix(drvName, drvExtension)); drvName.resize(drvName.size() - drvExtension.size()); auto h = std::get(hashDerivationModulo(*store, drv, true)); auto outPath = store->makeOutputPath("out", h, drvName); drv.env["out"] = store->printStorePath(outPath); drv.outputs.insert_or_assign("out", DerivationOutput { .output = DerivationOutputInputAddressed { .path = outPath } }); auto newDrvPath = store->printStorePath(writeDerivation(*store, drv)); debug("rewrote aggregate derivation %s -> %s", store->printStorePath(drvPath), newDrvPath); job["drvPath"] = newDrvPath; job["outputs"]["out"] = store->printStorePath(outPath); } job.erase("namedConstituents"); } std::cout << state->jobs.dump(2) << "\n"; }); }