Compare commits
19 Commits
nix-2.32
...
use-store-
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
02a65eb8b9 | ||
|
|
1381ee85d2 | ||
|
|
b767c82b6e | ||
|
|
19a1c5ff04 | ||
|
|
cde792e718 | ||
|
|
9b0d74ed84 | ||
|
|
a94c1aeac4 | ||
|
|
9df591a8dd | ||
|
|
8be9f4c938 | ||
|
|
181b52787e | ||
|
|
fdd70363d7 | ||
|
|
42cc55abf0 | ||
|
|
5c7e5b6465 | ||
|
|
89c504e2d9 | ||
|
|
2c3072aaee | ||
|
|
71796e7c8b | ||
|
|
022160809b | ||
|
|
262a6027e1 | ||
|
|
3246bb6807 |
@@ -85,6 +85,7 @@
|
|||||||
nix-util
|
nix-util
|
||||||
nix-store
|
nix-store
|
||||||
nix-main
|
nix-main
|
||||||
|
nix-cmd
|
||||||
nix-cli
|
nix-cli
|
||||||
;
|
;
|
||||||
nix-perl-bindings = nix.hydraJobs.perlBindings.${system};
|
nix-perl-bindings = nix.hydraJobs.perlBindings.${system};
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
|
|
||||||
mdbook serve \
|
exec mdbook serve \
|
||||||
--port 63332 \
|
--port 63332 \
|
||||||
--dest-dir ./.hydra-data/manual \
|
--dest-dir ./.hydra-data/manual \
|
||||||
./doc/manual/
|
./doc/manual/
|
||||||
|
|||||||
@@ -11,6 +11,7 @@
|
|||||||
, nix-util
|
, nix-util
|
||||||
, nix-store
|
, nix-store
|
||||||
, nix-main
|
, nix-main
|
||||||
|
, nix-cmd
|
||||||
, nix-cli
|
, nix-cli
|
||||||
, nix-perl-bindings
|
, nix-perl-bindings
|
||||||
, git
|
, git
|
||||||
@@ -178,6 +179,7 @@ stdenv.mkDerivation (finalAttrs: {
|
|||||||
nix-util
|
nix-util
|
||||||
nix-store
|
nix-store
|
||||||
nix-main
|
nix-main
|
||||||
|
nix-cmd
|
||||||
perlDeps
|
perlDeps
|
||||||
perl
|
perl
|
||||||
boost
|
boost
|
||||||
|
|||||||
213
src/hydra-build-step/hydra-build-step.cc
Normal file
213
src/hydra-build-step/hydra-build-step.cc
Normal file
@@ -0,0 +1,213 @@
|
|||||||
|
/* This is a helper program that performs a build step, i.e. a single
|
||||||
|
derivation. In addition to a derivation path, it takes three store
|
||||||
|
URLs as arguments:
|
||||||
|
|
||||||
|
* --store: The store that will hold the resulting store paths
|
||||||
|
(typically a binary cache).
|
||||||
|
|
||||||
|
* --eval-store: The store that holds the .drv files, as produced by
|
||||||
|
hydra-evaluator.
|
||||||
|
|
||||||
|
* --build-store: The store that performs the build (often a
|
||||||
|
SSHStore for remote builds).
|
||||||
|
|
||||||
|
The build log is written to the path indicated by --log-file.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "util.hh"
|
||||||
|
#include "shared.hh"
|
||||||
|
#include "common-eval-args.hh"
|
||||||
|
#include "store-api.hh"
|
||||||
|
#include "build-result.hh"
|
||||||
|
#include "derivations.hh"
|
||||||
|
#include "worker-protocol.hh"
|
||||||
|
|
||||||
|
#include <chrono>
|
||||||
|
|
||||||
|
using namespace nix;
|
||||||
|
|
||||||
|
// FIXME: cut&paste
|
||||||
|
static std::string_view getS(const std::vector<Logger::Field> & fields, size_t n)
|
||||||
|
{
|
||||||
|
assert(n < fields.size());
|
||||||
|
assert(fields[n].type == Logger::Field::tString);
|
||||||
|
return fields[n].s;
|
||||||
|
}
|
||||||
|
|
||||||
|
void mainWrapped(std::list<std::string> args)
|
||||||
|
{
|
||||||
|
verbosity = lvlError;
|
||||||
|
|
||||||
|
struct MyArgs : MixEvalArgs, MixCommonArgs, RootArgs
|
||||||
|
{
|
||||||
|
Path drvPath;
|
||||||
|
std::optional<std::string> buildStoreUrl;
|
||||||
|
std::optional<Path> logPath;
|
||||||
|
std::optional<uint64_t> maxOutputSize;
|
||||||
|
|
||||||
|
MyArgs() : MixCommonArgs("hydra-build-step")
|
||||||
|
{
|
||||||
|
expectArg("drv-path", &drvPath);
|
||||||
|
|
||||||
|
addFlag({
|
||||||
|
.longName = "build-store",
|
||||||
|
.description = "The Nix store to use for building the derivation.",
|
||||||
|
//.category = category,
|
||||||
|
.labels = {"store-url"},
|
||||||
|
.handler = {&buildStoreUrl},
|
||||||
|
});
|
||||||
|
|
||||||
|
addFlag({
|
||||||
|
.longName = "log-file",
|
||||||
|
.description = "The path to the build log.",
|
||||||
|
.labels = {"path"},
|
||||||
|
.handler = {&logPath},
|
||||||
|
});
|
||||||
|
|
||||||
|
addFlag({
|
||||||
|
.longName = "max-output-size",
|
||||||
|
.description = "Maximum size of the outputs.",
|
||||||
|
.labels = {"bytes"},
|
||||||
|
.handler = {&maxOutputSize},
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
/* A logger that intercepts all build log lines and writes them to
|
||||||
|
the log file. */
|
||||||
|
MyArgs myArgs;
|
||||||
|
myArgs.parseCmdline(args);
|
||||||
|
|
||||||
|
struct MyLogger : public Logger
|
||||||
|
{
|
||||||
|
Logger & prev;
|
||||||
|
AutoCloseFD logFile;
|
||||||
|
|
||||||
|
MyLogger(Logger & prev, Path logPath) : prev(prev)
|
||||||
|
{
|
||||||
|
logFile = open(logPath.c_str(), O_CREAT | O_TRUNC | O_WRONLY, 0666);
|
||||||
|
if (!logFile)
|
||||||
|
throw SysError("creating log file '%s'", logPath);
|
||||||
|
}
|
||||||
|
|
||||||
|
void log(Verbosity lvl, std::string_view s) override
|
||||||
|
{ prev.log(lvl, s); }
|
||||||
|
|
||||||
|
void logEI(const ErrorInfo & ei) override
|
||||||
|
{ prev.logEI(ei); }
|
||||||
|
|
||||||
|
void writeToStdout(std::string_view s) override
|
||||||
|
{ prev.writeToStdout(s); }
|
||||||
|
|
||||||
|
void result(ActivityId act, ResultType type, const Fields & fields) override
|
||||||
|
{
|
||||||
|
if (type == resBuildLogLine)
|
||||||
|
writeLine(logFile.get(), std::string(getS(fields, 0)));
|
||||||
|
else
|
||||||
|
prev.result(act, type, fields);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
auto destStore = openStore();
|
||||||
|
auto evalStore = myArgs.evalStoreUrl ? openStore(*myArgs.evalStoreUrl) : destStore;
|
||||||
|
auto buildStore = myArgs.buildStoreUrl ? openStore(*myArgs.buildStoreUrl) : destStore;
|
||||||
|
|
||||||
|
auto drvPath = evalStore->parseStorePath(myArgs.drvPath);
|
||||||
|
|
||||||
|
auto drv = evalStore->readDerivation(drvPath);
|
||||||
|
BasicDerivation basicDrv(drv);
|
||||||
|
|
||||||
|
uint64_t overhead = 0;
|
||||||
|
|
||||||
|
/* Gather the inputs. */
|
||||||
|
StorePathSet inputs;
|
||||||
|
|
||||||
|
for (auto & p : drv.inputSrcs)
|
||||||
|
inputs.insert(p);
|
||||||
|
|
||||||
|
for (auto & [drvPath, node] : drv.inputDrvs.map) {
|
||||||
|
auto drv2 = evalStore->readDerivation(drvPath);
|
||||||
|
for (auto & name : node.value) {
|
||||||
|
if (auto i = get(drv2.outputs, name)) {
|
||||||
|
auto outPath = i->path(*evalStore, drv2.name, name);
|
||||||
|
inputs.insert(*outPath);
|
||||||
|
basicDrv.inputSrcs.insert(*outPath);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Ensure that the inputs exist in the destination store (so that
|
||||||
|
the builder can substitute them from the destination
|
||||||
|
store). This is a no-op for regular stores, but for the binary
|
||||||
|
cache store, this will copy the inputs to the binary cache from
|
||||||
|
the local store. */
|
||||||
|
{
|
||||||
|
auto now1 = std::chrono::steady_clock::now();
|
||||||
|
|
||||||
|
debug("sending closure of '%s' to '%s'",
|
||||||
|
evalStore->printStorePath(drvPath), destStore->getUri());
|
||||||
|
|
||||||
|
if (evalStore != destStore)
|
||||||
|
copyClosure(*evalStore, *destStore, drv.inputSrcs, NoRepair, NoCheckSigs);
|
||||||
|
|
||||||
|
copyClosure(*destStore, *buildStore, inputs, NoRepair, NoCheckSigs, Substitute);
|
||||||
|
|
||||||
|
auto now2 = std::chrono::steady_clock::now();
|
||||||
|
|
||||||
|
overhead += std::chrono::duration_cast<std::chrono::milliseconds>(now2 - now1).count();
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Perform the build. */
|
||||||
|
if (myArgs.logPath)
|
||||||
|
logger = new MyLogger(*logger, *myArgs.logPath);
|
||||||
|
|
||||||
|
auto buildResult = buildStore->buildDerivation(drvPath, basicDrv);
|
||||||
|
|
||||||
|
/* Copy the output paths from the build store to the destination
|
||||||
|
store. */
|
||||||
|
size_t totalNarSize = 0;
|
||||||
|
|
||||||
|
if (buildResult.success()) {
|
||||||
|
|
||||||
|
std::map<StorePath, ValidPathInfo> infos;
|
||||||
|
StorePathSet outputs;
|
||||||
|
for (auto & [output, realisation] : buildResult.builtOutputs) {
|
||||||
|
auto info = buildStore->queryPathInfo(realisation.outPath);
|
||||||
|
totalNarSize += info->narSize;
|
||||||
|
infos.insert_or_assign(info->path, *info);
|
||||||
|
outputs.insert(info->path);
|
||||||
|
}
|
||||||
|
|
||||||
|
if ((!myArgs.maxOutputSize || totalNarSize <= *myArgs.maxOutputSize)
|
||||||
|
&& buildStore != destStore)
|
||||||
|
{
|
||||||
|
debug("copying outputs of '%s' from '%s' (%d bytes)",
|
||||||
|
buildStore->printStorePath(drvPath), buildStore->getUri(), totalNarSize);
|
||||||
|
|
||||||
|
auto now1 = std::chrono::steady_clock::now();
|
||||||
|
|
||||||
|
copyPaths(*buildStore, *destStore, outputs, NoRepair, NoCheckSigs);
|
||||||
|
|
||||||
|
auto now2 = std::chrono::steady_clock::now();
|
||||||
|
|
||||||
|
overhead += std::chrono::duration_cast<std::chrono::milliseconds>(now2 - now1).count();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
FdSink to { STDOUT_FILENO };
|
||||||
|
WorkerProto::WriteConn wconn {
|
||||||
|
.to = to,
|
||||||
|
// Hardcode latest version because we are deploying hydra
|
||||||
|
// itself atomically
|
||||||
|
.version = PROTOCOL_VERSION,
|
||||||
|
};
|
||||||
|
WorkerProto::write(*evalStore, wconn, buildResult);
|
||||||
|
}
|
||||||
|
|
||||||
|
int main(int argc, char * * argv)
|
||||||
|
{
|
||||||
|
return handleExceptions(argv[0], [&]() {
|
||||||
|
initNix();
|
||||||
|
mainWrapped(argvToStrings(argc, argv));
|
||||||
|
});
|
||||||
|
}
|
||||||
14
src/hydra-build-step/meson.build
Normal file
14
src/hydra-build-step/meson.build
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
srcs = files(
|
||||||
|
'hydra-build-step.cc',
|
||||||
|
)
|
||||||
|
|
||||||
|
hydra_build_step = executable('hydra-build-step',
|
||||||
|
'hydra-build-step.cc',
|
||||||
|
srcs,
|
||||||
|
dependencies: [
|
||||||
|
libhydra_dep,
|
||||||
|
nix_dep,
|
||||||
|
dependency('nix-cmd', required: true)
|
||||||
|
],
|
||||||
|
install: true,
|
||||||
|
)
|
||||||
@@ -1,292 +1,27 @@
|
|||||||
#include <algorithm>
|
#include <math.h>
|
||||||
#include <cmath>
|
|
||||||
|
|
||||||
#include <sys/types.h>
|
|
||||||
#include <sys/stat.h>
|
|
||||||
#include <fcntl.h>
|
|
||||||
|
|
||||||
#include "build-result.hh"
|
#include "build-result.hh"
|
||||||
#include "path.hh"
|
|
||||||
#include "legacy-ssh-store.hh"
|
|
||||||
#include "serve-protocol.hh"
|
#include "serve-protocol.hh"
|
||||||
#include "state.hh"
|
#include "state.hh"
|
||||||
#include "current-process.hh"
|
#include "current-process.hh"
|
||||||
#include "processes.hh"
|
#include "processes.hh"
|
||||||
#include "util.hh"
|
#include "util.hh"
|
||||||
#include "ssh.hh"
|
|
||||||
#include "finally.hh"
|
#include "finally.hh"
|
||||||
#include "url.hh"
|
#include "url.hh"
|
||||||
|
#include "worker-protocol.hh"
|
||||||
|
|
||||||
using namespace nix;
|
using namespace nix;
|
||||||
|
|
||||||
bool ::Machine::isLocalhost() const
|
|
||||||
{
|
|
||||||
return storeUri.params.empty() && std::visit(overloaded {
|
|
||||||
[](const StoreReference::Auto &) {
|
|
||||||
return true;
|
|
||||||
},
|
|
||||||
[](const StoreReference::Specified & s) {
|
|
||||||
return
|
|
||||||
(s.scheme == "local" || s.scheme == "unix") ||
|
|
||||||
((s.scheme == "ssh" || s.scheme == "ssh-ng") &&
|
|
||||||
s.authority == "localhost");
|
|
||||||
},
|
|
||||||
}, storeUri.variant);
|
|
||||||
}
|
|
||||||
|
|
||||||
namespace nix::build_remote {
|
namespace nix::build_remote {
|
||||||
|
|
||||||
static void copyClosureTo(
|
static Path createLogFileDir(const std::string & logDir, const StorePath & drvPath)
|
||||||
::Machine::Connection & conn,
|
|
||||||
Store & destStore,
|
|
||||||
const StorePathSet & paths,
|
|
||||||
SubstituteFlag useSubstitutes = NoSubstitute)
|
|
||||||
{
|
|
||||||
StorePathSet closure;
|
|
||||||
destStore.computeFSClosure(paths, closure);
|
|
||||||
|
|
||||||
/* Send the "query valid paths" command with the "lock" option
|
|
||||||
enabled. This prevents a race where the remote host
|
|
||||||
garbage-collect paths that are already there. Optionally, ask
|
|
||||||
the remote host to substitute missing paths. */
|
|
||||||
// FIXME: substitute output pollutes our build log
|
|
||||||
/* Get back the set of paths that are already valid on the remote
|
|
||||||
host. */
|
|
||||||
auto present = conn.store->queryValidPaths(
|
|
||||||
closure, true, useSubstitutes);
|
|
||||||
|
|
||||||
if (present.size() == closure.size()) return;
|
|
||||||
|
|
||||||
auto sorted = destStore.topoSortPaths(closure);
|
|
||||||
|
|
||||||
StorePathSet missing;
|
|
||||||
for (auto i = sorted.rbegin(); i != sorted.rend(); ++i)
|
|
||||||
if (!present.count(*i)) missing.insert(*i);
|
|
||||||
|
|
||||||
printMsg(lvlDebug, "sending %d missing paths", missing.size());
|
|
||||||
|
|
||||||
std::unique_lock<std::timed_mutex> sendLock(conn.machine->state->sendLock,
|
|
||||||
std::chrono::seconds(600));
|
|
||||||
|
|
||||||
conn.store->addMultipleToStoreLegacy(destStore, missing);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
// FIXME: use Store::topoSortPaths().
|
|
||||||
static StorePaths reverseTopoSortPaths(const std::map<StorePath, UnkeyedValidPathInfo> & paths)
|
|
||||||
{
|
|
||||||
StorePaths sorted;
|
|
||||||
StorePathSet visited;
|
|
||||||
|
|
||||||
std::function<void(const StorePath & path)> dfsVisit;
|
|
||||||
|
|
||||||
dfsVisit = [&](const StorePath & path) {
|
|
||||||
if (!visited.insert(path).second) return;
|
|
||||||
|
|
||||||
auto info = paths.find(path);
|
|
||||||
auto references = info == paths.end() ? StorePathSet() : info->second.references;
|
|
||||||
|
|
||||||
for (auto & i : references)
|
|
||||||
/* Don't traverse into paths that don't exist. That can
|
|
||||||
happen due to substitutes for non-existent paths. */
|
|
||||||
if (i != path && paths.count(i))
|
|
||||||
dfsVisit(i);
|
|
||||||
|
|
||||||
sorted.push_back(path);
|
|
||||||
};
|
|
||||||
|
|
||||||
for (auto & i : paths)
|
|
||||||
dfsVisit(i.first);
|
|
||||||
|
|
||||||
return sorted;
|
|
||||||
}
|
|
||||||
|
|
||||||
static std::pair<Path, AutoCloseFD> openLogFile(const std::string & logDir, const StorePath & drvPath)
|
|
||||||
{
|
{
|
||||||
std::string base(drvPath.to_string());
|
std::string base(drvPath.to_string());
|
||||||
auto logFile = logDir + "/" + std::string(base, 0, 2) + "/" + std::string(base, 2);
|
auto logFile = logDir + "/" + std::string(base, 0, 2) + "/" + std::string(base, 2);
|
||||||
|
|
||||||
createDirs(dirOf(logFile));
|
createDirs(dirOf(logFile));
|
||||||
|
|
||||||
AutoCloseFD logFD = open(logFile.c_str(), O_CREAT | O_TRUNC | O_WRONLY, 0666);
|
return logFile;
|
||||||
if (!logFD) throw SysError("creating log file ‘%s’", logFile);
|
|
||||||
|
|
||||||
return {std::move(logFile), std::move(logFD)};
|
|
||||||
}
|
|
||||||
|
|
||||||
static BasicDerivation sendInputs(
|
|
||||||
State & state,
|
|
||||||
Step & step,
|
|
||||||
Store & localStore,
|
|
||||||
Store & destStore,
|
|
||||||
::Machine::Connection & conn,
|
|
||||||
unsigned int & overhead,
|
|
||||||
counter & nrStepsWaiting,
|
|
||||||
counter & nrStepsCopyingTo
|
|
||||||
)
|
|
||||||
{
|
|
||||||
/* Replace the input derivations by their output paths to send a
|
|
||||||
minimal closure to the builder.
|
|
||||||
|
|
||||||
`tryResolve` currently does *not* rewrite input addresses, so it
|
|
||||||
is safe to do this in all cases. (It should probably have a mode
|
|
||||||
to do that, however, but we would not use it here.)
|
|
||||||
*/
|
|
||||||
BasicDerivation basicDrv = ({
|
|
||||||
auto maybeBasicDrv = step.drv->tryResolve(destStore, &localStore);
|
|
||||||
if (!maybeBasicDrv)
|
|
||||||
throw Error(
|
|
||||||
"the derivation '%s' can’t be resolved. It’s probably "
|
|
||||||
"missing some outputs",
|
|
||||||
localStore.printStorePath(step.drvPath));
|
|
||||||
*maybeBasicDrv;
|
|
||||||
});
|
|
||||||
|
|
||||||
/* Ensure that the inputs exist in the destination store. This is
|
|
||||||
a no-op for regular stores, but for the binary cache store,
|
|
||||||
this will copy the inputs to the binary cache from the local
|
|
||||||
store. */
|
|
||||||
if (&localStore != &destStore) {
|
|
||||||
copyClosure(localStore, destStore,
|
|
||||||
step.drv->inputSrcs,
|
|
||||||
NoRepair, NoCheckSigs, NoSubstitute);
|
|
||||||
}
|
|
||||||
|
|
||||||
{
|
|
||||||
auto mc1 = std::make_shared<MaintainCount<counter>>(nrStepsWaiting);
|
|
||||||
mc1.reset();
|
|
||||||
MaintainCount<counter> mc2(nrStepsCopyingTo);
|
|
||||||
|
|
||||||
printMsg(lvlDebug, "sending closure of ‘%s’ to ‘%s’",
|
|
||||||
localStore.printStorePath(step.drvPath), conn.machine->storeUri.render());
|
|
||||||
|
|
||||||
auto now1 = std::chrono::steady_clock::now();
|
|
||||||
|
|
||||||
/* Copy the input closure. */
|
|
||||||
if (conn.machine->isLocalhost()) {
|
|
||||||
StorePathSet closure;
|
|
||||||
destStore.computeFSClosure(basicDrv.inputSrcs, closure);
|
|
||||||
copyPaths(destStore, localStore, closure, NoRepair, NoCheckSigs, NoSubstitute);
|
|
||||||
} else {
|
|
||||||
copyClosureTo(conn, destStore, basicDrv.inputSrcs, Substitute);
|
|
||||||
}
|
|
||||||
|
|
||||||
auto now2 = std::chrono::steady_clock::now();
|
|
||||||
|
|
||||||
overhead += std::chrono::duration_cast<std::chrono::milliseconds>(now2 - now1).count();
|
|
||||||
}
|
|
||||||
|
|
||||||
return basicDrv;
|
|
||||||
}
|
|
||||||
|
|
||||||
static BuildResult performBuild(
|
|
||||||
::Machine::Connection & conn,
|
|
||||||
Store & localStore,
|
|
||||||
StorePath drvPath,
|
|
||||||
const BasicDerivation & drv,
|
|
||||||
const ServeProto::BuildOptions & options,
|
|
||||||
counter & nrStepsBuilding
|
|
||||||
)
|
|
||||||
{
|
|
||||||
auto kont = conn.store->buildDerivationAsync(drvPath, drv, options);
|
|
||||||
|
|
||||||
BuildResult result;
|
|
||||||
|
|
||||||
time_t startTime, stopTime;
|
|
||||||
|
|
||||||
startTime = time(0);
|
|
||||||
{
|
|
||||||
MaintainCount<counter> mc(nrStepsBuilding);
|
|
||||||
result = kont();
|
|
||||||
// Without proper call-once functions, we need to manually
|
|
||||||
// delete after calling.
|
|
||||||
kont = {};
|
|
||||||
}
|
|
||||||
stopTime = time(0);
|
|
||||||
|
|
||||||
if (!result.startTime) {
|
|
||||||
// If the builder gave `startTime = 0`, use our measurements
|
|
||||||
// instead of the builder's.
|
|
||||||
//
|
|
||||||
// Note: this represents the duration of a single round, rather
|
|
||||||
// than all rounds.
|
|
||||||
result.startTime = startTime;
|
|
||||||
result.stopTime = stopTime;
|
|
||||||
}
|
|
||||||
|
|
||||||
// If the protocol was too old to give us `builtOutputs`, initialize
|
|
||||||
// it manually by introspecting the derivation.
|
|
||||||
if (GET_PROTOCOL_MINOR(conn.store->getProtocol()) < 6)
|
|
||||||
{
|
|
||||||
// If the remote is too old to handle CA derivations, we can’t get this
|
|
||||||
// far anyways
|
|
||||||
assert(drv.type().hasKnownOutputPaths());
|
|
||||||
DerivationOutputsAndOptPaths drvOutputs = drv.outputsAndOptPaths(localStore);
|
|
||||||
// Since this a `BasicDerivation`, `staticOutputHashes` will not
|
|
||||||
// do any real work.
|
|
||||||
auto outputHashes = staticOutputHashes(localStore, drv);
|
|
||||||
for (auto & [outputName, output] : drvOutputs) {
|
|
||||||
auto outputPath = output.second;
|
|
||||||
// We’ve just asserted that the output paths of the derivation
|
|
||||||
// were known
|
|
||||||
assert(outputPath);
|
|
||||||
auto outputHash = outputHashes.at(outputName);
|
|
||||||
auto drvOutput = DrvOutput { outputHash, outputName };
|
|
||||||
result.builtOutputs.insert_or_assign(
|
|
||||||
std::move(outputName),
|
|
||||||
Realisation { drvOutput, *outputPath });
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return result;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void copyPathFromRemote(
|
|
||||||
::Machine::Connection & conn,
|
|
||||||
NarMemberDatas & narMembers,
|
|
||||||
Store & localStore,
|
|
||||||
Store & destStore,
|
|
||||||
const ValidPathInfo & info
|
|
||||||
)
|
|
||||||
{
|
|
||||||
/* Receive the NAR from the remote and add it to the
|
|
||||||
destination store. Meanwhile, extract all the info from the
|
|
||||||
NAR that getBuildOutput() needs. */
|
|
||||||
auto source2 = sinkToSource([&](Sink & sink)
|
|
||||||
{
|
|
||||||
/* Note: we should only send the command to dump the store
|
|
||||||
path to the remote if the NAR is actually going to get read
|
|
||||||
by the destination store, which won't happen if this path
|
|
||||||
is already valid on the destination store. Since this
|
|
||||||
lambda function only gets executed if someone tries to read
|
|
||||||
from source2, we will send the command from here rather
|
|
||||||
than outside the lambda. */
|
|
||||||
conn.store->narFromPath(info.path, [&](Source & source) {
|
|
||||||
TeeSource tee{source, sink};
|
|
||||||
extractNarData(tee, conn.store->printStorePath(info.path), narMembers);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
destStore.addToStore(info, *source2, NoRepair, NoCheckSigs);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void copyPathsFromRemote(
|
|
||||||
::Machine::Connection & conn,
|
|
||||||
NarMemberDatas & narMembers,
|
|
||||||
Store & localStore,
|
|
||||||
Store & destStore,
|
|
||||||
const std::map<StorePath, UnkeyedValidPathInfo> & infos
|
|
||||||
)
|
|
||||||
{
|
|
||||||
auto pathsSorted = reverseTopoSortPaths(infos);
|
|
||||||
|
|
||||||
for (auto & path : pathsSorted) {
|
|
||||||
auto & info = infos.find(path)->second;
|
|
||||||
copyPathFromRemote(
|
|
||||||
conn, narMembers, localStore, destStore,
|
|
||||||
ValidPathInfo { path, info });
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
@@ -295,11 +30,14 @@ static void copyPathsFromRemote(
|
|||||||
|
|
||||||
void RemoteResult::updateWithBuildResult(const nix::BuildResult & buildResult)
|
void RemoteResult::updateWithBuildResult(const nix::BuildResult & buildResult)
|
||||||
{
|
{
|
||||||
startTime = buildResult.startTime;
|
// FIXME: make RemoteResult inherit BuildResult.
|
||||||
stopTime = buildResult.stopTime;
|
|
||||||
timesBuilt = buildResult.timesBuilt;
|
timesBuilt = buildResult.timesBuilt;
|
||||||
errorMsg = buildResult.errorMsg;
|
errorMsg = buildResult.errorMsg;
|
||||||
isNonDeterministic = buildResult.isNonDeterministic;
|
isNonDeterministic = buildResult.isNonDeterministic;
|
||||||
|
if (buildResult.startTime && buildResult.stopTime) {
|
||||||
|
startTime = buildResult.startTime;
|
||||||
|
stopTime = buildResult.stopTime;
|
||||||
|
}
|
||||||
|
|
||||||
switch ((BuildResult::Status) buildResult.status) {
|
switch ((BuildResult::Status) buildResult.status) {
|
||||||
case BuildResult::Built:
|
case BuildResult::Built:
|
||||||
@@ -358,43 +96,40 @@ void State::buildRemote(ref<Store> destStore,
|
|||||||
{
|
{
|
||||||
assert(BuildResult::TimedOut == 8);
|
assert(BuildResult::TimedOut == 8);
|
||||||
|
|
||||||
auto [logFile, logFD] = build_remote::openLogFile(logDir, step->drvPath);
|
result.logFile = build_remote::createLogFileDir(logDir, step->drvPath);
|
||||||
AutoDelete logFileDel(logFile, false);
|
|
||||||
result.logFile = logFile;
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
|
|
||||||
updateStep(ssConnecting);
|
updateStep(ssBuilding);
|
||||||
|
result.startTime = time(0);
|
||||||
|
|
||||||
// FIXME: rewrite to use Store.
|
auto buildStoreUrl = machine->completeStoreReference().render();
|
||||||
::Machine::Connection conn {
|
|
||||||
.machine = machine,
|
|
||||||
.store = [&]{
|
|
||||||
auto * pSpecified = std::get_if<StoreReference::Specified>(&machine->storeUri.variant);
|
|
||||||
if (!pSpecified || pSpecified->scheme != "ssh") {
|
|
||||||
throw Error("Currently, only (legacy-)ssh stores are supported!");
|
|
||||||
}
|
|
||||||
|
|
||||||
auto remoteStore = machine->openStore().dynamic_pointer_cast<LegacySSHStore>();
|
Strings args = {
|
||||||
assert(remoteStore);
|
localStore->printStorePath(step->drvPath),
|
||||||
|
"--store", destStore->getUri(),
|
||||||
remoteStore->connPipeSize = 1024 * 1024;
|
"--eval-store", localStore->getUri(),
|
||||||
|
"--build-store", buildStoreUrl,
|
||||||
if (machine->isLocalhost()) {
|
"--max-silent-time", std::to_string(buildOptions.maxSilentTime),
|
||||||
auto rp_new = remoteStore->remoteProgram.get();
|
"--timeout", std::to_string(buildOptions.buildTimeout),
|
||||||
rp_new.push_back("--builders");
|
"--max-build-log-size", std::to_string(buildOptions.maxLogSize),
|
||||||
rp_new.push_back("");
|
"--max-output-size", std::to_string(maxOutputSize),
|
||||||
const_cast<nix::Setting<Strings> &>(remoteStore->remoteProgram).assign(rp_new);
|
"--repeat", std::to_string(buildOptions.nrRepeats),
|
||||||
}
|
"--log-file", result.logFile,
|
||||||
remoteStore->extraSshArgs = {
|
// FIXME: step->isDeterministic
|
||||||
"-a", "-oBatchMode=yes", "-oConnectTimeout=60", "-oTCPKeepAlive=yes"
|
|
||||||
};
|
|
||||||
const_cast<nix::Setting<int> &>(remoteStore->logFD).assign(logFD.get());
|
|
||||||
|
|
||||||
return nix::ref{remoteStore};
|
|
||||||
}(),
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
// FIXME: set pid for cancellation
|
||||||
|
|
||||||
|
auto [status, childStdout] = [&]() {
|
||||||
|
MaintainCount<counter> mc(nrStepsBuilding);
|
||||||
|
return runProgram({
|
||||||
|
.program = "hydra-build-step",
|
||||||
|
.args = std::move(args),
|
||||||
|
});
|
||||||
|
}();
|
||||||
|
|
||||||
|
#if 0
|
||||||
{
|
{
|
||||||
auto activeStepState(activeStep->state_.lock());
|
auto activeStepState(activeStep->state_.lock());
|
||||||
if (activeStepState->cancelled) throw Error("step cancelled");
|
if (activeStepState->cancelled) throw Error("step cancelled");
|
||||||
@@ -412,53 +147,32 @@ void State::buildRemote(ref<Store> destStore,
|
|||||||
possibility that we end up killing another
|
possibility that we end up killing another
|
||||||
process. Meh. */
|
process. Meh. */
|
||||||
});
|
});
|
||||||
|
#endif
|
||||||
|
|
||||||
Finally updateStats([&]() {
|
result.stopTime = time(0);
|
||||||
auto stats = conn.store->getConnectionStats();
|
|
||||||
bytesReceived += stats.bytesReceived;
|
|
||||||
bytesSent += stats.bytesSent;
|
|
||||||
});
|
|
||||||
|
|
||||||
|
if (!statusOk(status))
|
||||||
|
throw ExecError(status, fmt("hydra-build-step %s with output:\n%s", statusToString(status), stdout));
|
||||||
|
|
||||||
|
/* The build was executed successfully, so clear the failure
|
||||||
|
count for this machine. */
|
||||||
{
|
{
|
||||||
auto info(machine->state->connectInfo.lock());
|
auto info(machine->state->connectInfo.lock());
|
||||||
info->consecutiveFailures = 0;
|
info->consecutiveFailures = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Gather the inputs. If the remote side is Nix <= 1.9, we have to
|
StringSource from { childStdout };
|
||||||
copy the entire closure of ‘drvPath’, as well as the required
|
/* Read the BuildResult from the child. */
|
||||||
outputs of the input derivations. On Nix > 1.9, we only need to
|
WorkerProto::ReadConn rconn {
|
||||||
copy the immediate sources of the derivation and the required
|
.from = from,
|
||||||
outputs of the input derivations. */
|
// Hardcode latest version because we are deploying hydra
|
||||||
updateStep(ssSendingInputs);
|
// itself atomically
|
||||||
BasicDerivation resolvedDrv = build_remote::sendInputs(*this, *step, *localStore, *destStore, conn, result.overhead, nrStepsWaiting, nrStepsCopyingTo);
|
.version = PROTOCOL_VERSION,
|
||||||
|
};
|
||||||
|
result.overhead += readNum<uint64_t>(rconn.from);
|
||||||
|
auto totalNarSize = readNum<uint64_t>(rconn.from);
|
||||||
|
auto buildResult = WorkerProto::Serialise<BuildResult>::read(*localStore, rconn);
|
||||||
|
|
||||||
logFileDel.cancel();
|
|
||||||
|
|
||||||
/* Truncate the log to get rid of messages about substitutions
|
|
||||||
etc. on the remote system. */
|
|
||||||
if (lseek(logFD.get(), SEEK_SET, 0) != 0)
|
|
||||||
throw SysError("seeking to the start of log file ‘%s’", result.logFile);
|
|
||||||
|
|
||||||
if (ftruncate(logFD.get(), 0) == -1)
|
|
||||||
throw SysError("truncating log file ‘%s’", result.logFile);
|
|
||||||
|
|
||||||
logFD = -1;
|
|
||||||
|
|
||||||
/* Do the build. */
|
|
||||||
printMsg(lvlDebug, "building ‘%s’ on ‘%s’",
|
|
||||||
localStore->printStorePath(step->drvPath),
|
|
||||||
machine->storeUri.render());
|
|
||||||
|
|
||||||
updateStep(ssBuilding);
|
|
||||||
|
|
||||||
BuildResult buildResult = build_remote::performBuild(
|
|
||||||
conn,
|
|
||||||
*localStore,
|
|
||||||
step->drvPath,
|
|
||||||
resolvedDrv,
|
|
||||||
buildOptions,
|
|
||||||
nrStepsBuilding
|
|
||||||
);
|
|
||||||
|
|
||||||
result.updateWithBuildResult(buildResult);
|
result.updateWithBuildResult(buildResult);
|
||||||
|
|
||||||
@@ -466,6 +180,13 @@ void State::buildRemote(ref<Store> destStore,
|
|||||||
|
|
||||||
result.errorMsg = "";
|
result.errorMsg = "";
|
||||||
|
|
||||||
|
/* If the NAR size limit was exceeded, then hydra-build-step
|
||||||
|
will not have copied the output paths. */
|
||||||
|
if (totalNarSize > maxOutputSize) {
|
||||||
|
result.stepStatus = bsNarSizeLimitExceeded;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
/* If the path was substituted or already valid, then we didn't
|
/* If the path was substituted or already valid, then we didn't
|
||||||
get a build log. */
|
get a build log. */
|
||||||
if (result.isCached) {
|
if (result.isCached) {
|
||||||
@@ -475,59 +196,6 @@ void State::buildRemote(ref<Store> destStore,
|
|||||||
result.logFile = "";
|
result.logFile = "";
|
||||||
}
|
}
|
||||||
|
|
||||||
StorePathSet outputs;
|
|
||||||
for (auto & [_, realisation] : buildResult.builtOutputs)
|
|
||||||
outputs.insert(realisation.outPath);
|
|
||||||
|
|
||||||
/* Copy the output paths. */
|
|
||||||
if (!machine->isLocalhost() || localStore != std::shared_ptr<Store>(destStore)) {
|
|
||||||
updateStep(ssReceivingOutputs);
|
|
||||||
|
|
||||||
MaintainCount<counter> mc(nrStepsCopyingFrom);
|
|
||||||
|
|
||||||
auto now1 = std::chrono::steady_clock::now();
|
|
||||||
|
|
||||||
auto infos = conn.store->queryPathInfosUncached(outputs);
|
|
||||||
|
|
||||||
size_t totalNarSize = 0;
|
|
||||||
for (auto & [_, info] : infos) totalNarSize += info.narSize;
|
|
||||||
|
|
||||||
if (totalNarSize > maxOutputSize) {
|
|
||||||
result.stepStatus = bsNarSizeLimitExceeded;
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Copy each path. */
|
|
||||||
printMsg(lvlDebug, "copying outputs of ‘%s’ from ‘%s’ (%d bytes)",
|
|
||||||
localStore->printStorePath(step->drvPath), machine->storeUri.render(), totalNarSize);
|
|
||||||
|
|
||||||
build_remote::copyPathsFromRemote(conn, narMembers, *localStore, *destStore, infos);
|
|
||||||
auto now2 = std::chrono::steady_clock::now();
|
|
||||||
|
|
||||||
result.overhead += std::chrono::duration_cast<std::chrono::milliseconds>(now2 - now1).count();
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Register the outputs of the newly built drv */
|
|
||||||
if (experimentalFeatureSettings.isEnabled(Xp::CaDerivations)) {
|
|
||||||
auto outputHashes = staticOutputHashes(*localStore, *step->drv);
|
|
||||||
for (auto & [outputName, realisation] : buildResult.builtOutputs) {
|
|
||||||
// Register the resolved drv output
|
|
||||||
destStore->registerDrvOutput(realisation);
|
|
||||||
|
|
||||||
// Also register the unresolved one
|
|
||||||
auto unresolvedRealisation = realisation;
|
|
||||||
unresolvedRealisation.signatures.clear();
|
|
||||||
unresolvedRealisation.id.drvHash = outputHashes.at(outputName);
|
|
||||||
destStore->registerDrvOutput(unresolvedRealisation);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Shut down the connection done by RAII.
|
|
||||||
|
|
||||||
Only difference is kill() instead of wait() (i.e. send signal
|
|
||||||
then wait())
|
|
||||||
*/
|
|
||||||
|
|
||||||
} catch (Error & e) {
|
} catch (Error & e) {
|
||||||
/* Disable this machine until a certain period of time has
|
/* Disable this machine until a certain period of time has
|
||||||
passed. This period increases on every consecutive
|
passed. This period increases on every consecutive
|
||||||
|
|||||||
@@ -3,6 +3,7 @@
|
|||||||
#include "state.hh"
|
#include "state.hh"
|
||||||
#include "hydra-build-result.hh"
|
#include "hydra-build-result.hh"
|
||||||
#include "finally.hh"
|
#include "finally.hh"
|
||||||
|
#include "terminal.hh"
|
||||||
#include "binary-cache-store.hh"
|
#include "binary-cache-store.hh"
|
||||||
|
|
||||||
using namespace nix;
|
using namespace nix;
|
||||||
@@ -219,7 +220,7 @@ State::StepResult State::doBuildStep(nix::ref<Store> destStore,
|
|||||||
result.canRetry = false;
|
result.canRetry = false;
|
||||||
} else {
|
} else {
|
||||||
result.stepStatus = bsAborted;
|
result.stepStatus = bsAborted;
|
||||||
result.errorMsg = e.msg();
|
result.errorMsg = filterANSIEscapes(e.msg(), true);
|
||||||
result.canRetry = true;
|
result.canRetry = true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -550,12 +550,11 @@ void State::dumpStatus(Connection & conn)
|
|||||||
{"nrQueuedBuilds", builds.lock()->size()},
|
{"nrQueuedBuilds", builds.lock()->size()},
|
||||||
{"nrActiveSteps", activeSteps_.lock()->size()},
|
{"nrActiveSteps", activeSteps_.lock()->size()},
|
||||||
{"nrStepsBuilding", nrStepsBuilding.load()},
|
{"nrStepsBuilding", nrStepsBuilding.load()},
|
||||||
|
#if 0
|
||||||
{"nrStepsCopyingTo", nrStepsCopyingTo.load()},
|
{"nrStepsCopyingTo", nrStepsCopyingTo.load()},
|
||||||
{"nrStepsCopyingFrom", nrStepsCopyingFrom.load()},
|
{"nrStepsCopyingFrom", nrStepsCopyingFrom.load()},
|
||||||
{"nrStepsWaiting", nrStepsWaiting.load()},
|
#endif
|
||||||
{"nrUnsupportedSteps", nrUnsupportedSteps.load()},
|
{"nrUnsupportedSteps", nrUnsupportedSteps.load()},
|
||||||
{"bytesSent", bytesSent.load()},
|
|
||||||
{"bytesReceived", bytesReceived.load()},
|
|
||||||
{"nrBuildsRead", nrBuildsRead.load()},
|
{"nrBuildsRead", nrBuildsRead.load()},
|
||||||
{"buildReadTimeMs", buildReadTimeMs.load()},
|
{"buildReadTimeMs", buildReadTimeMs.load()},
|
||||||
{"buildReadTimeAvgMs", nrBuildsRead == 0 ? 0.0 : (float) buildReadTimeMs / nrBuildsRead},
|
{"buildReadTimeAvgMs", nrBuildsRead == 0 ? 0.0 : (float) buildReadTimeMs / nrBuildsRead},
|
||||||
|
|||||||
@@ -20,7 +20,7 @@
|
|||||||
#include "store-api.hh"
|
#include "store-api.hh"
|
||||||
#include "sync.hh"
|
#include "sync.hh"
|
||||||
#include "nar-extractor.hh"
|
#include "nar-extractor.hh"
|
||||||
#include "legacy-ssh-store.hh"
|
#include "serve-protocol.hh"
|
||||||
#include "machines.hh"
|
#include "machines.hh"
|
||||||
|
|
||||||
|
|
||||||
@@ -288,14 +288,6 @@ struct Machine : nix::Machine
|
|||||||
}
|
}
|
||||||
|
|
||||||
bool isLocalhost() const;
|
bool isLocalhost() const;
|
||||||
|
|
||||||
// A connection to a machine
|
|
||||||
struct Connection {
|
|
||||||
// Backpointer to the machine
|
|
||||||
ptr machine;
|
|
||||||
// Opened store
|
|
||||||
nix::ref<nix::LegacySSHStore> store;
|
|
||||||
};
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
@@ -360,9 +352,10 @@ private:
|
|||||||
counter nrStepsStarted{0};
|
counter nrStepsStarted{0};
|
||||||
counter nrStepsDone{0};
|
counter nrStepsDone{0};
|
||||||
counter nrStepsBuilding{0};
|
counter nrStepsBuilding{0};
|
||||||
|
#if 0
|
||||||
counter nrStepsCopyingTo{0};
|
counter nrStepsCopyingTo{0};
|
||||||
counter nrStepsCopyingFrom{0};
|
counter nrStepsCopyingFrom{0};
|
||||||
counter nrStepsWaiting{0};
|
#endif
|
||||||
counter nrUnsupportedSteps{0};
|
counter nrUnsupportedSteps{0};
|
||||||
counter nrRetries{0};
|
counter nrRetries{0};
|
||||||
counter maxNrRetries{0};
|
counter maxNrRetries{0};
|
||||||
@@ -371,8 +364,6 @@ private:
|
|||||||
counter nrQueueWakeups{0};
|
counter nrQueueWakeups{0};
|
||||||
counter nrDispatcherWakeups{0};
|
counter nrDispatcherWakeups{0};
|
||||||
counter dispatchTimeMs{0};
|
counter dispatchTimeMs{0};
|
||||||
counter bytesSent{0};
|
|
||||||
counter bytesReceived{0};
|
|
||||||
counter nrActiveDbUpdates{0};
|
counter nrActiveDbUpdates{0};
|
||||||
|
|
||||||
/* Specific build to do for --build-one (testing only). */
|
/* Specific build to do for --build-one (testing only). */
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
# Native code
|
# Native code
|
||||||
subdir('libhydra')
|
subdir('libhydra')
|
||||||
|
subdir('hydra-build-step')
|
||||||
subdir('hydra-evaluator')
|
subdir('hydra-evaluator')
|
||||||
subdir('hydra-queue-runner')
|
subdir('hydra-queue-runner')
|
||||||
|
|
||||||
|
|||||||
@@ -28,6 +28,7 @@ testenv.prepend('PERL5LIB',
|
|||||||
)
|
)
|
||||||
testenv.prepend('PATH',
|
testenv.prepend('PATH',
|
||||||
fs.parent(find_program('nix').full_path()),
|
fs.parent(find_program('nix').full_path()),
|
||||||
|
fs.parent(hydra_build_step.full_path()),
|
||||||
fs.parent(hydra_evaluator.full_path()),
|
fs.parent(hydra_evaluator.full_path()),
|
||||||
fs.parent(hydra_queue_runner.full_path()),
|
fs.parent(hydra_queue_runner.full_path()),
|
||||||
meson.project_source_root() / 'src/script',
|
meson.project_source_root() / 'src/script',
|
||||||
|
|||||||
Reference in New Issue
Block a user