2020-04-06 21:57:28 +00:00
|
|
|
#include "filetransfer.hh"
|
2015-04-09 10:12:50 +00:00
|
|
|
#include "util.hh"
|
|
|
|
#include "globals.hh"
|
2015-05-05 15:09:42 +00:00
|
|
|
#include "store-api.hh"
|
2017-02-14 13:20:00 +00:00
|
|
|
#include "s3.hh"
|
2017-03-13 13:40:15 +00:00
|
|
|
#include "compression.hh"
|
2018-03-27 22:01:47 +00:00
|
|
|
#include "finally.hh"
|
2020-09-21 16:40:11 +00:00
|
|
|
#include "callback.hh"
|
2017-03-13 13:40:15 +00:00
|
|
|
|
2021-04-23 06:30:05 +00:00
|
|
|
#if ENABLE_S3
|
2017-03-03 21:12:17 +00:00
|
|
|
#include <aws/core/client/ClientConfiguration.h>
|
|
|
|
#endif
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
#include <unistd.h>
|
|
|
|
#include <fcntl.h>
|
|
|
|
|
2015-04-09 10:12:50 +00:00
|
|
|
#include <curl/curl.h>
|
|
|
|
|
2018-01-31 21:34:19 +00:00
|
|
|
#include <algorithm>
|
2019-07-10 17:46:15 +00:00
|
|
|
#include <cmath>
|
2018-01-31 21:34:19 +00:00
|
|
|
#include <cstring>
|
|
|
|
#include <iostream>
|
|
|
|
#include <queue>
|
2019-07-10 17:46:15 +00:00
|
|
|
#include <random>
|
2018-01-31 21:34:19 +00:00
|
|
|
#include <thread>
|
2020-06-17 13:18:10 +00:00
|
|
|
#include <regex>
|
2015-10-07 15:31:50 +00:00
|
|
|
|
2017-10-30 09:16:19 +00:00
|
|
|
using namespace std::string_literals;
|
|
|
|
|
2015-04-09 10:12:50 +00:00
|
|
|
namespace nix {
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
FileTransferSettings fileTransferSettings;
|
2018-03-27 16:54:02 +00:00
|
|
|
|
2020-10-06 11:36:55 +00:00
|
|
|
static GlobalConfig::Register rFileTransferSettings(&fileTransferSettings);
|
2019-11-06 16:30:48 +00:00
|
|
|
|
2016-04-14 14:27:48 +00:00
|
|
|
std::string resolveUri(const std::string & uri)
|
|
|
|
{
|
|
|
|
if (uri.compare(0, 8, "channel:") == 0)
|
|
|
|
return "https://nixos.org/channels/" + std::string(uri, 8) + "/nixexprs.tar.xz";
|
|
|
|
else
|
|
|
|
return uri;
|
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
struct curlFileTransfer : public FileTransfer
|
2015-04-09 10:12:50 +00:00
|
|
|
{
|
2016-09-14 14:00:40 +00:00
|
|
|
CURLM * curlm = 0;
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2019-07-10 17:46:15 +00:00
|
|
|
std::random_device rd;
|
|
|
|
std::mt19937 mt19937;
|
|
|
|
|
2020-04-06 20:59:36 +00:00
|
|
|
struct TransferItem : public std::enable_shared_from_this<TransferItem>
|
2015-04-09 10:12:50 +00:00
|
|
|
{
|
2020-04-06 21:43:43 +00:00
|
|
|
curlFileTransfer & fileTransfer;
|
|
|
|
FileTransferRequest request;
|
|
|
|
FileTransferResult result;
|
2017-05-16 14:09:57 +00:00
|
|
|
Activity act;
|
2016-09-16 16:54:14 +00:00
|
|
|
bool done = false; // whether either the success or failure function has been called
|
2020-04-06 21:43:43 +00:00
|
|
|
Callback<FileTransferResult> callback;
|
2016-09-14 14:00:40 +00:00
|
|
|
CURL * req = 0;
|
|
|
|
bool active = false; // whether the handle has been added to the multi object
|
2020-06-09 12:20:22 +00:00
|
|
|
std::string statusMsg;
|
2016-09-14 14:00:40 +00:00
|
|
|
|
2019-07-10 17:46:15 +00:00
|
|
|
unsigned int attempt = 0;
|
|
|
|
|
|
|
|
/* Don't start this download until the specified time point
|
|
|
|
has been reached. */
|
|
|
|
std::chrono::steady_clock::time_point embargo;
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
struct curl_slist * requestHeaders = 0;
|
|
|
|
|
2017-03-13 13:40:15 +00:00
|
|
|
std::string encoding;
|
|
|
|
|
2019-07-10 21:12:17 +00:00
|
|
|
bool acceptRanges = false;
|
|
|
|
|
|
|
|
curl_off_t writtenToSink = 0;
|
|
|
|
|
2020-06-17 16:58:59 +00:00
|
|
|
inline static const std::set<long> successfulStatuses {200, 201, 204, 206, 304, 0 /* other protocol */};
|
2020-06-16 09:51:34 +00:00
|
|
|
/* Get the HTTP status code, or 0 for other protocols. */
|
|
|
|
long getHTTPStatus()
|
|
|
|
{
|
|
|
|
long httpStatus = 0;
|
|
|
|
long protocol = 0;
|
|
|
|
curl_easy_getinfo(req, CURLINFO_PROTOCOL, &protocol);
|
|
|
|
if (protocol == CURLPROTO_HTTP || protocol == CURLPROTO_HTTPS)
|
|
|
|
curl_easy_getinfo(req, CURLINFO_RESPONSE_CODE, &httpStatus);
|
|
|
|
return httpStatus;
|
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
TransferItem(curlFileTransfer & fileTransfer,
|
|
|
|
const FileTransferRequest & request,
|
|
|
|
Callback<FileTransferResult> && callback)
|
|
|
|
: fileTransfer(fileTransfer)
|
2017-08-14 18:14:55 +00:00
|
|
|
, request(request)
|
2020-04-06 21:43:43 +00:00
|
|
|
, act(*logger, lvlTalkative, actFileTransfer,
|
2018-06-01 12:14:22 +00:00
|
|
|
fmt(request.data ? "uploading '%s'" : "downloading '%s'", request.uri),
|
|
|
|
{request.uri}, request.parentAct)
|
2019-09-03 10:51:35 +00:00
|
|
|
, callback(std::move(callback))
|
2020-12-02 13:00:43 +00:00
|
|
|
, finalSink([this](std::string_view data) {
|
2019-07-10 21:12:17 +00:00
|
|
|
if (this->request.dataCallback) {
|
2020-06-16 09:51:34 +00:00
|
|
|
auto httpStatus = getHTTPStatus();
|
2020-04-08 13:18:41 +00:00
|
|
|
|
|
|
|
/* Only write data to the sink if this is a
|
|
|
|
successful response. */
|
2020-06-17 16:58:59 +00:00
|
|
|
if (successfulStatuses.count(httpStatus)) {
|
2020-12-02 13:00:43 +00:00
|
|
|
writtenToSink += data.size();
|
|
|
|
this->request.dataCallback(data);
|
2020-04-08 13:18:41 +00:00
|
|
|
}
|
2019-07-10 21:12:17 +00:00
|
|
|
} else
|
2020-12-02 13:00:43 +00:00
|
|
|
this->result.data->append(data);
|
2018-08-06 13:40:29 +00:00
|
|
|
})
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
if (!request.expectedETag.empty())
|
|
|
|
requestHeaders = curl_slist_append(requestHeaders, ("If-None-Match: " + request.expectedETag).c_str());
|
2018-01-26 19:12:30 +00:00
|
|
|
if (!request.mimeType.empty())
|
|
|
|
requestHeaders = curl_slist_append(requestHeaders, ("Content-Type: " + request.mimeType).c_str());
|
2020-06-17 19:08:59 +00:00
|
|
|
for (auto it = request.headers.begin(); it != request.headers.end(); ++it){
|
|
|
|
requestHeaders = curl_slist_append(requestHeaders, fmt("%s: %s", it->first, it->second).c_str());
|
|
|
|
}
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2020-04-06 20:59:36 +00:00
|
|
|
~TransferItem()
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
if (req) {
|
|
|
|
if (active)
|
2020-04-06 21:43:43 +00:00
|
|
|
curl_multi_remove_handle(fileTransfer.curlm, req);
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_easy_cleanup(req);
|
|
|
|
}
|
|
|
|
if (requestHeaders) curl_slist_free_all(requestHeaders);
|
|
|
|
try {
|
|
|
|
if (!done)
|
2020-06-16 19:14:11 +00:00
|
|
|
fail(FileTransferError(Interrupted, nullptr, "download of '%s' was interrupted", request.uri));
|
2016-09-14 14:00:40 +00:00
|
|
|
} catch (...) {
|
|
|
|
ignoreException();
|
|
|
|
}
|
|
|
|
}
|
2016-02-29 17:15:20 +00:00
|
|
|
|
2018-08-06 13:40:29 +00:00
|
|
|
void failEx(std::exception_ptr ex)
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
2016-09-16 16:54:14 +00:00
|
|
|
assert(!done);
|
2016-09-14 14:00:40 +00:00
|
|
|
done = true;
|
2018-08-06 13:40:29 +00:00
|
|
|
callback.rethrow(ex);
|
|
|
|
}
|
|
|
|
|
|
|
|
template<class T>
|
|
|
|
void fail(const T & e)
|
|
|
|
{
|
|
|
|
failEx(std::make_exception_ptr(e));
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
|
2018-08-06 13:40:29 +00:00
|
|
|
LambdaSink finalSink;
|
2019-12-10 08:47:38 +00:00
|
|
|
std::shared_ptr<FinishSink> decompressionSink;
|
2020-07-15 16:58:38 +00:00
|
|
|
std::optional<StringSink> errorSink;
|
2018-08-06 13:40:29 +00:00
|
|
|
|
|
|
|
std::exception_ptr writeException;
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
size_t writeCallback(void * contents, size_t size, size_t nmemb)
|
|
|
|
{
|
2018-08-06 13:40:29 +00:00
|
|
|
try {
|
|
|
|
size_t realSize = size * nmemb;
|
|
|
|
result.bodySize += realSize;
|
|
|
|
|
2020-06-16 19:14:11 +00:00
|
|
|
if (!decompressionSink) {
|
2018-08-06 13:40:29 +00:00
|
|
|
decompressionSink = makeDecompressionSink(encoding, finalSink);
|
2020-06-16 19:14:11 +00:00
|
|
|
if (! successfulStatuses.count(getHTTPStatus())) {
|
|
|
|
// In this case we want to construct a TeeSink, to keep
|
|
|
|
// the response around (which we figure won't be big
|
|
|
|
// like an actual download should be) to improve error
|
|
|
|
// messages.
|
2020-07-15 16:58:38 +00:00
|
|
|
errorSink = StringSink { };
|
2020-06-16 19:14:11 +00:00
|
|
|
}
|
|
|
|
}
|
2018-08-06 13:40:29 +00:00
|
|
|
|
2020-07-15 16:58:38 +00:00
|
|
|
if (errorSink)
|
2020-12-02 13:00:43 +00:00
|
|
|
(*errorSink)({(char *) contents, realSize});
|
|
|
|
(*decompressionSink)({(char *) contents, realSize});
|
2018-08-06 13:40:29 +00:00
|
|
|
|
|
|
|
return realSize;
|
|
|
|
} catch (...) {
|
|
|
|
writeException = std::current_exception();
|
|
|
|
return 0;
|
|
|
|
}
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static size_t writeCallbackWrapper(void * contents, size_t size, size_t nmemb, void * userp)
|
|
|
|
{
|
2020-04-06 20:59:36 +00:00
|
|
|
return ((TransferItem *) userp)->writeCallback(contents, size, nmemb);
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
size_t headerCallback(void * contents, size_t size, size_t nmemb)
|
|
|
|
{
|
|
|
|
size_t realSize = size * nmemb;
|
|
|
|
std::string line((char *) contents, realSize);
|
2017-07-30 11:27:57 +00:00
|
|
|
printMsg(lvlVomit, format("got header for '%s': %s") % request.uri % trim(line));
|
2020-06-09 12:20:22 +00:00
|
|
|
static std::regex statusLine("HTTP/[^ ]+ +[0-9]+(.*)", std::regex::extended | std::regex::icase);
|
|
|
|
std::smatch match;
|
|
|
|
if (std::regex_match(line, match, statusLine)) {
|
2016-09-14 14:00:40 +00:00
|
|
|
result.etag = "";
|
|
|
|
result.data = std::make_shared<std::string>();
|
2018-08-06 09:31:14 +00:00
|
|
|
result.bodySize = 0;
|
2020-06-09 12:20:22 +00:00
|
|
|
statusMsg = trim(match[1]);
|
2019-07-10 21:12:17 +00:00
|
|
|
acceptRanges = false;
|
2017-03-13 13:40:15 +00:00
|
|
|
encoding = "";
|
2016-09-14 14:00:40 +00:00
|
|
|
} else {
|
|
|
|
auto i = line.find(':');
|
|
|
|
if (i != string::npos) {
|
|
|
|
string name = toLower(trim(string(line, 0, i)));
|
|
|
|
if (name == "etag") {
|
|
|
|
result.etag = trim(string(line, i + 1));
|
|
|
|
/* Hack to work around a GitHub bug: it sends
|
|
|
|
ETags, but ignores If-None-Match. So if we get
|
|
|
|
the expected ETag on a 200 response, then shut
|
|
|
|
down the connection because we already have the
|
|
|
|
data. */
|
2020-06-09 12:05:15 +00:00
|
|
|
long httpStatus = 0;
|
|
|
|
curl_easy_getinfo(req, CURLINFO_RESPONSE_CODE, &httpStatus);
|
|
|
|
if (result.etag == request.expectedETag && httpStatus == 200) {
|
2016-09-14 14:00:40 +00:00
|
|
|
debug(format("shutting down on 200 HTTP response with expected ETag"));
|
|
|
|
return 0;
|
|
|
|
}
|
2017-03-13 13:40:15 +00:00
|
|
|
} else if (name == "content-encoding")
|
2019-07-10 21:12:17 +00:00
|
|
|
encoding = trim(string(line, i + 1));
|
|
|
|
else if (name == "accept-ranges" && toLower(trim(std::string(line, i + 1))) == "bytes")
|
|
|
|
acceptRanges = true;
|
2015-04-09 10:12:50 +00:00
|
|
|
}
|
|
|
|
}
|
2016-09-14 14:00:40 +00:00
|
|
|
return realSize;
|
2015-04-09 10:12:50 +00:00
|
|
|
}
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
static size_t headerCallbackWrapper(void * contents, size_t size, size_t nmemb, void * userp)
|
|
|
|
{
|
2020-04-06 20:59:36 +00:00
|
|
|
return ((TransferItem *) userp)->headerCallback(contents, size, nmemb);
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
2016-02-29 17:15:20 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
int progressCallback(double dltotal, double dlnow)
|
|
|
|
{
|
2017-11-30 17:38:12 +00:00
|
|
|
try {
|
|
|
|
act.progress(dlnow, dltotal);
|
|
|
|
} catch (nix::Interrupted &) {
|
|
|
|
assert(_isInterrupted);
|
|
|
|
}
|
2016-09-14 14:00:40 +00:00
|
|
|
return _isInterrupted;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int progressCallbackWrapper(void * userp, double dltotal, double dlnow, double ultotal, double ulnow)
|
|
|
|
{
|
2020-04-06 20:59:36 +00:00
|
|
|
return ((TransferItem *) userp)->progressCallback(dltotal, dlnow);
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
|
2017-01-24 12:15:24 +00:00
|
|
|
static int debugCallback(CURL * handle, curl_infotype type, char * data, size_t size, void * userptr)
|
|
|
|
{
|
|
|
|
if (type == CURLINFO_TEXT)
|
|
|
|
vomit("curl: %s", chomp(std::string(data, size)));
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-01-26 19:12:30 +00:00
|
|
|
size_t readOffset = 0;
|
2018-06-01 15:53:28 +00:00
|
|
|
size_t readCallback(char *buffer, size_t size, size_t nitems)
|
2018-01-26 19:12:30 +00:00
|
|
|
{
|
|
|
|
if (readOffset == request.data->length())
|
|
|
|
return 0;
|
|
|
|
auto count = std::min(size * nitems, request.data->length() - readOffset);
|
2018-03-15 03:53:43 +00:00
|
|
|
assert(count);
|
2018-01-26 19:12:30 +00:00
|
|
|
memcpy(buffer, request.data->data() + readOffset, count);
|
|
|
|
readOffset += count;
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
2018-06-01 15:53:28 +00:00
|
|
|
static size_t readCallbackWrapper(char *buffer, size_t size, size_t nitems, void * userp)
|
2018-01-26 19:12:30 +00:00
|
|
|
{
|
2020-04-06 20:59:36 +00:00
|
|
|
return ((TransferItem *) userp)->readCallback(buffer, size, nitems);
|
2018-01-26 19:12:30 +00:00
|
|
|
}
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
void init()
|
|
|
|
{
|
|
|
|
if (!req) req = curl_easy_init();
|
|
|
|
|
|
|
|
curl_easy_reset(req);
|
2017-01-24 12:15:24 +00:00
|
|
|
|
|
|
|
if (verbosity >= lvlVomit) {
|
|
|
|
curl_easy_setopt(req, CURLOPT_VERBOSE, 1);
|
2020-04-06 20:59:36 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_DEBUGFUNCTION, TransferItem::debugCallback);
|
2017-01-24 12:15:24 +00:00
|
|
|
}
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_URL, request.uri.c_str());
|
|
|
|
curl_easy_setopt(req, CURLOPT_FOLLOWLOCATION, 1L);
|
2018-06-18 08:36:19 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_MAXREDIRS, 10);
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_NOSIGNAL, 1);
|
2017-05-11 13:09:09 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_USERAGENT,
|
|
|
|
("curl/" LIBCURL_VERSION " Nix/" + nixVersion +
|
2020-04-06 21:43:43 +00:00
|
|
|
(fileTransferSettings.userAgentSuffix != "" ? " " + fileTransferSettings.userAgentSuffix.get() : "")).c_str());
|
2016-10-14 12:44:15 +00:00
|
|
|
#if LIBCURL_VERSION_NUM >= 0x072b00
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_PIPEWAIT, 1);
|
2016-10-04 12:43:23 +00:00
|
|
|
#endif
|
2016-10-14 12:44:15 +00:00
|
|
|
#if LIBCURL_VERSION_NUM >= 0x072f00
|
2020-04-06 21:43:43 +00:00
|
|
|
if (fileTransferSettings.enableHttp2)
|
2016-09-14 14:34:37 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_HTTP_VERSION, CURL_HTTP_VERSION_2TLS);
|
2019-06-28 14:13:22 +00:00
|
|
|
else
|
|
|
|
curl_easy_setopt(req, CURLOPT_HTTP_VERSION, CURL_HTTP_VERSION_1_1);
|
2016-10-04 12:43:23 +00:00
|
|
|
#endif
|
2020-04-06 20:59:36 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_WRITEFUNCTION, TransferItem::writeCallbackWrapper);
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_WRITEDATA, this);
|
2020-04-06 20:59:36 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_HEADERFUNCTION, TransferItem::headerCallbackWrapper);
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_HEADERDATA, this);
|
|
|
|
|
|
|
|
curl_easy_setopt(req, CURLOPT_PROGRESSFUNCTION, progressCallbackWrapper);
|
|
|
|
curl_easy_setopt(req, CURLOPT_PROGRESSDATA, this);
|
|
|
|
curl_easy_setopt(req, CURLOPT_NOPROGRESS, 0);
|
|
|
|
|
|
|
|
curl_easy_setopt(req, CURLOPT_HTTPHEADER, requestHeaders);
|
|
|
|
|
|
|
|
if (request.head)
|
|
|
|
curl_easy_setopt(req, CURLOPT_NOBODY, 1);
|
|
|
|
|
2018-01-26 19:12:30 +00:00
|
|
|
if (request.data) {
|
|
|
|
curl_easy_setopt(req, CURLOPT_UPLOAD, 1L);
|
|
|
|
curl_easy_setopt(req, CURLOPT_READFUNCTION, readCallbackWrapper);
|
|
|
|
curl_easy_setopt(req, CURLOPT_READDATA, this);
|
|
|
|
curl_easy_setopt(req, CURLOPT_INFILESIZE_LARGE, (curl_off_t) request.data->length());
|
|
|
|
}
|
|
|
|
|
2017-06-12 14:44:43 +00:00
|
|
|
if (request.verifyTLS) {
|
2019-10-12 13:00:19 +00:00
|
|
|
debug("verify TLS: Nix CA file = '%s'", settings.caFile);
|
2017-06-12 14:44:43 +00:00
|
|
|
if (settings.caFile != "")
|
|
|
|
curl_easy_setopt(req, CURLOPT_CAINFO, settings.caFile.c_str());
|
|
|
|
} else {
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_SSL_VERIFYPEER, 0);
|
|
|
|
curl_easy_setopt(req, CURLOPT_SSL_VERIFYHOST, 0);
|
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_CONNECTTIMEOUT, fileTransferSettings.connectTimeout.get());
|
2017-04-19 12:54:52 +00:00
|
|
|
|
2017-09-01 14:51:26 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_LOW_SPEED_LIMIT, 1L);
|
2020-04-06 21:43:43 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_LOW_SPEED_TIME, fileTransferSettings.stalledDownloadTimeout.get());
|
2017-09-01 14:51:26 +00:00
|
|
|
|
2017-02-09 17:16:09 +00:00
|
|
|
/* If no file exist in the specified path, curl continues to work
|
2017-02-16 13:46:36 +00:00
|
|
|
anyway as if netrc support was disabled. */
|
2017-04-13 18:53:23 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_NETRC_FILE, settings.netrcFile.get().c_str());
|
2017-02-01 12:37:34 +00:00
|
|
|
curl_easy_setopt(req, CURLOPT_NETRC, CURL_NETRC_OPTIONAL);
|
|
|
|
|
2019-07-10 21:12:17 +00:00
|
|
|
if (writtenToSink)
|
|
|
|
curl_easy_setopt(req, CURLOPT_RESUME_FROM_LARGE, writtenToSink);
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
result.data = std::make_shared<std::string>();
|
2018-08-06 09:31:14 +00:00
|
|
|
result.bodySize = 0;
|
2015-10-07 15:31:50 +00:00
|
|
|
}
|
2015-05-05 12:39:48 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
void finish(CURLcode code)
|
|
|
|
{
|
2020-06-16 09:51:34 +00:00
|
|
|
auto httpStatus = getHTTPStatus();
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2019-05-22 21:36:29 +00:00
|
|
|
char * effectiveUriCStr;
|
|
|
|
curl_easy_getinfo(req, CURLINFO_EFFECTIVE_URL, &effectiveUriCStr);
|
|
|
|
if (effectiveUriCStr)
|
|
|
|
result.effectiveUri = effectiveUriCStr;
|
2016-02-29 17:15:20 +00:00
|
|
|
|
2018-06-05 12:37:26 +00:00
|
|
|
debug("finished %s of '%s'; curl status = %d, HTTP status = %d, body = %d bytes",
|
2018-08-06 09:31:14 +00:00
|
|
|
request.verb(), request.uri, code, httpStatus, result.bodySize);
|
2016-02-29 17:15:20 +00:00
|
|
|
|
2019-06-24 19:06:37 +00:00
|
|
|
if (decompressionSink) {
|
|
|
|
try {
|
|
|
|
decompressionSink->finish();
|
|
|
|
} catch (...) {
|
|
|
|
writeException = std::current_exception();
|
|
|
|
}
|
|
|
|
}
|
2018-08-06 13:40:29 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
if (code == CURLE_WRITE_ERROR && result.etag == request.expectedETag) {
|
|
|
|
code = CURLE_OK;
|
|
|
|
httpStatus = 304;
|
|
|
|
}
|
|
|
|
|
2018-08-06 13:40:29 +00:00
|
|
|
if (writeException)
|
|
|
|
failEx(writeException);
|
|
|
|
|
2020-06-17 16:58:59 +00:00
|
|
|
else if (code == CURLE_OK && successfulStatuses.count(httpStatus))
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
result.cached = httpStatus == 304;
|
2021-01-22 20:46:40 +00:00
|
|
|
|
|
|
|
// In 2021, GitHub responds to If-None-Match with 304,
|
|
|
|
// but omits ETag. We just use the If-None-Match etag
|
|
|
|
// since 304 implies they are the same.
|
|
|
|
if (httpStatus == 304 && result.etag == "")
|
|
|
|
result.etag = request.expectedETag;
|
|
|
|
|
2019-09-03 10:51:14 +00:00
|
|
|
act.progress(result.bodySize, result.bodySize);
|
2016-09-14 14:00:40 +00:00
|
|
|
done = true;
|
2019-09-03 10:51:14 +00:00
|
|
|
callback(std::move(result));
|
2018-08-06 13:40:29 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
else {
|
2017-09-25 19:36:01 +00:00
|
|
|
// We treat most errors as transient, but won't retry when hopeless
|
|
|
|
Error err = Transient;
|
|
|
|
|
2018-06-05 14:03:32 +00:00
|
|
|
if (httpStatus == 404 || httpStatus == 410 || code == CURLE_FILE_COULDNT_READ_FILE) {
|
2017-09-25 19:36:01 +00:00
|
|
|
// The file is definitely not there
|
|
|
|
err = NotFound;
|
|
|
|
} else if (httpStatus == 401 || httpStatus == 403 || httpStatus == 407) {
|
|
|
|
// Don't retry on authentication/authorization failures
|
|
|
|
err = Forbidden;
|
2020-01-29 10:47:39 +00:00
|
|
|
} else if (httpStatus >= 400 && httpStatus < 500 && httpStatus != 408 && httpStatus != 429) {
|
2017-09-25 19:36:01 +00:00
|
|
|
// Most 4xx errors are client errors and are probably not worth retrying:
|
|
|
|
// * 408 means the server timed out waiting for us, so we try again
|
2020-01-29 10:47:39 +00:00
|
|
|
// * 429 means too many requests, so we retry (with a delay)
|
2017-09-25 19:36:01 +00:00
|
|
|
err = Misc;
|
|
|
|
} else if (httpStatus == 501 || httpStatus == 505 || httpStatus == 511) {
|
|
|
|
// Let's treat most 5xx (server) errors as transient, except for a handful:
|
|
|
|
// * 501 not implemented
|
|
|
|
// * 505 http version not supported
|
|
|
|
// * 511 we're behind a captive portal
|
|
|
|
err = Misc;
|
|
|
|
} else {
|
|
|
|
// Don't bother retrying on certain cURL errors either
|
|
|
|
switch (code) {
|
|
|
|
case CURLE_FAILED_INIT:
|
2018-01-19 13:05:08 +00:00
|
|
|
case CURLE_URL_MALFORMAT:
|
2017-09-25 19:36:01 +00:00
|
|
|
case CURLE_NOT_BUILT_IN:
|
|
|
|
case CURLE_REMOTE_ACCESS_DENIED:
|
|
|
|
case CURLE_FILE_COULDNT_READ_FILE:
|
|
|
|
case CURLE_FUNCTION_NOT_FOUND:
|
|
|
|
case CURLE_ABORTED_BY_CALLBACK:
|
|
|
|
case CURLE_BAD_FUNCTION_ARGUMENT:
|
|
|
|
case CURLE_INTERFACE_FAILED:
|
|
|
|
case CURLE_UNKNOWN_OPTION:
|
2018-03-16 11:03:18 +00:00
|
|
|
case CURLE_SSL_CACERT_BADFILE:
|
2018-06-18 08:36:19 +00:00
|
|
|
case CURLE_TOO_MANY_REDIRECTS:
|
2018-08-06 13:40:29 +00:00
|
|
|
case CURLE_WRITE_ERROR:
|
2020-04-04 19:27:39 +00:00
|
|
|
case CURLE_UNSUPPORTED_PROTOCOL:
|
2018-01-29 15:22:59 +00:00
|
|
|
err = Misc;
|
|
|
|
break;
|
2017-09-25 19:36:01 +00:00
|
|
|
default: // Shut up warnings
|
2018-01-29 15:22:59 +00:00
|
|
|
break;
|
2017-09-25 19:36:01 +00:00
|
|
|
}
|
|
|
|
}
|
2016-02-29 17:15:20 +00:00
|
|
|
|
2019-07-10 17:46:15 +00:00
|
|
|
attempt++;
|
|
|
|
|
2020-06-16 19:14:11 +00:00
|
|
|
std::shared_ptr<std::string> response;
|
2020-07-15 16:58:38 +00:00
|
|
|
if (errorSink)
|
|
|
|
response = errorSink->s;
|
2019-07-10 17:46:15 +00:00
|
|
|
auto exc =
|
2016-09-16 16:54:14 +00:00
|
|
|
code == CURLE_ABORTED_BY_CALLBACK && _isInterrupted
|
2020-06-18 18:08:02 +00:00
|
|
|
? FileTransferError(Interrupted, response, "%s of '%s' was interrupted", request.verb(), request.uri)
|
2016-09-16 16:54:14 +00:00
|
|
|
: httpStatus != 0
|
2020-04-06 21:43:43 +00:00
|
|
|
? FileTransferError(err,
|
2020-07-16 16:04:10 +00:00
|
|
|
response,
|
2020-06-09 12:20:22 +00:00
|
|
|
fmt("unable to %s '%s': HTTP error %d ('%s')",
|
|
|
|
request.verb(), request.uri, httpStatus, statusMsg)
|
2018-06-01 12:14:22 +00:00
|
|
|
+ (code == CURLE_OK ? "" : fmt(" (curl error: %s)", curl_easy_strerror(code)))
|
|
|
|
)
|
2020-04-06 21:43:43 +00:00
|
|
|
: FileTransferError(err,
|
2020-07-16 16:04:10 +00:00
|
|
|
response,
|
2018-06-01 12:14:22 +00:00
|
|
|
fmt("unable to %s '%s': %s (%d)",
|
2019-07-10 17:46:15 +00:00
|
|
|
request.verb(), request.uri, curl_easy_strerror(code), code));
|
|
|
|
|
|
|
|
/* If this is a transient error, then maybe retry the
|
2019-07-10 21:12:17 +00:00
|
|
|
download after a while. If we're writing to a
|
|
|
|
sink, we can only retry if the server supports
|
|
|
|
ranged requests. */
|
|
|
|
if (err == Transient
|
|
|
|
&& attempt < request.tries
|
|
|
|
&& (!this->request.dataCallback
|
|
|
|
|| writtenToSink == 0
|
|
|
|
|| (acceptRanges && encoding.empty())))
|
|
|
|
{
|
2020-04-06 21:43:43 +00:00
|
|
|
int ms = request.baseRetryTimeMs * std::pow(2.0f, attempt - 1 + std::uniform_real_distribution<>(0.0, 0.5)(fileTransfer.mt19937));
|
2019-07-10 21:12:17 +00:00
|
|
|
if (writtenToSink)
|
|
|
|
warn("%s; retrying from offset %d in %d ms", exc.what(), writtenToSink, ms);
|
|
|
|
else
|
|
|
|
warn("%s; retrying in %d ms", exc.what(), ms);
|
2019-07-10 17:46:15 +00:00
|
|
|
embargo = std::chrono::steady_clock::now() + std::chrono::milliseconds(ms);
|
2020-04-06 21:43:43 +00:00
|
|
|
fileTransfer.enqueueItem(shared_from_this());
|
2019-07-10 17:46:15 +00:00
|
|
|
}
|
|
|
|
else
|
|
|
|
fail(exc);
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
struct State
|
|
|
|
{
|
2019-07-10 17:46:15 +00:00
|
|
|
struct EmbargoComparator {
|
2020-04-06 20:59:36 +00:00
|
|
|
bool operator() (const std::shared_ptr<TransferItem> & i1, const std::shared_ptr<TransferItem> & i2) {
|
2019-07-10 17:46:15 +00:00
|
|
|
return i1->embargo > i2->embargo;
|
|
|
|
}
|
|
|
|
};
|
2016-09-14 14:00:40 +00:00
|
|
|
bool quit = false;
|
2020-04-06 20:59:36 +00:00
|
|
|
std::priority_queue<std::shared_ptr<TransferItem>, std::vector<std::shared_ptr<TransferItem>>, EmbargoComparator> incoming;
|
2016-09-14 14:00:40 +00:00
|
|
|
};
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
Sync<State> state_;
|
2015-05-05 12:39:48 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
/* We can't use a std::condition_variable to wake up the curl
|
|
|
|
thread, because it only monitors file descriptors. So use a
|
|
|
|
pipe instead. */
|
|
|
|
Pipe wakeupPipe;
|
2016-02-22 12:13:19 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
std::thread workerThread;
|
2015-10-21 13:03:29 +00:00
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
curlFileTransfer()
|
2019-07-10 17:46:15 +00:00
|
|
|
: mt19937(rd())
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
static std::once_flag globalInit;
|
|
|
|
std::call_once(globalInit, curl_global_init, CURL_GLOBAL_ALL);
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
curlm = curl_multi_init();
|
2015-10-21 12:59:01 +00:00
|
|
|
|
2017-12-10 01:02:21 +00:00
|
|
|
#if LIBCURL_VERSION_NUM >= 0x072b00 // Multiplex requires >= 7.43.0
|
2016-09-14 14:00:40 +00:00
|
|
|
curl_multi_setopt(curlm, CURLMOPT_PIPELINING, CURLPIPE_MULTIPLEX);
|
2016-10-04 12:43:23 +00:00
|
|
|
#endif
|
2017-12-10 01:02:21 +00:00
|
|
|
#if LIBCURL_VERSION_NUM >= 0x071e00 // Max connections requires >= 7.30.0
|
2016-09-14 14:34:37 +00:00
|
|
|
curl_multi_setopt(curlm, CURLMOPT_MAX_TOTAL_CONNECTIONS,
|
2020-04-06 21:43:43 +00:00
|
|
|
fileTransferSettings.httpConnections.get());
|
2017-12-10 01:02:21 +00:00
|
|
|
#endif
|
2016-09-14 14:34:37 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
wakeupPipe.create();
|
|
|
|
fcntl(wakeupPipe.readSide.get(), F_SETFL, O_NONBLOCK);
|
|
|
|
|
|
|
|
workerThread = std::thread([&]() { workerThreadEntry(); });
|
|
|
|
}
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
~curlFileTransfer()
|
2017-01-17 17:21:02 +00:00
|
|
|
{
|
|
|
|
stopWorkerThread();
|
|
|
|
|
|
|
|
workerThread.join();
|
|
|
|
|
|
|
|
if (curlm) curl_multi_cleanup(curlm);
|
|
|
|
}
|
|
|
|
|
|
|
|
void stopWorkerThread()
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
/* Signal the worker thread to exit. */
|
|
|
|
{
|
|
|
|
auto state(state_.lock());
|
|
|
|
state->quit = true;
|
2015-04-09 10:12:50 +00:00
|
|
|
}
|
2017-01-17 17:21:02 +00:00
|
|
|
writeFull(wakeupPipe.writeSide.get(), " ", false);
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
2016-02-29 17:15:20 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
void workerThreadMain()
|
|
|
|
{
|
2017-01-17 17:21:02 +00:00
|
|
|
/* Cause this thread to be notified on SIGINT. */
|
|
|
|
auto callback = createInterruptCallback([&]() {
|
|
|
|
stopWorkerThread();
|
|
|
|
});
|
|
|
|
|
2020-04-06 20:59:36 +00:00
|
|
|
std::map<CURL *, std::shared_ptr<TransferItem>> items;
|
2016-09-14 14:00:40 +00:00
|
|
|
|
2016-10-19 13:02:38 +00:00
|
|
|
bool quit = false;
|
2016-09-14 14:00:40 +00:00
|
|
|
|
|
|
|
std::chrono::steady_clock::time_point nextWakeup;
|
|
|
|
|
|
|
|
while (!quit) {
|
|
|
|
checkInterrupt();
|
|
|
|
|
|
|
|
/* Let curl do its thing. */
|
|
|
|
int running;
|
|
|
|
CURLMcode mc = curl_multi_perform(curlm, &running);
|
|
|
|
if (mc != CURLM_OK)
|
2020-04-21 23:07:07 +00:00
|
|
|
throw nix::Error("unexpected error from curl_multi_perform(): %s", curl_multi_strerror(mc));
|
2016-09-14 14:00:40 +00:00
|
|
|
|
|
|
|
/* Set the promises of any finished requests. */
|
|
|
|
CURLMsg * msg;
|
|
|
|
int left;
|
|
|
|
while ((msg = curl_multi_info_read(curlm, &left))) {
|
|
|
|
if (msg->msg == CURLMSG_DONE) {
|
|
|
|
auto i = items.find(msg->easy_handle);
|
|
|
|
assert(i != items.end());
|
|
|
|
i->second->finish(msg->data.result);
|
|
|
|
curl_multi_remove_handle(curlm, i->second->req);
|
|
|
|
i->second->active = false;
|
|
|
|
items.erase(i);
|
|
|
|
}
|
|
|
|
}
|
2015-10-07 15:31:50 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
/* Wait for activity, including wakeup events. */
|
|
|
|
int numfds = 0;
|
|
|
|
struct curl_waitfd extraFDs[1];
|
|
|
|
extraFDs[0].fd = wakeupPipe.readSide.get();
|
|
|
|
extraFDs[0].events = CURL_WAIT_POLLIN;
|
|
|
|
extraFDs[0].revents = 0;
|
2018-05-24 21:23:05 +00:00
|
|
|
long maxSleepTimeMs = items.empty() ? 10000 : 100;
|
2016-09-14 14:00:40 +00:00
|
|
|
auto sleepTimeMs =
|
|
|
|
nextWakeup != std::chrono::steady_clock::time_point()
|
|
|
|
? std::max(0, (int) std::chrono::duration_cast<std::chrono::milliseconds>(nextWakeup - std::chrono::steady_clock::now()).count())
|
2018-05-24 21:23:05 +00:00
|
|
|
: maxSleepTimeMs;
|
2017-01-24 12:15:24 +00:00
|
|
|
vomit("download thread waiting for %d ms", sleepTimeMs);
|
2016-09-14 14:00:40 +00:00
|
|
|
mc = curl_multi_wait(curlm, extraFDs, 1, sleepTimeMs, &numfds);
|
|
|
|
if (mc != CURLM_OK)
|
2020-04-21 23:07:07 +00:00
|
|
|
throw nix::Error("unexpected error from curl_multi_wait(): %s", curl_multi_strerror(mc));
|
2016-09-14 14:00:40 +00:00
|
|
|
|
|
|
|
nextWakeup = std::chrono::steady_clock::time_point();
|
|
|
|
|
2019-07-10 17:46:15 +00:00
|
|
|
/* Add new curl requests from the incoming requests queue,
|
|
|
|
except for requests that are embargoed (waiting for a
|
|
|
|
retry timeout to expire). */
|
2016-09-14 14:00:40 +00:00
|
|
|
if (extraFDs[0].revents & CURL_WAIT_POLLIN) {
|
|
|
|
char buf[1024];
|
|
|
|
auto res = read(extraFDs[0].fd, buf, sizeof(buf));
|
|
|
|
if (res == -1 && errno != EINTR)
|
|
|
|
throw SysError("reading curl wakeup socket");
|
|
|
|
}
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2020-04-06 20:59:36 +00:00
|
|
|
std::vector<std::shared_ptr<TransferItem>> incoming;
|
2019-07-10 17:46:15 +00:00
|
|
|
auto now = std::chrono::steady_clock::now();
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
auto state(state_.lock());
|
2019-07-10 17:46:15 +00:00
|
|
|
while (!state->incoming.empty()) {
|
|
|
|
auto item = state->incoming.top();
|
|
|
|
if (item->embargo <= now) {
|
|
|
|
incoming.push_back(item);
|
|
|
|
state->incoming.pop();
|
|
|
|
} else {
|
|
|
|
if (nextWakeup == std::chrono::steady_clock::time_point()
|
|
|
|
|| item->embargo < nextWakeup)
|
|
|
|
nextWakeup = item->embargo;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2016-09-14 14:00:40 +00:00
|
|
|
quit = state->quit;
|
|
|
|
}
|
2016-08-11 15:34:43 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
for (auto & item : incoming) {
|
2018-06-05 12:37:26 +00:00
|
|
|
debug("starting %s of %s", item->request.verb(), item->request.uri);
|
2016-09-14 14:00:40 +00:00
|
|
|
item->init();
|
|
|
|
curl_multi_add_handle(curlm, item->req);
|
|
|
|
item->active = true;
|
|
|
|
items[item->req] = item;
|
|
|
|
}
|
|
|
|
}
|
2015-04-09 10:12:50 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
debug("download thread shutting down");
|
2015-04-09 10:12:50 +00:00
|
|
|
}
|
2016-02-29 17:15:20 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
void workerThreadEntry()
|
2016-02-29 17:15:20 +00:00
|
|
|
{
|
2016-09-14 14:00:40 +00:00
|
|
|
try {
|
|
|
|
workerThreadMain();
|
2016-09-16 16:54:14 +00:00
|
|
|
} catch (nix::Interrupted & e) {
|
2016-09-14 14:00:40 +00:00
|
|
|
} catch (std::exception & e) {
|
2021-01-20 23:27:36 +00:00
|
|
|
printError("unexpected error in download thread: %s", e.what());
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
2016-08-10 14:06:33 +00:00
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
auto state(state_.lock());
|
2019-07-10 17:46:15 +00:00
|
|
|
while (!state->incoming.empty()) state->incoming.pop();
|
2016-09-14 14:00:40 +00:00
|
|
|
state->quit = true;
|
2016-08-10 14:06:33 +00:00
|
|
|
}
|
2016-02-29 17:15:20 +00:00
|
|
|
}
|
2016-09-14 14:00:40 +00:00
|
|
|
|
2020-04-06 20:59:36 +00:00
|
|
|
void enqueueItem(std::shared_ptr<TransferItem> item)
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
2018-06-05 12:44:26 +00:00
|
|
|
if (item->request.data
|
|
|
|
&& !hasPrefix(item->request.uri, "http://")
|
|
|
|
&& !hasPrefix(item->request.uri, "https://"))
|
|
|
|
throw nix::Error("uploading to '%s' is not supported", item->request.uri);
|
|
|
|
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
|
|
|
auto state(state_.lock());
|
|
|
|
if (state->quit)
|
|
|
|
throw nix::Error("cannot enqueue download request because the download thread is shutting down");
|
2019-07-10 17:46:15 +00:00
|
|
|
state->incoming.push(item);
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
writeFull(wakeupPipe.writeSide.get(), " ");
|
|
|
|
}
|
|
|
|
|
2021-04-23 06:30:05 +00:00
|
|
|
#if ENABLE_S3
|
2019-02-25 16:56:52 +00:00
|
|
|
std::tuple<std::string, std::string, Store::Params> parseS3Uri(std::string uri)
|
|
|
|
{
|
|
|
|
auto [path, params] = splitUriAndParams(uri);
|
|
|
|
|
|
|
|
auto slash = path.find('/', 5); // 5 is the length of "s3://" prefix
|
2019-02-26 10:04:18 +00:00
|
|
|
if (slash == std::string::npos)
|
|
|
|
throw nix::Error("bad S3 URI '%s'", path);
|
2019-02-25 16:56:52 +00:00
|
|
|
|
|
|
|
std::string bucketName(path, 5, slash - 5);
|
|
|
|
std::string key(path, slash + 1);
|
|
|
|
|
|
|
|
return {bucketName, key, params};
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
void enqueueFileTransfer(const FileTransferRequest & request,
|
|
|
|
Callback<FileTransferResult> callback) override
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
2017-02-14 13:20:00 +00:00
|
|
|
/* Ugly hack to support s3:// URIs. */
|
|
|
|
if (hasPrefix(request.uri, "s3://")) {
|
|
|
|
// FIXME: do this on a worker thread
|
2018-03-27 20:16:01 +00:00
|
|
|
try {
|
2021-04-23 06:30:05 +00:00
|
|
|
#if ENABLE_S3
|
2019-02-25 16:56:52 +00:00
|
|
|
auto [bucketName, key, params] = parseS3Uri(request.uri);
|
2019-02-21 13:26:50 +00:00
|
|
|
|
2019-12-05 18:11:09 +00:00
|
|
|
std::string profile = get(params, "profile").value_or("");
|
|
|
|
std::string region = get(params, "region").value_or(Aws::Region::US_EAST_1);
|
|
|
|
std::string scheme = get(params, "scheme").value_or("");
|
|
|
|
std::string endpoint = get(params, "endpoint").value_or("");
|
2019-02-21 13:26:50 +00:00
|
|
|
|
|
|
|
S3Helper s3Helper(profile, region, scheme, endpoint);
|
|
|
|
|
2017-02-14 13:20:00 +00:00
|
|
|
// FIXME: implement ETag
|
|
|
|
auto s3Res = s3Helper.getObject(bucketName, key);
|
2020-04-06 21:43:43 +00:00
|
|
|
FileTransferResult res;
|
2017-02-14 13:20:00 +00:00
|
|
|
if (!s3Res.data)
|
2020-06-18 18:08:02 +00:00
|
|
|
throw FileTransferError(NotFound, nullptr, "S3 object '%s' does not exist", request.uri);
|
2017-02-14 13:20:00 +00:00
|
|
|
res.data = s3Res.data;
|
2018-03-27 20:16:01 +00:00
|
|
|
callback(std::move(res));
|
2017-02-14 13:20:00 +00:00
|
|
|
#else
|
2017-07-30 11:27:57 +00:00
|
|
|
throw nix::Error("cannot download '%s' because Nix is not built with S3 support", request.uri);
|
2017-02-14 13:20:00 +00:00
|
|
|
#endif
|
2018-03-27 20:16:01 +00:00
|
|
|
} catch (...) { callback.rethrow(); }
|
2017-02-14 13:20:00 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2020-04-06 20:59:36 +00:00
|
|
|
enqueueItem(std::make_shared<TransferItem>(*this, request, std::move(callback)));
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
2015-04-09 10:12:50 +00:00
|
|
|
};
|
|
|
|
|
2021-10-07 21:58:02 +00:00
|
|
|
ref<curlFileTransfer> makeCurlFileTransfer()
|
|
|
|
{
|
|
|
|
return make_ref<curlFileTransfer>();
|
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
ref<FileTransfer> getFileTransfer()
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
2021-10-07 21:58:02 +00:00
|
|
|
static ref<curlFileTransfer> fileTransfer = makeCurlFileTransfer();
|
|
|
|
|
|
|
|
// this has to be done in its own scope to make sure that the lock is released
|
|
|
|
// before creating a new fileTransfer instance.
|
|
|
|
auto needsRecreation = [&]() -> bool {
|
|
|
|
auto state = fileTransfer->state_.lock();
|
|
|
|
return state->quit;
|
|
|
|
};
|
|
|
|
|
|
|
|
if (needsRecreation()) {
|
|
|
|
fileTransfer = makeCurlFileTransfer();
|
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
return fileTransfer;
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
ref<FileTransfer> makeFileTransfer()
|
2015-04-09 10:12:50 +00:00
|
|
|
{
|
2021-10-07 21:58:02 +00:00
|
|
|
return makeCurlFileTransfer();
|
2015-04-09 10:12:50 +00:00
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
std::future<FileTransferResult> FileTransfer::enqueueFileTransfer(const FileTransferRequest & request)
|
2016-09-16 16:54:14 +00:00
|
|
|
{
|
2020-04-06 21:43:43 +00:00
|
|
|
auto promise = std::make_shared<std::promise<FileTransferResult>>();
|
|
|
|
enqueueFileTransfer(request,
|
|
|
|
{[promise](std::future<FileTransferResult> fut) {
|
2018-03-27 20:16:01 +00:00
|
|
|
try {
|
|
|
|
promise->set_value(fut.get());
|
|
|
|
} catch (...) {
|
|
|
|
promise->set_exception(std::current_exception());
|
|
|
|
}
|
|
|
|
}});
|
2016-09-16 16:54:14 +00:00
|
|
|
return promise->get_future();
|
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
FileTransferResult FileTransfer::download(const FileTransferRequest & request)
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
2020-04-06 21:43:43 +00:00
|
|
|
return enqueueFileTransfer(request).get();
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
FileTransferResult FileTransfer::upload(const FileTransferRequest & request)
|
2016-09-14 14:00:40 +00:00
|
|
|
{
|
2020-04-06 21:34:31 +00:00
|
|
|
/* Note: this method is the same as download, but helps in readability */
|
2020-04-06 21:43:43 +00:00
|
|
|
return enqueueFileTransfer(request).get();
|
2016-09-14 14:00:40 +00:00
|
|
|
}
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
void FileTransfer::download(FileTransferRequest && request, Sink & sink)
|
2018-03-27 22:01:47 +00:00
|
|
|
{
|
|
|
|
/* Note: we can't call 'sink' via request.dataCallback, because
|
2020-04-06 21:43:43 +00:00
|
|
|
that would cause the sink to execute on the fileTransfer
|
2018-03-27 22:01:47 +00:00
|
|
|
thread. If 'sink' is a coroutine, this will fail. Also, if the
|
|
|
|
sink is expensive (e.g. one that does decompression and writing
|
|
|
|
to the Nix store), it would stall the download thread too much.
|
|
|
|
Therefore we use a buffer to communicate data between the
|
|
|
|
download thread and the calling thread. */
|
|
|
|
|
|
|
|
struct State {
|
|
|
|
bool quit = false;
|
|
|
|
std::exception_ptr exc;
|
|
|
|
std::string data;
|
|
|
|
std::condition_variable avail, request;
|
|
|
|
};
|
|
|
|
|
|
|
|
auto _state = std::make_shared<Sync<State>>();
|
|
|
|
|
|
|
|
/* In case of an exception, wake up the download thread. FIXME:
|
|
|
|
abort the download request. */
|
|
|
|
Finally finally([&]() {
|
|
|
|
auto state(_state->lock());
|
|
|
|
state->quit = true;
|
|
|
|
state->request.notify_one();
|
|
|
|
});
|
|
|
|
|
2020-12-02 13:00:43 +00:00
|
|
|
request.dataCallback = [_state](std::string_view data) {
|
2018-03-27 22:01:47 +00:00
|
|
|
|
|
|
|
auto state(_state->lock());
|
|
|
|
|
|
|
|
if (state->quit) return;
|
|
|
|
|
|
|
|
/* If the buffer is full, then go to sleep until the calling
|
|
|
|
thread wakes us up (i.e. when it has removed data from the
|
2018-09-26 19:43:17 +00:00
|
|
|
buffer). We don't wait forever to prevent stalling the
|
|
|
|
download thread. (Hopefully sleeping will throttle the
|
|
|
|
sender.) */
|
|
|
|
if (state->data.size() > 1024 * 1024) {
|
2018-03-27 22:01:47 +00:00
|
|
|
debug("download buffer is full; going to sleep");
|
2018-09-26 19:43:17 +00:00
|
|
|
state.wait_for(state->request, std::chrono::seconds(10));
|
2018-03-27 22:01:47 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Append data to the buffer and wake up the calling
|
|
|
|
thread. */
|
2020-12-02 13:00:43 +00:00
|
|
|
state->data.append(data);
|
2018-03-27 22:01:47 +00:00
|
|
|
state->avail.notify_one();
|
|
|
|
};
|
|
|
|
|
2020-04-06 21:43:43 +00:00
|
|
|
enqueueFileTransfer(request,
|
|
|
|
{[_state](std::future<FileTransferResult> fut) {
|
2018-03-27 22:01:47 +00:00
|
|
|
auto state(_state->lock());
|
|
|
|
state->quit = true;
|
|
|
|
try {
|
|
|
|
fut.get();
|
|
|
|
} catch (...) {
|
|
|
|
state->exc = std::current_exception();
|
|
|
|
}
|
|
|
|
state->avail.notify_one();
|
|
|
|
state->request.notify_one();
|
|
|
|
}});
|
|
|
|
|
|
|
|
while (true) {
|
|
|
|
checkInterrupt();
|
|
|
|
|
2018-09-26 19:43:17 +00:00
|
|
|
std::string chunk;
|
|
|
|
|
|
|
|
/* Grab data if available, otherwise wait for the download
|
|
|
|
thread to wake us up. */
|
|
|
|
{
|
|
|
|
auto state(_state->lock());
|
|
|
|
|
|
|
|
while (state->data.empty()) {
|
2018-07-31 15:26:04 +00:00
|
|
|
|
2018-09-26 19:43:17 +00:00
|
|
|
if (state->quit) {
|
|
|
|
if (state->exc) std::rethrow_exception(state->exc);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
state.wait(state->avail);
|
2018-07-31 15:26:04 +00:00
|
|
|
}
|
|
|
|
|
2018-09-26 19:43:17 +00:00
|
|
|
chunk = std::move(state->data);
|
2018-03-27 22:01:47 +00:00
|
|
|
|
|
|
|
state->request.notify_one();
|
|
|
|
}
|
2018-09-26 19:43:17 +00:00
|
|
|
|
|
|
|
/* Flush the data to the sink and wake up the download thread
|
|
|
|
if it's blocked on a full buffer. We don't hold the state
|
|
|
|
lock while doing this to prevent blocking the download
|
|
|
|
thread if sink() takes a long time. */
|
2020-12-02 13:00:43 +00:00
|
|
|
sink(chunk);
|
2018-03-27 22:01:47 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-06-18 14:48:45 +00:00
|
|
|
template<typename... Args>
|
|
|
|
FileTransferError::FileTransferError(FileTransfer::Error error, std::shared_ptr<string> response, const Args & ... args)
|
|
|
|
: Error(args...), error(error), response(response)
|
|
|
|
{
|
|
|
|
const auto hf = hintfmt(args...);
|
2020-07-20 19:56:52 +00:00
|
|
|
// FIXME: Due to https://github.com/NixOS/nix/issues/3841 we don't know how
|
|
|
|
// to print different messages for different verbosity levels. For now
|
|
|
|
// we add some heuristics for detecting when we want to show the response.
|
2021-01-20 23:27:36 +00:00
|
|
|
if (response && (response->size() < 1024 || response->find("<html>") != string::npos))
|
2021-01-27 13:04:49 +00:00
|
|
|
err.msg = hintfmt("%1%\n\nresponse body:\n\n%2%", normaltxt(hf.str()), chomp(*response));
|
2021-01-20 23:27:36 +00:00
|
|
|
else
|
|
|
|
err.msg = hf;
|
2020-06-18 14:48:45 +00:00
|
|
|
}
|
|
|
|
|
2015-05-06 12:54:31 +00:00
|
|
|
bool isUri(const string & s)
|
|
|
|
{
|
2016-04-14 14:27:48 +00:00
|
|
|
if (s.compare(0, 8, "channel:") == 0) return true;
|
2015-05-06 12:54:31 +00:00
|
|
|
size_t pos = s.find("://");
|
|
|
|
if (pos == string::npos) return false;
|
|
|
|
string scheme(s, 0, pos);
|
2017-11-01 15:32:53 +00:00
|
|
|
return scheme == "http" || scheme == "https" || scheme == "file" || scheme == "channel" || scheme == "git" || scheme == "s3" || scheme == "ssh";
|
2015-05-06 12:54:31 +00:00
|
|
|
}
|
|
|
|
|
2019-07-10 17:46:15 +00:00
|
|
|
|
2015-04-09 10:12:50 +00:00
|
|
|
}
|