diff options
author | Lennart Poettering <lennart@poettering.net> | 2014-12-19 01:59:52 +0100 |
---|---|---|
committer | Lennart Poettering <lennart@poettering.net> | 2014-12-19 02:08:14 +0100 |
commit | 72648326ea6d3e68cdb0b5890df737047d031a41 (patch) | |
tree | 951f4f2d744a03f24f67b03b1b7d59e3c2d0eb5a /src | |
parent | util: add generalization of verb parsing for command line tools (diff) | |
download | systemd-72648326ea6d3e68cdb0b5890df737047d031a41.tar.xz systemd-72648326ea6d3e68cdb0b5890df737047d031a41.zip |
import: add new minimal tool "systemd-import" for pulling down foreign containers and install them locally
This adds a simply but powerful tool for downloading container images
from the most popular container solution used today. Use it like
this:
# systemd-import pull-dck mattdm/fedora
# systemd-nspawn -M fedora
This will donwload the layers for "mattdm/fedora", and make them
available locally as /var/lib/container/fedora.
The tool is pretty complete, as long as it's only about pulling down
images, or updating them. Pushing or searching is not supported yet.
Diffstat (limited to '')
l--------- | src/import/Makefile | 1 | ||||
-rw-r--r-- | src/import/aufs-util.c | 73 | ||||
-rw-r--r-- | src/import/aufs-util.h | 24 | ||||
-rw-r--r-- | src/import/curl-util.c | 415 | ||||
-rw-r--r-- | src/import/curl-util.h | 56 | ||||
-rw-r--r-- | src/import/import-dck.c | 1156 | ||||
-rw-r--r-- | src/import/import-dck.h | 39 | ||||
-rw-r--r-- | src/import/import.c | 213 | ||||
-rw-r--r-- | src/shared/util.h | 2 |
9 files changed, 1979 insertions, 0 deletions
diff --git a/src/import/Makefile b/src/import/Makefile new file mode 120000 index 0000000000..d0b0e8e008 --- /dev/null +++ b/src/import/Makefile @@ -0,0 +1 @@ +../Makefile
\ No newline at end of file diff --git a/src/import/aufs-util.c b/src/import/aufs-util.c new file mode 100644 index 0000000000..c1301cdb4a --- /dev/null +++ b/src/import/aufs-util.c @@ -0,0 +1,73 @@ +/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ + +/*** + This file is part of systemd. + + Copyright 2014 Lennart Poettering + + systemd is free software; you can redistribute it and/or modify it + under the terms of the GNU Lesser General Public License as published by + the Free Software Foundation; either version 2.1 of the License, or + (at your option) any later version. + + systemd is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public License + along with systemd; If not, see <http://www.gnu.org/licenses/>. +***/ + +#include <ftw.h> + +#include "util.h" +#include "aufs-util.h" + +static int nftw_cb( + const char *fpath, + const struct stat *sb, + int flag, + struct FTW *ftwbuf) { + + const char *fn, *original; + char *p; + int r; + + fn = fpath + ftwbuf->base; + + /* We remove all whiteout files, and all whiteouts */ + + original = startswith(fn, ".wh."); + if (!original) + return FTW_CONTINUE; + + log_debug("Removing whiteout indicator %s.", fpath); + r = rm_rf_dangerous(fpath, false, true, false); + if (r < 0) + return FTW_STOP; + + if (!startswith(fn, ".wh..wh.")) { + + p = alloca(ftwbuf->base + strlen(original)); + strcpy(mempcpy(p, fpath, ftwbuf->base), original); + + log_debug("Removing deleted file %s.", p); + r = rm_rf_dangerous(p, false, true, false); + if (r < 0) + return FTW_STOP; + } + + return FTW_CONTINUE; +} + +int aufs_resolve(const char *path) { + int r; + + errno = 0; + r = nftw(path, nftw_cb, 64, FTW_MOUNT|FTW_PHYS|FTW_ACTIONRETVAL); + if (r == FTW_STOP) + return errno ? -errno : -EIO; + + return 0; +} diff --git a/src/import/aufs-util.h b/src/import/aufs-util.h new file mode 100644 index 0000000000..712fb81cee --- /dev/null +++ b/src/import/aufs-util.h @@ -0,0 +1,24 @@ +/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ + +#pragma once + +/*** + This file is part of systemd. + + Copyright 2014 Lennart Poettering + + systemd is free software; you can redistribute it and/or modify it + under the terms of the GNU Lesser General Public License as published by + the Free Software Foundation; either version 2.1 of the License, or + (at your option) any later version. + + systemd is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public License + along with systemd; If not, see <http://www.gnu.org/licenses/>. +***/ + +int aufs_resolve(const char *path); diff --git a/src/import/curl-util.c b/src/import/curl-util.c new file mode 100644 index 0000000000..eaaebaef42 --- /dev/null +++ b/src/import/curl-util.c @@ -0,0 +1,415 @@ +/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ + +/*** + This file is part of systemd. + + Copyright 2014 Lennart Poettering + + systemd is free software; you can redistribute it and/or modify it + under the terms of the GNU Lesser General Public License as published by + the Free Software Foundation; either version 2.1 of the License, or + (at your option) any later version. + + systemd is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public License + along with systemd; If not, see <http://www.gnu.org/licenses/>. +***/ + +#include "curl-util.h" + +static void curl_glue_check_finished(CurlGlue *g) { + CURLMsg *msg; + int k = 0; + + assert(g); + + msg = curl_multi_info_read(g->curl, &k); + if (!msg) + return; + + if (msg->msg != CURLMSG_DONE) + return; + + if (g->on_finished) + g->on_finished(g, msg->easy_handle, msg->data.result); +} + +static int curl_glue_on_io(sd_event_source *s, int fd, uint32_t revents, void *userdata) { + CurlGlue *g = userdata; + int action, k = 0, translated_fd; + + assert(s); + assert(g); + + translated_fd = PTR_TO_INT(hashmap_get(g->translate_fds, INT_TO_PTR(fd+1))); + assert(translated_fd > 0); + translated_fd--; + + if ((revents & (EPOLLIN|EPOLLOUT)) == (EPOLLIN|EPOLLOUT)) + action = CURL_POLL_INOUT; + else if (revents & EPOLLIN) + action = CURL_POLL_IN; + else if (revents & EPOLLOUT) + action = CURL_POLL_OUT; + else + action = 0; + + if (curl_multi_socket_action(g->curl, translated_fd, action, &k) < 0) { + log_debug("Failed to propagate IO event."); + return -EINVAL; + } + + curl_glue_check_finished(g); + return 0; +} + +static int curl_glue_socket_callback(CURLM *curl, curl_socket_t s, int action, void *userdata, void *socketp) { + sd_event_source *io; + CurlGlue *g = userdata; + uint32_t events = 0; + int r; + + assert(curl); + assert(g); + + io = hashmap_get(g->ios, INT_TO_PTR(s+1)); + + if (action == CURL_POLL_REMOVE) { + if (io) { + int fd; + + fd = sd_event_source_get_io_fd(io); + assert(fd >= 0); + + sd_event_source_set_enabled(io, SD_EVENT_OFF); + sd_event_source_unref(io); + + hashmap_remove(g->ios, INT_TO_PTR(s+1)); + hashmap_remove(g->translate_fds, INT_TO_PTR(fd+1)); + + safe_close(fd); + } + + return 0; + } + + r = hashmap_ensure_allocated(&g->ios, &trivial_hash_ops); + if (r < 0) { + log_oom(); + return -1; + } + + r = hashmap_ensure_allocated(&g->translate_fds, &trivial_hash_ops); + if (r < 0) { + log_oom(); + return -1; + } + + if (action == CURL_POLL_IN) + events = EPOLLIN; + else if (action == CURL_POLL_OUT) + events = EPOLLOUT; + else if (action == CURL_POLL_INOUT) + events = EPOLLIN|EPOLLOUT; + + if (io) { + if (sd_event_source_set_io_events(io, events) < 0) + return -1; + + if (sd_event_source_set_enabled(io, SD_EVENT_ON) < 0) + return -1; + } else { + _cleanup_close_ int fd = -1; + + /* When curl needs to remove an fd from us it closes + * the fd first, and only then calls into us. This is + * nasty, since we cannot pass the fd on to epoll() + * anymore. Hence, duplicate the fds here, and keep a + * copy for epoll which we control after use. */ + + fd = fcntl(s, F_DUPFD_CLOEXEC, 3); + if (fd < 0) + return -1; + + if (sd_event_add_io(g->event, &io, fd, events, curl_glue_on_io, g) < 0) + return -1; + + sd_event_source_set_description(io, "curl-io"); + + r = hashmap_put(g->ios, INT_TO_PTR(s+1), io); + if (r < 0) { + log_oom(); + sd_event_source_unref(io); + return -1; + } + + r = hashmap_put(g->translate_fds, INT_TO_PTR(fd+1), INT_TO_PTR(s+1)); + if (r < 0) { + log_oom(); + hashmap_remove(g->ios, INT_TO_PTR(s+1)); + sd_event_source_unref(io); + return -1; + } + + fd = -1; + } + + return 0; +} + +static int curl_glue_on_timer(sd_event_source *s, uint64_t usec, void *userdata) { + CurlGlue *g = userdata; + int k = 0; + + assert(s); + assert(g); + + if (curl_multi_socket_action(g->curl, CURL_SOCKET_TIMEOUT, 0, &k) != CURLM_OK) { + log_debug("Failed to propagate timeout."); + return -EINVAL; + } + + curl_glue_check_finished(g); + return 0; +} + +static int curl_glue_timer_callback(CURLM *curl, long timeout_ms, void *userdata) { + CurlGlue *g = userdata; + usec_t usec; + + assert(curl); + assert(g); + + if (timeout_ms < 0) { + if (g->timer) { + if (sd_event_source_set_enabled(g->timer, SD_EVENT_OFF) < 0) + return -1; + } + + return 0; + } + + usec = now(clock_boottime_or_monotonic()) + (usec_t) timeout_ms * USEC_PER_MSEC + USEC_PER_MSEC - 1; + + if (g->timer) { + if (sd_event_source_set_time(g->timer, usec) < 0) + return -1; + + if (sd_event_source_set_enabled(g->timer, SD_EVENT_ONESHOT) < 0) + return -1; + } else { + if (sd_event_add_time(g->event, &g->timer, clock_boottime_or_monotonic(), usec, 0, curl_glue_on_timer, g) < 0) + return -1; + + sd_event_source_set_description(g->timer, "curl-timer"); + } + + return 0; +} + +CurlGlue *curl_glue_unref(CurlGlue *g) { + sd_event_source *io; + + if (!g) + return NULL; + + if (g->curl) + curl_multi_cleanup(g->curl); + + while ((io = hashmap_steal_first(g->ios))) { + int fd; + + fd = sd_event_source_get_io_fd(io); + assert(fd >= 0); + + hashmap_remove(g->translate_fds, INT_TO_PTR(fd+1)); + + safe_close(fd); + sd_event_source_unref(io); + } + + hashmap_free(g->ios); + + sd_event_source_unref(g->timer); + sd_event_unref(g->event); + + return NULL; +} + +int curl_glue_new(CurlGlue **glue, sd_event *event) { + _cleanup_(curl_glue_unrefp) CurlGlue *g = NULL; + int r; + + g = new0(CurlGlue, 1); + if (!g) + return -ENOMEM; + + if (event) + g->event = sd_event_ref(event); + else { + r = sd_event_default(&g->event); + if (r < 0) + return r; + } + + g->curl = curl_multi_init(); + if (!g->curl) + return -ENOMEM; + + if (curl_multi_setopt(g->curl, CURLMOPT_SOCKETDATA, g) != CURLM_OK) + return -EINVAL; + + if (curl_multi_setopt(g->curl, CURLMOPT_SOCKETFUNCTION, curl_glue_socket_callback) != CURLM_OK) + return -EINVAL; + + if (curl_multi_setopt(g->curl, CURLMOPT_TIMERDATA, g) != CURLM_OK) + return -EINVAL; + + if (curl_multi_setopt(g->curl, CURLMOPT_TIMERFUNCTION, curl_glue_timer_callback) != CURLM_OK) + return -EINVAL; + + *glue = g; + g = NULL; + + return 0; +} + +int curl_glue_make(CURL **ret, const char *url, void *userdata) { + const char *useragent; + CURL *c; + int r; + + assert(ret); + assert(url); + + c = curl_easy_init(); + if (!c) + return -ENOMEM; + + /* curl_easy_setopt(c, CURLOPT_VERBOSE, 1L); */ + + if (curl_easy_setopt(c, CURLOPT_URL, url) != CURLE_OK) { + r = -EIO; + goto fail; + } + + if (curl_easy_setopt(c, CURLOPT_PRIVATE, userdata) != CURLE_OK) { + r = -EIO; + goto fail; + } + + useragent = strappenda(program_invocation_short_name, "/" PACKAGE_VERSION); + if (curl_easy_setopt(c, CURLOPT_USERAGENT, useragent) != CURLE_OK) { + r = -EIO; + goto fail; + } + + if (curl_easy_setopt(c, CURLOPT_FOLLOWLOCATION, 1L) != CURLE_OK) { + r = -EIO; + goto fail; + } + + *ret = c; + return 0; + +fail: + curl_easy_cleanup(c); + return r; +} + +int curl_glue_add(CurlGlue *g, CURL *c) { + assert(g); + assert(c); + + if (curl_multi_add_handle(g->curl, c) != CURLM_OK) + return -EIO; + + return 0; +} + +void curl_glue_remove_and_free(CurlGlue *g, CURL *c) { + assert(g); + + if (!c) + return; + + if (g->curl) + curl_multi_remove_handle(g->curl, c); + + curl_easy_cleanup(c); +} + +struct curl_slist *curl_slist_new(const char *first, ...) { + struct curl_slist *l; + va_list ap; + + if (!first) + return NULL; + + l = curl_slist_append(NULL, first); + if (!l) + return NULL; + + va_start(ap, first); + + for (;;) { + struct curl_slist *n; + const char *i; + + i = va_arg(ap, const char*); + if (!i) + break; + + n = curl_slist_append(l, i); + if (!n) { + va_end(ap); + curl_slist_free_all(l); + return NULL; + } + + l = n; + } + + va_end(ap); + return l; +} + +int curl_header_strdup(const void *contents, size_t sz, const char *field, char **value) { + const char *p = contents; + size_t l; + char *s; + + l = strlen(field); + if (sz < l) + return 0; + + if (memcmp(p, field, l) != 0) + return 0; + + p += l; + sz -= l; + + if (memchr(p, 0, sz)) + return 0; + + /* Skip over preceeding whitespace */ + while (sz > 0 && strchr(WHITESPACE, p[0])) { + p++; + sz--; + } + + /* Truncate trailing whitespace*/ + while (sz > 0 && strchr(WHITESPACE, p[sz-1])) + sz--; + + s = strndup(p, sz); + if (!s) + return -ENOMEM; + + *value = s; + return 1; +} diff --git a/src/import/curl-util.h b/src/import/curl-util.h new file mode 100644 index 0000000000..5a7550deb9 --- /dev/null +++ b/src/import/curl-util.h @@ -0,0 +1,56 @@ +/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ + +#pragma once + +/*** + This file is part of systemd. + + Copyright 2014 Lennart Poettering + + systemd is free software; you can redistribute it and/or modify it + under the terms of the GNU Lesser General Public License as published by + the Free Software Foundation; either version 2.1 of the License, or + (at your option) any later version. + + systemd is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public License + along with systemd; If not, see <http://www.gnu.org/licenses/>. +***/ + +#include <sys/types.h> +#include <curl/curl.h> + +#include "hashmap.h" +#include "sd-event.h" + +typedef struct CurlGlue CurlGlue; + +struct CurlGlue { + sd_event *event; + CURLM *curl; + sd_event_source *timer; + Hashmap *ios; + Hashmap *translate_fds; + + void (*on_finished)(CurlGlue *g, CURL *curl, CURLcode code); + void *userdata; +}; + +int curl_glue_new(CurlGlue **glue, sd_event *event); +CurlGlue* curl_glue_unref(CurlGlue *glue); + +DEFINE_TRIVIAL_CLEANUP_FUNC(CurlGlue*, curl_glue_unref); + +int curl_glue_make(CURL **ret, const char *url, void *userdata); +int curl_glue_add(CurlGlue *g, CURL *c); +void curl_glue_remove_and_free(CurlGlue *g, CURL *c); + +struct curl_slist *curl_slist_new(const char *first, ...) _sentinel_; +int curl_header_strdup(const void *contents, size_t sz, const char *field, char **value); + +DEFINE_TRIVIAL_CLEANUP_FUNC(CURL*, curl_easy_cleanup); +DEFINE_TRIVIAL_CLEANUP_FUNC(struct curl_slist*, curl_slist_free_all); diff --git a/src/import/import-dck.c b/src/import/import-dck.c new file mode 100644 index 0000000000..7c83ff722b --- /dev/null +++ b/src/import/import-dck.c @@ -0,0 +1,1156 @@ +/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ + +/*** + This file is part of systemd. + + Copyright 2014 Lennart Poettering + + systemd is free software; you can redistribute it and/or modify it + under the terms of the GNU Lesser General Public License as published by + the Free Software Foundation; either version 2.1 of the License, or + (at your option) any later version. + + systemd is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public License + along with systemd; If not, see <http://www.gnu.org/licenses/>. +***/ + +#include <curl/curl.h> +#include <sys/prctl.h> + +#include "hashmap.h" +#include "set.h" +#include "json.h" +#include "strv.h" +#include "curl-util.h" +#include "import-dck.h" +#include "btrfs-util.h" +#include "aufs-util.h" + +/* TODO: + - convert json bits + - man page + - fall back to btrfs loop pool device +*/ + +typedef struct DckImportJob DckImportJob; +typedef struct DckImportName DckImportName; + +typedef enum DckImportJobType { + DCK_IMPORT_JOB_IMAGES, + DCK_IMPORT_JOB_TAGS, + DCK_IMPORT_JOB_ANCESTRY, + DCK_IMPORT_JOB_JSON, + DCK_IMPORT_JOB_LAYER, +} DckImportJobType; + +struct DckImportJob { + DckImport *import; + DckImportJobType type; + bool done; + + char *url; + + Set *needed_by; /* DckImport Name objects */ + + CURL *curl; + struct curl_slist *request_header; + void *payload; + size_t payload_size; + + char *response_token; + char **response_registries; + + char *temp_path; + char *final_path; + + pid_t tar_pid; + FILE *tar_stream; +}; + +struct DckImportName { + DckImport *import; + + char *name; + char *tag; + char *id; + char *local; + + DckImportJob *job_images, *job_tags, *job_ancestry, *job_json, *job_layer; + + char **ancestry; + unsigned current_ancestry; + + bool force_local; +}; + +struct DckImport { + sd_event *event; + CurlGlue *glue; + + Hashmap *names; + Hashmap *jobs; + + dck_import_on_finished on_finished; + void *userdata; +}; + +#define PROTOCOL_PREFIX "https://" +#define INDEX_HOST "index.do" /* the URL we get the data from */ "cker.io" + +#define HEADER_TOKEN "X-Do" /* the HTTP header for the auth token */ "cker-Token:" +#define HEADER_REGISTRY "X-Do" /*the HTTP header for the registry */ "cker-Endpoints:" + +#define PAYLOAD_MAX (16*1024*1024) +#define LAYERS_MAX 2048 + +static int dck_import_name_add_job(DckImportName *name, DckImportJobType type, const char *url, DckImportJob **ret); + +static DckImportJob *dck_import_job_unref(DckImportJob *job) { + if (!job) + return NULL; + + if (job->import) + curl_glue_remove_and_free(job->import->glue, job->curl); + curl_slist_free_all(job->request_header); + + if (job->tar_stream) + fclose(job->tar_stream); + + free(job->final_path); + + if (job->temp_path) { + btrfs_subvol_remove(job->temp_path); + free(job->temp_path); + } + + set_free(job->needed_by); + + if (job->tar_pid > 0) + kill(job->tar_pid, SIGTERM); + + free(job->url); + free(job->payload); + free(job->response_token); + strv_free(job->response_registries); + + free(job); + + return NULL; +} + +static DckImportName *dck_import_name_unref(DckImportName *name) { + if (!name) + return NULL; + + if (name->job_images) + set_remove(name->job_images->needed_by, name); + + if (name->job_tags) + set_remove(name->job_tags->needed_by, name); + + if (name->job_ancestry) + set_remove(name->job_ancestry->needed_by, name); + + if (name->job_json) + set_remove(name->job_json->needed_by, name); + + if (name->job_layer) + set_remove(name->job_layer->needed_by, name); + + free(name->name); + free(name->id); + free(name->tag); + free(name->local); + + strv_free(name->ancestry); + free(name); + + return NULL; +} + +DEFINE_TRIVIAL_CLEANUP_FUNC(DckImportJob*, dck_import_job_unref); +DEFINE_TRIVIAL_CLEANUP_FUNC(DckImportName*, dck_import_name_unref); + +static void dck_import_finish(DckImport *import, int error) { + assert(import); + + if (import->on_finished) + import->on_finished(import, error, import->userdata); + else + sd_event_exit(import->event, error); +} + +static int parse_id(const void *payload, size_t size, char **ret) { + _cleanup_free_ char *buf = NULL, *id = NULL, *other = NULL; + union json_value v = {}; + void *json_state = NULL; + const char *p; + int t; + + assert(payload); + assert(ret); + + if (size <= 0) + return -EBADMSG; + + if (memchr(payload, 0, size)) + return -EBADMSG; + + buf = strndup(payload, size); + if (!buf) + return -ENOMEM; + + p = buf; + t = json_tokenize(&p, &id, &v, &json_state, NULL); + if (t < 0) + return t; + if (t != JSON_STRING) + return -EBADMSG; + + t = json_tokenize(&p, &other, &v, &json_state, NULL); + if (t < 0) + return t; + if (t != JSON_END) + return -EBADMSG; + + if (!dck_id_is_valid(id)) + return -EBADMSG; + + *ret = id; + id = NULL; + + return 0; +} + +static int parse_ancestry(const void *payload, size_t size, char ***ret) { + _cleanup_free_ char *buf = NULL; + void *json_state = NULL; + const char *p; + enum { + STATE_BEGIN, + STATE_ITEM, + STATE_COMMA, + STATE_END, + } state = STATE_BEGIN; + _cleanup_strv_free_ char **l = NULL; + size_t n = 0, allocated = 0; + + if (size <= 0) + return -EBADMSG; + + if (memchr(payload, 0, size)) + return -EBADMSG; + + buf = strndup(payload, size); + if (!buf) + return -ENOMEM; + + p = buf; + for (;;) { + _cleanup_free_ char *str; + union json_value v = {}; + int t; + + t = json_tokenize(&p, &str, &v, &json_state, NULL); + if (t < 0) + return t; + + switch (state) { + + case STATE_BEGIN: + if (t == JSON_ARRAY_OPEN) + state = STATE_ITEM; + else + return -EBADMSG; + + break; + + case STATE_ITEM: + if (t == JSON_STRING) { + if (!dck_id_is_valid(str)) + return -EBADMSG; + + if (n+1 > LAYERS_MAX) + return -EFBIG; + + if (!GREEDY_REALLOC(l, allocated, n + 2)) + return -ENOMEM; + + l[n++] = str; + str = NULL; + l[n] = NULL; + + state = STATE_COMMA; + + } else if (t == JSON_ARRAY_CLOSE) + state = STATE_END; + else + return -EBADMSG; + + break; + + case STATE_COMMA: + if (t == JSON_COMMA) + state = STATE_ITEM; + else if (t == JSON_ARRAY_CLOSE) + state = STATE_END; + else + return -EBADMSG; + break; + + case STATE_END: + if (t == JSON_END) { + + if (strv_isempty(l)) + return -EBADMSG; + + if (!strv_is_uniq(l)) + return -EBADMSG; + + l = strv_reverse(l); + + *ret = l; + l = NULL; + return 0; + } else + return -EBADMSG; + } + + } +} + +static const char *dck_import_name_current_layer(DckImportName *name) { + assert(name); + + if (strv_isempty(name->ancestry)) + return NULL; + + return name->ancestry[name->current_ancestry]; +} + +static const char *dck_import_name_current_base_layer(DckImportName *name) { + assert(name); + + if (strv_isempty(name->ancestry)) + return NULL; + + if (name->current_ancestry <= 0) + return NULL; + + return name->ancestry[name->current_ancestry-1]; +} + +static char** dck_import_name_get_registries(DckImportName *name) { + assert(name); + + if (!name->job_images) + return NULL; + + if (!name->job_images->done) + return NULL; + + if (strv_isempty(name->job_images->response_registries)) + return NULL; + + return name->job_images->response_registries; +} + +static const char*dck_import_name_get_token(DckImportName *name) { + assert(name); + + if (!name->job_images) + return NULL; + + if (!name->job_images->done) + return NULL; + + return name->job_images->response_token; +} + +static void dck_import_name_maybe_finish(DckImportName *name) { + int r; + + assert(name); + + if (!name->job_images || !name->job_images->done) + return; + + if (!name->job_ancestry || !name->job_ancestry->done) + return; + + if (!name->job_json || !name->job_json->done) + return; + + if (name->job_layer && !name->job_json->done) + return; + + if (dck_import_name_current_layer(name)) + return; + + if (name->local) { + const char *p, *q; + + assert(name->id); + + p = strappenda("/var/lib/container/", name->local); + q = strappenda("/var/lib/container/.dck-", name->id); + + if (name->force_local) { + (void) btrfs_subvol_remove(p); + (void) rm_rf(p, false, true, false); + } + + r = btrfs_subvol_snapshot(q, p, false, false); + if (r < 0) { + log_error_errno(r, "Failed to snapshot final image: %m"); + dck_import_finish(name->import, r); + return; + } + + log_info("Created new image %s.", p); + } + + dck_import_finish(name->import, 0); +} + +static int dck_import_job_run_tar(DckImportJob *job) { + _cleanup_close_pair_ int pipefd[2] = { -1, -1 }; + bool gzip; + + assert(job); + + /* A stream to run tar on? */ + if (!job->temp_path) + return 0; + + if (job->tar_stream) + return 0; + + /* Maybe fork off tar, if we have enough to figure out that + * something is gzip compressed or not */ + + if (job->payload_size < 2) + return 0; + + /* Detect gzip signature */ + gzip = ((uint8_t*) job->payload)[0] == 0x1f && + ((uint8_t*) job->payload)[1] == 0x8b; + + assert(!job->tar_stream); + assert(job->tar_pid <= 0); + + if (pipe2(pipefd, O_CLOEXEC) < 0) + return log_error_errno(errno, "Failed to create pipe for tar: %m"); + + job->tar_pid = fork(); + if (job->tar_pid < 0) + return log_error_errno(errno, "Failed to fork off tar: %m"); + if (job->tar_pid == 0) { + int null_fd; + + reset_all_signal_handlers(); + reset_signal_mask(); + assert_se(prctl(PR_SET_PDEATHSIG, SIGTERM) == 0); + + pipefd[1] = safe_close(pipefd[1]); + + if (dup2(pipefd[0], STDIN_FILENO) != STDIN_FILENO) { + log_error_errno(errno, "Failed to dup2() fd: %m"); + _exit(EXIT_FAILURE); + } + + if (pipefd[0] != STDIN_FILENO) + safe_close(pipefd[0]); + if (pipefd[1] != STDIN_FILENO) + safe_close(pipefd[1]); + + null_fd = open("/dev/null", O_WRONLY|O_NOCTTY); + if (null_fd < 0) { + log_error_errno(errno, "Failed to open /dev/null: %m"); + _exit(EXIT_FAILURE); + } + + if (dup2(null_fd, STDOUT_FILENO) != STDOUT_FILENO) { + log_error_errno(errno, "Failed to dup2() fd: %m"); + _exit(EXIT_FAILURE); + } + + if (null_fd != STDOUT_FILENO) + safe_close(null_fd); + + execlp("tar", "tar", "-C", job->temp_path, gzip ? "-xz" : "-x", NULL); + _exit(EXIT_FAILURE); + } + + pipefd[0] = safe_close(pipefd[0]); + + job->tar_stream = fdopen(pipefd[1], "w"); + if (!job->tar_stream) + return log_error_errno(errno, "Failed to allocate tar stream: %m"); + + pipefd[1] = -1; + + if (fwrite(job->payload, 1, job->payload_size, job->tar_stream) != job->payload_size) + return log_error_errno(errno, "Couldn't write payload: %m"); + + free(job->payload); + job->payload = NULL; + job->payload_size = 0; + + return 0; +} + +static int dck_import_name_pull_layer(DckImportName *name) { + _cleanup_free_ char *path = NULL, *temp = NULL; + const char *url, *layer = NULL, *base = NULL; + char **rg; + int r; + + assert(name); + + if (name->job_layer) { + set_remove(name->job_layer->needed_by, name); + name->job_layer = NULL; + } + + for (;;) { + layer = dck_import_name_current_layer(name); + if (!layer) { + dck_import_name_maybe_finish(name); + return 0; + } + + path = strjoin("/var/lib/container/.dck-", layer, NULL); + if (!path) + return log_oom(); + + if (laccess(path, F_OK) < 0) { + if (errno == ENOENT) + break; + + return log_error_errno(errno, "Failed to check for container: %m"); + } + + log_info("Layer %s already exists, skipping.", layer); + + name->current_ancestry++; + + free(path); + path = NULL; + } + + rg = dck_import_name_get_registries(name); + assert(rg && rg[0]); + + url = strappenda(PROTOCOL_PREFIX, rg[0], "/v1/images/", layer, "/layer"); + r = dck_import_name_add_job(name, DCK_IMPORT_JOB_LAYER, url, &name->job_layer); + if (r < 0) { + log_error_errno(r, "Failed to issue HTTP request: %m"); + return r; + } + if (r == 0) /* Already downloading this one? */ + return 0; + + log_info("Pulling layer %s...", layer); + + r = tempfn_random(path, &temp); + if (r < 0) + return log_oom(); + + base = dck_import_name_current_base_layer(name); + if (base) { + const char *base_path; + + base_path = strappend("/var/lib/container/.dck-", base); + r = btrfs_subvol_snapshot(base_path, temp, false, true); + } else + r = btrfs_subvol_make(temp); + + if (r < 0) + return log_error_errno(r, "Failed to make btrfs subvolume %s", temp); + + name->job_layer->final_path = path; + name->job_layer->temp_path = temp; + path = temp = NULL; + + return 0; +} + +static void dck_import_name_job_finished(DckImportName *name, DckImportJob *job) { + int r; + + assert(name); + assert(job); + + if (name->job_images == job) { + const char *url; + char **rg; + + assert(!name->job_tags); + assert(!name->job_ancestry); + assert(!name->job_json); + assert(!name->job_layer); + + rg = dck_import_name_get_registries(name); + if (strv_isempty(rg)) { + log_error("Didn't get registry information."); + r = -EBADMSG; + goto fail; + } + + log_info("Index lookup succeeded, directed to registry %s.", rg[0]); + + url = strappenda(PROTOCOL_PREFIX, rg[0], "/v1/repositories/", name->name, "/tags/", name->tag); + + r = dck_import_name_add_job(name, DCK_IMPORT_JOB_TAGS, url, &name->job_tags); + if (r < 0) { + log_error_errno(r, "Failed to issue HTTP request: %m"); + goto fail; + } + + } else if (name->job_tags == job) { + const char *url; + char *id = NULL, **rg; + + assert(!name->job_ancestry); + assert(!name->job_json); + assert(!name->job_layer); + + r = parse_id(job->payload, job->payload_size, &id); + if (r < 0) { + log_error_errno(r, "Failed to parse JSON id."); + goto fail; + } + + free(name->id); + name->id = id; + + rg = dck_import_name_get_registries(name); + assert(rg && rg[0]); + + log_info("Tag lookup succeeded, resolved to layer %s.", name->id); + + url = strappenda(PROTOCOL_PREFIX, rg[0], "/v1/images/", name->id, "/ancestry"); + r = dck_import_name_add_job(name, DCK_IMPORT_JOB_ANCESTRY, url, &name->job_ancestry); + if (r < 0) { + log_error_errno(r, "Failed to issue HTTP request: %m"); + goto fail; + } + + url = strappenda(PROTOCOL_PREFIX, rg[0], "/v1/images/", name->id, "/json"); + r = dck_import_name_add_job(name, DCK_IMPORT_JOB_JSON, url, &name->job_json); + if (r < 0) { + log_error_errno(r, "Failed to issue HTTP request: %m"); + goto fail; + } + + } else if (name->job_ancestry == job) { + char **ancestry = NULL, **i; + unsigned n; + + r = parse_ancestry(job->payload, job->payload_size, &ancestry); + if (r < 0) { + log_error_errno(r, "Failed to parse JSON id."); + goto fail; + } + + n = strv_length(ancestry); + if (n <= 0 || !streq(ancestry[n-1], name->id)) { + log_error("Ancestry doesn't end in main layer."); + r = -EBADMSG; + goto fail; + } + + log_info("Ancestor lookup succeeded, requires layers:\n"); + STRV_FOREACH(i, ancestry) + log_info("\t%s", *i); + + strv_free(name->ancestry); + name->ancestry = ancestry; + + name->current_ancestry = 0; + r = dck_import_name_pull_layer(name); + if (r < 0) + goto fail; + + } else if (name->job_json == job) { + + dck_import_name_maybe_finish(name); + + } else if (name->job_layer == job) { + + name->current_ancestry ++; + r = dck_import_name_pull_layer(name); + if (r < 0) + goto fail; + + } else + assert_not_reached("Got finished event for unknown curl object"); + + return; + +fail: + dck_import_finish(name->import, r); +} + +static void dck_import_curl_on_finished(CurlGlue *g, CURL *curl, CURLcode result) { + DckImportJob *job = NULL; + CURLcode code; + DckImportName *n; + long status; + Iterator i; + int r; + + if (curl_easy_getinfo(curl, CURLINFO_PRIVATE, &job) != CURLE_OK) + return; + + if (!job) + return; + + job->done = true; + + if (result != CURLE_OK) { + log_error("Transfer failed: %s", curl_easy_strerror(code)); + r = -EIO; + goto fail; + } + + code = curl_easy_getinfo(curl, CURLINFO_RESPONSE_CODE, &status); + if (code != CURLE_OK) { + log_error("Failed to retrieve response code: %s", curl_easy_strerror(code)); + r = -EIO; + goto fail; + } else if (status >= 300) { + log_error("HTTP request to %s failed with code %li.", job->url, status); + r = -EIO; + goto fail; + } else if (status < 200) { + log_error("HTTP request to %s finished with unexpected code %li.", job->url, status); + r = -EIO; + goto fail; + } + + switch (job->type) { + + case DCK_IMPORT_JOB_LAYER: { + siginfo_t si; + + if (!job->tar_stream) { + log_error("Downloaded layer too short."); + r = -EIO; + goto fail; + } + + fclose(job->tar_stream); + job->tar_stream = NULL; + + assert(job->tar_pid > 0); + + r = wait_for_terminate(job->tar_pid, &si); + if (r < 0) { + log_error_errno(r, "Failed to wait for tar process: %m"); + goto fail; + } + + job->tar_pid = 0; + + if (si.si_code != CLD_EXITED || si.si_status != EXIT_SUCCESS) { + log_error_errno(r, "tar failed abnormally."); + r = -EIO; + goto fail; + } + + r = aufs_resolve(job->temp_path); + if (r < 0) { + log_error_errno(r, "Couldn't resolve aufs whiteouts: %m"); + goto fail; + } + + r = btrfs_subvol_read_only(job->temp_path, true); + if (r < 0) { + log_error_errno(r, "Failed to mark snapshot read-only: %m"); + goto fail; + } + + if (rename(job->temp_path, job->final_path) < 0) { + log_error_errno(r, "Failed to rename snapshot: %m"); + goto fail; + } + + log_info("Completed writing to layer %s", job->final_path); + break; + } + + default: + ; + } + + SET_FOREACH(n, job->needed_by, i) + dck_import_name_job_finished(n, job); + + return; + +fail: + dck_import_finish(job->import, r); +} + +static size_t dck_import_job_write_callback(void *contents, size_t size, size_t nmemb, void *userdata) { + DckImportJob *j = userdata; + size_t sz = size * nmemb; + char *p; + int r; + + assert(contents); + assert(j); + + if (j->tar_stream) { + size_t l; + + l = fwrite(contents, size, nmemb, j->tar_stream); + if (l != nmemb) { + r = -errno; + goto fail; + } + + return l; + } + + if (j->payload_size + sz > PAYLOAD_MAX) { + r = -EFBIG; + goto fail; + } + + p = realloc(j->payload, j->payload_size + sz); + if (!p) { + r = -ENOMEM; + goto fail; + } + + memcpy(p + j->payload_size, contents, sz); + j->payload_size += sz; + j->payload = p; + + r = dck_import_job_run_tar(j); + if (r < 0) + goto fail; + + return sz; + +fail: + dck_import_finish(j->import, r); + return 0; +} + +static size_t dck_import_job_header_callback(void *contents, size_t size, size_t nmemb, void *userdata) { + _cleanup_free_ char *registry = NULL; + size_t sz = size * nmemb; + DckImportJob *j = userdata; + char *token; + int r; + + assert(contents); + assert(j); + + r = curl_header_strdup(contents, sz, HEADER_TOKEN, &token); + if (r < 0) { + log_oom(); + goto fail; + } + if (r > 0) { + free(j->response_token); + j->response_token = token; + } + + r = curl_header_strdup(contents, sz, HEADER_REGISTRY, ®istry); + if (r < 0) { + log_oom(); + goto fail; + } + if (r > 0) { + char **l, **i; + + l = strv_split(registry, ","); + if (!l) { + r = log_oom(); + goto fail; + } + + STRV_FOREACH(i, l) { + if (!hostname_is_valid(*i)) { + log_error("Registry hostname is not valid."); + strv_free(l); + r = -EBADMSG; + goto fail; + } + } + + strv_free(j->response_registries); + j->response_registries = l; + } + + return sz; + +fail: + dck_import_finish(j->import, r); + return 0; +} + +static int dck_import_name_add_job(DckImportName *name, DckImportJobType type, const char *url, DckImportJob **ret) { + _cleanup_(dck_import_job_unrefp) DckImportJob *j = NULL; + DckImportJob *f = NULL; + const char *t, *token; + int r; + + assert(name); + assert(url); + assert(ret); + + log_info("Getting %s.", url); + f = hashmap_get(name->import->jobs, url); + if (f) { + if (f->type != type) + return -EINVAL; + + r = set_put(f->needed_by, name); + if (r < 0) + return r; + + return 0; + } + + r = hashmap_ensure_allocated(&name->import->jobs, &string_hash_ops); + if (r < 0) + return r; + + j = new0(DckImportJob, 1); + if (!j) + return -ENOMEM; + + j->import = name->import; + j->type = type; + j->url = strdup(url); + if (!j->url) + return -ENOMEM; + + r = set_ensure_allocated(&j->needed_by, &trivial_hash_ops); + if (r < 0) + return r; + + r = curl_glue_make(&j->curl, j->url, j); + if (r < 0) + return r; + + token = dck_import_name_get_token(name); + if (token) + t = strappenda("Authorization: Token ", token); + else + t = HEADER_TOKEN " true"; + + j->request_header = curl_slist_new("Accept: application/json", t, NULL); + if (!j->request_header) + return -ENOMEM; + + if (curl_easy_setopt(j->curl, CURLOPT_HTTPHEADER, j->request_header) != CURLE_OK) + return -EIO; + + if (curl_easy_setopt(j->curl, CURLOPT_WRITEFUNCTION, dck_import_job_write_callback) != CURLE_OK) + return -EIO; + + if (curl_easy_setopt(j->curl, CURLOPT_WRITEDATA, j) != CURLE_OK) + return -EIO; + + if (curl_easy_setopt(j->curl, CURLOPT_HEADERFUNCTION, dck_import_job_header_callback) != CURLE_OK) + return -EIO; + + if (curl_easy_setopt(j->curl, CURLOPT_HEADERDATA, j) != CURLE_OK) + return -EIO; + + r = curl_glue_add(name->import->glue, j->curl); + if (r < 0) + return r; + + r = hashmap_put(name->import->jobs, j->url, j); + if (r < 0) + return r; + + r = set_put(j->needed_by, name); + if (r < 0) { + hashmap_remove(name->import->jobs, url); + return r; + } + + *ret = j; + j = NULL; + + return 1; +} + +static int dck_import_name_begin(DckImportName *name) { + const char *url; + + assert(name); + assert(!name->job_images); + + url = strappenda(PROTOCOL_PREFIX, INDEX_HOST, "/v1/repositories/", name->name, "/images"); + + return dck_import_name_add_job(name, DCK_IMPORT_JOB_IMAGES, url, &name->job_images); +} + +int dck_import_new(DckImport **import, sd_event *event, dck_import_on_finished on_finished, void *userdata) { + _cleanup_(dck_import_unrefp) DckImport *i = NULL; + int r; + + assert(import); + + i = new0(DckImport, 1); + if (!i) + return -ENOMEM; + + i->on_finished = on_finished; + i->userdata = userdata; + + if (event) + i->event = sd_event_ref(event); + else { + r = sd_event_default(&i->event); + if (r < 0) + return r; + } + + r = curl_glue_new(&i->glue, i->event); + if (r < 0) + return r; + + i->glue->on_finished = dck_import_curl_on_finished; + i->glue->userdata = i; + + *import = i; + i = NULL; + + return 0; +} + +DckImport* dck_import_unref(DckImport *import) { + DckImportName *n; + DckImportJob *j; + + if (!import) + return NULL; + + while ((n = hashmap_steal_first(import->names))) + dck_import_name_unref(n); + hashmap_free(import->names); + + while ((j = hashmap_steal_first(import->jobs))) + dck_import_job_unref(j); + hashmap_free(import->jobs); + + curl_glue_unref(import->glue); + sd_event_unref(import->event); + + free(import); + return NULL; +} + +int dck_import_cancel(DckImport *import, const char *name) { + DckImportName *n; + + assert(import); + assert(name); + + n = hashmap_remove(import->names, name); + if (!n) + return 0; + + dck_import_name_unref(n); + return 1; +} + +int dck_import_pull(DckImport *import, const char *name, const char *tag, const char *local, bool force_local) { + _cleanup_(dck_import_name_unrefp) DckImportName *n = NULL; + int r; + + assert(import); + assert(dck_name_is_valid(name)); + assert(dck_tag_is_valid(tag)); + assert(!local || machine_name_is_valid(local)); + + if (hashmap_get(import->names, name)) + return -EEXIST; + + r = hashmap_ensure_allocated(&import->names, &string_hash_ops); + if (r < 0) + return r; + + n = new0(DckImportName, 1); + if (!n) + return -ENOMEM; + + n->import = import; + + n->name = strdup(name); + if (!n->name) + return -ENOMEM; + + n->tag = strdup(tag); + if (!n->tag) + return -ENOMEM; + + if (local) { + n->local = strdup(local); + if (!n->local) + return -ENOMEM; + n->force_local = force_local; + } + + r = hashmap_put(import->names, name, n); + if (r < 0) + return r; + + r = dck_import_name_begin(n); + if (r < 0) { + dck_import_cancel(import, n->name); + n = NULL; + return r; + } + + n = NULL; + + return 0; +} + +bool dck_name_is_valid(const char *name) { + const char *slash, *p; + + if (isempty(name)) + return false; + + slash = strchr(name, '/'); + if (!slash) + return false; + + if (!filename_is_valid(slash + 1)) + return false; + + p = strndupa(name, slash - name); + if (!filename_is_valid(p)) + return false; + + return true; +} + +bool dck_id_is_valid(const char *id) { + + if (!filename_is_valid(id)) + return false; + + if (!in_charset(id, "0123456789abcdef")) + return false; + + return true; +} diff --git a/src/import/import-dck.h b/src/import/import-dck.h new file mode 100644 index 0000000000..cd2d27c7ee --- /dev/null +++ b/src/import/import-dck.h @@ -0,0 +1,39 @@ +/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ + +/*** + This file is part of systemd. + + Copyright 2014 Lennart Poettering + + systemd is free software; you can redistribute it and/or modify it + under the terms of the GNU Lesser General Public License as published by + the Free Software Foundation; either version 2.1 of the License, or + (at your option) any later version. + + systemd is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public License + along with systemd; If not, see <http://www.gnu.org/licenses/>. +***/ + +#include "sd-event.h" +#include "util.h" + +typedef struct DckImport DckImport; + +typedef void (*dck_import_on_finished)(DckImport *import, int error, void *userdata); + +int dck_import_new(DckImport **import, sd_event *event, dck_import_on_finished on_finished, void *userdata); +DckImport* dck_import_unref(DckImport *import); + +DEFINE_TRIVIAL_CLEANUP_FUNC(DckImport*, dck_import_unref); + +int dck_import_pull(DckImport *import, const char *name, const char *tag, const char *local, bool force_local); +int dck_import_cancel(DckImport *import, const char *name); + +bool dck_name_is_valid(const char *name); +bool dck_id_is_valid(const char *id); +#define dck_tag_is_valid(tag) filename_is_valid(tag) diff --git a/src/import/import.c b/src/import/import.c new file mode 100644 index 0000000000..9bade38c3a --- /dev/null +++ b/src/import/import.c @@ -0,0 +1,213 @@ +/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ + +/*** + This file is part of systemd. + + Copyright 2014 Lennart Poettering + + systemd is free software; you can redistribute it and/or modify it + under the terms of the GNU Lesser General Public License as published by + the Free Software Foundation; either version 2.1 of the License, or + (at your option) any later version. + + systemd is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public License + along with systemd; If not, see <http://www.gnu.org/licenses/>. +***/ + +#include <getopt.h> + +#include "sd-event.h" +#include "event-util.h" +#include "verbs.h" +#include "build.h" +#include "import-dck.h" + +static bool arg_force = false; + +static void on_finished(DckImport *import, int error, void *userdata) { + sd_event *event = userdata; + assert(import); + + if (error == 0) + log_info("Operation completed successfully."); + else + log_info_errno(error, "Operation failed: %m"); + + sd_event_exit(event, error); +} + +static int pull_dck(int argc, char *argv[], void *userdata) { + _cleanup_(dck_import_unrefp) DckImport *import = NULL; + _cleanup_event_unref_ sd_event *event = NULL; + const char *name, *tag, *local; + int r; + + tag = strchr(argv[1], ':'); + if (tag) { + name = strndupa(argv[1], tag - argv[1]); + tag++; + } else { + name = argv[1]; + tag = "latest"; + } + + if (argc >= 3) + local = argv[2]; + else { + local = strchr(name, '/'); + if (local) + local++; + else + local = name; + } + + if (streq(local, "-") || isempty(local)) + local = NULL; + + if (!dck_name_is_valid(name)) { + log_error("Remote name '%s' is not valid.", name); + return -EINVAL; + } + + if (!dck_tag_is_valid(tag)) { + log_error("Tag name '%s' is not valid.", tag); + return -EINVAL; + } + + if (local) { + const char *p; + + if (!machine_name_is_valid(tag)) { + log_error("Local image name '%s' is not valid.", local); + return -EINVAL; + } + + p = strappenda("/var/lib/container/", local); + if (laccess(p, F_OK) >= 0) { + if (!arg_force) { + log_info("Image '%s' already exists.", local); + return 0; + } + } else if (errno != ENOENT) + return log_error_errno(errno, "Can't check if image '%s' already exists: %m", local); + + log_info("Pulling '%s' with tag '%s', saving as '%s'.", name, tag, local); + } else + log_info("Pulling '%s' with tag '%s'.", name, tag); + + r = sd_event_default(&event); + if (r < 0) + return log_error_errno(r, "Failed to allocate event loop: %m"); + + assert_se(sigprocmask_many(SIG_BLOCK, SIGTERM, SIGINT, -1) == 0); + sd_event_add_signal(event, NULL, SIGTERM, NULL, NULL); + sd_event_add_signal(event, NULL, SIGINT, NULL, NULL); + + r = dck_import_new(&import, event, on_finished, event); + if (r < 0) + return log_error_errno(r, "Failed to allocate importer: %m"); + + r = dck_import_pull(import, name, tag, local, arg_force); + if (r < 0) + return log_error_errno(r, "Failed to pull image: %m"); + + r = sd_event_loop(event); + if (r < 0) + return log_error_errno(r, "Failed to run event loop: %m"); + + log_info("Exiting."); + + return 0; +} + +static int help(int argc, char *argv[], void *userdata) { + + printf("%s [OPTIONS...] {COMMAND} ...\n\n" + "Import container or virtual machine image.\n\n" + " -h --help Show this help\n" + " --version Show package version\n" + " --force Force creation of image\n\n" + "Commands:\n" + " pull-dck REMOTE [NAME] Download an image\n", + program_invocation_short_name); + + return 0; +} + +static int parse_argv(int argc, char *argv[]) { + + enum { + ARG_VERSION = 0x100, + ARG_FORCE, + }; + + static const struct option options[] = { + { "help", no_argument, NULL, 'h' }, + { "version", no_argument, NULL, ARG_VERSION }, + { "force", no_argument, NULL, ARG_FORCE }, + {} + }; + + int c; + + assert(argc >= 0); + assert(argv); + + while ((c = getopt_long(argc, argv, "h", options, NULL)) >= 0) + + switch (c) { + + case 'h': + return help(argc, argv, NULL); + + case ARG_VERSION: + puts(PACKAGE_STRING); + puts(SYSTEMD_FEATURES); + return 0; + + case ARG_FORCE: + arg_force = true; + break; + + case '?': + return -EINVAL; + + default: + assert_not_reached("Unhandled option"); + } + + return 1; +} + +static int import_main(int argc, char *argv[]) { + + const Verb verbs[] = { + { "help", VERB_ANY, VERB_ANY, 0, help }, + { "pull-dck", 2, 3, 0, pull_dck }, + {} + }; + + return dispatch_verb(argc, argv, verbs, NULL); +} + +int main(int argc, char *argv[]) { + int r; + + setlocale(LC_ALL, ""); + log_parse_environment(); + log_open(); + + r = parse_argv(argc, argv); + if (r <= 0) + goto finish; + + r = import_main(argc, argv); + +finish: + return r < 0 ? EXIT_FAILURE : EXIT_SUCCESS; +} diff --git a/src/shared/util.h b/src/shared/util.h index a8ccf20bb8..96b8c1bcc0 100644 --- a/src/shared/util.h +++ b/src/shared/util.h @@ -1040,3 +1040,5 @@ int sethostname_idempotent(const char *s); for ((e) = (struct inotify_event*) (buffer); \ (uint8_t*) (e) < (uint8_t*) (buffer) + (sz); \ (e) = (struct inotify_event*) ((uint8_t*) (e) + sizeof(struct inotify_event) + (e)->len)) + +#define laccess(path, mode) faccessat(AT_FDCWD, (path), (mode), AT_SYMLINK_NOFOLLOW) |