mirror of
https://github.com/moparisthebest/pacman
synced 2024-11-11 11:55:12 -05:00
Merge branch 'download'
This commit is contained in:
commit
cff36093f3
31
configure.ac
31
configure.ac
@ -93,10 +93,10 @@ AC_ARG_WITH(openssl,
|
|||||||
AS_HELP_STRING([--with-openssl], [use OpenSSL crypto implementations instead of internal routines]),
|
AS_HELP_STRING([--with-openssl], [use OpenSSL crypto implementations instead of internal routines]),
|
||||||
[], [with_openssl=check])
|
[], [with_openssl=check])
|
||||||
|
|
||||||
# Help line for libfetch
|
# Help line for libcurl
|
||||||
AC_ARG_WITH(fetch,
|
AC_ARG_WITH(curl,
|
||||||
AS_HELP_STRING([--with-fetch], [use libfetch as an internal downloader]),
|
AS_HELP_STRING([--with-curl], [use libcurl as an internal downloader]),
|
||||||
[], [with_fetch=check])
|
[], [with_curl=check])
|
||||||
|
|
||||||
# Help line for documentation
|
# Help line for documentation
|
||||||
AC_ARG_ENABLE(doc,
|
AC_ARG_ENABLE(doc,
|
||||||
@ -149,24 +149,17 @@ AS_IF([test "x$with_openssl" != "xno"],
|
|||||||
AC_MSG_RESULT(no))
|
AC_MSG_RESULT(no))
|
||||||
AM_CONDITIONAL([HAVE_LIBSSL], [test "x$ac_cv_lib_ssl_MD5_Final" = "xyes"])
|
AM_CONDITIONAL([HAVE_LIBSSL], [test "x$ac_cv_lib_ssl_MD5_Final" = "xyes"])
|
||||||
|
|
||||||
# Enable or disable usage of libfetch
|
# Enable or disable usage of libcurl
|
||||||
AC_MSG_CHECKING(whether to link with libfetch)
|
AC_MSG_CHECKING(whether to link with libcurl)
|
||||||
AS_IF([test "x$with_fetch" != "xno"],
|
AS_IF([test "x$with_curl" != "xno"],
|
||||||
[AC_MSG_RESULT(yes)
|
[AC_MSG_RESULT(yes)
|
||||||
AC_CHECK_LIB([fetch], [fetchParseURL], ,
|
AC_CHECK_LIB([curl], [curl_easy_perform], ,
|
||||||
[if test "x$with_fetch" != "xcheck"; then
|
[if test "x$with_curl" != "xcheck"; then
|
||||||
AC_MSG_FAILURE([--with-fetch was given, but -lfetch was not found])
|
AC_MSG_FAILURE([--with-curl was given, but -lcurl was not found])
|
||||||
fi],
|
fi],
|
||||||
[-lcrypto -ldl])
|
[-lcurl])],
|
||||||
# Check if libfetch supports connnection caching which we use
|
|
||||||
AS_IF([test "x$ac_cv_lib_fetch_fetchParseURL" = "xyes"],
|
|
||||||
[AC_CHECK_DECL(fetchConnectionCacheInit, ,
|
|
||||||
AC_MSG_ERROR([libfetch must be version 2.28 or greater]),
|
|
||||||
[#include <fetch.h>])
|
|
||||||
])
|
|
||||||
],
|
|
||||||
AC_MSG_RESULT(no))
|
AC_MSG_RESULT(no))
|
||||||
AM_CONDITIONAL([HAVE_LIBFETCH], [test "x$ac_cv_lib_fetch_fetchParseURL" = "xyes"])
|
AM_CONDITIONAL([HAVE_LIBCURL], [test "x$ac_cv_lib_curl_curl_easy_perform" = "xyes"])
|
||||||
|
|
||||||
# Checks for header files.
|
# Checks for header files.
|
||||||
AC_CHECK_HEADERS([fcntl.h glob.h libintl.h locale.h mntent.h string.h \
|
AC_CHECK_HEADERS([fcntl.h glob.h libintl.h locale.h mntent.h string.h \
|
||||||
|
@ -23,9 +23,8 @@
|
|||||||
|
|
||||||
#include "config.h"
|
#include "config.h"
|
||||||
|
|
||||||
/* connection caching setup */
|
#ifdef HAVE_LIBCURL
|
||||||
#ifdef HAVE_LIBFETCH
|
#include <curl/curl.h>
|
||||||
#include <fetch.h>
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* libalpm */
|
/* libalpm */
|
||||||
@ -65,8 +64,9 @@ int SYMEXPORT alpm_initialize(void)
|
|||||||
bindtextdomain("libalpm", LOCALEDIR);
|
bindtextdomain("libalpm", LOCALEDIR);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef HAVE_LIBFETCH
|
#ifdef HAVE_LIBCURL
|
||||||
fetchConnectionCacheInit(5, 1);
|
curl_global_init(CURL_GLOBAL_SSL);
|
||||||
|
handle->curl = curl_easy_init();
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
return(0);
|
return(0);
|
||||||
@ -88,8 +88,8 @@ int SYMEXPORT alpm_release(void)
|
|||||||
_alpm_handle_free(handle);
|
_alpm_handle_free(handle);
|
||||||
handle = NULL;
|
handle = NULL;
|
||||||
|
|
||||||
#ifdef HAVE_LIBFETCH
|
#ifdef HAVE_LIBCURL
|
||||||
fetchConnectionCacheClose();
|
curl_global_cleanup();
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
return(0);
|
return(0);
|
||||||
|
@ -535,6 +535,7 @@ enum _pmerrno_t {
|
|||||||
/* External library errors */
|
/* External library errors */
|
||||||
PM_ERR_LIBARCHIVE,
|
PM_ERR_LIBARCHIVE,
|
||||||
PM_ERR_LIBFETCH,
|
PM_ERR_LIBFETCH,
|
||||||
|
PM_ERR_LIBCURL,
|
||||||
PM_ERR_EXTERNAL_DOWNLOAD
|
PM_ERR_EXTERNAL_DOWNLOAD
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -35,8 +35,8 @@
|
|||||||
#include <sys/param.h> /* MAXHOSTNAMELEN */
|
#include <sys/param.h> /* MAXHOSTNAMELEN */
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef HAVE_LIBFETCH
|
#ifdef HAVE_LIBCURL
|
||||||
#include <fetch.h>
|
#include <curl/curl.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* libalpm */
|
/* libalpm */
|
||||||
@ -55,7 +55,7 @@ static char *get_filename(const char *url) {
|
|||||||
return(filename);
|
return(filename);
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef HAVE_LIBFETCH
|
#ifdef HAVE_LIBCURL
|
||||||
static char *get_destfile(const char *path, const char *filename) {
|
static char *get_destfile(const char *path, const char *filename) {
|
||||||
char *destfile;
|
char *destfile;
|
||||||
/* len = localpath len + filename len + null */
|
/* len = localpath len + filename len + null */
|
||||||
@ -76,14 +76,8 @@ static char *get_tempfile(const char *path, const char *filename) {
|
|||||||
return(tempfile);
|
return(tempfile);
|
||||||
}
|
}
|
||||||
|
|
||||||
static const char *gethost(struct url *fileurl)
|
#define check_stop() if(dload_interrupted) { ret = -1; goto cleanup; }
|
||||||
{
|
enum sighandlers { OLD = 0, NEW = 1 };
|
||||||
const char *host = _("disk");
|
|
||||||
if(strcmp(SCHEME_FILE, fileurl->scheme) != 0) {
|
|
||||||
host = fileurl->host;
|
|
||||||
}
|
|
||||||
return(host);
|
|
||||||
}
|
|
||||||
|
|
||||||
int dload_interrupted;
|
int dload_interrupted;
|
||||||
static void inthandler(int signum)
|
static void inthandler(int signum)
|
||||||
@ -91,59 +85,107 @@ static void inthandler(int signum)
|
|||||||
dload_interrupted = 1;
|
dload_interrupted = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
#define check_stop() if(dload_interrupted) { ret = -1; goto cleanup; }
|
static int curl_progress(void *filename, double dltotal, double dlnow,
|
||||||
enum sighandlers { OLD = 0, NEW = 1 };
|
double ultotal, double ulnow) {
|
||||||
|
|
||||||
static int download_internal(const char *url, const char *localpath,
|
/* unused parameters */
|
||||||
|
(void)ultotal;
|
||||||
|
(void)ulnow;
|
||||||
|
|
||||||
|
if(dltotal == 0) {
|
||||||
|
return(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
if(dload_interrupted) {
|
||||||
|
return(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
handle->dlcb((const char*)filename, (long)dlnow, (long)dltotal);
|
||||||
|
|
||||||
|
return(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int curl_gethost(const char *url, char *buffer) {
|
||||||
|
int hostlen;
|
||||||
|
char *p;
|
||||||
|
|
||||||
|
if(strncmp(url, "file://", 7) == 0) {
|
||||||
|
strcpy(buffer, _("disk"));
|
||||||
|
} else {
|
||||||
|
p = strstr(url, "//");
|
||||||
|
if(!p) {
|
||||||
|
return(1);
|
||||||
|
}
|
||||||
|
p += 2; /* jump over the found // */
|
||||||
|
hostlen = strcspn(p, "/");
|
||||||
|
if(hostlen > 255) {
|
||||||
|
/* buffer overflow imminent */
|
||||||
|
_alpm_log(PM_LOG_ERROR, _("buffer overflow detected"));
|
||||||
|
return(1);
|
||||||
|
}
|
||||||
|
snprintf(buffer, hostlen + 1, "%s", p);
|
||||||
|
}
|
||||||
|
|
||||||
|
return(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int curl_download_internal(const char *url, const char *localpath,
|
||||||
int force) {
|
int force) {
|
||||||
|
int ret = -1;
|
||||||
FILE *localf = NULL;
|
FILE *localf = NULL;
|
||||||
|
char *destfile, *filename, *tempfile;
|
||||||
|
char hostname[256]; /* RFC1123 states applications should support this length */
|
||||||
struct stat st;
|
struct stat st;
|
||||||
int ret = 0;
|
long httpresp, timecond, remote_time, local_time;
|
||||||
off_t dl_thisfile = 0;
|
double remote_size, bytes_dl;
|
||||||
ssize_t nread = 0;
|
|
||||||
char *tempfile, *destfile, *filename;
|
|
||||||
struct sigaction sig_pipe[2], sig_int[2];
|
struct sigaction sig_pipe[2], sig_int[2];
|
||||||
|
|
||||||
off_t local_size = 0;
|
|
||||||
time_t local_time = 0;
|
|
||||||
|
|
||||||
struct url *fileurl;
|
|
||||||
struct url_stat ust;
|
|
||||||
fetchIO *dlf = NULL;
|
|
||||||
|
|
||||||
char buffer[PM_DLBUF_LEN];
|
|
||||||
|
|
||||||
filename = get_filename(url);
|
filename = get_filename(url);
|
||||||
if(!filename) {
|
if(!filename || curl_gethost(url, hostname) != 0) {
|
||||||
_alpm_log(PM_LOG_ERROR, _("url '%s' is invalid\n"), url);
|
_alpm_log(PM_LOG_ERROR, _("url '%s' is invalid\n"), url);
|
||||||
RET_ERR(PM_ERR_SERVER_BAD_URL, -1);
|
RET_ERR(PM_ERR_SERVER_BAD_URL, -1);
|
||||||
}
|
}
|
||||||
|
|
||||||
fileurl = fetchParseURL(url);
|
|
||||||
if(!fileurl) {
|
|
||||||
_alpm_log(PM_LOG_ERROR, _("url '%s' is invalid\n"), url);
|
|
||||||
RET_ERR(PM_ERR_LIBFETCH, -1);
|
|
||||||
}
|
|
||||||
|
|
||||||
destfile = get_destfile(localpath, filename);
|
destfile = get_destfile(localpath, filename);
|
||||||
tempfile = get_tempfile(localpath, filename);
|
tempfile = get_tempfile(localpath, filename);
|
||||||
|
|
||||||
if(stat(tempfile, &st) == 0 && S_ISREG(st.st_mode) && st.st_size > 0) {
|
/* the curl_easy handle is initialized with the alpm handle, so we only need
|
||||||
_alpm_log(PM_LOG_DEBUG, "tempfile found, attempting continuation\n");
|
* to reset the curl handle set parameters for each time it's used. */
|
||||||
local_time = fileurl->last_modified = st.st_mtime;
|
curl_easy_reset(handle->curl);
|
||||||
local_size = fileurl->offset = (off_t)st.st_size;
|
curl_easy_setopt(handle->curl, CURLOPT_URL, url);
|
||||||
dl_thisfile = st.st_size;
|
curl_easy_setopt(handle->curl, CURLOPT_FAILONERROR, 1L);
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_ENCODING, "deflate, gzip");
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_CONNECTTIMEOUT, 10L);
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_FILETIME, 1L);
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_NOPROGRESS, 0L);
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_FOLLOWLOCATION, 1L);
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_PROGRESSFUNCTION, curl_progress);
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_PROGRESSDATA, filename);
|
||||||
|
|
||||||
|
if(!force && stat(destfile, &st) == 0) {
|
||||||
|
/* assume its a sync, so we're starting from scratch. but, only download
|
||||||
|
* our local is out of date. */
|
||||||
|
local_time = (long)st.st_mtime;
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_TIMECONDITION, CURL_TIMECOND_IFMODSINCE);
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_TIMEVALUE, local_time);
|
||||||
|
} else if(stat(tempfile, &st) == 0 && st.st_size > 0) {
|
||||||
|
/* assume its a partial package download. we do not support resuming of
|
||||||
|
* transfers on partially downloaded sync DBs. */
|
||||||
localf = fopen(tempfile, "ab");
|
localf = fopen(tempfile, "ab");
|
||||||
} else if(!force && stat(destfile, &st) == 0 && S_ISREG(st.st_mode) && st.st_size > 0) {
|
curl_easy_setopt(handle->curl, CURLOPT_RESUME_FROM, (long)st.st_size);
|
||||||
_alpm_log(PM_LOG_DEBUG, "destfile found, using mtime only\n");
|
_alpm_log(PM_LOG_DEBUG, "tempfile found, attempting continuation");
|
||||||
local_time = fileurl->last_modified = st.st_mtime;
|
|
||||||
local_size = /* no fu->off here */ (off_t)st.st_size;
|
|
||||||
} else {
|
|
||||||
_alpm_log(PM_LOG_DEBUG, "no file found matching criteria, starting from scratch\n");
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* pass the raw filename for passing to the callback function */
|
/* no destfile and no tempfile. start from scratch */
|
||||||
_alpm_log(PM_LOG_DEBUG, "using '%s' for download progress\n", filename);
|
if(localf == NULL) {
|
||||||
|
localf = fopen(tempfile, "wb");
|
||||||
|
if(localf == NULL) {
|
||||||
|
goto cleanup;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* this has to be set _after_ figuring out which file we're opening */
|
||||||
|
curl_easy_setopt(handle->curl, CURLOPT_WRITEDATA, localf);
|
||||||
|
|
||||||
/* print proxy info for debug purposes */
|
/* print proxy info for debug purposes */
|
||||||
_alpm_log(PM_LOG_DEBUG, "HTTP_PROXY: %s\n", getenv("HTTP_PROXY"));
|
_alpm_log(PM_LOG_DEBUG, "HTTP_PROXY: %s\n", getenv("HTTP_PROXY"));
|
||||||
@ -151,9 +193,6 @@ static int download_internal(const char *url, const char *localpath,
|
|||||||
_alpm_log(PM_LOG_DEBUG, "FTP_PROXY: %s\n", getenv("FTP_PROXY"));
|
_alpm_log(PM_LOG_DEBUG, "FTP_PROXY: %s\n", getenv("FTP_PROXY"));
|
||||||
_alpm_log(PM_LOG_DEBUG, "ftp_proxy: %s\n", getenv("ftp_proxy"));
|
_alpm_log(PM_LOG_DEBUG, "ftp_proxy: %s\n", getenv("ftp_proxy"));
|
||||||
|
|
||||||
/* 10s timeout */
|
|
||||||
fetchTimeout = 10;
|
|
||||||
|
|
||||||
/* ignore any SIGPIPE signals- these may occur if our FTP socket dies or
|
/* ignore any SIGPIPE signals- these may occur if our FTP socket dies or
|
||||||
* something along those lines. Store the old signal handler first. */
|
* something along those lines. Store the old signal handler first. */
|
||||||
sig_pipe[NEW].sa_handler = SIG_IGN;
|
sig_pipe[NEW].sa_handler = SIG_IGN;
|
||||||
@ -169,146 +208,60 @@ static int download_internal(const char *url, const char *localpath,
|
|||||||
sigaction(SIGINT, NULL, &sig_int[OLD]);
|
sigaction(SIGINT, NULL, &sig_int[OLD]);
|
||||||
sigaction(SIGINT, &sig_int[NEW], NULL);
|
sigaction(SIGINT, &sig_int[NEW], NULL);
|
||||||
|
|
||||||
/* NOTE: libfetch does not reset the error code, be sure to do it before
|
|
||||||
* calls into the library */
|
|
||||||
|
|
||||||
/* TODO: if we call fetchStat() and get a redirect (disabling automagic
|
|
||||||
* redirect following), we should repeat the file locator stuff and get a new
|
|
||||||
* filename rather than only base if off the first URL, and then verify
|
|
||||||
* get_filename() didn't return ''. Of course, libfetch might not even allow
|
|
||||||
* us to even get that URL...FS#22645. This would allow us to download things
|
|
||||||
* without totally puking like
|
|
||||||
* http://www.archlinux.org/packages/community/x86_64/exim/download/ */
|
|
||||||
|
|
||||||
/* find out the remote size *and* mtime in one go. there is a lot of
|
|
||||||
* trouble in trying to do both size and "if-modified-since" logic in a
|
|
||||||
* non-stat request, so avoid it. */
|
|
||||||
fetchLastErrCode = 0;
|
|
||||||
if(fetchStat(fileurl, &ust, "") == -1) {
|
|
||||||
pm_errno = PM_ERR_LIBFETCH;
|
|
||||||
_alpm_log(PM_LOG_ERROR, _("failed retrieving file '%s' from %s : %s\n"),
|
|
||||||
filename, gethost(fileurl), fetchLastErrString);
|
|
||||||
ret = -1;
|
|
||||||
goto cleanup;
|
|
||||||
}
|
|
||||||
check_stop();
|
|
||||||
|
|
||||||
_alpm_log(PM_LOG_DEBUG, "ust.mtime: %ld local_time: %ld compare: %ld\n",
|
|
||||||
ust.mtime, local_time, local_time - ust.mtime);
|
|
||||||
_alpm_log(PM_LOG_DEBUG, "ust.size: %jd local_size: %jd compare: %jd\n",
|
|
||||||
(intmax_t)ust.size, (intmax_t)local_size, (intmax_t)(local_size - ust.size));
|
|
||||||
if(!force && ust.mtime && ust.mtime == local_time
|
|
||||||
&& ust.size && ust.size == local_size) {
|
|
||||||
/* the remote time and size values agreed with what we have, so move on
|
|
||||||
* because there is nothing more to do. */
|
|
||||||
_alpm_log(PM_LOG_DEBUG, "files are identical, skipping %s\n", filename);
|
|
||||||
ret = 1;
|
|
||||||
goto cleanup;
|
|
||||||
}
|
|
||||||
if(!ust.mtime || ust.mtime != local_time) {
|
|
||||||
_alpm_log(PM_LOG_DEBUG, "mtimes were different or unavailable, downloading %s from beginning\n", filename);
|
|
||||||
fileurl->offset = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
fetchLastErrCode = 0;
|
|
||||||
dlf = fetchGet(fileurl, "");
|
|
||||||
check_stop();
|
|
||||||
|
|
||||||
if(fetchLastErrCode != 0 || dlf == NULL) {
|
|
||||||
pm_errno = PM_ERR_LIBFETCH;
|
|
||||||
_alpm_log(PM_LOG_ERROR, _("failed retrieving file '%s' from %s : %s\n"),
|
|
||||||
filename, gethost(fileurl), fetchLastErrString);
|
|
||||||
ret = -1;
|
|
||||||
goto cleanup;
|
|
||||||
} else {
|
|
||||||
_alpm_log(PM_LOG_DEBUG, "connected to %s successfully\n", fileurl->host);
|
|
||||||
}
|
|
||||||
|
|
||||||
if(localf && fileurl->offset == 0) {
|
|
||||||
_alpm_log(PM_LOG_WARNING, _("resuming download of %s not possible; starting over\n"), filename);
|
|
||||||
fclose(localf);
|
|
||||||
localf = NULL;
|
|
||||||
} else if(fileurl->offset) {
|
|
||||||
_alpm_log(PM_LOG_DEBUG, "resuming download at position %jd\n", (intmax_t)fileurl->offset);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
if(localf == NULL) {
|
|
||||||
_alpm_rmrf(tempfile);
|
|
||||||
fileurl->offset = (off_t)0;
|
|
||||||
dl_thisfile = 0;
|
|
||||||
localf = fopen(tempfile, "wb");
|
|
||||||
if(localf == NULL) { /* still null? */
|
|
||||||
pm_errno = PM_ERR_RETRIEVE;
|
|
||||||
_alpm_log(PM_LOG_ERROR, _("error writing to file '%s': %s\n"),
|
|
||||||
tempfile, strerror(errno));
|
|
||||||
ret = -1;
|
|
||||||
goto cleanup;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Progress 0 - initialize */
|
/* Progress 0 - initialize */
|
||||||
if(handle->dlcb) {
|
if(handle->dlcb) {
|
||||||
handle->dlcb(filename, 0, ust.size);
|
handle->dlcb(filename, 0, 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
while((nread = fetchIO_read(dlf, buffer, PM_DLBUF_LEN)) > 0) {
|
/* perform transfer */
|
||||||
check_stop();
|
handle->curlerr = curl_easy_perform(handle->curl);
|
||||||
size_t nwritten = 0;
|
|
||||||
nwritten = fwrite(buffer, 1, (size_t)nread, localf);
|
|
||||||
if((nwritten != (size_t)nread) || ferror(localf)) {
|
|
||||||
pm_errno = PM_ERR_RETRIEVE;
|
|
||||||
_alpm_log(PM_LOG_ERROR, _("error writing to file '%s': %s\n"),
|
|
||||||
tempfile, strerror(errno));
|
|
||||||
ret = -1;
|
|
||||||
goto cleanup;
|
|
||||||
}
|
|
||||||
dl_thisfile += nread;
|
|
||||||
|
|
||||||
if(handle->dlcb) {
|
/* retrieve info about the state of the transfer */
|
||||||
handle->dlcb(filename, dl_thisfile, ust.size);
|
curl_easy_getinfo(handle->curl, CURLINFO_HTTP_CODE, &httpresp);
|
||||||
}
|
curl_easy_getinfo(handle->curl, CURLINFO_FILETIME, &remote_time);
|
||||||
}
|
curl_easy_getinfo(handle->curl, CURLINFO_CONTENT_LENGTH_DOWNLOAD, &remote_size);
|
||||||
|
curl_easy_getinfo(handle->curl, CURLINFO_SIZE_DOWNLOAD, &bytes_dl);
|
||||||
|
curl_easy_getinfo(handle->curl, CURLINFO_CONDITION_UNMET, &timecond);
|
||||||
|
|
||||||
/* did the transfer complete normally? */
|
/* time condition was met and we didn't download anything. we need to
|
||||||
if (nread == -1) {
|
* clean up the 0 byte .part file that's left behind. */
|
||||||
/* not PM_ERR_LIBFETCH here because libfetch error string might be empty */
|
if(bytes_dl == 0 && timecond == 1) {
|
||||||
pm_errno = PM_ERR_RETRIEVE;
|
ret = 1;
|
||||||
_alpm_log(PM_LOG_ERROR, _("failed retrieving file '%s' from %s\n"),
|
unlink(tempfile);
|
||||||
filename, gethost(fileurl));
|
|
||||||
ret = -1;
|
|
||||||
goto cleanup;
|
goto cleanup;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (ust.size != -1 && dl_thisfile < ust.size) {
|
if(handle->curlerr == CURLE_ABORTED_BY_CALLBACK) {
|
||||||
|
goto cleanup;
|
||||||
|
} else if(handle->curlerr != CURLE_OK) {
|
||||||
|
pm_errno = PM_ERR_LIBCURL;
|
||||||
|
_alpm_log(PM_LOG_ERROR, _("failed retrieving file '%s' from %s : %s\n"),
|
||||||
|
filename, hostname, curl_easy_strerror(handle->curlerr));
|
||||||
|
unlink(tempfile);
|
||||||
|
goto cleanup;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* remote_size isn't necessarily the full size of the file, just what the
|
||||||
|
* server reported as remaining to download. compare it to what curl reported
|
||||||
|
* as actually being transferred during curl_easy_perform() */
|
||||||
|
if((remote_size != -1 && bytes_dl != -1) && bytes_dl != remote_size) {
|
||||||
pm_errno = PM_ERR_RETRIEVE;
|
pm_errno = PM_ERR_RETRIEVE;
|
||||||
_alpm_log(PM_LOG_ERROR, _("%s appears to be truncated: %jd/%jd bytes\n"),
|
_alpm_log(PM_LOG_ERROR, _("%s appears to be truncated: %jd/%jd bytes\n"),
|
||||||
filename, (intmax_t)dl_thisfile, (intmax_t)ust.size);
|
filename, (intmax_t)bytes_dl, (intmax_t)remote_size);
|
||||||
ret = -1;
|
|
||||||
goto cleanup;
|
goto cleanup;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* probably safer to close the file descriptors now before renaming the file,
|
|
||||||
* for example to make sure the buffers are flushed.
|
|
||||||
*/
|
|
||||||
fclose(localf);
|
fclose(localf);
|
||||||
localf = NULL;
|
localf = NULL;
|
||||||
fetchIO_close(dlf);
|
|
||||||
dlf = NULL;
|
|
||||||
|
|
||||||
/* set the times on the file to the same as that of the remote file */
|
/* set the times on the file to the same as that of the remote file */
|
||||||
if(ust.mtime) {
|
if(remote_time != -1) {
|
||||||
struct timeval tv[2];
|
struct timeval tv[2];
|
||||||
memset(&tv, 0, sizeof(tv));
|
memset(&tv, 0, sizeof(tv));
|
||||||
tv[0].tv_sec = ust.atime;
|
tv[0].tv_sec = tv[1].tv_sec = remote_time;
|
||||||
tv[1].tv_sec = ust.mtime;
|
|
||||||
utimes(tempfile, tv);
|
utimes(tempfile, tv);
|
||||||
}
|
}
|
||||||
if(rename(tempfile, destfile)) {
|
rename(tempfile, destfile);
|
||||||
_alpm_log(PM_LOG_ERROR, _("could not rename %s to %s (%s)\n"),
|
|
||||||
tempfile, destfile, strerror(errno));
|
|
||||||
ret = -1;
|
|
||||||
}
|
|
||||||
ret = 0;
|
ret = 0;
|
||||||
|
|
||||||
cleanup:
|
cleanup:
|
||||||
@ -318,19 +271,14 @@ cleanup:
|
|||||||
/* if we still had a local file open, we got interrupted. set the mtimes on
|
/* if we still had a local file open, we got interrupted. set the mtimes on
|
||||||
* the file accordingly. */
|
* the file accordingly. */
|
||||||
fflush(localf);
|
fflush(localf);
|
||||||
if(ust.mtime) {
|
if(remote_time != -1) {
|
||||||
struct timeval tv[2];
|
struct timeval tv[2];
|
||||||
memset(&tv, 0, sizeof(tv));
|
memset(&tv, 0, sizeof(tv));
|
||||||
tv[0].tv_sec = ust.atime;
|
tv[0].tv_sec = tv[1].tv_sec = remote_time;
|
||||||
tv[1].tv_sec = ust.mtime;
|
|
||||||
futimes(fileno(localf), tv);
|
futimes(fileno(localf), tv);
|
||||||
}
|
}
|
||||||
fclose(localf);
|
fclose(localf);
|
||||||
}
|
}
|
||||||
if(dlf != NULL) {
|
|
||||||
fetchIO_close(dlf);
|
|
||||||
}
|
|
||||||
fetchFreeURL(fileurl);
|
|
||||||
|
|
||||||
/* restore the old signal handlers */
|
/* restore the old signal handlers */
|
||||||
sigaction(SIGINT, &sig_int[OLD], NULL);
|
sigaction(SIGINT, &sig_int[OLD], NULL);
|
||||||
@ -347,8 +295,8 @@ cleanup:
|
|||||||
static int download(const char *url, const char *localpath,
|
static int download(const char *url, const char *localpath,
|
||||||
int force) {
|
int force) {
|
||||||
if(handle->fetchcb == NULL) {
|
if(handle->fetchcb == NULL) {
|
||||||
#ifdef HAVE_LIBFETCH
|
#ifdef HAVE_LIBCURL
|
||||||
return(download_internal(url, localpath, force));
|
return(curl_download_internal(url, localpath, force));
|
||||||
#else
|
#else
|
||||||
RET_ERR(PM_ERR_EXTERNAL_DOWNLOAD, -1);
|
RET_ERR(PM_ERR_EXTERNAL_DOWNLOAD, -1);
|
||||||
#endif
|
#endif
|
||||||
|
@ -25,8 +25,6 @@
|
|||||||
|
|
||||||
#include <time.h>
|
#include <time.h>
|
||||||
|
|
||||||
#define PM_DLBUF_LEN (1024 * 16)
|
|
||||||
|
|
||||||
int _alpm_download_single_file(const char *filename,
|
int _alpm_download_single_file(const char *filename,
|
||||||
alpm_list_t *servers, const char *localpath,
|
alpm_list_t *servers, const char *localpath,
|
||||||
int force);
|
int force);
|
||||||
|
@ -20,21 +20,14 @@
|
|||||||
|
|
||||||
#include "config.h"
|
#include "config.h"
|
||||||
|
|
||||||
/* TODO: needed for the libfetch stuff, unfortunately- we should kill it */
|
#ifdef HAVE_LIBCURL
|
||||||
#include <stdio.h>
|
#include <curl/curl.h>
|
||||||
/* the following two are needed for FreeBSD's libfetch */
|
|
||||||
#include <limits.h> /* PATH_MAX */
|
|
||||||
#if defined(HAVE_SYS_PARAM_H)
|
|
||||||
#include <sys/param.h> /* MAXHOSTNAMELEN */
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifdef HAVE_LIBFETCH
|
|
||||||
#include <fetch.h> /* fetchLastErrString */
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* libalpm */
|
/* libalpm */
|
||||||
#include "util.h"
|
#include "util.h"
|
||||||
#include "alpm.h"
|
#include "alpm.h"
|
||||||
|
#include "handle.h"
|
||||||
|
|
||||||
const char SYMEXPORT *alpm_strerrorlast(void)
|
const char SYMEXPORT *alpm_strerrorlast(void)
|
||||||
{
|
{
|
||||||
@ -147,9 +140,9 @@ const char SYMEXPORT *alpm_strerror(int err)
|
|||||||
* requires the archive struct, so we can't. Just use a generic
|
* requires the archive struct, so we can't. Just use a generic
|
||||||
* error string instead. */
|
* error string instead. */
|
||||||
return _("libarchive error");
|
return _("libarchive error");
|
||||||
case PM_ERR_LIBFETCH:
|
case PM_ERR_LIBCURL:
|
||||||
#ifdef HAVE_LIBFETCH
|
#ifdef HAVE_LIBCURL
|
||||||
return fetchLastErrString;
|
return(curl_easy_strerror(handle->curlerr));
|
||||||
#else
|
#else
|
||||||
/* obviously shouldn't get here... */
|
/* obviously shouldn't get here... */
|
||||||
return _("download library error");
|
return _("download library error");
|
||||||
|
@ -71,6 +71,11 @@ void _alpm_handle_free(pmhandle_t *handle)
|
|||||||
closelog();
|
closelog();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#ifdef HAVE_LIBCURL
|
||||||
|
/* release curl handle */
|
||||||
|
curl_easy_cleanup(handle->curl);
|
||||||
|
#endif
|
||||||
|
|
||||||
/* free memory */
|
/* free memory */
|
||||||
_alpm_trans_free(handle->trans);
|
_alpm_trans_free(handle->trans);
|
||||||
FREE(handle->root);
|
FREE(handle->root);
|
||||||
@ -85,6 +90,7 @@ void _alpm_handle_free(pmhandle_t *handle)
|
|||||||
FREELIST(handle->ignorepkg);
|
FREELIST(handle->ignorepkg);
|
||||||
FREELIST(handle->ignoregrp);
|
FREELIST(handle->ignoregrp);
|
||||||
FREE(handle);
|
FREE(handle);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
alpm_cb_log SYMEXPORT alpm_option_get_logcb()
|
alpm_cb_log SYMEXPORT alpm_option_get_logcb()
|
||||||
|
@ -29,6 +29,10 @@
|
|||||||
#include "alpm.h"
|
#include "alpm.h"
|
||||||
#include "trans.h"
|
#include "trans.h"
|
||||||
|
|
||||||
|
#ifdef HAVE_LIBCURL
|
||||||
|
#include <curl/curl.h>
|
||||||
|
#endif
|
||||||
|
|
||||||
typedef struct _pmhandle_t {
|
typedef struct _pmhandle_t {
|
||||||
/* internal usage */
|
/* internal usage */
|
||||||
pmdb_t *db_local; /* local db pointer */
|
pmdb_t *db_local; /* local db pointer */
|
||||||
@ -37,6 +41,12 @@ typedef struct _pmhandle_t {
|
|||||||
FILE *lckstream; /* lock file stream pointer if one exists */
|
FILE *lckstream; /* lock file stream pointer if one exists */
|
||||||
pmtrans_t *trans;
|
pmtrans_t *trans;
|
||||||
|
|
||||||
|
#ifdef HAVE_LIBCURL
|
||||||
|
/* libcurl handle */
|
||||||
|
CURL *curl; /* reusable curl_easy handle */
|
||||||
|
CURLcode curlerr; /* last error produced by curl */
|
||||||
|
#endif
|
||||||
|
|
||||||
/* callback functions */
|
/* callback functions */
|
||||||
alpm_cb_log logcb; /* Log callback function */
|
alpm_cb_log logcb; /* Log callback function */
|
||||||
alpm_cb_download dlcb; /* Download callback function */
|
alpm_cb_download dlcb; /* Download callback function */
|
||||||
|
Loading…
Reference in New Issue
Block a user