1
0
mirror of https://github.com/moparisthebest/wget synced 2024-07-03 16:38:41 -04:00

Add space after function names

This commit is contained in:
Tim Rühsen 2014-11-26 12:39:47 +01:00
parent f0e7326c20
commit d87fdecd55
2 changed files with 16 additions and 12 deletions

View File

@ -1,6 +1,10 @@
2014-11-26 Tim Ruehsen <tim.ruehsen@gmx.de> 2014-11-26 Tim Ruehsen <tim.ruehsen@gmx.de>
* src/recur.c: Fix blacklisting of URLs to download * recur.c: Add space after function names
2014-11-26 Tim Ruehsen <tim.ruehsen@gmx.de>
* recur.c: Fix blacklisting of URLs to download
Fixes a reported crash and prevents multiple downloads of the Fixes a reported crash and prevents multiple downloads of the
same file in case the URL is escaped in different ways. same file in case the URL is escaped in different ways.

View File

@ -160,16 +160,16 @@ url_dequeue (struct url_queue *queue, struct iri **i,
return true; return true;
} }
static void blacklist_add(struct hash_table *blacklist, const char *url) static void blacklist_add (struct hash_table *blacklist, const char *url)
{ {
char *url_unescaped = xstrdup(url); char *url_unescaped = xstrdup (url);
url_unescape (url_unescaped); url_unescape (url_unescaped);
string_set_add (blacklist, url_unescaped); string_set_add (blacklist, url_unescaped);
xfree (url_unescaped); xfree (url_unescaped);
} }
static int blacklist_contains(struct hash_table *blacklist, const char *url) static int blacklist_contains (struct hash_table *blacklist, const char *url)
{ {
char *url_unescaped = xstrdup(url); char *url_unescaped = xstrdup(url);
int ret; int ret;
@ -241,7 +241,7 @@ retrieve_tree (struct url *start_url_parsed, struct iri *pi)
just URL so we enqueue the canonical form of the URL. */ just URL so we enqueue the canonical form of the URL. */
url_enqueue (queue, i, xstrdup (start_url_parsed->url), NULL, 0, true, url_enqueue (queue, i, xstrdup (start_url_parsed->url), NULL, 0, true,
false); false);
blacklist_add(blacklist, start_url_parsed->url); blacklist_add (blacklist, start_url_parsed->url);
while (1) while (1)
{ {
@ -332,7 +332,7 @@ retrieve_tree (struct url *start_url_parsed, struct iri *pi)
else else
/* Make sure that the old pre-redirect form gets /* Make sure that the old pre-redirect form gets
blacklisted. */ blacklisted. */
blacklist_add(blacklist, url); blacklist_add (blacklist, url);
} }
xfree (url); xfree (url);
@ -343,7 +343,7 @@ retrieve_tree (struct url *start_url_parsed, struct iri *pi)
xfree (url); xfree (url);
url = xstrdup (url_parsed->url); url = xstrdup (url_parsed->url);
} }
url_free(url_parsed); url_free (url_parsed);
} }
if (opt.spider) if (opt.spider)
@ -425,7 +425,7 @@ retrieve_tree (struct url *start_url_parsed, struct iri *pi)
/* We blacklist the URL we have enqueued, because we /* We blacklist the URL we have enqueued, because we
don't want to enqueue (and hence download) the don't want to enqueue (and hence download) the
same URL twice. */ same URL twice. */
blacklist_add(blacklist, child->url->url); blacklist_add (blacklist, child->url->url);
} }
} }
@ -497,7 +497,7 @@ retrieve_tree (struct url *start_url_parsed, struct iri *pi)
URL is to be descended to. This is only ever called from URL is to be descended to. This is only ever called from
retrieve_tree, but is in a separate function for clarity. retrieve_tree, but is in a separate function for clarity.
The most expensive checks (such as those for robots) are memorized The most expensive checks (such as those for robots) are memoized
by storing these URLs to BLACKLIST. This may or may not help. It by storing these URLs to BLACKLIST. This may or may not help. It
will help if those URLs are encountered many times. */ will help if those URLs are encountered many times. */
@ -512,7 +512,7 @@ download_child_p (const struct urlpos *upos, struct url *parent, int depth,
DEBUGP (("Deciding whether to enqueue \"%s\".\n", url)); DEBUGP (("Deciding whether to enqueue \"%s\".\n", url));
if (blacklist_contains(blacklist, url)) if (blacklist_contains (blacklist, url))
{ {
if (opt.spider) if (opt.spider)
{ {
@ -691,7 +691,7 @@ download_child_p (const struct urlpos *upos, struct url *parent, int depth,
if (!res_match_path (specs, u->path)) if (!res_match_path (specs, u->path))
{ {
DEBUGP (("Not following %s because robots.txt forbids it.\n", url)); DEBUGP (("Not following %s because robots.txt forbids it.\n", url));
blacklist_add(blacklist, url); blacklist_add (blacklist, url);
goto out; goto out;
} }
} }
@ -734,7 +734,7 @@ descend_redirect_p (const char *redirected, struct url *orig_parsed, int depth,
start_url_parsed, blacklist, iri); start_url_parsed, blacklist, iri);
if (success) if (success)
blacklist_add(blacklist, upos->url->url); blacklist_add (blacklist, upos->url->url);
else else
DEBUGP (("Redirection \"%s\" failed the test.\n", redirected)); DEBUGP (("Redirection \"%s\" failed the test.\n", redirected));