1
0
mirror of https://github.com/moparisthebest/wget synced 2024-07-03 16:38:41 -04:00

[svn] Applied Edward Sabol's patch.

This commit is contained in:
hniksic 2000-03-02 05:28:31 -08:00
parent 3f97db2c27
commit f4f8e83327
2 changed files with 13 additions and 5 deletions

View File

@ -1,3 +1,9 @@
1998-11-03 Edward J. Sabol <sabol@alderaan.gsfc.nasa.gov>
* recur.c (recursive_retrieve): If a finite maximum depth is
specified, and we're are already at that depth, don't download the
HTML file for parsing.
2000-03-01 Dan Harkless <dan-wget@dilvish.speed.net>
* ftp.c (ftp_loop_internal): Call new downloaded_file() function,

View File

@ -221,9 +221,9 @@ recursive_retrieve (const char *file, const char *this_url)
Addendum: If the URL is FTP, and it is to be loaded, only the
domain and suffix settings are "stronger".
Note that .html and (yuck) .htm will get loaded
regardless of suffix rules (but that is remedied later with
unlink).
Note that .html and (yuck) .htm will get loaded regardless of
suffix rules (but that is remedied later with unlink) unless
the depth equals the maximum depth.
More time- and memory- consuming tests should be put later on
the list. */
@ -306,11 +306,13 @@ recursive_retrieve (const char *file, const char *this_url)
b) it is "htm"
If the file *is* supposed to be HTML, it will *not* be
subject to acc/rej rules. That's why the `!'. */
subject to acc/rej rules, unless a finite maximum depth has
been specified and the current depth is the maximum depth. */
if (!
(!*u->file
|| (((suf = suffix (constr)) != NULL)
&& (!strcmp (suf, "html") || !strcmp (suf, "htm")))))
&& ((!strcmp (suf, "html") || !strcmp (suf, "htm"))
&& ((opt.reclevel != 0) && (depth != opt.reclevel))))))
{
if (!acceptable (u->file))
{