1
0
mirror of https://github.com/moparisthebest/curl synced 2024-08-13 17:03:50 -04:00

wording: avoid blacklist/whitelist stereotypes

Instead of discussing if there's value or meaning (implied or not) in
the colors, let's use words without the same possibly negative
associations.

Closes #5546
This commit is contained in:
Daniel Stenberg 2020-06-09 16:08:11 +02:00
parent f54b6c4bc2
commit eab2f95c0d
No known key found for this signature in database
GPG Key ID: 5CC908FDB71E12C2
17 changed files with 67 additions and 70 deletions

View File

@ -346,7 +346,7 @@ before_script:
script:
- ./scripts/travis/script.sh || travis_terminate 1
# whitelist branches to avoid testing feature branches twice (as branch and as pull request)
# select branches to avoid testing feature branches twice (as branch and as pull request)
branches:
only:
- master

View File

@ -285,7 +285,7 @@ test_script:
echo APPVEYOR_API_URL=%APPVEYOR_API_URL% &&
bash.exe -e -l -c "cd /c/projects/curl/tests && ./runtests.pl -a -b$(($(echo '%APPVEYOR_API_URL%' | cut -d'/' -f3 | cut -d':' -f2)+1)) -p !flaky %DISABLED_TESTS%" )
# whitelist branches to avoid testing feature branches twice (as branch and as pull request)
# select branches to avoid testing feature branches twice (as branch and as pull request)
branches:
only:
- master

View File

@ -9,7 +9,7 @@ check that it adheres to our [Source Code Style guide](CODE_STYLE.md).
## Command line options
`-W[file]` whitelists that file and excludes it from being checked. Helpful
`-W[file]` skip that file and excludes it from being checked. Helpful
when, for example, one of the files is generated.
`-D[dir]` directory name to prepend to file names when accessing them.
@ -158,5 +158,5 @@ instances are ignored and nothing extra.
This is a method we've transitioned away from. Use inline ignores as far as
possible.
Make a `checksrc.whitelist` file in the directory of the source code with the
Make a `checksrc.skip` file in the directory of the source code with the
false positive, and include the full offending line into this file.

View File

@ -78,7 +78,7 @@
depending on the OS or build configuration. The --ssl-no-revoke option was
introduced in 7.44.0 to disable revocation checking but currently is only
supported for Schannel (the native Windows SSL library), with an exception
in the case of Windows' Untrusted Publishers blacklist which it seems can't
in the case of Windows' Untrusted Publishers block list which it seems can't
be bypassed. This option may have broader support to accommodate other SSL
backends in the future.

View File

@ -5,7 +5,7 @@
.\" * | (__| |_| | _ <| |___
.\" * \___|\___/|_| \_\_____|
.\" *
.\" * Copyright (C) 1998 - 2019, Daniel Stenberg, <daniel@haxx.se>, et al.
.\" * Copyright (C) 1998 - 2020, Daniel Stenberg, <daniel@haxx.se>, et al.
.\" *
.\" * This software is licensed as described in the file COPYING, which
.\" * you should have received as part of this distribution. The terms
@ -22,7 +22,7 @@
.\"
.TH CURLMOPT_PIPELINING_SERVER_BL 3 "4 Nov 2014" "libcurl 7.39.0" "curl_multi_setopt options"
.SH NAME
CURLMOPT_PIPELINING_SERVER_BL \- pipelining server blacklist
CURLMOPT_PIPELINING_SERVER_BL \- pipelining server block list
.SH SYNOPSIS
#include <curl/curl.h>
@ -31,28 +31,28 @@ CURLMcode curl_multi_setopt(CURLM *handle, CURLMOPT_PIPELINING_SERVER_BL, char *
No function since pipelining was removed in 7.62.0.
Pass a \fBservers\fP array of char *, ending with a NULL entry. This is a list
of server types prefixes (in the Server: HTTP header) that are blacklisted
from pipelining, i.e server types that are known to not support HTTP
of server types prefixes (in the Server: HTTP header) that are blocked from
pipelining, i.e server types that are known to not support HTTP
pipelining. The array is copied by libcurl.
Note that the comparison matches if the Server: header begins with the string
in the blacklist, i.e "Server: Ninja 1.2.3" and "Server: Ninja 1.4.0" can
both be blacklisted by having "Ninja" in the backlist.
in the block list, i.e "Server: Ninja 1.2.3" and "Server: Ninja 1.4.0" can
both be blocked by having "Ninja" in the list.
Pass a NULL pointer to clear the blacklist.
Pass a NULL pointer to clear the block list.
.SH DEFAULT
The default value is NULL, which means that there is no blacklist.
The default value is NULL, which means that there is no block list.
.SH PROTOCOLS
.SH EXAMPLE
.nf
server_blacklist[] =
char *server_block_list[] =
{
"Microsoft-IIS/6.0",
"nginx/0.8.54",
NULL
};
curl_multi_setopt(m, CURLMOPT_PIPELINING_SERVER_BL, server_blacklist);
curl_multi_setopt(m, CURLMOPT_PIPELINING_SERVER_BL, server_block_list);
.fi
.SH AVAILABILITY
Added in 7.30.0

View File

@ -5,7 +5,7 @@
.\" * | (__| |_| | _ <| |___
.\" * \___|\___/|_| \_\_____|
.\" *
.\" * Copyright (C) 1998 - 2019, Daniel Stenberg, <daniel@haxx.se>, et al.
.\" * Copyright (C) 1998 - 2020, Daniel Stenberg, <daniel@haxx.se>, et al.
.\" *
.\" * This software is licensed as described in the file COPYING, which
.\" * you should have received as part of this distribution. The terms
@ -22,7 +22,7 @@
.\"
.TH CURLMOPT_PIPELINING_SITE_BL 3 "4 Nov 2014" "libcurl 7.39.0" "curl_multi_setopt options"
.SH NAME
CURLMOPT_PIPELINING_SITE_BL \- pipelining host blacklist
CURLMOPT_PIPELINING_SITE_BL \- pipelining host block list
.SH SYNOPSIS
#include <curl/curl.h>
@ -31,24 +31,24 @@ CURLMcode curl_multi_setopt(CURLM *handle, CURLMOPT_PIPELINING_SITE_BL, char **h
No function since pipelining was removed in 7.62.0.
Pass a \fBhosts\fP array of char *, ending with a NULL entry. This is a list
of sites that are blacklisted from pipelining, i.e sites that are known to not
of sites that are blocked from pipelining, i.e sites that are known to not
support HTTP pipelining. The array is copied by libcurl.
Pass a NULL pointer to clear the blacklist.
Pass a NULL pointer to clear the block list.
.SH DEFAULT
The default value is NULL, which means that there is no blacklist.
The default value is NULL, which means that there is no block list.
.SH PROTOCOLS
HTTP(S)
.SH EXAMPLE
.nf
site_blacklist[] =
char *site_block_list[] =
{
"www.haxx.se",
"www.example.com:1234",
NULL
};
curl_multi_setopt(m, CURLMOPT_PIPELINING_SITE_BL, site_blacklist);
curl_multi_setopt(m, CURLMOPT_PIPELINING_SITE_BL, site_block_list);
.fi
.SH AVAILABILITY
Added in 7.30.0

View File

@ -5,7 +5,7 @@
.\" * | (__| |_| | _ <| |___
.\" * \___|\___/|_| \_\_____|
.\" *
.\" * Copyright (C) 1998 - 2019, Daniel Stenberg, <daniel@haxx.se>, et al.
.\" * Copyright (C) 1998 - 2020, Daniel Stenberg, <daniel@haxx.se>, et al.
.\" *
.\" * This software is licensed as described in the file COPYING, which
.\" * you should have received as part of this distribution. The terms
@ -65,7 +65,7 @@ additional \fIsetsockopt(2)\fP calls can of course be done on the socket at
the user's discretion. A \fICURL_SOCKET_BAD\fP return value from the callback
function will signal an unrecoverable error to libcurl and it will return
\fICURLE_COULDNT_CONNECT\fP from the function that triggered this callback.
This return code can be used for IP address blacklisting.
This return code can be used for IP address block listing.
If you want to pass in a socket with an already established connection, pass
the socket back with this callback and then use

View File

@ -42,7 +42,7 @@ supported for DarwinSSL, NSS and OpenSSL.
Tells libcurl to disable certificate revocation checks for those SSL backends
where such behavior is present. This option is only supported for Schannel
(the native Windows SSL library), with an exception in the case of Windows'
Untrusted Publishers blacklist which it seems can't be bypassed. (Added in
Untrusted Publishers block list which it seems can't be bypassed. (Added in
7.44.0)
.IP CURLSSLOPT_NO_PARTIALCHAIN
Tells libcurl to not accept "partial" certificate chains, which it otherwise

View File

@ -42,7 +42,7 @@ supported for DarwinSSL, NSS and OpenSSL.
Tells libcurl to disable certificate revocation checks for those SSL backends
where such behavior is present. This option is only supported for Schannel
(the native Windows SSL library), with an exception in the case of Windows'
Untrusted Publishers blacklist which it seems can't be bypassed. (Added in
Untrusted Publishers block list which it seems can't be bypassed. (Added in
7.44.0)
.IP CURLSSLOPT_NO_PARTIALCHAIN
Tells libcurl to not accept "partial" certificate chains, which it otherwise

View File

@ -377,12 +377,10 @@ typedef enum {
will not be considered for pipelining */
CURLOPT(CURLMOPT_CHUNK_LENGTH_PENALTY_SIZE, CURLOPTTYPE_OFF_T, 10),
/* a list of site names(+port) that are blacklisted from
pipelining */
/* a list of site names(+port) that are blocked from pipelining */
CURLOPT(CURLMOPT_PIPELINING_SITE_BL, CURLOPTTYPE_OBJECTPOINT, 11),
/* a list of server types that are blacklisted from
pipelining */
/* a list of server types that are blocked from pipelining */
CURLOPT(CURLMOPT_PIPELINING_SERVER_BL, CURLOPTTYPE_OBJECTPOINT, 12),
/* maximum number of open connections in total */

View File

@ -31,14 +31,14 @@ my $warnings = 0;
my $swarnings = 0;
my $errors = 0;
my $serrors = 0;
my $suppressed; # whitelisted problems
my $suppressed; # skipped problems
my $file;
my $dir=".";
my $wlist="";
my @alist;
my $windows_os = $^O eq 'MSWin32' || $^O eq 'cygwin' || $^O eq 'msys';
my $verbose;
my %whitelist;
my %skiplist;
my %ignore;
my %ignore_set;
@ -84,12 +84,12 @@ my %warnings = (
'TYPEDEFSTRUCT' => 'typedefed struct',
);
sub readwhitelist {
open(W, "<$dir/checksrc.whitelist") or return;
sub readskiplist {
open(W, "<$dir/checksrc.skip") or return;
my @all=<W>;
for(@all) {
$windows_os ? $_ =~ s/\r?\n$// : chomp;
$whitelist{$_}=1;
$skiplist{$_}=1;
}
close(W);
}
@ -142,8 +142,8 @@ sub checkwarn {
# print STDERR "Dev! there's no description for $name!\n";
#}
# checksrc.whitelist
if($whitelist{$line}) {
# checksrc.skip
if($skiplist{$line}) {
$nowarn = 1;
}
# !checksrc! controlled
@ -228,7 +228,7 @@ if(!$file) {
print " -A[rule] Accept this violation, can be used multiple times\n";
print " -D[DIR] Directory to prepend file names\n";
print " -h Show help output\n";
print " -W[file] Whitelist the given file - ignore all its flaws\n";
print " -W[file] Skip the given file - ignore all its flaws\n";
print " -i<n> Indent spaces. Default: 2\n";
print " -m<n> Maximum line length. Default: 79\n";
print "\nDetects and warns for these problems:\n";
@ -238,7 +238,7 @@ if(!$file) {
exit;
}
readwhitelist();
readskiplist();
readlocalfile();
do {

View File

@ -3698,9 +3698,8 @@ CURLcode Curl_http_readwrite_headers(struct Curl_easy *data,
else if(conn->httpversion == 20 ||
(k->upgr101 == UPGR101_REQUESTED && k->httpcode == 101)) {
DEBUGF(infof(data, "HTTP/2 found, allow multiplexing\n"));
/* HTTP/2 cannot blacklist multiplexing since it is a core
functionality of the protocol */
/* HTTP/2 cannot avoid multiplexing since it is a core functionality
of the protocol */
conn->bundle->multiuse = BUNDLE_MULTIPLEX;
}
else if(conn->httpversion >= 11 &&

View File

@ -28,7 +28,7 @@
#
# regexes of files to not scan
my @whitelist=(
my @skiplist=(
'^tests\/data\/test(\d+)$', # test case data
'^docs\/cmdline-opts\/[a-z]+(.*)\.d$', # curl.1 pieces
'(\/|^)[A-Z0-9_.-]+$', # all uppercase file name, possibly with dot and dash
@ -168,10 +168,10 @@ else {
for my $f (@all) {
chomp $f;
my $skipped = 0;
for my $skip (@whitelist) {
for my $skip (@skiplist) {
#print "$f matches $skip ?\n";
if($f =~ /$skip/) {
$whitelisted++;
$skiplisted++;
$skipped = 1;
#print "$f: SKIPPED ($skip)\n";
last;
@ -186,6 +186,6 @@ for my $f (@all) {
print STDERR "$missing files have no copyright\n" if($missing);
print STDERR "$wrong files have wrong copyright year\n" if ($wrong);
print STDERR "$whitelisted files are whitelisted\n" if ($whitelisted);
print STDERR "$skiplisted files are skipped\n" if ($skiplisted);
exit 1 if($missing || $wrong);

View File

@ -20,7 +20,7 @@ lib1550
</tool>
<name>
verify setting pipeling blacklisting options
verify setting pipeling blocklisting options
</name>
<command>
http://%HOSTIP:%NOLISTENPORT/1550

View File

@ -40,13 +40,13 @@ http
lib1900
</tool>
<name>
HTTP GET using pipelining, blacklisted site
HTTP GET using pipelining, blocklisted site
</name>
<command>
http://%HOSTIP:%HTTPPIPEPORT/ log/urls1901.txt
</command>
<file name="log/urls1901.txt">
blacklist_site 127.0.0.1:%HTTPPIPEPORT
blocklist_site 127.0.0.1:%HTTPPIPEPORT
0 1k.txt
1000 100k.txt
0 1k.txt

View File

@ -5,7 +5,7 @@
* | (__| |_| | _ <| |___
* \___|\___/|_| \_\_____|
*
* Copyright (C) 2013 - 2019, Linus Nielsen Feltzing, <linus@haxx.se>
* Copyright (C) 2013 - 2020, Linus Nielsen Feltzing, <linus@haxx.se>
*
* This software is licensed as described in the file COPYING, which
* you should have received as part of this distribution. The terms
@ -27,16 +27,16 @@
#define TEST_HANG_TIMEOUT 60 * 1000
#define MAX_URLS 200
#define MAX_BLACKLIST 20
#define MAX_BLOCKLIST 20
static int urltime[MAX_URLS];
static char *urlstring[MAX_URLS];
static CURL *handles[MAX_URLS];
static char *site_blacklist[MAX_BLACKLIST];
static char *server_blacklist[MAX_BLACKLIST];
static char *site_blocklist[MAX_BLOCKLIST];
static char *server_blocklist[MAX_BLOCKLIST];
static int num_handles;
static int blacklist_num_servers;
static int blacklist_num_sites;
static int blocklist_num_servers;
static int blocklist_num_sites;
static size_t
write_callback(void *contents, size_t size, size_t nmemb, void *userp)
@ -55,8 +55,8 @@ static int parse_url_file(const char *filename)
char buf[200];
num_handles = 0;
blacklist_num_sites = 0;
blacklist_num_servers = 0;
blocklist_num_sites = 0;
blocklist_num_servers = 0;
f = fopen(filename, "rb");
if(!f)
@ -70,9 +70,9 @@ static int parse_url_file(const char *filename)
continue;
}
if(fscanf(f, "blacklist_site %199s\n", buf)) {
site_blacklist[blacklist_num_sites] = strdup(buf);
blacklist_num_sites++;
if(fscanf(f, "blocklist_site %199s\n", buf)) {
site_blocklist[blocklist_num_sites] = strdup(buf);
blocklist_num_sites++;
continue;
}
@ -80,8 +80,8 @@ static int parse_url_file(const char *filename)
}
fclose(f);
site_blacklist[blacklist_num_sites] = NULL;
server_blacklist[blacklist_num_servers] = NULL;
site_blocklist[blocklist_num_sites] = NULL;
server_blocklist[blocklist_num_servers] = NULL;
return num_handles;
}
@ -91,11 +91,11 @@ static void free_urls(void)
for(i = 0; i < num_handles; i++) {
Curl_safefree(urlstring[i]);
}
for(i = 0; i < blacklist_num_servers; i++) {
Curl_safefree(server_blacklist[i]);
for(i = 0; i < blocklist_num_servers; i++) {
Curl_safefree(server_blocklist[i]);
}
for(i = 0; i < blacklist_num_sites; i++) {
Curl_safefree(site_blacklist[i]);
for(i = 0; i < blocklist_num_sites; i++) {
Curl_safefree(site_blocklist[i]);
}
}
@ -159,8 +159,8 @@ int test(char *URL)
multi_setopt(m, CURLMOPT_CONTENT_LENGTH_PENALTY_SIZE, 15000L);
multi_setopt(m, CURLMOPT_CHUNK_LENGTH_PENALTY_SIZE, 10000L);
multi_setopt(m, CURLMOPT_PIPELINING_SITE_BL, site_blacklist);
multi_setopt(m, CURLMOPT_PIPELINING_SERVER_BL, server_blacklist);
multi_setopt(m, CURLMOPT_PIPELINING_SITE_BL, site_blocklist);
multi_setopt(m, CURLMOPT_PIPELINING_SERVER_BL, server_blocklist);
last_handle_add = tutil_tvnow();

View File

@ -6,7 +6,7 @@
# | (__| |_| | _ <| |___
# \___|\___/|_| \_\_____|
#
# Copyright (C) 2016 - 2019, Daniel Stenberg, <daniel@haxx.se>, et al.
# Copyright (C) 2016 - 2020, Daniel Stenberg, <daniel@haxx.se>, et al.
#
# This software is licensed as described in the file COPYING, which
# you should have received as part of this distribution. The terms
@ -130,7 +130,7 @@ scanmanpage("$root/docs/libcurl/curl_easy_setopt.3", @curlopt);
scanmanpage("$root/docs/libcurl/curl_easy_getinfo.3", @curlinfo);
scanmanpage("$root/docs/libcurl/curl_multi_setopt.3", @curlmopt);
# using this hash array, we can whitelist specific options
# using this hash array, we can skip specific options
my %opts = (
# pretend these --no options exists in tool_getparam.c
'--no-alpn' => 1,