mirror of
https://github.com/moparisthebest/curl
synced 2024-12-22 16:18:48 -05:00
2ae6c47d5d
It was pointed out that the special case libcurl did for 416 was incorrect and wrong. 416 is not really different to other errors so the response body must be handled like for other errors/http responses. Reported by: Chris Smowton Bug: http://curl.haxx.se/bug/view.cgi?id=3076808
3785 lines
122 KiB
C
3785 lines
122 KiB
C
/***************************************************************************
|
|
* _ _ ____ _
|
|
* Project ___| | | | _ \| |
|
|
* / __| | | | |_) | |
|
|
* | (__| |_| | _ <| |___
|
|
* \___|\___/|_| \_\_____|
|
|
*
|
|
* Copyright (C) 1998 - 2010, Daniel Stenberg, <daniel@haxx.se>, et al.
|
|
*
|
|
* This software is licensed as described in the file COPYING, which
|
|
* you should have received as part of this distribution. The terms
|
|
* are also available at http://curl.haxx.se/docs/copyright.html.
|
|
*
|
|
* You may opt to use, copy, modify, merge, publish, distribute and/or sell
|
|
* copies of the Software, and permit persons to whom the Software is
|
|
* furnished to do so, under the terms of the COPYING file.
|
|
*
|
|
* This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
|
|
* KIND, either express or implied.
|
|
*
|
|
***************************************************************************/
|
|
|
|
#include "setup.h"
|
|
|
|
#ifndef CURL_DISABLE_HTTP
|
|
/* -- WIN32 approved -- */
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <stdarg.h>
|
|
#include <stdlib.h>
|
|
#include <ctype.h>
|
|
|
|
#ifdef WIN32
|
|
#include <time.h>
|
|
#include <io.h>
|
|
#else
|
|
#ifdef HAVE_SYS_SOCKET_H
|
|
#include <sys/socket.h>
|
|
#endif
|
|
#ifdef HAVE_NETINET_IN_H
|
|
#include <netinet/in.h>
|
|
#endif
|
|
#ifdef HAVE_SYS_TIME_H
|
|
#include <sys/time.h>
|
|
#endif
|
|
|
|
#ifdef HAVE_TIME_H
|
|
#ifdef TIME_WITH_SYS_TIME
|
|
#include <time.h>
|
|
#endif
|
|
#endif
|
|
|
|
#ifdef HAVE_UNISTD_H
|
|
#include <unistd.h>
|
|
#endif
|
|
#ifdef HAVE_NETDB_H
|
|
#include <netdb.h>
|
|
#endif
|
|
#ifdef HAVE_ARPA_INET_H
|
|
#include <arpa/inet.h>
|
|
#endif
|
|
#ifdef HAVE_NET_IF_H
|
|
#include <net/if.h>
|
|
#endif
|
|
#ifdef HAVE_SYS_IOCTL_H
|
|
#include <sys/ioctl.h>
|
|
#endif
|
|
|
|
#ifdef HAVE_SYS_PARAM_H
|
|
#include <sys/param.h>
|
|
#endif
|
|
|
|
#endif
|
|
|
|
#include "urldata.h"
|
|
#include <curl/curl.h>
|
|
#include "transfer.h"
|
|
#include "sendf.h"
|
|
#include "easyif.h" /* for Curl_convert_... prototypes */
|
|
#include "formdata.h"
|
|
#include "progress.h"
|
|
#include "curl_base64.h"
|
|
#include "cookie.h"
|
|
#include "strequal.h"
|
|
#include "sslgen.h"
|
|
#include "http_digest.h"
|
|
#include "http_ntlm.h"
|
|
#include "http_negotiate.h"
|
|
#include "url.h"
|
|
#include "share.h"
|
|
#include "hostip.h"
|
|
#include "http.h"
|
|
#include "curl_memory.h"
|
|
#include "select.h"
|
|
#include "parsedate.h" /* for the week day and month names */
|
|
#include "strtoofft.h"
|
|
#include "multiif.h"
|
|
#include "rawstr.h"
|
|
#include "content_encoding.h"
|
|
#include "rtsp.h"
|
|
|
|
#define _MPRINTF_REPLACE /* use our functions only */
|
|
#include <curl/mprintf.h>
|
|
|
|
/* The last #include file should be: */
|
|
#include "memdebug.h"
|
|
|
|
/* Default proxy timeout in milliseconds */
|
|
#define PROXY_TIMEOUT (3600*1000)
|
|
|
|
/*
|
|
* Forward declarations.
|
|
*/
|
|
|
|
static int http_getsock_do(struct connectdata *conn,
|
|
curl_socket_t *socks,
|
|
int numsocks);
|
|
#ifdef USE_SSL
|
|
static CURLcode https_connecting(struct connectdata *conn, bool *done);
|
|
static int https_getsock(struct connectdata *conn,
|
|
curl_socket_t *socks,
|
|
int numsocks);
|
|
#else
|
|
#define https_connecting(x,y) CURLE_COULDNT_CONNECT
|
|
#endif
|
|
|
|
/*
|
|
* HTTP handler interface.
|
|
*/
|
|
const struct Curl_handler Curl_handler_http = {
|
|
"HTTP", /* scheme */
|
|
ZERO_NULL, /* setup_connection */
|
|
Curl_http, /* do_it */
|
|
Curl_http_done, /* done */
|
|
ZERO_NULL, /* do_more */
|
|
Curl_http_connect, /* connect_it */
|
|
ZERO_NULL, /* connecting */
|
|
ZERO_NULL, /* doing */
|
|
ZERO_NULL, /* proto_getsock */
|
|
http_getsock_do, /* doing_getsock */
|
|
ZERO_NULL, /* perform_getsock */
|
|
ZERO_NULL, /* disconnect */
|
|
PORT_HTTP, /* defport */
|
|
PROT_HTTP, /* protocol */
|
|
};
|
|
|
|
#ifdef USE_SSL
|
|
/*
|
|
* HTTPS handler interface.
|
|
*/
|
|
const struct Curl_handler Curl_handler_https = {
|
|
"HTTPS", /* scheme */
|
|
ZERO_NULL, /* setup_connection */
|
|
Curl_http, /* do_it */
|
|
Curl_http_done, /* done */
|
|
ZERO_NULL, /* do_more */
|
|
Curl_http_connect, /* connect_it */
|
|
https_connecting, /* connecting */
|
|
ZERO_NULL, /* doing */
|
|
https_getsock, /* proto_getsock */
|
|
http_getsock_do, /* doing_getsock */
|
|
ZERO_NULL, /* perform_getsock */
|
|
ZERO_NULL, /* disconnect */
|
|
PORT_HTTPS, /* defport */
|
|
PROT_HTTP | PROT_HTTPS | PROT_SSL /* protocol */
|
|
};
|
|
#endif
|
|
|
|
|
|
/*
|
|
* checkheaders() checks the linked list of custom HTTP headers for a
|
|
* particular header (prefix).
|
|
*
|
|
* Returns a pointer to the first matching header or NULL if none matched.
|
|
*/
|
|
char *Curl_checkheaders(struct SessionHandle *data, const char *thisheader)
|
|
{
|
|
struct curl_slist *head;
|
|
size_t thislen = strlen(thisheader);
|
|
|
|
for(head = data->set.headers; head; head=head->next) {
|
|
if(Curl_raw_nequal(head->data, thisheader, thislen))
|
|
return head->data;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* Strip off leading and trailing whitespace from the value in the
|
|
* given HTTP header line and return a strdupped copy. Returns NULL in
|
|
* case of allocation failure. Returns an empty string if the header value
|
|
* consists entirely of whitespace.
|
|
*/
|
|
char *Curl_copy_header_value(const char *h)
|
|
{
|
|
const char *start;
|
|
const char *end;
|
|
char *value;
|
|
size_t len;
|
|
|
|
DEBUGASSERT(h);
|
|
|
|
/* Find the end of the header name */
|
|
while (*h && (*h != ':'))
|
|
++h;
|
|
|
|
if (*h)
|
|
/* Skip over colon */
|
|
++h;
|
|
|
|
/* Find the first non-space letter */
|
|
start = h;
|
|
while(*start && ISSPACE(*start))
|
|
start++;
|
|
|
|
/* data is in the host encoding so
|
|
use '\r' and '\n' instead of 0x0d and 0x0a */
|
|
end = strchr(start, '\r');
|
|
if(!end)
|
|
end = strchr(start, '\n');
|
|
if(!end)
|
|
end = strchr(start, '\0');
|
|
if(!end)
|
|
return NULL;
|
|
|
|
/* skip all trailing space letters */
|
|
while((end > start) && ISSPACE(*end))
|
|
end--;
|
|
|
|
/* get length of the type */
|
|
len = end-start+1;
|
|
|
|
value = malloc(len + 1);
|
|
if(!value)
|
|
return NULL;
|
|
|
|
memcpy(value, start, len);
|
|
value[len] = 0; /* zero terminate */
|
|
|
|
return value;
|
|
}
|
|
|
|
/*
|
|
* http_output_basic() sets up an Authorization: header (or the proxy version)
|
|
* for HTTP Basic authentication.
|
|
*
|
|
* Returns CURLcode.
|
|
*/
|
|
static CURLcode http_output_basic(struct connectdata *conn, bool proxy)
|
|
{
|
|
char *authorization;
|
|
struct SessionHandle *data=conn->data;
|
|
char **userp;
|
|
const char *user;
|
|
const char *pwd;
|
|
|
|
if(proxy) {
|
|
userp = &conn->allocptr.proxyuserpwd;
|
|
user = conn->proxyuser;
|
|
pwd = conn->proxypasswd;
|
|
}
|
|
else {
|
|
userp = &conn->allocptr.userpwd;
|
|
user = conn->user;
|
|
pwd = conn->passwd;
|
|
}
|
|
|
|
snprintf(data->state.buffer, sizeof(data->state.buffer), "%s:%s", user, pwd);
|
|
if(Curl_base64_encode(data, data->state.buffer,
|
|
strlen(data->state.buffer),
|
|
&authorization) > 0) {
|
|
if(*userp)
|
|
free(*userp);
|
|
*userp = aprintf( "%sAuthorization: Basic %s\r\n",
|
|
proxy?"Proxy-":"",
|
|
authorization);
|
|
free(authorization);
|
|
if(!*userp)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
else
|
|
return CURLE_OUT_OF_MEMORY;
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/* pickoneauth() selects the most favourable authentication method from the
|
|
* ones available and the ones we want.
|
|
*
|
|
* return TRUE if one was picked
|
|
*/
|
|
static bool pickoneauth(struct auth *pick)
|
|
{
|
|
bool picked;
|
|
/* only deal with authentication we want */
|
|
long avail = pick->avail & pick->want;
|
|
picked = TRUE;
|
|
|
|
/* The order of these checks is highly relevant, as this will be the order
|
|
of preference in case of the existence of multiple accepted types. */
|
|
if(avail & CURLAUTH_GSSNEGOTIATE)
|
|
pick->picked = CURLAUTH_GSSNEGOTIATE;
|
|
else if(avail & CURLAUTH_DIGEST)
|
|
pick->picked = CURLAUTH_DIGEST;
|
|
else if(avail & CURLAUTH_NTLM)
|
|
pick->picked = CURLAUTH_NTLM;
|
|
else if(avail & CURLAUTH_BASIC)
|
|
pick->picked = CURLAUTH_BASIC;
|
|
else {
|
|
pick->picked = CURLAUTH_PICKNONE; /* we select to use nothing */
|
|
picked = FALSE;
|
|
}
|
|
pick->avail = CURLAUTH_NONE; /* clear it here */
|
|
|
|
return picked;
|
|
}
|
|
|
|
/*
|
|
* Curl_http_perhapsrewind()
|
|
*
|
|
* If we are doing POST or PUT {
|
|
* If we have more data to send {
|
|
* If we are doing NTLM {
|
|
* Keep sending since we must not disconnect
|
|
* }
|
|
* else {
|
|
* If there is more than just a little data left to send, close
|
|
* the current connection by force.
|
|
* }
|
|
* }
|
|
* If we have sent any data {
|
|
* If we don't have track of all the data {
|
|
* call app to tell it to rewind
|
|
* }
|
|
* else {
|
|
* rewind internally so that the operation can restart fine
|
|
* }
|
|
* }
|
|
* }
|
|
*/
|
|
CURLcode Curl_http_perhapsrewind(struct connectdata *conn)
|
|
{
|
|
struct SessionHandle *data = conn->data;
|
|
struct HTTP *http = data->state.proto.http;
|
|
curl_off_t bytessent;
|
|
curl_off_t expectsend = -1; /* default is unknown */
|
|
|
|
if(!http || !(conn->protocol & PROT_HTTP))
|
|
/* If this is still NULL, we have not reach very far and we can
|
|
safely skip this rewinding stuff, or this is attempted to get used
|
|
when HTTP isn't activated */
|
|
return CURLE_OK;
|
|
|
|
switch(data->set.httpreq) {
|
|
case HTTPREQ_GET:
|
|
case HTTPREQ_HEAD:
|
|
return CURLE_OK;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
bytessent = http->writebytecount;
|
|
|
|
if(conn->bits.authneg)
|
|
/* This is a state where we are known to be negotiating and we don't send
|
|
any data then. */
|
|
expectsend = 0;
|
|
else {
|
|
/* figure out how much data we are expected to send */
|
|
switch(data->set.httpreq) {
|
|
case HTTPREQ_POST:
|
|
if(data->set.postfieldsize != -1)
|
|
expectsend = data->set.postfieldsize;
|
|
else if(data->set.postfields)
|
|
expectsend = (curl_off_t)strlen(data->set.postfields);
|
|
break;
|
|
case HTTPREQ_PUT:
|
|
if(data->set.infilesize != -1)
|
|
expectsend = data->set.infilesize;
|
|
break;
|
|
case HTTPREQ_POST_FORM:
|
|
expectsend = http->postsize;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
conn->bits.rewindaftersend = FALSE; /* default */
|
|
|
|
if((expectsend == -1) || (expectsend > bytessent)) {
|
|
/* There is still data left to send */
|
|
if((data->state.authproxy.picked == CURLAUTH_NTLM) ||
|
|
(data->state.authhost.picked == CURLAUTH_NTLM)) {
|
|
if(((expectsend - bytessent) < 2000) ||
|
|
(conn->ntlm.state != NTLMSTATE_NONE)) {
|
|
/* The NTLM-negotiation has started *OR* there is just a little (<2K)
|
|
data left to send, keep on sending. */
|
|
|
|
/* rewind data when completely done sending! */
|
|
if(!conn->bits.authneg)
|
|
conn->bits.rewindaftersend = TRUE;
|
|
|
|
return CURLE_OK;
|
|
}
|
|
if(conn->bits.close)
|
|
/* this is already marked to get closed */
|
|
return CURLE_OK;
|
|
|
|
infof(data, "NTLM send, close instead of sending %" FORMAT_OFF_T
|
|
" bytes\n", (curl_off_t)(expectsend - bytessent));
|
|
}
|
|
|
|
/* This is not NTLM or NTLM with many bytes left to send: close
|
|
*/
|
|
conn->bits.close = TRUE;
|
|
data->req.size = 0; /* don't download any more than 0 bytes */
|
|
|
|
/* There still is data left to send, but this connection is marked for
|
|
closure so we can safely do the rewind right now */
|
|
}
|
|
|
|
if(bytessent)
|
|
/* we rewind now at once since if we already sent something */
|
|
return Curl_readrewind(conn);
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/*
|
|
* Curl_http_auth_act() gets called when all HTTP headers have been received
|
|
* and it checks what authentication methods that are available and decides
|
|
* which one (if any) to use. It will set 'newurl' if an auth method was
|
|
* picked.
|
|
*/
|
|
|
|
CURLcode Curl_http_auth_act(struct connectdata *conn)
|
|
{
|
|
struct SessionHandle *data = conn->data;
|
|
bool pickhost = FALSE;
|
|
bool pickproxy = FALSE;
|
|
CURLcode code = CURLE_OK;
|
|
|
|
if(100 <= data->req.httpcode && 199 >= data->req.httpcode)
|
|
/* this is a transient response code, ignore */
|
|
return CURLE_OK;
|
|
|
|
if(data->state.authproblem)
|
|
return data->set.http_fail_on_error?CURLE_HTTP_RETURNED_ERROR:CURLE_OK;
|
|
|
|
if(conn->bits.user_passwd &&
|
|
((data->req.httpcode == 401) ||
|
|
(conn->bits.authneg && data->req.httpcode < 300))) {
|
|
pickhost = pickoneauth(&data->state.authhost);
|
|
if(!pickhost)
|
|
data->state.authproblem = TRUE;
|
|
}
|
|
if(conn->bits.proxy_user_passwd &&
|
|
((data->req.httpcode == 407) ||
|
|
(conn->bits.authneg && data->req.httpcode < 300))) {
|
|
pickproxy = pickoneauth(&data->state.authproxy);
|
|
if(!pickproxy)
|
|
data->state.authproblem = TRUE;
|
|
}
|
|
|
|
if(pickhost || pickproxy) {
|
|
/* In case this is GSS auth, the newurl field is already allocated so
|
|
we must make sure to free it before allocating a new one. As figured
|
|
out in bug #2284386 */
|
|
Curl_safefree(data->req.newurl);
|
|
data->req.newurl = strdup(data->change.url); /* clone URL */
|
|
if(!data->req.newurl)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
|
|
if((data->set.httpreq != HTTPREQ_GET) &&
|
|
(data->set.httpreq != HTTPREQ_HEAD) &&
|
|
!conn->bits.rewindaftersend) {
|
|
code = Curl_http_perhapsrewind(conn);
|
|
if(code)
|
|
return code;
|
|
}
|
|
}
|
|
|
|
else if((data->req.httpcode < 300) &&
|
|
(!data->state.authhost.done) &&
|
|
conn->bits.authneg) {
|
|
/* no (known) authentication available,
|
|
authentication is not "done" yet and
|
|
no authentication seems to be required and
|
|
we didn't try HEAD or GET */
|
|
if((data->set.httpreq != HTTPREQ_GET) &&
|
|
(data->set.httpreq != HTTPREQ_HEAD)) {
|
|
data->req.newurl = strdup(data->change.url); /* clone URL */
|
|
if(!data->req.newurl)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
data->state.authhost.done = TRUE;
|
|
}
|
|
}
|
|
if(Curl_http_should_fail(conn)) {
|
|
failf (data, "The requested URL returned error: %d",
|
|
data->req.httpcode);
|
|
code = CURLE_HTTP_RETURNED_ERROR;
|
|
}
|
|
|
|
return code;
|
|
}
|
|
|
|
|
|
/*
|
|
* Output the correct authentication header depending on the auth type
|
|
* and whether or not it is to a proxy.
|
|
*/
|
|
static CURLcode
|
|
output_auth_headers(struct connectdata *conn,
|
|
struct auth *authstatus,
|
|
const char *request,
|
|
const char *path,
|
|
bool proxy)
|
|
{
|
|
struct SessionHandle *data = conn->data;
|
|
const char *auth=NULL;
|
|
CURLcode result = CURLE_OK;
|
|
#ifdef HAVE_GSSAPI
|
|
struct negotiatedata *negdata = proxy?
|
|
&data->state.proxyneg:&data->state.negotiate;
|
|
#endif
|
|
|
|
#ifdef CURL_DISABLE_CRYPTO_AUTH
|
|
(void)request;
|
|
(void)path;
|
|
#endif
|
|
|
|
#ifdef HAVE_GSSAPI
|
|
if((authstatus->picked == CURLAUTH_GSSNEGOTIATE) &&
|
|
negdata->context && !GSS_ERROR(negdata->status)) {
|
|
auth="GSS-Negotiate";
|
|
result = Curl_output_negotiate(conn, proxy);
|
|
if(result)
|
|
return result;
|
|
authstatus->done = TRUE;
|
|
negdata->state = GSS_AUTHSENT;
|
|
}
|
|
else
|
|
#endif
|
|
#ifdef USE_NTLM
|
|
if(authstatus->picked == CURLAUTH_NTLM) {
|
|
auth="NTLM";
|
|
result = Curl_output_ntlm(conn, proxy);
|
|
if(result)
|
|
return result;
|
|
}
|
|
else
|
|
#endif
|
|
#ifndef CURL_DISABLE_CRYPTO_AUTH
|
|
if(authstatus->picked == CURLAUTH_DIGEST) {
|
|
auth="Digest";
|
|
result = Curl_output_digest(conn,
|
|
proxy,
|
|
(const unsigned char *)request,
|
|
(const unsigned char *)path);
|
|
if(result)
|
|
return result;
|
|
}
|
|
else
|
|
#endif
|
|
if(authstatus->picked == CURLAUTH_BASIC) {
|
|
/* Basic */
|
|
if((proxy && conn->bits.proxy_user_passwd &&
|
|
!Curl_checkheaders(data, "Proxy-authorization:")) ||
|
|
(!proxy && conn->bits.user_passwd &&
|
|
!Curl_checkheaders(data, "Authorization:"))) {
|
|
auth="Basic";
|
|
result = http_output_basic(conn, proxy);
|
|
if(result)
|
|
return result;
|
|
}
|
|
/* NOTE: this function should set 'done' TRUE, as the other auth
|
|
functions work that way */
|
|
authstatus->done = TRUE;
|
|
}
|
|
|
|
if(auth) {
|
|
infof(data, "%s auth using %s with user '%s'\n",
|
|
proxy?"Proxy":"Server", auth,
|
|
proxy?(conn->proxyuser?conn->proxyuser:""):
|
|
(conn->user?conn->user:""));
|
|
authstatus->multi = (bool)(!authstatus->done);
|
|
}
|
|
else
|
|
authstatus->multi = FALSE;
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/**
|
|
* Curl_http_output_auth() setups the authentication headers for the
|
|
* host/proxy and the correct authentication
|
|
* method. conn->data->state.authdone is set to TRUE when authentication is
|
|
* done.
|
|
*
|
|
* @param conn all information about the current connection
|
|
* @param request pointer to the request keyword
|
|
* @param path pointer to the requested path
|
|
* @param proxytunnel boolean if this is the request setting up a "proxy
|
|
* tunnel"
|
|
*
|
|
* @returns CURLcode
|
|
*/
|
|
static CURLcode
|
|
http_output_auth(struct connectdata *conn,
|
|
const char *request,
|
|
const char *path,
|
|
bool proxytunnel) /* TRUE if this is the request setting
|
|
up the proxy tunnel */
|
|
{
|
|
CURLcode result = CURLE_OK;
|
|
struct SessionHandle *data = conn->data;
|
|
struct auth *authhost;
|
|
struct auth *authproxy;
|
|
|
|
DEBUGASSERT(data);
|
|
|
|
authhost = &data->state.authhost;
|
|
authproxy = &data->state.authproxy;
|
|
|
|
if((conn->bits.httpproxy && conn->bits.proxy_user_passwd) ||
|
|
conn->bits.user_passwd)
|
|
/* continue please */ ;
|
|
else {
|
|
authhost->done = TRUE;
|
|
authproxy->done = TRUE;
|
|
return CURLE_OK; /* no authentication with no user or password */
|
|
}
|
|
|
|
if(authhost->want && !authhost->picked)
|
|
/* The app has selected one or more methods, but none has been picked
|
|
so far by a server round-trip. Then we set the picked one to the
|
|
want one, and if this is one single bit it'll be used instantly. */
|
|
authhost->picked = authhost->want;
|
|
|
|
if(authproxy->want && !authproxy->picked)
|
|
/* The app has selected one or more methods, but none has been picked so
|
|
far by a proxy round-trip. Then we set the picked one to the want one,
|
|
and if this is one single bit it'll be used instantly. */
|
|
authproxy->picked = authproxy->want;
|
|
|
|
#ifndef CURL_DISABLE_PROXY
|
|
/* Send proxy authentication header if needed */
|
|
if(conn->bits.httpproxy &&
|
|
(conn->bits.tunnel_proxy == proxytunnel)) {
|
|
result = output_auth_headers(conn, authproxy, request, path, TRUE);
|
|
if(result)
|
|
return result;
|
|
}
|
|
else
|
|
#else
|
|
(void)proxytunnel;
|
|
#endif /* CURL_DISABLE_PROXY */
|
|
/* we have no proxy so let's pretend we're done authenticating
|
|
with it */
|
|
authproxy->done = TRUE;
|
|
|
|
/* To prevent the user+password to get sent to other than the original
|
|
host due to a location-follow, we do some weirdo checks here */
|
|
if(!data->state.this_is_a_follow ||
|
|
conn->bits.netrc ||
|
|
!data->state.first_host ||
|
|
data->set.http_disable_hostname_check_before_authentication ||
|
|
Curl_raw_equal(data->state.first_host, conn->host.name)) {
|
|
result = output_auth_headers(conn, authhost, request, path, FALSE);
|
|
}
|
|
else
|
|
authhost->done = TRUE;
|
|
|
|
return result;
|
|
}
|
|
|
|
|
|
/*
|
|
* Curl_http_input_auth() deals with Proxy-Authenticate: and WWW-Authenticate:
|
|
* headers. They are dealt with both in the transfer.c main loop and in the
|
|
* proxy CONNECT loop.
|
|
*/
|
|
|
|
CURLcode Curl_http_input_auth(struct connectdata *conn,
|
|
int httpcode,
|
|
const char *header) /* the first non-space */
|
|
{
|
|
/*
|
|
* This resource requires authentication
|
|
*/
|
|
struct SessionHandle *data = conn->data;
|
|
|
|
long *availp;
|
|
const char *start;
|
|
struct auth *authp;
|
|
|
|
if(httpcode == 407) {
|
|
start = header+strlen("Proxy-authenticate:");
|
|
availp = &data->info.proxyauthavail;
|
|
authp = &data->state.authproxy;
|
|
}
|
|
else {
|
|
start = header+strlen("WWW-Authenticate:");
|
|
availp = &data->info.httpauthavail;
|
|
authp = &data->state.authhost;
|
|
}
|
|
|
|
/* pass all white spaces */
|
|
while(*start && ISSPACE(*start))
|
|
start++;
|
|
|
|
/*
|
|
* Here we check if we want the specific single authentication (using ==) and
|
|
* if we do, we initiate usage of it.
|
|
*
|
|
* If the provided authentication is wanted as one out of several accepted
|
|
* types (using &), we OR this authentication type to the authavail
|
|
* variable.
|
|
*
|
|
* Note:
|
|
*
|
|
* ->picked is first set to the 'want' value (one or more bits) before the
|
|
* request is sent, and then it is again set _after_ all response 401/407
|
|
* headers have been received but then only to a single preferred method
|
|
* (bit).
|
|
*
|
|
*/
|
|
|
|
#ifdef HAVE_GSSAPI
|
|
if(checkprefix("GSS-Negotiate", start) ||
|
|
checkprefix("Negotiate", start)) {
|
|
int neg;
|
|
*availp |= CURLAUTH_GSSNEGOTIATE;
|
|
authp->avail |= CURLAUTH_GSSNEGOTIATE;
|
|
|
|
if(data->state.negotiate.state == GSS_AUTHSENT) {
|
|
/* if we sent GSS authentication in the outgoing request and we get this
|
|
back, we're in trouble */
|
|
infof(data, "Authentication problem. Ignoring this.\n");
|
|
data->state.authproblem = TRUE;
|
|
}
|
|
else {
|
|
neg = Curl_input_negotiate(conn, (bool)(httpcode == 407), start);
|
|
if(neg == 0) {
|
|
DEBUGASSERT(!data->req.newurl);
|
|
data->req.newurl = strdup(data->change.url);
|
|
if(!data->req.newurl)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
data->state.authproblem = FALSE;
|
|
/* we received GSS auth info and we dealt with it fine */
|
|
data->state.negotiate.state = GSS_AUTHRECV;
|
|
}
|
|
else {
|
|
data->state.authproblem = TRUE;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
#endif
|
|
#ifdef USE_NTLM
|
|
/* NTLM support requires the SSL crypto libs */
|
|
if(checkprefix("NTLM", start)) {
|
|
*availp |= CURLAUTH_NTLM;
|
|
authp->avail |= CURLAUTH_NTLM;
|
|
if(authp->picked == CURLAUTH_NTLM) {
|
|
/* NTLM authentication is picked and activated */
|
|
CURLntlm ntlm =
|
|
Curl_input_ntlm(conn, (bool)(httpcode == 407), start);
|
|
|
|
if(CURLNTLM_BAD != ntlm)
|
|
data->state.authproblem = FALSE;
|
|
else {
|
|
infof(data, "Authentication problem. Ignoring this.\n");
|
|
data->state.authproblem = TRUE;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
#endif
|
|
#ifndef CURL_DISABLE_CRYPTO_AUTH
|
|
if(checkprefix("Digest", start)) {
|
|
if((authp->avail & CURLAUTH_DIGEST) != 0) {
|
|
infof(data, "Ignoring duplicate digest auth header.\n");
|
|
}
|
|
else {
|
|
CURLdigest dig;
|
|
*availp |= CURLAUTH_DIGEST;
|
|
authp->avail |= CURLAUTH_DIGEST;
|
|
|
|
/* We call this function on input Digest headers even if Digest
|
|
* authentication isn't activated yet, as we need to store the
|
|
* incoming data from this header in case we are gonna use Digest. */
|
|
dig = Curl_input_digest(conn, (bool)(httpcode == 407), start);
|
|
|
|
if(CURLDIGEST_FINE != dig) {
|
|
infof(data, "Authentication problem. Ignoring this.\n");
|
|
data->state.authproblem = TRUE;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
#endif
|
|
if(checkprefix("Basic", start)) {
|
|
*availp |= CURLAUTH_BASIC;
|
|
authp->avail |= CURLAUTH_BASIC;
|
|
if(authp->picked == CURLAUTH_BASIC) {
|
|
/* We asked for Basic authentication but got a 40X back
|
|
anyway, which basically means our name+password isn't
|
|
valid. */
|
|
authp->avail = CURLAUTH_NONE;
|
|
infof(data, "Authentication problem. Ignoring this.\n");
|
|
data->state.authproblem = TRUE;
|
|
}
|
|
}
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/**
|
|
* Curl_http_should_fail() determines whether an HTTP response has gotten us
|
|
* into an error state or not.
|
|
*
|
|
* @param conn all information about the current connection
|
|
*
|
|
* @retval 0 communications should continue
|
|
*
|
|
* @retval 1 communications should not continue
|
|
*/
|
|
int Curl_http_should_fail(struct connectdata *conn)
|
|
{
|
|
struct SessionHandle *data;
|
|
int httpcode;
|
|
|
|
DEBUGASSERT(conn);
|
|
data = conn->data;
|
|
DEBUGASSERT(data);
|
|
|
|
httpcode = data->req.httpcode;
|
|
|
|
/*
|
|
** If we haven't been asked to fail on error,
|
|
** don't fail.
|
|
*/
|
|
if(!data->set.http_fail_on_error)
|
|
return 0;
|
|
|
|
/*
|
|
** Any code < 400 is never terminal.
|
|
*/
|
|
if(httpcode < 400)
|
|
return 0;
|
|
|
|
if(data->state.resume_from &&
|
|
(data->set.httpreq==HTTPREQ_GET) &&
|
|
(httpcode == 416)) {
|
|
/* "Requested Range Not Satisfiable", just proceed and
|
|
pretend this is no error */
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
** Any code >= 400 that's not 401 or 407 is always
|
|
** a terminal error
|
|
*/
|
|
if((httpcode != 401) &&
|
|
(httpcode != 407))
|
|
return 1;
|
|
|
|
/*
|
|
** All we have left to deal with is 401 and 407
|
|
*/
|
|
DEBUGASSERT((httpcode == 401) || (httpcode == 407));
|
|
|
|
/*
|
|
** Examine the current authentication state to see if this
|
|
** is an error. The idea is for this function to get
|
|
** called after processing all the headers in a response
|
|
** message. So, if we've been to asked to authenticate a
|
|
** particular stage, and we've done it, we're OK. But, if
|
|
** we're already completely authenticated, it's not OK to
|
|
** get another 401 or 407.
|
|
**
|
|
** It is possible for authentication to go stale such that
|
|
** the client needs to reauthenticate. Once that info is
|
|
** available, use it here.
|
|
*/
|
|
#if 0 /* set to 1 when debugging this functionality */
|
|
infof(data,"%s: authstage = %d\n",__FUNCTION__,data->state.authstage);
|
|
infof(data,"%s: authwant = 0x%08x\n",__FUNCTION__,data->state.authwant);
|
|
infof(data,"%s: authavail = 0x%08x\n",__FUNCTION__,data->state.authavail);
|
|
infof(data,"%s: httpcode = %d\n",__FUNCTION__,k->httpcode);
|
|
infof(data,"%s: authdone = %d\n",__FUNCTION__,data->state.authdone);
|
|
infof(data,"%s: newurl = %s\n",__FUNCTION__,data->req.newurl ?
|
|
data->req.newurl : "(null)");
|
|
infof(data,"%s: authproblem = %d\n",__FUNCTION__,data->state.authproblem);
|
|
#endif
|
|
|
|
/*
|
|
** Either we're not authenticating, or we're supposed to
|
|
** be authenticating something else. This is an error.
|
|
*/
|
|
if((httpcode == 401) && !conn->bits.user_passwd)
|
|
return TRUE;
|
|
if((httpcode == 407) && !conn->bits.proxy_user_passwd)
|
|
return TRUE;
|
|
|
|
return data->state.authproblem;
|
|
}
|
|
|
|
/*
|
|
* readmoredata() is a "fread() emulation" to provide POST and/or request
|
|
* data. It is used when a huge POST is to be made and the entire chunk wasn't
|
|
* sent in the first send(). This function will then be called from the
|
|
* transfer.c loop when more data is to be sent to the peer.
|
|
*
|
|
* Returns the amount of bytes it filled the buffer with.
|
|
*/
|
|
static size_t readmoredata(char *buffer,
|
|
size_t size,
|
|
size_t nitems,
|
|
void *userp)
|
|
{
|
|
struct connectdata *conn = (struct connectdata *)userp;
|
|
struct HTTP *http = conn->data->state.proto.http;
|
|
size_t fullsize = size * nitems;
|
|
|
|
if(0 == http->postsize)
|
|
/* nothing to return */
|
|
return 0;
|
|
|
|
/* make sure that a HTTP request is never sent away chunked! */
|
|
conn->data->req.forbidchunk = (bool)(http->sending == HTTPSEND_REQUEST);
|
|
|
|
if(http->postsize <= (curl_off_t)fullsize) {
|
|
memcpy(buffer, http->postdata, (size_t)http->postsize);
|
|
fullsize = (size_t)http->postsize;
|
|
|
|
if(http->backup.postsize) {
|
|
/* move backup data into focus and continue on that */
|
|
http->postdata = http->backup.postdata;
|
|
http->postsize = http->backup.postsize;
|
|
conn->fread_func = http->backup.fread_func;
|
|
conn->fread_in = http->backup.fread_in;
|
|
|
|
http->sending++; /* move one step up */
|
|
|
|
http->backup.postsize=0;
|
|
}
|
|
else
|
|
http->postsize = 0;
|
|
|
|
return fullsize;
|
|
}
|
|
|
|
memcpy(buffer, http->postdata, fullsize);
|
|
http->postdata += fullsize;
|
|
http->postsize -= fullsize;
|
|
|
|
return fullsize;
|
|
}
|
|
|
|
/* ------------------------------------------------------------------------- */
|
|
/* add_buffer functions */
|
|
|
|
/*
|
|
* Curl_add_buffer_init() sets up and returns a fine buffer struct
|
|
*/
|
|
Curl_send_buffer *Curl_add_buffer_init(void)
|
|
{
|
|
return calloc(1, sizeof(Curl_send_buffer));
|
|
}
|
|
|
|
/*
|
|
* Curl_add_buffer_send() sends a header buffer and frees all associated
|
|
* memory. Body data may be appended to the header data if desired.
|
|
*
|
|
* Returns CURLcode
|
|
*/
|
|
CURLcode Curl_add_buffer_send(Curl_send_buffer *in,
|
|
struct connectdata *conn,
|
|
|
|
/* add the number of sent bytes to this
|
|
counter */
|
|
long *bytes_written,
|
|
|
|
/* how much of the buffer contains body data */
|
|
size_t included_body_bytes,
|
|
int socketindex)
|
|
|
|
{
|
|
ssize_t amount;
|
|
CURLcode res;
|
|
char *ptr;
|
|
size_t size;
|
|
struct HTTP *http = conn->data->state.proto.http;
|
|
size_t sendsize;
|
|
curl_socket_t sockfd;
|
|
size_t headersize;
|
|
|
|
DEBUGASSERT(socketindex <= SECONDARYSOCKET);
|
|
|
|
sockfd = conn->sock[socketindex];
|
|
|
|
/* The looping below is required since we use non-blocking sockets, but due
|
|
to the circumstances we will just loop and try again and again etc */
|
|
|
|
ptr = in->buffer;
|
|
size = in->size_used;
|
|
|
|
headersize = size - included_body_bytes; /* the initial part that isn't body
|
|
is header */
|
|
|
|
DEBUGASSERT(size > included_body_bytes);
|
|
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
res = Curl_convert_to_network(conn->data, ptr, headersize);
|
|
/* Curl_convert_to_network calls failf if unsuccessful */
|
|
if(res != CURLE_OK) {
|
|
/* conversion failed, free memory and return to the caller */
|
|
if(in->buffer)
|
|
free(in->buffer);
|
|
free(in);
|
|
return res;
|
|
}
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
|
|
if(conn->protocol & PROT_HTTPS) {
|
|
/* We never send more than CURL_MAX_WRITE_SIZE bytes in one single chunk
|
|
when we speak HTTPS, as if only a fraction of it is sent now, this data
|
|
needs to fit into the normal read-callback buffer later on and that
|
|
buffer is using this size.
|
|
*/
|
|
|
|
sendsize= (size > CURL_MAX_WRITE_SIZE)?CURL_MAX_WRITE_SIZE:size;
|
|
|
|
/* OpenSSL is very picky and we must send the SAME buffer pointer to the
|
|
library when we attempt to re-send this buffer. Sending the same data
|
|
is not enough, we must use the exact same address. For this reason, we
|
|
must copy the data to the uploadbuffer first, since that is the buffer
|
|
we will be using if this send is retried later.
|
|
*/
|
|
memcpy(conn->data->state.uploadbuffer, ptr, sendsize);
|
|
ptr = conn->data->state.uploadbuffer;
|
|
}
|
|
else
|
|
sendsize = size;
|
|
|
|
res = Curl_write(conn, sockfd, ptr, sendsize, &amount);
|
|
|
|
if(CURLE_OK == res) {
|
|
/*
|
|
* Note that we may not send the entire chunk at once, and we have a set
|
|
* number of data bytes at the end of the big buffer (out of which we may
|
|
* only send away a part).
|
|
*/
|
|
/* how much of the header that was sent */
|
|
size_t headlen = (size_t)amount>headersize?headersize:(size_t)amount;
|
|
size_t bodylen = amount - headlen;
|
|
|
|
if(conn->data->set.verbose) {
|
|
/* this data _may_ contain binary stuff */
|
|
Curl_debug(conn->data, CURLINFO_HEADER_OUT, ptr, headlen, conn);
|
|
if((size_t)amount > headlen) {
|
|
/* there was body data sent beyond the initial header part, pass that
|
|
on to the debug callback too */
|
|
Curl_debug(conn->data, CURLINFO_DATA_OUT,
|
|
ptr+headlen, bodylen, conn);
|
|
}
|
|
}
|
|
if(bodylen)
|
|
/* since we sent a piece of the body here, up the byte counter for it
|
|
accordingly */
|
|
http->writebytecount += bodylen;
|
|
|
|
/* 'amount' can never be a very large value here so typecasting it so a
|
|
signed 31 bit value should not cause problems even if ssize_t is
|
|
64bit */
|
|
*bytes_written += (long)amount;
|
|
|
|
if(http) {
|
|
if((size_t)amount != size) {
|
|
/* The whole request could not be sent in one system call. We must
|
|
queue it up and send it later when we get the chance. We must not
|
|
loop here and wait until it might work again. */
|
|
|
|
size -= amount;
|
|
|
|
ptr = in->buffer + amount;
|
|
|
|
/* backup the currently set pointers */
|
|
http->backup.fread_func = conn->fread_func;
|
|
http->backup.fread_in = conn->fread_in;
|
|
http->backup.postdata = http->postdata;
|
|
http->backup.postsize = http->postsize;
|
|
|
|
/* set the new pointers for the request-sending */
|
|
conn->fread_func = (curl_read_callback)readmoredata;
|
|
conn->fread_in = (void *)conn;
|
|
http->postdata = ptr;
|
|
http->postsize = (curl_off_t)size;
|
|
|
|
http->send_buffer = in;
|
|
http->sending = HTTPSEND_REQUEST;
|
|
|
|
return CURLE_OK;
|
|
}
|
|
http->sending = HTTPSEND_BODY;
|
|
/* the full buffer was sent, clean up and return */
|
|
}
|
|
else {
|
|
if((size_t)amount != size)
|
|
/* We have no continue-send mechanism now, fail. This can only happen
|
|
when this function is used from the CONNECT sending function. We
|
|
currently (stupidly) assume that the whole request is always sent
|
|
away in the first single chunk.
|
|
|
|
This needs FIXing.
|
|
*/
|
|
return CURLE_SEND_ERROR;
|
|
else
|
|
conn->writechannel_inuse = FALSE;
|
|
}
|
|
}
|
|
if(in->buffer)
|
|
free(in->buffer);
|
|
free(in);
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
/*
|
|
* add_bufferf() add the formatted input to the buffer.
|
|
*/
|
|
CURLcode Curl_add_bufferf(Curl_send_buffer *in, const char *fmt, ...)
|
|
{
|
|
char *s;
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
s = vaprintf(fmt, ap); /* this allocs a new string to append */
|
|
va_end(ap);
|
|
|
|
if(s) {
|
|
CURLcode result = Curl_add_buffer(in, s, strlen(s));
|
|
free(s);
|
|
return result;
|
|
}
|
|
/* If we failed, we cleanup the whole buffer and return error */
|
|
if(in->buffer)
|
|
free(in->buffer);
|
|
free(in);
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
/*
|
|
* add_buffer() appends a memory chunk to the existing buffer
|
|
*/
|
|
CURLcode Curl_add_buffer(Curl_send_buffer *in, const void *inptr, size_t size)
|
|
{
|
|
char *new_rb;
|
|
size_t new_size;
|
|
|
|
if(~size < in->size_used) {
|
|
/* If resulting used size of send buffer would wrap size_t, cleanup
|
|
the whole buffer and return error. Otherwise the required buffer
|
|
size will fit into a single allocatable memory chunk */
|
|
Curl_safefree(in->buffer);
|
|
free(in);
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if(!in->buffer ||
|
|
((in->size_used + size) > (in->size_max - 1))) {
|
|
|
|
/* If current buffer size isn't enough to hold the result, use a
|
|
buffer size that doubles the required size. If this new size
|
|
would wrap size_t, then just use the largest possible one */
|
|
|
|
if((size > (size_t)-1/2) || (in->size_used > (size_t)-1/2) ||
|
|
(~(size*2) < (in->size_used*2)))
|
|
new_size = (size_t)-1;
|
|
else
|
|
new_size = (in->size_used+size)*2;
|
|
|
|
if(in->buffer)
|
|
/* we have a buffer, enlarge the existing one */
|
|
new_rb = realloc(in->buffer, new_size);
|
|
else
|
|
/* create a new buffer */
|
|
new_rb = malloc(new_size);
|
|
|
|
if(!new_rb) {
|
|
/* If we failed, we cleanup the whole buffer and return error */
|
|
Curl_safefree(in->buffer);
|
|
free(in);
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
in->buffer = new_rb;
|
|
in->size_max = new_size;
|
|
}
|
|
memcpy(&in->buffer[in->size_used], inptr, size);
|
|
|
|
in->size_used += size;
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/* end of the add_buffer functions */
|
|
/* ------------------------------------------------------------------------- */
|
|
|
|
|
|
|
|
/*
|
|
* Curl_compareheader()
|
|
*
|
|
* Returns TRUE if 'headerline' contains the 'header' with given 'content'.
|
|
* Pass headers WITH the colon.
|
|
*/
|
|
bool
|
|
Curl_compareheader(const char *headerline, /* line to check */
|
|
const char *header, /* header keyword _with_ colon */
|
|
const char *content) /* content string to find */
|
|
{
|
|
/* RFC2616, section 4.2 says: "Each header field consists of a name followed
|
|
* by a colon (":") and the field value. Field names are case-insensitive.
|
|
* The field value MAY be preceded by any amount of LWS, though a single SP
|
|
* is preferred." */
|
|
|
|
size_t hlen = strlen(header);
|
|
size_t clen;
|
|
size_t len;
|
|
const char *start;
|
|
const char *end;
|
|
|
|
if(!Curl_raw_nequal(headerline, header, hlen))
|
|
return FALSE; /* doesn't start with header */
|
|
|
|
/* pass the header */
|
|
start = &headerline[hlen];
|
|
|
|
/* pass all white spaces */
|
|
while(*start && ISSPACE(*start))
|
|
start++;
|
|
|
|
/* find the end of the header line */
|
|
end = strchr(start, '\r'); /* lines end with CRLF */
|
|
if(!end) {
|
|
/* in case there's a non-standard compliant line here */
|
|
end = strchr(start, '\n');
|
|
|
|
if(!end)
|
|
/* hm, there's no line ending here, use the zero byte! */
|
|
end = strchr(start, '\0');
|
|
}
|
|
|
|
len = end-start; /* length of the content part of the input line */
|
|
clen = strlen(content); /* length of the word to find */
|
|
|
|
/* find the content string in the rest of the line */
|
|
for(;len>=clen;len--, start++) {
|
|
if(Curl_raw_nequal(start, content, clen))
|
|
return TRUE; /* match! */
|
|
}
|
|
|
|
return FALSE; /* no match */
|
|
}
|
|
|
|
#ifndef CURL_DISABLE_PROXY
|
|
/*
|
|
* Curl_proxyCONNECT() requires that we're connected to a HTTP proxy. This
|
|
* function will issue the necessary commands to get a seamless tunnel through
|
|
* this proxy. After that, the socket can be used just as a normal socket.
|
|
*
|
|
* This badly needs to be rewritten. CONNECT should be sent and dealt with
|
|
* like any ordinary HTTP request, and not specially crafted like this. This
|
|
* function only remains here like this for now since the rewrite is a bit too
|
|
* much work to do at the moment.
|
|
*
|
|
* This function is BLOCKING which is nasty for all multi interface using apps.
|
|
*/
|
|
|
|
CURLcode Curl_proxyCONNECT(struct connectdata *conn,
|
|
int sockindex,
|
|
const char *hostname,
|
|
unsigned short remote_port)
|
|
{
|
|
int subversion=0;
|
|
struct SessionHandle *data=conn->data;
|
|
struct SingleRequest *k = &data->req;
|
|
CURLcode result;
|
|
long timeout =
|
|
data->set.timeout?data->set.timeout:PROXY_TIMEOUT; /* in milliseconds */
|
|
curl_socket_t tunnelsocket = conn->sock[sockindex];
|
|
curl_off_t cl=0;
|
|
bool closeConnection = FALSE;
|
|
bool chunked_encoding = FALSE;
|
|
long check;
|
|
|
|
#define SELECT_OK 0
|
|
#define SELECT_ERROR 1
|
|
#define SELECT_TIMEOUT 2
|
|
int error = SELECT_OK;
|
|
|
|
conn->bits.proxy_connect_closed = FALSE;
|
|
|
|
do {
|
|
if(!conn->bits.tunnel_connecting) { /* BEGIN CONNECT PHASE */
|
|
char *host_port;
|
|
Curl_send_buffer *req_buffer;
|
|
|
|
infof(data, "Establish HTTP proxy tunnel to %s:%hu\n",
|
|
hostname, remote_port);
|
|
|
|
if(data->req.newurl) {
|
|
/* This only happens if we've looped here due to authentication
|
|
reasons, and we don't really use the newly cloned URL here
|
|
then. Just free() it. */
|
|
free(data->req.newurl);
|
|
data->req.newurl = NULL;
|
|
}
|
|
|
|
/* initialize a dynamic send-buffer */
|
|
req_buffer = Curl_add_buffer_init();
|
|
|
|
if(!req_buffer)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
|
|
host_port = aprintf("%s:%hu", hostname, remote_port);
|
|
if(!host_port) {
|
|
free(req_buffer);
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
/* Setup the proxy-authorization header, if any */
|
|
result = http_output_auth(conn, "CONNECT", host_port, TRUE);
|
|
|
|
if(CURLE_OK == result) {
|
|
char *host=(char *)"";
|
|
const char *proxyconn="";
|
|
const char *useragent="";
|
|
const char *http = (conn->proxytype == CURLPROXY_HTTP_1_0) ?
|
|
"1.0" : "1.1";
|
|
|
|
if(!Curl_checkheaders(data, "Host:")) {
|
|
host = aprintf("Host: %s\r\n", host_port);
|
|
if(!host) {
|
|
free(req_buffer);
|
|
free(host_port);
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
}
|
|
if(!Curl_checkheaders(data, "Proxy-Connection:"))
|
|
proxyconn = "Proxy-Connection: Keep-Alive\r\n";
|
|
|
|
if(!Curl_checkheaders(data, "User-Agent:") &&
|
|
data->set.str[STRING_USERAGENT])
|
|
useragent = conn->allocptr.uagent;
|
|
|
|
/* Send the connect request to the proxy */
|
|
/* BLOCKING */
|
|
result =
|
|
Curl_add_bufferf(req_buffer,
|
|
"CONNECT %s:%hu HTTP/%s\r\n"
|
|
"%s" /* Host: */
|
|
"%s" /* Proxy-Authorization */
|
|
"%s" /* User-Agent */
|
|
"%s", /* Proxy-Connection */
|
|
hostname, remote_port, http,
|
|
host,
|
|
conn->allocptr.proxyuserpwd?
|
|
conn->allocptr.proxyuserpwd:"",
|
|
useragent,
|
|
proxyconn);
|
|
|
|
if(host && *host)
|
|
free(host);
|
|
|
|
if(CURLE_OK == result)
|
|
result = Curl_add_custom_headers(conn, req_buffer);
|
|
|
|
if(CURLE_OK == result)
|
|
/* CRLF terminate the request */
|
|
result = Curl_add_bufferf(req_buffer, "\r\n");
|
|
|
|
if(CURLE_OK == result) {
|
|
/* Now send off the request */
|
|
result = Curl_add_buffer_send(req_buffer, conn,
|
|
&data->info.request_size, 0, sockindex);
|
|
}
|
|
req_buffer = NULL;
|
|
if(result)
|
|
failf(data, "Failed sending CONNECT to proxy");
|
|
}
|
|
free(host_port);
|
|
Curl_safefree(req_buffer);
|
|
if(result)
|
|
return result;
|
|
|
|
conn->bits.tunnel_connecting = TRUE;
|
|
} /* END CONNECT PHASE */
|
|
|
|
/* now we've issued the CONNECT and we're waiting to hear back -
|
|
we try not to block here in multi-mode because that might be a LONG
|
|
wait if the proxy cannot connect-through to the remote host. */
|
|
|
|
/* if timeout is requested, find out how much remaining time we have */
|
|
check = timeout - /* timeout time */
|
|
Curl_tvdiff(Curl_tvnow(), conn->now); /* spent time */
|
|
if(check <= 0) {
|
|
failf(data, "Proxy CONNECT aborted due to timeout");
|
|
return CURLE_RECV_ERROR;
|
|
}
|
|
|
|
/* if we're in multi-mode and we would block, return instead for a retry */
|
|
if(Curl_if_multi == data->state.used_interface) {
|
|
if(0 == Curl_socket_ready(tunnelsocket, CURL_SOCKET_BAD, 0))
|
|
/* return so we'll be called again polling-style */
|
|
return CURLE_OK;
|
|
else {
|
|
DEBUGF(infof(data,
|
|
"Multi mode finished polling for response from "
|
|
"proxy CONNECT."));
|
|
}
|
|
}
|
|
else {
|
|
DEBUGF(infof(data, "Easy mode waiting response from proxy CONNECT."));
|
|
}
|
|
|
|
/* at this point, either:
|
|
1) we're in easy-mode and so it's okay to block waiting for a CONNECT
|
|
response
|
|
2) we're in multi-mode and we didn't block - it's either an error or we
|
|
now have some data waiting.
|
|
In any case, the tunnel_connecting phase is over. */
|
|
conn->bits.tunnel_connecting = FALSE;
|
|
|
|
{ /* BEGIN NEGOTIATION PHASE */
|
|
size_t nread; /* total size read */
|
|
int perline; /* count bytes per line */
|
|
int keepon=TRUE;
|
|
ssize_t gotbytes;
|
|
char *ptr;
|
|
char *line_start;
|
|
|
|
ptr=data->state.buffer;
|
|
line_start = ptr;
|
|
|
|
nread=0;
|
|
perline=0;
|
|
keepon=TRUE;
|
|
|
|
while((nread<BUFSIZE) && (keepon && !error)) {
|
|
|
|
/* if timeout is requested, find out how much remaining time we have */
|
|
check = timeout - /* timeout time */
|
|
Curl_tvdiff(Curl_tvnow(), conn->now); /* spent time */
|
|
if(check <= 0) {
|
|
failf(data, "Proxy CONNECT aborted due to timeout");
|
|
error = SELECT_TIMEOUT; /* already too little time */
|
|
break;
|
|
}
|
|
|
|
/* loop every second at least, less if the timeout is near */
|
|
switch (Curl_socket_ready(tunnelsocket, CURL_SOCKET_BAD,
|
|
check<1000L?(int)check:1000)) {
|
|
case -1: /* select() error, stop reading */
|
|
error = SELECT_ERROR;
|
|
failf(data, "Proxy CONNECT aborted due to select/poll error");
|
|
break;
|
|
case 0: /* timeout */
|
|
break;
|
|
default:
|
|
DEBUGASSERT(ptr+BUFSIZE-nread <= data->state.buffer+BUFSIZE+1);
|
|
result = Curl_read(conn, tunnelsocket, ptr, BUFSIZE-nread, &gotbytes);
|
|
if(result==CURLE_AGAIN)
|
|
continue; /* go loop yourself */
|
|
else if(result)
|
|
keepon = FALSE;
|
|
else if(gotbytes <= 0) {
|
|
keepon = FALSE;
|
|
if(data->set.proxyauth && data->state.authproxy.avail) {
|
|
/* proxy auth was requested and there was proxy auth available,
|
|
then deem this as "mere" proxy disconnect */
|
|
conn->bits.proxy_connect_closed = TRUE;
|
|
}
|
|
else {
|
|
error = SELECT_ERROR;
|
|
failf(data, "Proxy CONNECT aborted");
|
|
}
|
|
}
|
|
else {
|
|
/*
|
|
* We got a whole chunk of data, which can be anything from one
|
|
* byte to a set of lines and possibly just a piece of the last
|
|
* line.
|
|
*/
|
|
int i;
|
|
|
|
nread += gotbytes;
|
|
|
|
if(keepon > TRUE) {
|
|
/* This means we are currently ignoring a response-body */
|
|
|
|
nread = 0; /* make next read start over in the read buffer */
|
|
ptr=data->state.buffer;
|
|
if(cl) {
|
|
/* A Content-Length based body: simply count down the counter
|
|
and make sure to break out of the loop when we're done! */
|
|
cl -= gotbytes;
|
|
if(cl<=0) {
|
|
keepon = FALSE;
|
|
break;
|
|
}
|
|
}
|
|
else {
|
|
/* chunked-encoded body, so we need to do the chunked dance
|
|
properly to know when the end of the body is reached */
|
|
CHUNKcode r;
|
|
ssize_t tookcareof=0;
|
|
|
|
/* now parse the chunked piece of data so that we can
|
|
properly tell when the stream ends */
|
|
r = Curl_httpchunk_read(conn, ptr, gotbytes, &tookcareof);
|
|
if(r == CHUNKE_STOP) {
|
|
/* we're done reading chunks! */
|
|
infof(data, "chunk reading DONE\n");
|
|
keepon = FALSE;
|
|
}
|
|
else
|
|
infof(data, "Read %zd bytes of chunk, continue\n",
|
|
tookcareof);
|
|
}
|
|
}
|
|
else
|
|
for(i = 0; i < gotbytes; ptr++, i++) {
|
|
perline++; /* amount of bytes in this line so far */
|
|
if(*ptr == 0x0a) {
|
|
char letter;
|
|
int writetype;
|
|
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
/* convert from the network encoding */
|
|
result = Curl_convert_from_network(data, line_start,
|
|
perline);
|
|
/* Curl_convert_from_network calls failf if unsuccessful */
|
|
if(result)
|
|
return result;
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
|
|
/* output debug if that is requested */
|
|
if(data->set.verbose)
|
|
Curl_debug(data, CURLINFO_HEADER_IN,
|
|
line_start, (size_t)perline, conn);
|
|
|
|
/* send the header to the callback */
|
|
writetype = CLIENTWRITE_HEADER;
|
|
if(data->set.include_header)
|
|
writetype |= CLIENTWRITE_BODY;
|
|
|
|
result = Curl_client_write(conn, writetype, line_start,
|
|
perline);
|
|
if(result)
|
|
return result;
|
|
|
|
/* Newlines are CRLF, so the CR is ignored as the line isn't
|
|
really terminated until the LF comes. Treat a following CR
|
|
as end-of-headers as well.*/
|
|
|
|
if(('\r' == line_start[0]) ||
|
|
('\n' == line_start[0])) {
|
|
/* end of response-headers from the proxy */
|
|
nread = 0; /* make next read start over in the read
|
|
buffer */
|
|
ptr=data->state.buffer;
|
|
if((407 == k->httpcode) && !data->state.authproblem) {
|
|
/* If we get a 407 response code with content length
|
|
when we have no auth problem, we must ignore the
|
|
whole response-body */
|
|
keepon = 2;
|
|
|
|
if(cl) {
|
|
|
|
infof(data, "Ignore %" FORMAT_OFF_T
|
|
" bytes of response-body\n", cl);
|
|
/* remove the remaining chunk of what we already
|
|
read */
|
|
cl -= (gotbytes - i);
|
|
|
|
if(cl<=0)
|
|
/* if the whole thing was already read, we are done!
|
|
*/
|
|
keepon=FALSE;
|
|
}
|
|
else if(chunked_encoding) {
|
|
CHUNKcode r;
|
|
/* We set ignorebody true here since the chunked
|
|
decoder function will acknowledge that. Pay
|
|
attention so that this is cleared again when this
|
|
function returns! */
|
|
k->ignorebody = TRUE;
|
|
infof(data, "%zd bytes of chunk left\n", gotbytes-i);
|
|
|
|
if(line_start[1] == '\n') {
|
|
/* this can only be a LF if the letter at index 0
|
|
was a CR */
|
|
line_start++;
|
|
i++;
|
|
}
|
|
|
|
/* now parse the chunked piece of data so that we can
|
|
properly tell when the stream ends */
|
|
r = Curl_httpchunk_read(conn, line_start+1,
|
|
gotbytes -i, &gotbytes);
|
|
if(r == CHUNKE_STOP) {
|
|
/* we're done reading chunks! */
|
|
infof(data, "chunk reading DONE\n");
|
|
keepon = FALSE;
|
|
}
|
|
else
|
|
infof(data, "Read %zd bytes of chunk, continue\n",
|
|
gotbytes);
|
|
}
|
|
else {
|
|
/* without content-length or chunked encoding, we
|
|
can't keep the connection alive since the close is
|
|
the end signal so we bail out at once instead */
|
|
keepon=FALSE;
|
|
}
|
|
}
|
|
else
|
|
keepon = FALSE;
|
|
break; /* breaks out of for-loop, not switch() */
|
|
}
|
|
|
|
/* keep a backup of the position we are about to blank */
|
|
letter = line_start[perline];
|
|
line_start[perline]=0; /* zero terminate the buffer */
|
|
if((checkprefix("WWW-Authenticate:", line_start) &&
|
|
(401 == k->httpcode)) ||
|
|
(checkprefix("Proxy-authenticate:", line_start) &&
|
|
(407 == k->httpcode))) {
|
|
result = Curl_http_input_auth(conn, k->httpcode,
|
|
line_start);
|
|
if(result)
|
|
return result;
|
|
}
|
|
else if(checkprefix("Content-Length:", line_start)) {
|
|
cl = curlx_strtoofft(line_start +
|
|
strlen("Content-Length:"), NULL, 10);
|
|
}
|
|
else if(Curl_compareheader(line_start,
|
|
"Connection:", "close"))
|
|
closeConnection = TRUE;
|
|
else if(Curl_compareheader(line_start,
|
|
"Transfer-Encoding:",
|
|
"chunked")) {
|
|
infof(data, "CONNECT responded chunked\n");
|
|
chunked_encoding = TRUE;
|
|
/* init our chunky engine */
|
|
Curl_httpchunk_init(conn);
|
|
}
|
|
else if(Curl_compareheader(line_start,
|
|
"Proxy-Connection:", "close"))
|
|
closeConnection = TRUE;
|
|
else if(2 == sscanf(line_start, "HTTP/1.%d %d",
|
|
&subversion,
|
|
&k->httpcode)) {
|
|
/* store the HTTP code from the proxy */
|
|
data->info.httpproxycode = k->httpcode;
|
|
}
|
|
/* put back the letter we blanked out before */
|
|
line_start[perline]= letter;
|
|
|
|
perline=0; /* line starts over here */
|
|
line_start = ptr+1; /* this skips the zero byte we wrote */
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
} /* switch */
|
|
if(Curl_pgrsUpdate(conn))
|
|
return CURLE_ABORTED_BY_CALLBACK;
|
|
} /* while there's buffer left and loop is requested */
|
|
|
|
if(error)
|
|
return CURLE_RECV_ERROR;
|
|
|
|
if(data->info.httpproxycode != 200) {
|
|
/* Deal with the possibly already received authenticate
|
|
headers. 'newurl' is set to a new URL if we must loop. */
|
|
result = Curl_http_auth_act(conn);
|
|
if(result)
|
|
return result;
|
|
|
|
if(conn->bits.close)
|
|
/* the connection has been marked for closure, most likely in the
|
|
Curl_http_auth_act() function and thus we can kill it at once
|
|
below
|
|
*/
|
|
closeConnection = TRUE;
|
|
}
|
|
|
|
if(closeConnection && data->req.newurl) {
|
|
/* Connection closed by server. Don't use it anymore */
|
|
sclose(conn->sock[sockindex]);
|
|
conn->sock[sockindex] = CURL_SOCKET_BAD;
|
|
break;
|
|
}
|
|
} /* END NEGOTIATION PHASE */
|
|
} while(data->req.newurl);
|
|
|
|
if(200 != data->req.httpcode) {
|
|
failf(data, "Received HTTP code %d from proxy after CONNECT",
|
|
data->req.httpcode);
|
|
|
|
if(closeConnection && data->req.newurl)
|
|
conn->bits.proxy_connect_closed = TRUE;
|
|
|
|
return CURLE_RECV_ERROR;
|
|
}
|
|
|
|
/* If a proxy-authorization header was used for the proxy, then we should
|
|
make sure that it isn't accidentally used for the document request
|
|
after we've connected. So let's free and clear it here. */
|
|
Curl_safefree(conn->allocptr.proxyuserpwd);
|
|
conn->allocptr.proxyuserpwd = NULL;
|
|
|
|
data->state.authproxy.done = TRUE;
|
|
|
|
infof (data, "Proxy replied OK to CONNECT request\n");
|
|
data->req.ignorebody = FALSE; /* put it (back) to non-ignore state */
|
|
return CURLE_OK;
|
|
}
|
|
#endif /* CURL_DISABLE_PROXY */
|
|
|
|
/*
|
|
* Curl_http_connect() performs HTTP stuff to do at connect-time, called from
|
|
* the generic Curl_connect().
|
|
*/
|
|
CURLcode Curl_http_connect(struct connectdata *conn, bool *done)
|
|
{
|
|
struct SessionHandle *data;
|
|
CURLcode result;
|
|
|
|
data=conn->data;
|
|
|
|
/* We default to persistent connections. We set this already in this connect
|
|
function to make the re-use checks properly be able to check this bit. */
|
|
conn->bits.close = FALSE;
|
|
|
|
#ifndef CURL_DISABLE_PROXY
|
|
/* If we are not using a proxy and we want a secure connection, perform SSL
|
|
* initialization & connection now. If using a proxy with https, then we
|
|
* must tell the proxy to CONNECT to the host we want to talk to. Only
|
|
* after the connect has occurred, can we start talking SSL
|
|
*/
|
|
if(conn->bits.tunnel_proxy && conn->bits.httpproxy) {
|
|
|
|
/* either SSL over proxy, or explicitly asked for */
|
|
result = Curl_proxyCONNECT(conn, FIRSTSOCKET,
|
|
conn->host.name,
|
|
conn->remote_port);
|
|
if(CURLE_OK != result)
|
|
return result;
|
|
}
|
|
|
|
if(conn->bits.tunnel_connecting) {
|
|
/* nothing else to do except wait right now - we're not done here. */
|
|
return CURLE_OK;
|
|
}
|
|
#endif /* CURL_DISABLE_PROXY */
|
|
|
|
if(conn->protocol & PROT_HTTPS) {
|
|
/* perform SSL initialization */
|
|
if(data->state.used_interface == Curl_if_multi) {
|
|
result = https_connecting(conn, done);
|
|
if(result)
|
|
return result;
|
|
}
|
|
else {
|
|
/* BLOCKING */
|
|
result = Curl_ssl_connect(conn, FIRSTSOCKET);
|
|
if(result)
|
|
return result;
|
|
*done = TRUE;
|
|
}
|
|
}
|
|
else {
|
|
*done = TRUE;
|
|
}
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/* this returns the socket to wait for in the DO and DOING state for the multi
|
|
interface and then we're always _sending_ a request and thus we wait for
|
|
the single socket to become writable only */
|
|
static int http_getsock_do(struct connectdata *conn,
|
|
curl_socket_t *socks,
|
|
int numsocks)
|
|
{
|
|
/* write mode */
|
|
(void)numsocks; /* unused, we trust it to be at least 1 */
|
|
socks[0] = conn->sock[FIRSTSOCKET];
|
|
return GETSOCK_WRITESOCK(0);
|
|
}
|
|
|
|
#ifdef USE_SSL
|
|
static CURLcode https_connecting(struct connectdata *conn, bool *done)
|
|
{
|
|
CURLcode result;
|
|
DEBUGASSERT((conn) && (conn->protocol & PROT_HTTPS));
|
|
|
|
/* perform SSL initialization for this socket */
|
|
result = Curl_ssl_connect_nonblocking(conn, FIRSTSOCKET, done);
|
|
if(result)
|
|
conn->bits.close = TRUE; /* a failed connection is marked for closure
|
|
to prevent (bad) re-use or similar */
|
|
return result;
|
|
}
|
|
#endif
|
|
|
|
#if defined(USE_SSLEAY) || defined(USE_GNUTLS)
|
|
/* This function is for OpenSSL and GnuTLS only. It should be made to query
|
|
the generic SSL layer instead. */
|
|
static int https_getsock(struct connectdata *conn,
|
|
curl_socket_t *socks,
|
|
int numsocks)
|
|
{
|
|
if(conn->protocol & PROT_HTTPS) {
|
|
struct ssl_connect_data *connssl = &conn->ssl[FIRSTSOCKET];
|
|
|
|
if(!numsocks)
|
|
return GETSOCK_BLANK;
|
|
|
|
if(connssl->connecting_state == ssl_connect_2_writing) {
|
|
/* write mode */
|
|
socks[0] = conn->sock[FIRSTSOCKET];
|
|
return GETSOCK_WRITESOCK(0);
|
|
}
|
|
else if(connssl->connecting_state == ssl_connect_2_reading) {
|
|
/* read mode */
|
|
socks[0] = conn->sock[FIRSTSOCKET];
|
|
return GETSOCK_READSOCK(0);
|
|
}
|
|
}
|
|
return CURLE_OK;
|
|
}
|
|
#else
|
|
#ifdef USE_NSS
|
|
static int https_getsock(struct connectdata *conn,
|
|
curl_socket_t *socks,
|
|
int numsocks)
|
|
{
|
|
(void)conn;
|
|
(void)socks;
|
|
(void)numsocks;
|
|
return GETSOCK_BLANK;
|
|
}
|
|
#else
|
|
#ifdef USE_QSOSSL
|
|
static int https_getsock(struct connectdata *conn,
|
|
curl_socket_t *socks,
|
|
int numsocks)
|
|
{
|
|
(void)conn;
|
|
(void)socks;
|
|
(void)numsocks;
|
|
return GETSOCK_BLANK;
|
|
}
|
|
#else
|
|
#ifdef USE_POLARSSL
|
|
static int https_getsock(struct connectdata *conn,
|
|
curl_socket_t *socks,
|
|
int numsocks)
|
|
{
|
|
(void)conn;
|
|
(void)socks;
|
|
(void)numsocks;
|
|
return GETSOCK_BLANK;
|
|
}
|
|
#endif
|
|
#endif
|
|
#endif
|
|
#endif
|
|
|
|
/*
|
|
* Curl_http_done() gets called from Curl_done() after a single HTTP request
|
|
* has been performed.
|
|
*/
|
|
|
|
CURLcode Curl_http_done(struct connectdata *conn,
|
|
CURLcode status, bool premature)
|
|
{
|
|
struct SessionHandle *data = conn->data;
|
|
struct HTTP *http =data->state.proto.http;
|
|
(void)premature; /* not used */
|
|
|
|
Curl_unencode_cleanup(conn);
|
|
|
|
/* set the proper values (possibly modified on POST) */
|
|
conn->fread_func = data->set.fread_func; /* restore */
|
|
conn->fread_in = data->set.in; /* restore */
|
|
conn->seek_func = data->set.seek_func; /* restore */
|
|
conn->seek_client = data->set.seek_client; /* restore */
|
|
|
|
if(http == NULL)
|
|
return CURLE_OK;
|
|
|
|
if(http->send_buffer) {
|
|
Curl_send_buffer *buff = http->send_buffer;
|
|
|
|
free(buff->buffer);
|
|
free(buff);
|
|
http->send_buffer = NULL; /* clear the pointer */
|
|
}
|
|
|
|
if(HTTPREQ_POST_FORM == data->set.httpreq) {
|
|
data->req.bytecount = http->readbytecount + http->writebytecount;
|
|
|
|
Curl_formclean(&http->sendit); /* Now free that whole lot */
|
|
if(http->form.fp) {
|
|
/* a file being uploaded was left opened, close it! */
|
|
fclose(http->form.fp);
|
|
http->form.fp = NULL;
|
|
}
|
|
}
|
|
else if(HTTPREQ_PUT == data->set.httpreq)
|
|
data->req.bytecount = http->readbytecount + http->writebytecount;
|
|
|
|
if(status != CURLE_OK)
|
|
return (status);
|
|
|
|
if(!premature && /* this check is pointless when DONE is called before the
|
|
entire operation is complete */
|
|
!conn->bits.retry &&
|
|
((http->readbytecount +
|
|
data->req.headerbytecount -
|
|
data->req.deductheadercount)) <= 0) {
|
|
/* If this connection isn't simply closed to be retried, AND nothing was
|
|
read from the HTTP server (that counts), this can't be right so we
|
|
return an error here */
|
|
failf(data, "Empty reply from server");
|
|
return CURLE_GOT_NOTHING;
|
|
}
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
|
|
/* Determine if we should use HTTP 1.1 for this request. Reasons to avoid it
|
|
are if the user specifically requested HTTP 1.0, if the server we are
|
|
connected to only supports 1.0, or if any server previously contacted to
|
|
handle this request only supports 1.0. */
|
|
static bool use_http_1_1(const struct SessionHandle *data,
|
|
const struct connectdata *conn)
|
|
{
|
|
return (bool)((data->set.httpversion == CURL_HTTP_VERSION_1_1) ||
|
|
((data->set.httpversion != CURL_HTTP_VERSION_1_0) &&
|
|
((conn->httpversion == 11) ||
|
|
((conn->httpversion != 10) &&
|
|
(data->state.httpversion != 10)))));
|
|
}
|
|
|
|
/* check and possibly add an Expect: header */
|
|
static CURLcode expect100(struct SessionHandle *data,
|
|
struct connectdata *conn,
|
|
Curl_send_buffer *req_buffer)
|
|
{
|
|
CURLcode result = CURLE_OK;
|
|
const char *ptr;
|
|
data->state.expect100header = FALSE; /* default to false unless it is set
|
|
to TRUE below */
|
|
if(use_http_1_1(data, conn)) {
|
|
/* if not doing HTTP 1.0 or disabled explicitly, we add a Expect:
|
|
100-continue to the headers which actually speeds up post operations
|
|
(as there is one packet coming back from the web server) */
|
|
ptr = Curl_checkheaders(data, "Expect:");
|
|
if (ptr) {
|
|
data->state.expect100header =
|
|
Curl_compareheader(ptr, "Expect:", "100-continue");
|
|
}
|
|
else {
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"Expect: 100-continue\r\n");
|
|
if(result == CURLE_OK)
|
|
data->state.expect100header = TRUE;
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
|
|
CURLcode Curl_add_custom_headers(struct connectdata *conn,
|
|
Curl_send_buffer *req_buffer)
|
|
{
|
|
char *ptr;
|
|
struct curl_slist *headers=conn->data->set.headers;
|
|
|
|
while(headers) {
|
|
ptr = strchr(headers->data, ':');
|
|
if(ptr) {
|
|
/* we require a colon for this to be a true header */
|
|
|
|
ptr++; /* pass the colon */
|
|
while(*ptr && ISSPACE(*ptr))
|
|
ptr++;
|
|
|
|
if(*ptr) {
|
|
/* only send this if the contents was non-blank */
|
|
|
|
if(conn->allocptr.host &&
|
|
/* a Host: header was sent already, don't pass on any custom Host:
|
|
header as that will produce *two* in the same request! */
|
|
checkprefix("Host:", headers->data))
|
|
;
|
|
else if(conn->data->set.httpreq == HTTPREQ_POST_FORM &&
|
|
/* this header (extended by formdata.c) is sent later */
|
|
checkprefix("Content-Type:", headers->data))
|
|
;
|
|
else if(conn->bits.authneg &&
|
|
/* while doing auth neg, don't allow the custom length since
|
|
we will force length zero then */
|
|
checkprefix("Content-Length", headers->data))
|
|
;
|
|
else {
|
|
CURLcode result = Curl_add_bufferf(req_buffer, "%s\r\n",
|
|
headers->data);
|
|
if(result)
|
|
return result;
|
|
}
|
|
}
|
|
}
|
|
headers = headers->next;
|
|
}
|
|
return CURLE_OK;
|
|
}
|
|
|
|
CURLcode Curl_add_timecondition(struct SessionHandle *data,
|
|
Curl_send_buffer *req_buffer)
|
|
{
|
|
struct tm *tm;
|
|
char *buf = data->state.buffer;
|
|
CURLcode result = CURLE_OK;
|
|
|
|
/* The If-Modified-Since header family should have their times set in
|
|
* GMT as RFC2616 defines: "All HTTP date/time stamps MUST be
|
|
* represented in Greenwich Mean Time (GMT), without exception. For the
|
|
* purposes of HTTP, GMT is exactly equal to UTC (Coordinated Universal
|
|
* Time)." (see page 20 of RFC2616).
|
|
*/
|
|
|
|
#ifdef HAVE_GMTIME_R
|
|
/* thread-safe version */
|
|
struct tm keeptime;
|
|
tm = (struct tm *)gmtime_r(&data->set.timevalue, &keeptime);
|
|
#else
|
|
tm = gmtime(&data->set.timevalue);
|
|
#endif
|
|
|
|
/* format: "Tue, 15 Nov 1994 12:45:26 GMT" */
|
|
snprintf(buf, BUFSIZE-1,
|
|
"%s, %02d %s %4d %02d:%02d:%02d GMT",
|
|
Curl_wkday[tm->tm_wday?tm->tm_wday-1:6],
|
|
tm->tm_mday,
|
|
Curl_month[tm->tm_mon],
|
|
tm->tm_year + 1900,
|
|
tm->tm_hour,
|
|
tm->tm_min,
|
|
tm->tm_sec);
|
|
|
|
switch(data->set.timecondition) {
|
|
case CURL_TIMECOND_IFMODSINCE:
|
|
default:
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"If-Modified-Since: %s\r\n", buf);
|
|
break;
|
|
case CURL_TIMECOND_IFUNMODSINCE:
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"If-Unmodified-Since: %s\r\n", buf);
|
|
break;
|
|
case CURL_TIMECOND_LASTMOD:
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"Last-Modified: %s\r\n", buf);
|
|
break;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
/*
|
|
* Curl_http() gets called from the generic Curl_do() function when a HTTP
|
|
* request is to be performed. This creates and sends a properly constructed
|
|
* HTTP request.
|
|
*/
|
|
CURLcode Curl_http(struct connectdata *conn, bool *done)
|
|
{
|
|
struct SessionHandle *data=conn->data;
|
|
CURLcode result=CURLE_OK;
|
|
struct HTTP *http;
|
|
const char *ppath = data->state.path;
|
|
bool paste_ftp_userpwd = FALSE;
|
|
char ftp_typecode[sizeof("/;type=?")] = "";
|
|
const char *host = conn->host.name;
|
|
const char *te = ""; /* transfer-encoding */
|
|
const char *ptr;
|
|
const char *request;
|
|
Curl_HttpReq httpreq = data->set.httpreq;
|
|
char *addcookies = NULL;
|
|
curl_off_t included_body = 0;
|
|
const char *httpstring;
|
|
Curl_send_buffer *req_buffer;
|
|
curl_off_t postsize; /* off_t type to be able to hold a large file size */
|
|
int seekerr = CURL_SEEKFUNC_OK;
|
|
|
|
/* Always consider the DO phase done after this function call, even if there
|
|
may be parts of the request that is not yet sent, since we can deal with
|
|
the rest of the request in the PERFORM phase. */
|
|
*done = TRUE;
|
|
|
|
/* If there already is a protocol-specific struct allocated for this
|
|
sessionhandle, deal with it */
|
|
Curl_reset_reqproto(conn);
|
|
|
|
if(!data->state.proto.http) {
|
|
/* Only allocate this struct if we don't already have it! */
|
|
|
|
http = calloc(1, sizeof(struct HTTP));
|
|
if(!http)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
data->state.proto.http = http;
|
|
}
|
|
else
|
|
http = data->state.proto.http;
|
|
|
|
if(!data->state.this_is_a_follow) {
|
|
/* this is not a followed location, get the original host name */
|
|
if(data->state.first_host)
|
|
/* Free to avoid leaking memory on multiple requests*/
|
|
free(data->state.first_host);
|
|
|
|
data->state.first_host = strdup(conn->host.name);
|
|
if(!data->state.first_host)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
if( (conn->protocol&(PROT_HTTP|PROT_FTP)) &&
|
|
data->set.upload) {
|
|
httpreq = HTTPREQ_PUT;
|
|
}
|
|
|
|
/* Now set the 'request' pointer to the proper request string */
|
|
if(data->set.str[STRING_CUSTOMREQUEST])
|
|
request = data->set.str[STRING_CUSTOMREQUEST];
|
|
else {
|
|
if(data->set.opt_no_body)
|
|
request = "HEAD";
|
|
else {
|
|
DEBUGASSERT((httpreq > HTTPREQ_NONE) && (httpreq < HTTPREQ_LAST));
|
|
switch(httpreq) {
|
|
case HTTPREQ_POST:
|
|
case HTTPREQ_POST_FORM:
|
|
request = "POST";
|
|
break;
|
|
case HTTPREQ_PUT:
|
|
request = "PUT";
|
|
break;
|
|
default: /* this should never happen */
|
|
case HTTPREQ_GET:
|
|
request = "GET";
|
|
break;
|
|
case HTTPREQ_HEAD:
|
|
request = "HEAD";
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* The User-Agent string might have been allocated in url.c already, because
|
|
it might have been used in the proxy connect, but if we have got a header
|
|
with the user-agent string specified, we erase the previously made string
|
|
here. */
|
|
if(Curl_checkheaders(data, "User-Agent:") && conn->allocptr.uagent) {
|
|
free(conn->allocptr.uagent);
|
|
conn->allocptr.uagent=NULL;
|
|
}
|
|
|
|
/* setup the authentication headers */
|
|
result = http_output_auth(conn, request, ppath, FALSE);
|
|
if(result)
|
|
return result;
|
|
|
|
if((data->state.authhost.multi || data->state.authproxy.multi) &&
|
|
(httpreq != HTTPREQ_GET) &&
|
|
(httpreq != HTTPREQ_HEAD)) {
|
|
/* Auth is required and we are not authenticated yet. Make a PUT or POST
|
|
with content-length zero as a "probe". */
|
|
conn->bits.authneg = TRUE;
|
|
}
|
|
else
|
|
conn->bits.authneg = FALSE;
|
|
|
|
Curl_safefree(conn->allocptr.ref);
|
|
if(data->change.referer && !Curl_checkheaders(data, "Referer:"))
|
|
conn->allocptr.ref = aprintf("Referer: %s\r\n", data->change.referer);
|
|
else
|
|
conn->allocptr.ref = NULL;
|
|
|
|
if(data->set.str[STRING_COOKIE] && !Curl_checkheaders(data, "Cookie:"))
|
|
addcookies = data->set.str[STRING_COOKIE];
|
|
|
|
if(!Curl_checkheaders(data, "Accept-Encoding:") &&
|
|
data->set.str[STRING_ENCODING]) {
|
|
Curl_safefree(conn->allocptr.accept_encoding);
|
|
conn->allocptr.accept_encoding =
|
|
aprintf("Accept-Encoding: %s\r\n", data->set.str[STRING_ENCODING]);
|
|
if(!conn->allocptr.accept_encoding)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
ptr = Curl_checkheaders(data, "Transfer-Encoding:");
|
|
if(ptr) {
|
|
/* Some kind of TE is requested, check if 'chunked' is chosen */
|
|
data->req.upload_chunky =
|
|
Curl_compareheader(ptr, "Transfer-Encoding:", "chunked");
|
|
}
|
|
else {
|
|
if((conn->protocol&PROT_HTTP) &&
|
|
data->set.upload &&
|
|
(data->set.infilesize == -1)) {
|
|
if(conn->bits.authneg)
|
|
/* don't enable chunked during auth neg */
|
|
;
|
|
else if(use_http_1_1(data, conn)) {
|
|
/* HTTP, upload, unknown file size and not HTTP 1.0 */
|
|
data->req.upload_chunky = TRUE;
|
|
}
|
|
else {
|
|
failf(data, "Chunky upload is not supported by HTTP 1.0");
|
|
return CURLE_UPLOAD_FAILED;
|
|
}
|
|
}
|
|
else {
|
|
/* else, no chunky upload */
|
|
data->req.upload_chunky = FALSE;
|
|
}
|
|
|
|
if(data->req.upload_chunky)
|
|
te = "Transfer-Encoding: chunked\r\n";
|
|
}
|
|
|
|
Curl_safefree(conn->allocptr.host);
|
|
|
|
ptr = Curl_checkheaders(data, "Host:");
|
|
if(ptr && (!data->state.this_is_a_follow ||
|
|
Curl_raw_equal(data->state.first_host, conn->host.name))) {
|
|
#if !defined(CURL_DISABLE_COOKIES)
|
|
/* If we have a given custom Host: header, we extract the host name in
|
|
order to possibly use it for cookie reasons later on. We only allow the
|
|
custom Host: header if this is NOT a redirect, as setting Host: in the
|
|
redirected request is being out on thin ice. Except if the host name
|
|
is the same as the first one! */
|
|
char *cookiehost = Curl_copy_header_value(ptr);
|
|
if (!cookiehost)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
if (!*cookiehost)
|
|
/* ignore empty data */
|
|
free(cookiehost);
|
|
else {
|
|
char *colon = strchr(cookiehost, ':');
|
|
if (colon)
|
|
*colon = 0; /* The host must not include an embedded port number */
|
|
Curl_safefree(conn->allocptr.cookiehost);
|
|
conn->allocptr.cookiehost = cookiehost;
|
|
}
|
|
#endif
|
|
|
|
conn->allocptr.host = NULL;
|
|
}
|
|
else {
|
|
/* When building Host: headers, we must put the host name within
|
|
[brackets] if the host name is a plain IPv6-address. RFC2732-style. */
|
|
|
|
if(((conn->protocol&PROT_HTTPS) && (conn->remote_port == PORT_HTTPS)) ||
|
|
(!(conn->protocol&PROT_HTTPS) && (conn->remote_port == PORT_HTTP)) )
|
|
/* if(HTTPS on port 443) OR (non-HTTPS on port 80) then don't include
|
|
the port number in the host string */
|
|
conn->allocptr.host = aprintf("Host: %s%s%s\r\n",
|
|
conn->bits.ipv6_ip?"[":"",
|
|
host,
|
|
conn->bits.ipv6_ip?"]":"");
|
|
else
|
|
conn->allocptr.host = aprintf("Host: %s%s%s:%hu\r\n",
|
|
conn->bits.ipv6_ip?"[":"",
|
|
host,
|
|
conn->bits.ipv6_ip?"]":"",
|
|
conn->remote_port);
|
|
|
|
if(!conn->allocptr.host)
|
|
/* without Host: we can't make a nice request */
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
#ifndef CURL_DISABLE_PROXY
|
|
if(conn->bits.httpproxy && !conn->bits.tunnel_proxy) {
|
|
/* Using a proxy but does not tunnel through it */
|
|
|
|
/* The path sent to the proxy is in fact the entire URL. But if the remote
|
|
host is a IDN-name, we must make sure that the request we produce only
|
|
uses the encoded host name! */
|
|
if(conn->host.dispname != conn->host.name) {
|
|
char *url = data->change.url;
|
|
ptr = strstr(url, conn->host.dispname);
|
|
if(ptr) {
|
|
/* This is where the display name starts in the URL, now replace this
|
|
part with the encoded name. TODO: This method of replacing the host
|
|
name is rather crude as I believe there's a slight risk that the
|
|
user has entered a user name or password that contain the host name
|
|
string. */
|
|
size_t currlen = strlen(conn->host.dispname);
|
|
size_t newlen = strlen(conn->host.name);
|
|
size_t urllen = strlen(url);
|
|
|
|
char *newurl;
|
|
|
|
newurl = malloc(urllen + newlen - currlen + 1);
|
|
if(newurl) {
|
|
/* copy the part before the host name */
|
|
memcpy(newurl, url, ptr - url);
|
|
/* append the new host name instead of the old */
|
|
memcpy(newurl + (ptr - url), conn->host.name, newlen);
|
|
/* append the piece after the host name */
|
|
memcpy(newurl + newlen + (ptr - url),
|
|
ptr + currlen, /* copy the trailing zero byte too */
|
|
urllen - (ptr-url) - currlen + 1);
|
|
if(data->change.url_alloc)
|
|
free(data->change.url);
|
|
data->change.url = newurl;
|
|
data->change.url_alloc = TRUE;
|
|
}
|
|
else
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
}
|
|
ppath = data->change.url;
|
|
if(checkprefix("ftp://", ppath)) {
|
|
if (data->set.proxy_transfer_mode) {
|
|
/* when doing ftp, append ;type=<a|i> if not present */
|
|
char *type = strstr(ppath, ";type=");
|
|
if(type && type[6] && type[7] == 0) {
|
|
switch (Curl_raw_toupper(type[6])) {
|
|
case 'A':
|
|
case 'D':
|
|
case 'I':
|
|
break;
|
|
default:
|
|
type = NULL;
|
|
}
|
|
}
|
|
if(!type) {
|
|
char *p = ftp_typecode;
|
|
/* avoid sending invalid URLs like ftp://example.com;type=i if the
|
|
* user specified ftp://example.com without the slash */
|
|
if (!*data->state.path && ppath[strlen(ppath) - 1] != '/') {
|
|
*p++ = '/';
|
|
}
|
|
snprintf(p, sizeof(ftp_typecode) - 1, ";type=%c",
|
|
data->set.prefer_ascii ? 'a' : 'i');
|
|
}
|
|
}
|
|
if (conn->bits.user_passwd && !conn->bits.userpwd_in_url)
|
|
paste_ftp_userpwd = TRUE;
|
|
}
|
|
}
|
|
#endif /* CURL_DISABLE_PROXY */
|
|
|
|
if(HTTPREQ_POST_FORM == httpreq) {
|
|
/* we must build the whole darned post sequence first, so that we have
|
|
a size of the whole shebang before we start to send it */
|
|
result = Curl_getFormData(&http->sendit, data->set.httppost,
|
|
Curl_checkheaders(data, "Content-Type:"),
|
|
&http->postsize);
|
|
if(CURLE_OK != result) {
|
|
/* Curl_getFormData() doesn't use failf() */
|
|
failf(data, "failed creating formpost data");
|
|
return result;
|
|
}
|
|
}
|
|
|
|
|
|
http->p_accept = Curl_checkheaders(data, "Accept:")?NULL:"Accept: */*\r\n";
|
|
|
|
if(( (HTTPREQ_POST == httpreq) ||
|
|
(HTTPREQ_POST_FORM == httpreq) ||
|
|
(HTTPREQ_PUT == httpreq) ) &&
|
|
data->state.resume_from) {
|
|
/**********************************************************************
|
|
* Resuming upload in HTTP means that we PUT or POST and that we have
|
|
* got a resume_from value set. The resume value has already created
|
|
* a Range: header that will be passed along. We need to "fast forward"
|
|
* the file the given number of bytes and decrease the assume upload
|
|
* file size before we continue this venture in the dark lands of HTTP.
|
|
*********************************************************************/
|
|
|
|
if(data->state.resume_from < 0 ) {
|
|
/*
|
|
* This is meant to get the size of the present remote-file by itself.
|
|
* We don't support this now. Bail out!
|
|
*/
|
|
data->state.resume_from = 0;
|
|
}
|
|
|
|
if(data->state.resume_from && !data->state.this_is_a_follow) {
|
|
/* do we still game? */
|
|
|
|
/* Now, let's read off the proper amount of bytes from the
|
|
input. */
|
|
if(conn->seek_func) {
|
|
seekerr = conn->seek_func(conn->seek_client, data->state.resume_from,
|
|
SEEK_SET);
|
|
}
|
|
|
|
if(seekerr != CURL_SEEKFUNC_OK) {
|
|
if(seekerr != CURL_SEEKFUNC_CANTSEEK) {
|
|
failf(data, "Could not seek stream");
|
|
return CURLE_READ_ERROR;
|
|
}
|
|
/* when seekerr == CURL_SEEKFUNC_CANTSEEK (can't seek to offset) */
|
|
else {
|
|
curl_off_t passed=0;
|
|
|
|
do {
|
|
size_t readthisamountnow = (size_t)(data->state.resume_from -
|
|
passed);
|
|
size_t actuallyread;
|
|
|
|
if(readthisamountnow > BUFSIZE)
|
|
readthisamountnow = BUFSIZE;
|
|
|
|
actuallyread = data->set.fread_func(data->state.buffer, 1,
|
|
(size_t)readthisamountnow,
|
|
data->set.in);
|
|
|
|
passed += actuallyread;
|
|
if(actuallyread != readthisamountnow) {
|
|
failf(data, "Could only read %" FORMAT_OFF_T
|
|
" bytes from the input",
|
|
passed);
|
|
return CURLE_READ_ERROR;
|
|
}
|
|
} while(passed != data->state.resume_from); /* loop until done */
|
|
}
|
|
}
|
|
|
|
/* now, decrease the size of the read */
|
|
if(data->set.infilesize>0) {
|
|
data->set.infilesize -= data->state.resume_from;
|
|
|
|
if(data->set.infilesize <= 0) {
|
|
failf(data, "File already completely uploaded");
|
|
return CURLE_PARTIAL_FILE;
|
|
}
|
|
}
|
|
/* we've passed, proceed as normal */
|
|
}
|
|
}
|
|
if(data->state.use_range) {
|
|
/*
|
|
* A range is selected. We use different headers whether we're downloading
|
|
* or uploading and we always let customized headers override our internal
|
|
* ones if any such are specified.
|
|
*/
|
|
if(((httpreq == HTTPREQ_GET) || (httpreq == HTTPREQ_HEAD)) &&
|
|
!Curl_checkheaders(data, "Range:")) {
|
|
/* if a line like this was already allocated, free the previous one */
|
|
if(conn->allocptr.rangeline)
|
|
free(conn->allocptr.rangeline);
|
|
conn->allocptr.rangeline = aprintf("Range: bytes=%s\r\n",
|
|
data->state.range);
|
|
}
|
|
else if((httpreq != HTTPREQ_GET) &&
|
|
!Curl_checkheaders(data, "Content-Range:")) {
|
|
|
|
/* if a line like this was already allocated, free the previous one */
|
|
if(conn->allocptr.rangeline)
|
|
free(conn->allocptr.rangeline);
|
|
|
|
if(data->set.set_resume_from < 0) {
|
|
/* Upload resume was asked for, but we don't know the size of the
|
|
remote part so we tell the server (and act accordingly) that we
|
|
upload the whole file (again) */
|
|
conn->allocptr.rangeline =
|
|
aprintf("Content-Range: bytes 0-%" FORMAT_OFF_T
|
|
"/%" FORMAT_OFF_T "\r\n",
|
|
data->set.infilesize - 1, data->set.infilesize);
|
|
|
|
}
|
|
else if(data->state.resume_from) {
|
|
/* This is because "resume" was selected */
|
|
curl_off_t total_expected_size=
|
|
data->state.resume_from + data->set.infilesize;
|
|
conn->allocptr.rangeline =
|
|
aprintf("Content-Range: bytes %s%" FORMAT_OFF_T
|
|
"/%" FORMAT_OFF_T "\r\n",
|
|
data->state.range, total_expected_size-1,
|
|
total_expected_size);
|
|
}
|
|
else {
|
|
/* Range was selected and then we just pass the incoming range and
|
|
append total size */
|
|
conn->allocptr.rangeline =
|
|
aprintf("Content-Range: bytes %s/%" FORMAT_OFF_T "\r\n",
|
|
data->state.range, data->set.infilesize);
|
|
}
|
|
if(!conn->allocptr.rangeline)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
}
|
|
|
|
/* Use 1.1 unless the user specifically asked for 1.0 or the server only
|
|
supports 1.0 */
|
|
httpstring= use_http_1_1(data, conn)?"1.1":"1.0";
|
|
|
|
/* initialize a dynamic send-buffer */
|
|
req_buffer = Curl_add_buffer_init();
|
|
|
|
if(!req_buffer)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
|
|
/* add the main request stuff */
|
|
/* GET/HEAD/POST/PUT */
|
|
result = Curl_add_bufferf(req_buffer, "%s ", request);
|
|
if (result)
|
|
return result;
|
|
|
|
/* url */
|
|
if (paste_ftp_userpwd)
|
|
result = Curl_add_bufferf(req_buffer, "ftp://%s:%s@%s",
|
|
conn->user, conn->passwd,
|
|
ppath + sizeof("ftp://") - 1);
|
|
else
|
|
result = Curl_add_buffer(req_buffer, ppath, strlen(ppath));
|
|
if (result)
|
|
return result;
|
|
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"%s" /* ftp typecode (;type=x) */
|
|
" HTTP/%s\r\n" /* HTTP version */
|
|
"%s" /* proxyuserpwd */
|
|
"%s" /* userpwd */
|
|
"%s" /* range */
|
|
"%s" /* user agent */
|
|
"%s" /* host */
|
|
"%s" /* accept */
|
|
"%s" /* accept-encoding */
|
|
"%s" /* referer */
|
|
"%s" /* Proxy-Connection */
|
|
"%s",/* transfer-encoding */
|
|
|
|
ftp_typecode,
|
|
httpstring,
|
|
conn->allocptr.proxyuserpwd?
|
|
conn->allocptr.proxyuserpwd:"",
|
|
conn->allocptr.userpwd?conn->allocptr.userpwd:"",
|
|
(data->state.use_range && conn->allocptr.rangeline)?
|
|
conn->allocptr.rangeline:"",
|
|
(data->set.str[STRING_USERAGENT] &&
|
|
*data->set.str[STRING_USERAGENT] && conn->allocptr.uagent)?
|
|
conn->allocptr.uagent:"",
|
|
(conn->allocptr.host?conn->allocptr.host:""), /* Host: host */
|
|
http->p_accept?http->p_accept:"",
|
|
(data->set.str[STRING_ENCODING] &&
|
|
*data->set.str[STRING_ENCODING] &&
|
|
conn->allocptr.accept_encoding)?
|
|
conn->allocptr.accept_encoding:"",
|
|
(data->change.referer && conn->allocptr.ref)?
|
|
conn->allocptr.ref:"" /* Referer: <data> */,
|
|
(conn->bits.httpproxy &&
|
|
!conn->bits.tunnel_proxy &&
|
|
!Curl_checkheaders(data, "Proxy-Connection:"))?
|
|
"Proxy-Connection: Keep-Alive\r\n":"",
|
|
te
|
|
);
|
|
|
|
/*
|
|
* Free userpwd now --- cannot reuse this for Negotiate and possibly NTLM
|
|
* with basic and digest, it will be freed anyway by the next request
|
|
*/
|
|
|
|
Curl_safefree (conn->allocptr.userpwd);
|
|
conn->allocptr.userpwd = NULL;
|
|
|
|
if(result)
|
|
return result;
|
|
|
|
#if !defined(CURL_DISABLE_COOKIES)
|
|
if(data->cookies || addcookies) {
|
|
struct Cookie *co=NULL; /* no cookies from start */
|
|
int count=0;
|
|
|
|
if(data->cookies) {
|
|
Curl_share_lock(data, CURL_LOCK_DATA_COOKIE, CURL_LOCK_ACCESS_SINGLE);
|
|
co = Curl_cookie_getlist(data->cookies,
|
|
conn->allocptr.cookiehost?
|
|
conn->allocptr.cookiehost:host,
|
|
data->state.path,
|
|
(bool)(conn->protocol&PROT_HTTPS?TRUE:FALSE));
|
|
Curl_share_unlock(data, CURL_LOCK_DATA_COOKIE);
|
|
}
|
|
if(co) {
|
|
struct Cookie *store=co;
|
|
/* now loop through all cookies that matched */
|
|
while(co) {
|
|
if(co->value) {
|
|
if(0 == count) {
|
|
result = Curl_add_bufferf(req_buffer, "Cookie: ");
|
|
if(result)
|
|
break;
|
|
}
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"%s%s=%s", count?"; ":"",
|
|
co->name, co->value);
|
|
if(result)
|
|
break;
|
|
count++;
|
|
}
|
|
co = co->next; /* next cookie please */
|
|
}
|
|
Curl_cookie_freelist(store, FALSE); /* free the cookie list */
|
|
}
|
|
if(addcookies && (CURLE_OK == result)) {
|
|
if(!count)
|
|
result = Curl_add_bufferf(req_buffer, "Cookie: ");
|
|
if(CURLE_OK == result) {
|
|
result = Curl_add_bufferf(req_buffer, "%s%s",
|
|
count?"; ":"",
|
|
addcookies);
|
|
count++;
|
|
}
|
|
}
|
|
if(count && (CURLE_OK == result))
|
|
result = Curl_add_buffer(req_buffer, "\r\n", 2);
|
|
|
|
if(result)
|
|
return result;
|
|
}
|
|
#endif
|
|
|
|
if(data->set.timecondition) {
|
|
result = Curl_add_timecondition(data, req_buffer);
|
|
if(result)
|
|
return result;
|
|
}
|
|
|
|
result = Curl_add_custom_headers(conn, req_buffer);
|
|
if(result)
|
|
return result;
|
|
|
|
http->postdata = NULL; /* nothing to post at this point */
|
|
Curl_pgrsSetUploadSize(data, 0); /* upload size is 0 atm */
|
|
|
|
/* If 'authdone' is FALSE, we must not set the write socket index to the
|
|
Curl_transfer() call below, as we're not ready to actually upload any
|
|
data yet. */
|
|
|
|
switch(httpreq) {
|
|
|
|
case HTTPREQ_POST_FORM:
|
|
if(!http->sendit || conn->bits.authneg) {
|
|
/* nothing to post! */
|
|
result = Curl_add_bufferf(req_buffer, "Content-Length: 0\r\n\r\n");
|
|
if(result)
|
|
return result;
|
|
|
|
result = Curl_add_buffer_send(req_buffer, conn,
|
|
&data->info.request_size, 0, FIRSTSOCKET);
|
|
if(result)
|
|
failf(data, "Failed sending POST request");
|
|
else
|
|
/* setup variables for the upcoming transfer */
|
|
Curl_setup_transfer(conn, FIRSTSOCKET, -1, TRUE, &http->readbytecount,
|
|
-1, NULL);
|
|
break;
|
|
}
|
|
|
|
if(Curl_FormInit(&http->form, http->sendit)) {
|
|
failf(data, "Internal HTTP POST error!");
|
|
return CURLE_HTTP_POST_ERROR;
|
|
}
|
|
|
|
/* Get the currently set callback function pointer and store that in the
|
|
form struct since we might want the actual user-provided callback later
|
|
on. The conn->fread_func pointer itself will be changed for the
|
|
multipart case to the function that returns a multipart formatted
|
|
stream. */
|
|
http->form.fread_func = conn->fread_func;
|
|
|
|
/* Set the read function to read from the generated form data */
|
|
conn->fread_func = (curl_read_callback)Curl_FormReader;
|
|
conn->fread_in = &http->form;
|
|
|
|
http->sending = HTTPSEND_BODY;
|
|
|
|
if(!data->req.upload_chunky) {
|
|
/* only add Content-Length if not uploading chunked */
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"Content-Length: %" FORMAT_OFF_T "\r\n",
|
|
http->postsize);
|
|
if(result)
|
|
return result;
|
|
}
|
|
|
|
result = expect100(data, conn, req_buffer);
|
|
if(result)
|
|
return result;
|
|
|
|
{
|
|
|
|
/* Get Content-Type: line from Curl_formpostheader.
|
|
*/
|
|
char *contentType;
|
|
size_t linelength=0;
|
|
contentType = Curl_formpostheader((void *)&http->form,
|
|
&linelength);
|
|
if(!contentType) {
|
|
failf(data, "Could not get Content-Type header line!");
|
|
return CURLE_HTTP_POST_ERROR;
|
|
}
|
|
|
|
result = Curl_add_buffer(req_buffer, contentType, linelength);
|
|
if(result)
|
|
return result;
|
|
}
|
|
|
|
/* make the request end in a true CRLF */
|
|
result = Curl_add_buffer(req_buffer, "\r\n", 2);
|
|
if(result)
|
|
return result;
|
|
|
|
/* set upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, http->postsize);
|
|
|
|
/* fire away the whole request to the server */
|
|
result = Curl_add_buffer_send(req_buffer, conn,
|
|
&data->info.request_size, 0, FIRSTSOCKET);
|
|
if(result)
|
|
failf(data, "Failed sending POST request");
|
|
else
|
|
/* setup variables for the upcoming transfer */
|
|
Curl_setup_transfer(conn, FIRSTSOCKET, -1, TRUE,
|
|
&http->readbytecount, FIRSTSOCKET,
|
|
&http->writebytecount);
|
|
|
|
if(result) {
|
|
Curl_formclean(&http->sendit); /* free that whole lot */
|
|
return result;
|
|
}
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
/* time to convert the form data... */
|
|
result = Curl_formconvert(data, http->sendit);
|
|
if(result) {
|
|
Curl_formclean(&http->sendit); /* free that whole lot */
|
|
return result;
|
|
}
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
break;
|
|
|
|
case HTTPREQ_PUT: /* Let's PUT the data to the server! */
|
|
|
|
if(conn->bits.authneg)
|
|
postsize = 0;
|
|
else
|
|
postsize = data->set.infilesize;
|
|
|
|
if((postsize != -1) && !data->req.upload_chunky) {
|
|
/* only add Content-Length if not uploading chunked */
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"Content-Length: %" FORMAT_OFF_T "\r\n",
|
|
postsize );
|
|
if(result)
|
|
return result;
|
|
}
|
|
|
|
result = expect100(data, conn, req_buffer);
|
|
if(result)
|
|
return result;
|
|
|
|
result = Curl_add_buffer(req_buffer, "\r\n", 2); /* end of headers */
|
|
if(result)
|
|
return result;
|
|
|
|
/* set the upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, postsize);
|
|
|
|
/* this sends the buffer and frees all the buffer resources */
|
|
result = Curl_add_buffer_send(req_buffer, conn,
|
|
&data->info.request_size, 0, FIRSTSOCKET);
|
|
if(result)
|
|
failf(data, "Failed sending PUT request");
|
|
else
|
|
/* prepare for transfer */
|
|
Curl_setup_transfer(conn, FIRSTSOCKET, -1, TRUE,
|
|
&http->readbytecount, postsize?FIRSTSOCKET:-1,
|
|
postsize?&http->writebytecount:NULL);
|
|
if(result)
|
|
return result;
|
|
break;
|
|
|
|
case HTTPREQ_POST:
|
|
/* this is the simple POST, using x-www-form-urlencoded style */
|
|
|
|
if(conn->bits.authneg)
|
|
postsize = 0;
|
|
else {
|
|
/* figure out the size of the postfields */
|
|
postsize = (data->set.postfieldsize != -1)?
|
|
data->set.postfieldsize:
|
|
(data->set.postfields? (curl_off_t)strlen(data->set.postfields):-1);
|
|
}
|
|
if(!data->req.upload_chunky) {
|
|
/* We only set Content-Length and allow a custom Content-Length if
|
|
we don't upload data chunked, as RFC2616 forbids us to set both
|
|
kinds of headers (Transfer-Encoding: chunked and Content-Length) */
|
|
|
|
if(conn->bits.authneg || !Curl_checkheaders(data, "Content-Length:")) {
|
|
/* we allow replacing this header if not during auth negotiation,
|
|
although it isn't very wise to actually set your own */
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"Content-Length: %" FORMAT_OFF_T"\r\n",
|
|
postsize);
|
|
if(result)
|
|
return result;
|
|
}
|
|
}
|
|
|
|
if(!Curl_checkheaders(data, "Content-Type:")) {
|
|
result = Curl_add_bufferf(req_buffer,
|
|
"Content-Type: application/"
|
|
"x-www-form-urlencoded\r\n");
|
|
if(result)
|
|
return result;
|
|
}
|
|
|
|
/* For really small posts we don't use Expect: headers at all, and for
|
|
the somewhat bigger ones we allow the app to disable it. Just make
|
|
sure that the expect100header is always set to the preferred value
|
|
here. */
|
|
ptr = Curl_checkheaders(data, "Expect:");
|
|
if(ptr) {
|
|
data->state.expect100header =
|
|
Curl_compareheader(ptr, "Expect:", "100-continue");
|
|
}
|
|
else if(postsize > TINY_INITIAL_POST_SIZE || postsize < 0) {
|
|
result = expect100(data, conn, req_buffer);
|
|
if(result)
|
|
return result;
|
|
}
|
|
else
|
|
data->state.expect100header = FALSE;
|
|
|
|
if(data->set.postfields) {
|
|
|
|
if(!data->state.expect100header &&
|
|
(postsize < MAX_INITIAL_POST_SIZE)) {
|
|
/* if we don't use expect: 100 AND
|
|
postsize is less than MAX_INITIAL_POST_SIZE
|
|
|
|
then append the post data to the HTTP request header. This limit
|
|
is no magic limit but only set to prevent really huge POSTs to
|
|
get the data duplicated with malloc() and family. */
|
|
|
|
result = Curl_add_buffer(req_buffer, "\r\n", 2); /* end of headers! */
|
|
if(result)
|
|
return result;
|
|
|
|
if(!data->req.upload_chunky) {
|
|
/* We're not sending it 'chunked', append it to the request
|
|
already now to reduce the number if send() calls */
|
|
result = Curl_add_buffer(req_buffer, data->set.postfields,
|
|
(size_t)postsize);
|
|
included_body = postsize;
|
|
}
|
|
else {
|
|
/* Append the POST data chunky-style */
|
|
result = Curl_add_bufferf(req_buffer, "%x\r\n", (int)postsize);
|
|
if(CURLE_OK == result)
|
|
result = Curl_add_buffer(req_buffer, data->set.postfields,
|
|
(size_t)postsize);
|
|
if(CURLE_OK == result)
|
|
result = Curl_add_buffer(req_buffer,
|
|
"\x0d\x0a\x30\x0d\x0a\x0d\x0a", 7);
|
|
/* CR LF 0 CR LF CR LF */
|
|
included_body = postsize + 7;
|
|
}
|
|
if(result)
|
|
return result;
|
|
/* Make sure the progress information is accurate */
|
|
Curl_pgrsSetUploadSize(data, postsize);
|
|
}
|
|
else {
|
|
/* A huge POST coming up, do data separate from the request */
|
|
http->postsize = postsize;
|
|
http->postdata = data->set.postfields;
|
|
|
|
http->sending = HTTPSEND_BODY;
|
|
|
|
conn->fread_func = (curl_read_callback)readmoredata;
|
|
conn->fread_in = (void *)conn;
|
|
|
|
/* set the upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, http->postsize);
|
|
|
|
result = Curl_add_buffer(req_buffer, "\r\n", 2); /* end of headers! */
|
|
if(result)
|
|
return result;
|
|
}
|
|
}
|
|
else {
|
|
result = Curl_add_buffer(req_buffer, "\r\n", 2); /* end of headers! */
|
|
if(result)
|
|
return result;
|
|
|
|
if(data->req.upload_chunky && conn->bits.authneg) {
|
|
/* Chunky upload is selected and we're negotiating auth still, send
|
|
end-of-data only */
|
|
result = Curl_add_buffer(req_buffer,
|
|
"\x0d\x0a\x30\x0d\x0a\x0d\x0a", 7);
|
|
/* CR LF 0 CR LF CR LF */
|
|
if(result)
|
|
return result;
|
|
}
|
|
|
|
else if(data->set.postfieldsize) {
|
|
/* set the upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, postsize?postsize:-1);
|
|
|
|
/* set the pointer to mark that we will send the post body using the
|
|
read callback, but only if we're not in authenticate
|
|
negotiation */
|
|
if(!conn->bits.authneg) {
|
|
http->postdata = (char *)&http->postdata;
|
|
http->postsize = postsize;
|
|
}
|
|
}
|
|
}
|
|
/* issue the request */
|
|
result = Curl_add_buffer_send(req_buffer, conn, &data->info.request_size,
|
|
(size_t)included_body, FIRSTSOCKET);
|
|
|
|
if(result)
|
|
failf(data, "Failed sending HTTP POST request");
|
|
else
|
|
Curl_setup_transfer(conn, FIRSTSOCKET, -1, TRUE,
|
|
&http->readbytecount, http->postdata?FIRSTSOCKET:-1,
|
|
http->postdata?&http->writebytecount:NULL);
|
|
break;
|
|
|
|
default:
|
|
result = Curl_add_buffer(req_buffer, "\r\n", 2);
|
|
if(result)
|
|
return result;
|
|
|
|
/* issue the request */
|
|
result = Curl_add_buffer_send(req_buffer, conn,
|
|
&data->info.request_size, 0, FIRSTSOCKET);
|
|
|
|
if(result)
|
|
failf(data, "Failed sending HTTP request");
|
|
else
|
|
/* HTTP GET/HEAD download: */
|
|
Curl_setup_transfer(conn, FIRSTSOCKET, -1, TRUE, &http->readbytecount,
|
|
http->postdata?FIRSTSOCKET:-1,
|
|
http->postdata?&http->writebytecount:NULL);
|
|
}
|
|
if(result)
|
|
return result;
|
|
|
|
if(http->writebytecount) {
|
|
/* if a request-body has been sent off, we make sure this progress is noted
|
|
properly */
|
|
Curl_pgrsSetUploadCounter(data, http->writebytecount);
|
|
if(Curl_pgrsUpdate(conn))
|
|
result = CURLE_ABORTED_BY_CALLBACK;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
/*
|
|
* checkhttpprefix()
|
|
*
|
|
* Returns TRUE if member of the list matches prefix of string
|
|
*/
|
|
static bool
|
|
checkhttpprefix(struct SessionHandle *data,
|
|
const char *s)
|
|
{
|
|
struct curl_slist *head = data->set.http200aliases;
|
|
bool rc = FALSE;
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
/* convert from the network encoding using a scratch area */
|
|
char *scratch = strdup(s);
|
|
if(NULL == scratch) {
|
|
failf (data, "Failed to allocate memory for conversion!");
|
|
return FALSE; /* can't return CURLE_OUT_OF_MEMORY so return FALSE */
|
|
}
|
|
if(CURLE_OK != Curl_convert_from_network(data, scratch, strlen(s)+1)) {
|
|
/* Curl_convert_from_network calls failf if unsuccessful */
|
|
free(scratch);
|
|
return FALSE; /* can't return CURLE_foobar so return FALSE */
|
|
}
|
|
s = scratch;
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
|
|
while(head) {
|
|
if(checkprefix(head->data, s)) {
|
|
rc = TRUE;
|
|
break;
|
|
}
|
|
head = head->next;
|
|
}
|
|
|
|
if((rc != TRUE) && (checkprefix("HTTP/", s))) {
|
|
rc = TRUE;
|
|
}
|
|
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
free(scratch);
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
return rc;
|
|
}
|
|
|
|
#ifndef CURL_DISABLE_RTSP
|
|
static bool
|
|
checkrtspprefix(struct SessionHandle *data,
|
|
const char *s)
|
|
{
|
|
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
/* convert from the network encoding using a scratch area */
|
|
char *scratch = strdup(s);
|
|
if(NULL == scratch) {
|
|
failf (data, "Failed to allocate memory for conversion!");
|
|
return FALSE; /* can't return CURLE_OUT_OF_MEMORY so return FALSE */
|
|
}
|
|
if(CURLE_OK != Curl_convert_from_network(data, scratch, strlen(s)+1)) {
|
|
/* Curl_convert_from_network calls failf if unsuccessful */
|
|
free(scratch);
|
|
return FALSE; /* can't return CURLE_foobar so return FALSE */
|
|
}
|
|
s = scratch;
|
|
#else
|
|
(void)data; /* unused */
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
if(checkprefix("RTSP/", s))
|
|
return TRUE;
|
|
else
|
|
return FALSE;
|
|
}
|
|
#endif /* CURL_DISABLE_RTSP */
|
|
|
|
static bool
|
|
checkprotoprefix(struct SessionHandle *data, struct connectdata *conn,
|
|
const char *s)
|
|
{
|
|
#ifndef CURL_DISABLE_RTSP
|
|
if(conn->protocol & PROT_RTSP)
|
|
return checkrtspprefix(data, s);
|
|
#else
|
|
(void)conn;
|
|
#endif /* CURL_DISABLE_RTSP */
|
|
|
|
return checkhttpprefix(data, s);
|
|
}
|
|
|
|
/*
|
|
* header_append() copies a chunk of data to the end of the already received
|
|
* header. We make sure that the full string fit in the allocated header
|
|
* buffer, or else we enlarge it.
|
|
*/
|
|
static CURLcode header_append(struct SessionHandle *data,
|
|
struct SingleRequest *k,
|
|
size_t length)
|
|
{
|
|
if(k->hbuflen + length >= data->state.headersize) {
|
|
/* We enlarge the header buffer as it is too small */
|
|
char *newbuff;
|
|
size_t hbufp_index;
|
|
size_t newsize;
|
|
|
|
if(k->hbuflen + length > CURL_MAX_HTTP_HEADER) {
|
|
/* The reason to have a max limit for this is to avoid the risk of a bad
|
|
server feeding libcurl with a never-ending header that will cause
|
|
reallocs infinitely */
|
|
failf (data, "Avoided giant realloc for header (max is %d)!",
|
|
CURL_MAX_HTTP_HEADER);
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
newsize=CURLMAX((k->hbuflen+ length)*3/2, data->state.headersize*2);
|
|
hbufp_index = k->hbufp - data->state.headerbuff;
|
|
newbuff = realloc(data->state.headerbuff, newsize);
|
|
if(!newbuff) {
|
|
failf (data, "Failed to alloc memory for big header!");
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
data->state.headersize=newsize;
|
|
data->state.headerbuff = newbuff;
|
|
k->hbufp = data->state.headerbuff + hbufp_index;
|
|
}
|
|
memcpy(k->hbufp, k->str_start, length);
|
|
k->hbufp += length;
|
|
k->hbuflen += length;
|
|
*k->hbufp = 0;
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
|
|
/*
|
|
* Read any HTTP header lines from the server and pass them to the client app.
|
|
*/
|
|
CURLcode Curl_http_readwrite_headers(struct SessionHandle *data,
|
|
struct connectdata *conn,
|
|
ssize_t *nread,
|
|
bool *stop_reading)
|
|
{
|
|
CURLcode result;
|
|
struct SingleRequest *k = &data->req;
|
|
|
|
/* header line within buffer loop */
|
|
do {
|
|
size_t rest_length;
|
|
size_t full_length;
|
|
int writetype;
|
|
|
|
/* str_start is start of line within buf */
|
|
k->str_start = k->str;
|
|
|
|
/* data is in network encoding so use 0x0a instead of '\n' */
|
|
k->end_ptr = memchr(k->str_start, 0x0a, *nread);
|
|
|
|
if(!k->end_ptr) {
|
|
/* Not a complete header line within buffer, append the data to
|
|
the end of the headerbuff. */
|
|
result = header_append(data, k, *nread);
|
|
if(result)
|
|
return result;
|
|
|
|
if(!k->headerline && (k->hbuflen>5)) {
|
|
/* make a first check that this looks like a protocol header */
|
|
if(!checkprotoprefix(data, conn, data->state.headerbuff)) {
|
|
/* this is not the beginning of a protocol first header line */
|
|
k->header = FALSE;
|
|
k->badheader = HEADER_ALLBAD;
|
|
break;
|
|
}
|
|
}
|
|
|
|
break; /* read more and try again */
|
|
}
|
|
|
|
/* decrease the size of the remaining (supposed) header line */
|
|
rest_length = (k->end_ptr - k->str)+1;
|
|
*nread -= (ssize_t)rest_length;
|
|
|
|
k->str = k->end_ptr + 1; /* move past new line */
|
|
|
|
full_length = k->str - k->str_start;
|
|
|
|
result = header_append(data, k, full_length);
|
|
if(result)
|
|
return result;
|
|
|
|
k->end_ptr = k->hbufp;
|
|
k->p = data->state.headerbuff;
|
|
|
|
/****
|
|
* We now have a FULL header line that p points to
|
|
*****/
|
|
|
|
if(!k->headerline) {
|
|
/* the first read header */
|
|
if((k->hbuflen>5) &&
|
|
!checkprotoprefix(data, conn, data->state.headerbuff)) {
|
|
/* this is not the beginning of a protocol first header line */
|
|
k->header = FALSE;
|
|
if(*nread)
|
|
/* since there's more, this is a partial bad header */
|
|
k->badheader = HEADER_PARTHEADER;
|
|
else {
|
|
/* this was all we read so it's all a bad header */
|
|
k->badheader = HEADER_ALLBAD;
|
|
*nread = (ssize_t)rest_length;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* headers are in network encoding so
|
|
use 0x0a and 0x0d instead of '\n' and '\r' */
|
|
if((0x0a == *k->p) || (0x0d == *k->p)) {
|
|
size_t headerlen;
|
|
/* Zero-length header line means end of headers! */
|
|
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
if(0x0d == *k->p) {
|
|
*k->p = '\r'; /* replace with CR in host encoding */
|
|
k->p++; /* pass the CR byte */
|
|
}
|
|
if(0x0a == *k->p) {
|
|
*k->p = '\n'; /* replace with LF in host encoding */
|
|
k->p++; /* pass the LF byte */
|
|
}
|
|
#else
|
|
if('\r' == *k->p)
|
|
k->p++; /* pass the \r byte */
|
|
if('\n' == *k->p)
|
|
k->p++; /* pass the \n byte */
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
|
|
if(100 <= k->httpcode && 199 >= k->httpcode) {
|
|
/*
|
|
* We have made a HTTP PUT or POST and this is 1.1-lingo
|
|
* that tells us that the server is OK with this and ready
|
|
* to receive the data.
|
|
* However, we'll get more headers now so we must get
|
|
* back into the header-parsing state!
|
|
*/
|
|
k->header = TRUE;
|
|
k->headerline = 0; /* restart the header line counter */
|
|
|
|
/* if we did wait for this do enable write now! */
|
|
if(k->exp100) {
|
|
k->exp100 = EXP100_SEND_DATA;
|
|
k->keepon |= KEEP_SEND;
|
|
}
|
|
}
|
|
else {
|
|
k->header = FALSE; /* no more header to parse! */
|
|
|
|
if((k->size == -1) && !k->chunk && !conn->bits.close &&
|
|
(conn->httpversion >= 11) && !(conn->protocol & PROT_RTSP)) {
|
|
/* On HTTP 1.1, when connection is not to get closed, but no
|
|
Content-Length nor Content-Encoding chunked have been
|
|
received, according to RFC2616 section 4.4 point 5, we
|
|
assume that the server will close the connection to
|
|
signal the end of the document. */
|
|
infof(data, "no chunk, no close, no size. Assume close to "
|
|
"signal end\n");
|
|
conn->bits.close = TRUE;
|
|
}
|
|
}
|
|
|
|
if(417 == k->httpcode) {
|
|
/*
|
|
* we got: "417 Expectation Failed" this means:
|
|
* we have made a HTTP call and our Expect Header
|
|
* seems to cause a problem => abort the write operations
|
|
* (or prevent them from starting).
|
|
*/
|
|
k->exp100 = EXP100_FAILED;
|
|
k->keepon &= ~KEEP_SEND;
|
|
}
|
|
|
|
/*
|
|
* When all the headers have been parsed, see if we should give
|
|
* up and return an error.
|
|
*/
|
|
if(Curl_http_should_fail(conn)) {
|
|
failf (data, "The requested URL returned error: %d",
|
|
k->httpcode);
|
|
return CURLE_HTTP_RETURNED_ERROR;
|
|
}
|
|
|
|
/* now, only output this if the header AND body are requested:
|
|
*/
|
|
writetype = CLIENTWRITE_HEADER;
|
|
if(data->set.include_header)
|
|
writetype |= CLIENTWRITE_BODY;
|
|
|
|
headerlen = k->p - data->state.headerbuff;
|
|
|
|
result = Curl_client_write(conn, writetype,
|
|
data->state.headerbuff,
|
|
headerlen);
|
|
if(result)
|
|
return result;
|
|
|
|
data->info.header_size += (long)headerlen;
|
|
data->req.headerbytecount += (long)headerlen;
|
|
|
|
data->req.deductheadercount =
|
|
(100 <= k->httpcode && 199 >= k->httpcode)?data->req.headerbytecount:0;
|
|
|
|
if(!*stop_reading) {
|
|
/* Curl_http_auth_act() checks what authentication methods
|
|
* that are available and decides which one (if any) to
|
|
* use. It will set 'newurl' if an auth method was picked. */
|
|
result = Curl_http_auth_act(conn);
|
|
|
|
if(result)
|
|
return result;
|
|
|
|
if(conn->bits.rewindaftersend) {
|
|
/* We rewind after a complete send, so thus we continue
|
|
sending now */
|
|
infof(data, "Keep sending data to get tossed away!\n");
|
|
k->keepon |= KEEP_SEND;
|
|
}
|
|
}
|
|
|
|
if(!k->header) {
|
|
/*
|
|
* really end-of-headers.
|
|
*
|
|
* If we requested a "no body", this is a good time to get
|
|
* out and return home.
|
|
*/
|
|
if(data->set.opt_no_body)
|
|
*stop_reading = TRUE;
|
|
else {
|
|
/* If we know the expected size of this document, we set the
|
|
maximum download size to the size of the expected
|
|
document or else, we won't know when to stop reading!
|
|
|
|
Note that we set the download maximum even if we read a
|
|
"Connection: close" header, to make sure that
|
|
"Content-Length: 0" still prevents us from attempting to
|
|
read the (missing) response-body.
|
|
*/
|
|
/* According to RFC2616 section 4.4, we MUST ignore
|
|
Content-Length: headers if we are now receiving data
|
|
using chunked Transfer-Encoding.
|
|
*/
|
|
if(k->chunk)
|
|
k->maxdownload = k->size = -1;
|
|
}
|
|
if(-1 != k->size) {
|
|
/* We do this operation even if no_body is true, since this
|
|
data might be retrieved later with curl_easy_getinfo()
|
|
and its CURLINFO_CONTENT_LENGTH_DOWNLOAD option. */
|
|
|
|
Curl_pgrsSetDownloadSize(data, k->size);
|
|
k->maxdownload = k->size;
|
|
}
|
|
|
|
/* If max download size is *zero* (nothing) we already
|
|
have nothing and can safely return ok now! */
|
|
if(0 == k->maxdownload)
|
|
*stop_reading = TRUE;
|
|
|
|
if(*stop_reading) {
|
|
/* we make sure that this socket isn't read more now */
|
|
k->keepon &= ~KEEP_RECV;
|
|
}
|
|
|
|
if(data->set.verbose)
|
|
Curl_debug(data, CURLINFO_HEADER_IN,
|
|
k->str_start, headerlen, conn);
|
|
break; /* exit header line loop */
|
|
}
|
|
|
|
/* We continue reading headers, so reset the line-based
|
|
header parsing variables hbufp && hbuflen */
|
|
k->hbufp = data->state.headerbuff;
|
|
k->hbuflen = 0;
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* Checks for special headers coming up.
|
|
*/
|
|
|
|
if(!k->headerline++) {
|
|
/* This is the first header, it MUST be the error code line
|
|
or else we consider this to be the body right away! */
|
|
int httpversion_major;
|
|
int rtspversion_major;
|
|
int nc = 0;
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
#define HEADER1 scratch
|
|
#define SCRATCHSIZE 21
|
|
CURLcode res;
|
|
char scratch[SCRATCHSIZE+1]; /* "HTTP/major.minor 123" */
|
|
/* We can't really convert this yet because we
|
|
don't know if it's the 1st header line or the body.
|
|
So we do a partial conversion into a scratch area,
|
|
leaving the data at k->p as-is.
|
|
*/
|
|
strncpy(&scratch[0], k->p, SCRATCHSIZE);
|
|
scratch[SCRATCHSIZE] = 0; /* null terminate */
|
|
res = Curl_convert_from_network(data,
|
|
&scratch[0],
|
|
SCRATCHSIZE);
|
|
if(CURLE_OK != res) {
|
|
/* Curl_convert_from_network calls failf if unsuccessful */
|
|
return res;
|
|
}
|
|
#else
|
|
#define HEADER1 k->p /* no conversion needed, just use k->p */
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
|
|
if(conn->protocol & PROT_HTTP) {
|
|
nc = sscanf(HEADER1,
|
|
" HTTP/%d.%d %3d",
|
|
&httpversion_major,
|
|
&conn->httpversion,
|
|
&k->httpcode);
|
|
if(nc==3) {
|
|
conn->httpversion += 10 * httpversion_major;
|
|
}
|
|
else {
|
|
/* this is the real world, not a Nirvana
|
|
NCSA 1.5.x returns this crap when asked for HTTP/1.1
|
|
*/
|
|
nc=sscanf(HEADER1, " HTTP %3d", &k->httpcode);
|
|
conn->httpversion = 10;
|
|
|
|
/* If user has set option HTTP200ALIASES,
|
|
compare header line against list of aliases
|
|
*/
|
|
if(!nc) {
|
|
if(checkhttpprefix(data, k->p)) {
|
|
nc = 1;
|
|
k->httpcode = 200;
|
|
conn->httpversion = 10;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else if(conn->protocol & PROT_RTSP) {
|
|
nc = sscanf(HEADER1,
|
|
" RTSP/%d.%d %3d",
|
|
&rtspversion_major,
|
|
&conn->rtspversion,
|
|
&k->httpcode);
|
|
if(nc==3) {
|
|
conn->rtspversion += 10 * rtspversion_major;
|
|
conn->httpversion = 11; /* For us, RTSP acts like HTTP 1.1 */
|
|
}
|
|
else {
|
|
/* TODO: do we care about the other cases here? */
|
|
nc = 0;
|
|
}
|
|
}
|
|
|
|
if(nc) {
|
|
data->info.httpcode = k->httpcode;
|
|
|
|
data->info.httpversion = conn->httpversion;
|
|
if (!data->state.httpversion ||
|
|
data->state.httpversion > conn->httpversion)
|
|
/* store the lowest server version we encounter */
|
|
data->state.httpversion = conn->httpversion;
|
|
|
|
/*
|
|
* This code executes as part of processing the header. As a
|
|
* result, it's not totally clear how to interpret the
|
|
* response code yet as that depends on what other headers may
|
|
* be present. 401 and 407 may be errors, but may be OK
|
|
* depending on how authentication is working. Other codes
|
|
* are definitely errors, so give up here.
|
|
*/
|
|
if(data->set.http_fail_on_error && (k->httpcode >= 400) &&
|
|
((k->httpcode != 401) || !conn->bits.user_passwd) &&
|
|
((k->httpcode != 407) || !conn->bits.proxy_user_passwd) ) {
|
|
|
|
if(data->state.resume_from &&
|
|
(data->set.httpreq==HTTPREQ_GET) &&
|
|
(k->httpcode == 416)) {
|
|
/* "Requested Range Not Satisfiable", just proceed and
|
|
pretend this is no error */
|
|
}
|
|
else {
|
|
/* serious error, go home! */
|
|
failf (data, "The requested URL returned error: %d",
|
|
k->httpcode);
|
|
return CURLE_HTTP_RETURNED_ERROR;
|
|
}
|
|
}
|
|
|
|
if(conn->httpversion == 10) {
|
|
/* Default action for HTTP/1.0 must be to close, unless
|
|
we get one of those fancy headers that tell us the
|
|
server keeps it open for us! */
|
|
infof(data, "HTTP 1.0, assume close after body\n");
|
|
conn->bits.close = TRUE;
|
|
}
|
|
else if(conn->httpversion >= 11 &&
|
|
!conn->bits.close) {
|
|
/* If HTTP version is >= 1.1 and connection is persistent
|
|
server supports pipelining. */
|
|
DEBUGF(infof(data,
|
|
"HTTP 1.1 or later with persistent connection, "
|
|
"pipelining supported\n"));
|
|
conn->server_supports_pipelining = TRUE;
|
|
}
|
|
|
|
switch(k->httpcode) {
|
|
case 204:
|
|
/* (quote from RFC2616, section 10.2.5): The server has
|
|
* fulfilled the request but does not need to return an
|
|
* entity-body ... The 204 response MUST NOT include a
|
|
* message-body, and thus is always terminated by the first
|
|
* empty line after the header fields. */
|
|
/* FALLTHROUGH */
|
|
case 304:
|
|
/* (quote from RFC2616, section 10.3.5): The 304 response
|
|
* MUST NOT contain a message-body, and thus is always
|
|
* terminated by the first empty line after the header
|
|
* fields. */
|
|
if(data->set.timecondition)
|
|
data->info.timecond = TRUE;
|
|
k->size=0;
|
|
k->maxdownload=0;
|
|
k->ignorecl = TRUE; /* ignore Content-Length headers */
|
|
break;
|
|
default:
|
|
/* nothing */
|
|
break;
|
|
}
|
|
}
|
|
else {
|
|
k->header = FALSE; /* this is not a header line */
|
|
break;
|
|
}
|
|
}
|
|
|
|
#ifdef CURL_DOES_CONVERSIONS
|
|
/* convert from the network encoding */
|
|
result = Curl_convert_from_network(data, k->p, strlen(k->p));
|
|
if(CURLE_OK != result) {
|
|
return(result);
|
|
}
|
|
/* Curl_convert_from_network calls failf if unsuccessful */
|
|
#endif /* CURL_DOES_CONVERSIONS */
|
|
|
|
/* Check for Content-Length: header lines to get size */
|
|
if(!k->ignorecl && !data->set.ignorecl &&
|
|
checkprefix("Content-Length:", k->p)) {
|
|
curl_off_t contentlength = curlx_strtoofft(k->p+15, NULL, 10);
|
|
if(data->set.max_filesize &&
|
|
contentlength > data->set.max_filesize) {
|
|
failf(data, "Maximum file size exceeded");
|
|
return CURLE_FILESIZE_EXCEEDED;
|
|
}
|
|
if(contentlength >= 0) {
|
|
k->size = contentlength;
|
|
k->maxdownload = k->size;
|
|
/* we set the progress download size already at this point
|
|
just to make it easier for apps/callbacks to extract this
|
|
info as soon as possible */
|
|
Curl_pgrsSetDownloadSize(data, k->size);
|
|
}
|
|
else {
|
|
/* Negative Content-Length is really odd, and we know it
|
|
happens for example when older Apache servers send large
|
|
files */
|
|
conn->bits.close = TRUE;
|
|
infof(data, "Negative content-length: %" FORMAT_OFF_T
|
|
", closing after transfer\n", contentlength);
|
|
}
|
|
}
|
|
/* check for Content-Type: header lines to get the MIME-type */
|
|
else if(checkprefix("Content-Type:", k->p)) {
|
|
char *contenttype = Curl_copy_header_value(k->p);
|
|
if (!contenttype)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
if (!*contenttype)
|
|
/* ignore empty data */
|
|
free(contenttype);
|
|
else {
|
|
Curl_safefree(data->info.contenttype);
|
|
data->info.contenttype = contenttype;
|
|
}
|
|
}
|
|
else if((conn->httpversion == 10) &&
|
|
conn->bits.httpproxy &&
|
|
Curl_compareheader(k->p,
|
|
"Proxy-Connection:", "keep-alive")) {
|
|
/*
|
|
* When a HTTP/1.0 reply comes when using a proxy, the
|
|
* 'Proxy-Connection: keep-alive' line tells us the
|
|
* connection will be kept alive for our pleasure.
|
|
* Default action for 1.0 is to close.
|
|
*/
|
|
conn->bits.close = FALSE; /* don't close when done */
|
|
infof(data, "HTTP/1.0 proxy connection set to keep alive!\n");
|
|
}
|
|
else if((conn->httpversion == 11) &&
|
|
conn->bits.httpproxy &&
|
|
Curl_compareheader(k->p,
|
|
"Proxy-Connection:", "close")) {
|
|
/*
|
|
* We get a HTTP/1.1 response from a proxy and it says it'll
|
|
* close down after this transfer.
|
|
*/
|
|
conn->bits.close = TRUE; /* close when done */
|
|
infof(data, "HTTP/1.1 proxy connection set close!\n");
|
|
}
|
|
else if((conn->httpversion == 10) &&
|
|
Curl_compareheader(k->p, "Connection:", "keep-alive")) {
|
|
/*
|
|
* A HTTP/1.0 reply with the 'Connection: keep-alive' line
|
|
* tells us the connection will be kept alive for our
|
|
* pleasure. Default action for 1.0 is to close.
|
|
*
|
|
* [RFC2068, section 19.7.1] */
|
|
conn->bits.close = FALSE; /* don't close when done */
|
|
infof(data, "HTTP/1.0 connection set to keep alive!\n");
|
|
}
|
|
else if(Curl_compareheader(k->p, "Connection:", "close")) {
|
|
/*
|
|
* [RFC 2616, section 8.1.2.1]
|
|
* "Connection: close" is HTTP/1.1 language and means that
|
|
* the connection will close when this request has been
|
|
* served.
|
|
*/
|
|
conn->bits.close = TRUE; /* close when done */
|
|
}
|
|
else if(Curl_compareheader(k->p, "Transfer-Encoding:", "chunked") &&
|
|
!(conn->protocol & PROT_RTSP)) {
|
|
/*
|
|
* [RFC 2616, section 3.6.1] A 'chunked' transfer encoding
|
|
* means that the server will send a series of "chunks". Each
|
|
* chunk starts with line with info (including size of the
|
|
* coming block) (terminated with CRLF), then a block of data
|
|
* with the previously mentioned size. There can be any amount
|
|
* of chunks, and a chunk-data set to zero signals the
|
|
* end-of-chunks. */
|
|
k->chunk = TRUE; /* chunks coming our way */
|
|
|
|
/* init our chunky engine */
|
|
Curl_httpchunk_init(conn);
|
|
}
|
|
else if(checkprefix("Content-Encoding:", k->p) &&
|
|
data->set.str[STRING_ENCODING]) {
|
|
/*
|
|
* Process Content-Encoding. Look for the values: identity,
|
|
* gzip, deflate, compress, x-gzip and x-compress. x-gzip and
|
|
* x-compress are the same as gzip and compress. (Sec 3.5 RFC
|
|
* 2616). zlib cannot handle compress. However, errors are
|
|
* handled further down when the response body is processed
|
|
*/
|
|
char *start;
|
|
|
|
/* Find the first non-space letter */
|
|
start = k->p + 17;
|
|
while(*start && ISSPACE(*start))
|
|
start++;
|
|
|
|
/* Record the content-encoding for later use */
|
|
if(checkprefix("identity", start))
|
|
k->content_encoding = IDENTITY;
|
|
else if(checkprefix("deflate", start))
|
|
k->content_encoding = DEFLATE;
|
|
else if(checkprefix("gzip", start)
|
|
|| checkprefix("x-gzip", start))
|
|
k->content_encoding = GZIP;
|
|
else if(checkprefix("compress", start)
|
|
|| checkprefix("x-compress", start))
|
|
k->content_encoding = COMPRESS;
|
|
}
|
|
else if(checkprefix("Content-Range:", k->p)) {
|
|
/* Content-Range: bytes [num]-
|
|
Content-Range: bytes: [num]-
|
|
Content-Range: [num]-
|
|
|
|
The second format was added since Sun's webserver
|
|
JavaWebServer/1.1.1 obviously sends the header this way!
|
|
The third added since some servers use that!
|
|
*/
|
|
|
|
char *ptr = k->p + 14;
|
|
|
|
/* Move forward until first digit */
|
|
while(*ptr && !ISDIGIT(*ptr))
|
|
ptr++;
|
|
|
|
k->offset = curlx_strtoofft(ptr, NULL, 10);
|
|
|
|
if(data->state.resume_from == k->offset)
|
|
/* we asked for a resume and we got it */
|
|
k->content_range = TRUE;
|
|
}
|
|
#if !defined(CURL_DISABLE_COOKIES)
|
|
else if(data->cookies &&
|
|
checkprefix("Set-Cookie:", k->p)) {
|
|
Curl_share_lock(data, CURL_LOCK_DATA_COOKIE,
|
|
CURL_LOCK_ACCESS_SINGLE);
|
|
Curl_cookie_add(data,
|
|
data->cookies, TRUE, k->p+11,
|
|
/* If there is a custom-set Host: name, use it
|
|
here, or else use real peer host name. */
|
|
conn->allocptr.cookiehost?
|
|
conn->allocptr.cookiehost:conn->host.name,
|
|
data->state.path);
|
|
Curl_share_unlock(data, CURL_LOCK_DATA_COOKIE);
|
|
}
|
|
#endif
|
|
else if(checkprefix("Last-Modified:", k->p) &&
|
|
(data->set.timecondition || data->set.get_filetime) ) {
|
|
time_t secs=time(NULL);
|
|
k->timeofdoc = curl_getdate(k->p+strlen("Last-Modified:"),
|
|
&secs);
|
|
if(data->set.get_filetime)
|
|
data->info.filetime = (long)k->timeofdoc;
|
|
}
|
|
else if((checkprefix("WWW-Authenticate:", k->p) &&
|
|
(401 == k->httpcode)) ||
|
|
(checkprefix("Proxy-authenticate:", k->p) &&
|
|
(407 == k->httpcode))) {
|
|
result = Curl_http_input_auth(conn, k->httpcode, k->p);
|
|
if(result)
|
|
return result;
|
|
}
|
|
else if((k->httpcode >= 300 && k->httpcode < 400) &&
|
|
checkprefix("Location:", k->p)) {
|
|
/* this is the URL that the server advises us to use instead */
|
|
char *location = Curl_copy_header_value(k->p);
|
|
if (!location)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
if (!*location)
|
|
/* ignore empty data */
|
|
free(location);
|
|
else {
|
|
DEBUGASSERT(!data->req.location);
|
|
data->req.location = location;
|
|
|
|
if(data->set.http_follow_location) {
|
|
DEBUGASSERT(!data->req.newurl);
|
|
data->req.newurl = strdup(data->req.location); /* clone */
|
|
if(!data->req.newurl)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
|
|
/* some cases of POST and PUT etc needs to rewind the data
|
|
stream at this point */
|
|
result = Curl_http_perhapsrewind(conn);
|
|
if(result)
|
|
return result;
|
|
}
|
|
}
|
|
}
|
|
#ifndef CURL_DISABLE_RTSP
|
|
else if(conn->protocol & PROT_RTSP) {
|
|
result = Curl_rtsp_parseheader(conn, k->p);
|
|
if(result)
|
|
return result;
|
|
}
|
|
#endif
|
|
/*
|
|
* End of header-checks. Write them to the client.
|
|
*/
|
|
|
|
writetype = CLIENTWRITE_HEADER;
|
|
if(data->set.include_header)
|
|
writetype |= CLIENTWRITE_BODY;
|
|
|
|
if(data->set.verbose)
|
|
Curl_debug(data, CURLINFO_HEADER_IN,
|
|
k->p, (size_t)k->hbuflen, conn);
|
|
|
|
result = Curl_client_write(conn, writetype, k->p, k->hbuflen);
|
|
if(result)
|
|
return result;
|
|
|
|
data->info.header_size += (long)k->hbuflen;
|
|
data->req.headerbytecount += (long)k->hbuflen;
|
|
|
|
/* reset hbufp pointer && hbuflen */
|
|
k->hbufp = data->state.headerbuff;
|
|
k->hbuflen = 0;
|
|
}
|
|
while(!*stop_reading && *k->str); /* header line within buffer */
|
|
|
|
/* We might have reached the end of the header part here, but
|
|
there might be a non-header part left in the end of the read
|
|
buffer. */
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
#endif /* CURL_DISABLE_HTTP */
|