mirror of
https://github.com/unknownworlds/NS.git
synced 2024-12-11 05:10:56 +00:00
19b458f8bc
git-svn-id: https://unknownworlds.svn.cloudforge.com/ns1@141 67975925-1194-0748-b3d5-c16f83f1a3a1
1212 lines
38 KiB
C
1212 lines
38 KiB
C
/***************************************************************************
|
|
* _ _ ____ _
|
|
* Project ___| | | | _ \| |
|
|
* / __| | | | |_) | |
|
|
* | (__| |_| | _ <| |___
|
|
* \___|\___/|_| \_\_____|
|
|
*
|
|
* Copyright (C) 1998 - 2003, Daniel Stenberg, <daniel@haxx.se>, et al.
|
|
*
|
|
* This software is licensed as described in the file COPYING, which
|
|
* you should have received as part of this distribution. The terms
|
|
* are also available at http://curl.haxx.se/docs/copyright.html.
|
|
*
|
|
* You may opt to use, copy, modify, merge, publish, distribute and/or sell
|
|
* copies of the Software, and permit persons to whom the Software is
|
|
* furnished to do so, under the terms of the COPYING file.
|
|
*
|
|
* This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
|
|
* KIND, either express or implied.
|
|
*
|
|
* $Id: http.c,v 1.128 2003/05/12 12:29:00 bagder Exp $
|
|
***************************************************************************/
|
|
|
|
#include "setup.h"
|
|
|
|
#ifndef CURL_DISABLE_HTTP
|
|
/* -- WIN32 approved -- */
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <stdarg.h>
|
|
#include <stdlib.h>
|
|
#include <sys/types.h>
|
|
#include <sys/stat.h>
|
|
|
|
#include <errno.h>
|
|
|
|
#if defined(WIN32) && !defined(__GNUC__) || defined(__MINGW32__)
|
|
#include <winsock.h>
|
|
#include <time.h>
|
|
#include <io.h>
|
|
#else
|
|
#ifdef HAVE_SYS_SOCKET_H
|
|
#include <sys/socket.h>
|
|
#endif
|
|
#ifdef HAVE_NETINET_IN_H
|
|
#include <netinet/in.h>
|
|
#endif
|
|
#include <sys/time.h>
|
|
|
|
#ifdef HAVE_TIME_H
|
|
#ifdef TIME_WITH_SYS_TIME
|
|
#include <time.h>
|
|
#endif
|
|
#endif
|
|
|
|
#include <sys/resource.h>
|
|
#ifdef HAVE_UNISTD_H
|
|
#include <unistd.h>
|
|
#endif
|
|
#include <netdb.h>
|
|
#ifdef HAVE_ARPA_INET_H
|
|
#include <arpa/inet.h>
|
|
#endif
|
|
#ifdef HAVE_NET_IF_H
|
|
#include <net/if.h>
|
|
#endif
|
|
#include <sys/ioctl.h>
|
|
#include <signal.h>
|
|
|
|
#ifdef HAVE_SYS_PARAM_H
|
|
#include <sys/param.h>
|
|
#endif
|
|
|
|
#ifdef HAVE_SYS_SELECT_H
|
|
#include <sys/select.h>
|
|
#endif
|
|
|
|
|
|
#endif
|
|
|
|
#include "urldata.h"
|
|
#include <curl/curl.h>
|
|
#include "transfer.h"
|
|
#include "sendf.h"
|
|
#include "formdata.h"
|
|
#include "progress.h"
|
|
#include "base64.h"
|
|
#include "cookie.h"
|
|
#include "strequal.h"
|
|
#include "ssluse.h"
|
|
|
|
#define _MPRINTF_REPLACE /* use our functions only */
|
|
#include <curl/mprintf.h>
|
|
|
|
/* The last #include file should be: */
|
|
#ifdef MALLOCDEBUG
|
|
#include "memdebug.h"
|
|
#endif
|
|
|
|
/* fread() emulation to provide POST and/or request data */
|
|
static int readmoredata(char *buffer,
|
|
size_t size,
|
|
size_t nitems,
|
|
void *userp)
|
|
{
|
|
struct connectdata *conn = (struct connectdata *)userp;
|
|
struct HTTP *http = conn->proto.http;
|
|
int fullsize = size * nitems;
|
|
|
|
if(0 == http->postsize)
|
|
/* nothing to return */
|
|
return 0;
|
|
|
|
/* make sure that a HTTP request is never sent away chunked! */
|
|
conn->bits.forbidchunk= (http->sending == HTTPSEND_REQUEST)?TRUE:FALSE;
|
|
|
|
if(http->postsize <= fullsize) {
|
|
memcpy(buffer, http->postdata, http->postsize);
|
|
fullsize = http->postsize;
|
|
|
|
if(http->backup.postsize) {
|
|
/* move backup data into focus and continue on that */
|
|
http->postdata = http->backup.postdata;
|
|
http->postsize = http->backup.postsize;
|
|
conn->fread = http->backup.fread;
|
|
conn->fread_in = http->backup.fread_in;
|
|
|
|
http->sending++; /* move one step up */
|
|
|
|
http->backup.postsize=0;
|
|
}
|
|
else
|
|
http->postsize = 0;
|
|
|
|
return fullsize;
|
|
}
|
|
|
|
memcpy(buffer, http->postdata, fullsize);
|
|
http->postdata += fullsize;
|
|
http->postsize -= fullsize;
|
|
|
|
return fullsize;
|
|
}
|
|
|
|
/* ------------------------------------------------------------------------- */
|
|
/*
|
|
* The add_buffer series of functions are used to build one large memory chunk
|
|
* from repeated function invokes. Used so that the entire HTTP request can
|
|
* be sent in one go.
|
|
*/
|
|
|
|
struct send_buffer {
|
|
char *buffer;
|
|
size_t size_max;
|
|
size_t size_used;
|
|
};
|
|
typedef struct send_buffer send_buffer;
|
|
|
|
static CURLcode
|
|
add_buffer(send_buffer *in, const void *inptr, size_t size);
|
|
|
|
/*
|
|
* add_buffer_init() returns a fine buffer struct
|
|
*/
|
|
static
|
|
send_buffer *add_buffer_init(void)
|
|
{
|
|
send_buffer *blonk;
|
|
blonk=(send_buffer *)malloc(sizeof(send_buffer));
|
|
if(blonk) {
|
|
memset(blonk, 0, sizeof(send_buffer));
|
|
return blonk;
|
|
}
|
|
return NULL; /* failed, go home */
|
|
}
|
|
|
|
/*
|
|
* add_buffer_send() sends a buffer and frees all associated memory.
|
|
*/
|
|
static
|
|
CURLcode add_buffer_send(send_buffer *in,
|
|
int sockfd,
|
|
struct connectdata *conn,
|
|
long *bytes_written) /* add the number of sent
|
|
bytes to this counter */
|
|
{
|
|
ssize_t amount;
|
|
CURLcode res;
|
|
char *ptr;
|
|
int size;
|
|
struct HTTP *http = conn->proto.http;
|
|
|
|
/* The looping below is required since we use non-blocking sockets, but due
|
|
to the circumstances we will just loop and try again and again etc */
|
|
|
|
ptr = in->buffer;
|
|
size = in->size_used;
|
|
|
|
res = Curl_write(conn, sockfd, ptr, size, &amount);
|
|
|
|
if(CURLE_OK == res) {
|
|
|
|
if(conn->data->set.verbose)
|
|
/* this data _may_ contain binary stuff */
|
|
Curl_debug(conn->data, CURLINFO_HEADER_OUT, ptr, amount);
|
|
|
|
*bytes_written += amount;
|
|
|
|
if(amount != size) {
|
|
/* The whole request could not be sent in one system call. We must queue
|
|
it up and send it later when we get the chance. We must not loop here
|
|
and wait until it might work again. */
|
|
|
|
size -= amount;
|
|
ptr += amount;
|
|
|
|
/* backup the currently set pointers */
|
|
http->backup.fread = conn->fread;
|
|
http->backup.fread_in = conn->fread_in;
|
|
http->backup.postdata = http->postdata;
|
|
http->backup.postsize = http->postsize;
|
|
|
|
/* set the new pointers for the request-sending */
|
|
conn->fread = (curl_read_callback)readmoredata;
|
|
conn->fread_in = (void *)conn;
|
|
http->postdata = ptr;
|
|
http->postsize = size;
|
|
|
|
http->send_buffer = in;
|
|
http->sending = HTTPSEND_REQUEST;
|
|
|
|
return CURLE_OK;
|
|
}
|
|
http->sending = HTTPSEND_BODY;
|
|
/* the full buffer was sent, clean up and return */
|
|
}
|
|
if(in->buffer)
|
|
free(in->buffer);
|
|
free(in);
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
/*
|
|
* add_bufferf() builds a buffer from the formatted input
|
|
*/
|
|
static
|
|
CURLcode add_bufferf(send_buffer *in, const char *fmt, ...)
|
|
{
|
|
CURLcode result = CURLE_OUT_OF_MEMORY;
|
|
char *s;
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
s = vaprintf(fmt, ap); /* this allocs a new string to append */
|
|
va_end(ap);
|
|
|
|
if(s) {
|
|
result = add_buffer(in, s, strlen(s));
|
|
free(s);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
/*
|
|
* add_buffer() appends a memory chunk to the existing one
|
|
*/
|
|
static
|
|
CURLcode add_buffer(send_buffer *in, const void *inptr, size_t size)
|
|
{
|
|
char *new_rb;
|
|
int new_size;
|
|
|
|
if(!in->buffer ||
|
|
((in->size_used + size) > (in->size_max - 1))) {
|
|
new_size = (in->size_used+size)*2;
|
|
if(in->buffer)
|
|
/* we have a buffer, enlarge the existing one */
|
|
new_rb = (char *)realloc(in->buffer, new_size);
|
|
else
|
|
/* create a new buffer */
|
|
new_rb = (char *)malloc(new_size);
|
|
|
|
if(!new_rb)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
|
|
in->buffer = new_rb;
|
|
in->size_max = new_size;
|
|
}
|
|
memcpy(&in->buffer[in->size_used], inptr, size);
|
|
|
|
in->size_used += size;
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/* end of the add_buffer functions */
|
|
/* ------------------------------------------------------------------------- */
|
|
|
|
/*
|
|
* Curl_compareheader()
|
|
*
|
|
* Returns TRUE if 'headerline' contains the 'header' with given 'content'.
|
|
* Pass headers WITH the colon.
|
|
*/
|
|
bool
|
|
Curl_compareheader(char *headerline, /* line to check */
|
|
const char *header, /* header keyword _with_ colon */
|
|
const char *content) /* content string to find */
|
|
{
|
|
/* RFC2616, section 4.2 says: "Each header field consists of a name followed
|
|
* by a colon (":") and the field value. Field names are case-insensitive.
|
|
* The field value MAY be preceded by any amount of LWS, though a single SP
|
|
* is preferred." */
|
|
|
|
size_t hlen = strlen(header);
|
|
size_t clen;
|
|
size_t len;
|
|
char *start;
|
|
char *end;
|
|
|
|
if(!strnequal(headerline, header, hlen))
|
|
return FALSE; /* doesn't start with header */
|
|
|
|
/* pass the header */
|
|
start = &headerline[hlen];
|
|
|
|
/* pass all white spaces */
|
|
while(*start && isspace((int)*start))
|
|
start++;
|
|
|
|
/* find the end of the header line */
|
|
end = strchr(start, '\r'); /* lines end with CRLF */
|
|
if(!end) {
|
|
/* in case there's a non-standard compliant line here */
|
|
end = strchr(start, '\n');
|
|
|
|
if(!end)
|
|
/* hm, there's no line ending here, use the zero byte! */
|
|
end = strchr(start, '\0');
|
|
}
|
|
|
|
len = end-start; /* length of the content part of the input line */
|
|
clen = strlen(content); /* length of the word to find */
|
|
|
|
/* find the content string in the rest of the line */
|
|
for(;len>=clen;len--, start++) {
|
|
if(strnequal(start, content, clen))
|
|
return TRUE; /* match! */
|
|
}
|
|
|
|
return FALSE; /* no match */
|
|
}
|
|
|
|
/*
|
|
* This function checks the linked list of custom HTTP headers for a particular
|
|
* header (prefix).
|
|
*/
|
|
static char *checkheaders(struct SessionHandle *data, const char *thisheader)
|
|
{
|
|
struct curl_slist *head;
|
|
size_t thislen = strlen(thisheader);
|
|
|
|
for(head = data->set.headers; head; head=head->next) {
|
|
if(strnequal(head->data, thisheader, thislen))
|
|
return head->data;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* ConnectHTTPProxyTunnel() requires that we're connected to a HTTP proxy. This
|
|
* function will issue the necessary commands to get a seamless tunnel through
|
|
* this proxy. After that, the socket can be used just as a normal socket.
|
|
*/
|
|
|
|
CURLcode Curl_ConnectHTTPProxyTunnel(struct connectdata *conn,
|
|
int tunnelsocket,
|
|
char *hostname, int remote_port)
|
|
{
|
|
int httperror=0;
|
|
int subversion=0;
|
|
struct SessionHandle *data=conn->data;
|
|
CURLcode result;
|
|
int res;
|
|
|
|
int nread; /* total size read */
|
|
int perline; /* count bytes per line */
|
|
bool keepon=TRUE;
|
|
ssize_t gotbytes;
|
|
char *ptr;
|
|
int timeout = 3600; /* default timeout in seconds */
|
|
struct timeval interval;
|
|
fd_set rkeepfd;
|
|
fd_set readfd;
|
|
char *line_start;
|
|
|
|
#define SELECT_OK 0
|
|
#define SELECT_ERROR 1
|
|
#define SELECT_TIMEOUT 2
|
|
int error = SELECT_OK;
|
|
|
|
infof(data, "Establish HTTP proxy tunnel to %s:%d\n", hostname, remote_port);
|
|
|
|
/* OK, now send the connect request to the proxy */
|
|
result =
|
|
Curl_sendf(tunnelsocket, conn,
|
|
"CONNECT %s:%d HTTP/1.0\015\012"
|
|
"%s"
|
|
"%s"
|
|
"\r\n",
|
|
hostname, remote_port,
|
|
(conn->bits.proxy_user_passwd)?conn->allocptr.proxyuserpwd:"",
|
|
(data->set.useragent?conn->allocptr.uagent:"")
|
|
);
|
|
if(result) {
|
|
failf(data, "Failed sending CONNECT to proxy");
|
|
return result;
|
|
}
|
|
|
|
/* Now, read the full reply we get from the proxy */
|
|
|
|
|
|
if(data->set.timeout) {
|
|
/* if timeout is requested, find out how much remaining time we have */
|
|
timeout = data->set.timeout - /* timeout time */
|
|
Curl_tvdiff(Curl_tvnow(), conn->now)/1000; /* spent time */
|
|
if(timeout <=0 ) {
|
|
failf(data, "Transfer aborted due to timeout");
|
|
return -SELECT_TIMEOUT; /* already too little time */
|
|
}
|
|
}
|
|
|
|
FD_ZERO (&readfd); /* clear it */
|
|
FD_SET (tunnelsocket, &readfd); /* read socket */
|
|
|
|
/* get this in a backup variable to be able to restore it on each lap in the
|
|
select() loop */
|
|
rkeepfd = readfd;
|
|
|
|
ptr=data->state.buffer;
|
|
line_start = ptr;
|
|
|
|
nread=0;
|
|
perline=0;
|
|
keepon=TRUE;
|
|
|
|
while((nread<BUFSIZE) && (keepon && !error)) {
|
|
readfd = rkeepfd; /* set every lap */
|
|
interval.tv_sec = timeout;
|
|
interval.tv_usec = 0;
|
|
|
|
switch (select (tunnelsocket+1, &readfd, NULL, NULL, &interval)) {
|
|
case -1: /* select() error, stop reading */
|
|
error = SELECT_ERROR;
|
|
failf(data, "Transfer aborted due to select() error");
|
|
break;
|
|
case 0: /* timeout */
|
|
error = SELECT_TIMEOUT;
|
|
failf(data, "Transfer aborted due to timeout");
|
|
break;
|
|
default:
|
|
/*
|
|
* This code previously didn't use the kerberos sec_read() code
|
|
* to read, but when we use Curl_read() it may do so. Do confirm
|
|
* that this is still ok and then remove this comment!
|
|
*/
|
|
res= Curl_read(conn, tunnelsocket, ptr, BUFSIZE-nread,
|
|
&gotbytes);
|
|
if(res< 0)
|
|
/* EWOULDBLOCK */
|
|
continue; /* go loop yourself */
|
|
else if(res)
|
|
keepon = FALSE;
|
|
else if(gotbytes <= 0) {
|
|
keepon = FALSE;
|
|
error = SELECT_ERROR;
|
|
failf(data, "Connection aborted");
|
|
}
|
|
else {
|
|
/* we got a whole chunk of data, which can be anything from one
|
|
* byte to a set of lines and possibly just a piece of the last
|
|
* line */
|
|
int i;
|
|
|
|
nread += gotbytes;
|
|
for(i = 0; i < gotbytes; ptr++, i++) {
|
|
perline++; /* amount of bytes in this line so far */
|
|
if(*ptr=='\n') {
|
|
/* a newline is CRLF in ftp-talk, so the CR is ignored as
|
|
the line isn't really terminated until the LF comes */
|
|
|
|
if('\r' == line_start[0]) {
|
|
/* end of headers */
|
|
keepon=FALSE;
|
|
break; /* breaks out of loop, not switch */
|
|
}
|
|
|
|
/* output debug output if that is requested */
|
|
if(data->set.verbose)
|
|
Curl_debug(data, CURLINFO_HEADER_IN, line_start, perline);
|
|
|
|
if(2 == sscanf(line_start, "HTTP/1.%d %d",
|
|
&subversion,
|
|
&httperror)) {
|
|
;
|
|
}
|
|
|
|
perline=0; /* line starts over here */
|
|
line_start = ptr+1;
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
} /* switch */
|
|
} /* while there's buffer left and loop is requested */
|
|
|
|
if(error)
|
|
return CURLE_RECV_ERROR;
|
|
|
|
if(200 != httperror) {
|
|
if(407 == httperror)
|
|
/* Added Nov 6 1998 */
|
|
failf(data, "Proxy requires authorization!");
|
|
else
|
|
failf(data, "Received error code %d from proxy", httperror);
|
|
return CURLE_RECV_ERROR;
|
|
}
|
|
|
|
infof (data, "Proxy replied to CONNECT request\n");
|
|
return CURLE_OK;
|
|
}
|
|
|
|
/*
|
|
* HTTP stuff to do at connect-time.
|
|
*/
|
|
CURLcode Curl_http_connect(struct connectdata *conn)
|
|
{
|
|
struct SessionHandle *data;
|
|
CURLcode result;
|
|
|
|
data=conn->data;
|
|
|
|
/* If we are not using a proxy and we want a secure connection,
|
|
* perform SSL initialization & connection now.
|
|
* If using a proxy with https, then we must tell the proxy to CONNECT
|
|
* us to the host we want to talk to. Only after the connect
|
|
* has occured, can we start talking SSL
|
|
*/
|
|
|
|
if(conn->bits.httpproxy &&
|
|
((conn->protocol & PROT_HTTPS) || data->set.tunnel_thru_httpproxy)) {
|
|
|
|
/* either HTTPS over proxy, OR explicitly asked for a tunnel */
|
|
result = Curl_ConnectHTTPProxyTunnel(conn, conn->firstsocket,
|
|
conn->hostname, conn->remote_port);
|
|
if(CURLE_OK != result)
|
|
return result;
|
|
}
|
|
|
|
if(conn->protocol & PROT_HTTPS) {
|
|
/* now, perform the SSL initialization for this socket */
|
|
result = Curl_SSLConnect(conn);
|
|
if(result)
|
|
return result;
|
|
}
|
|
|
|
if(conn->bits.user_passwd && !data->state.this_is_a_follow) {
|
|
/* Authorization: is requested, this is not a followed location, get the
|
|
original host name */
|
|
if (data->state.auth_host)
|
|
/* Free to avoid leaking memory on multiple requests*/
|
|
free(data->state.auth_host);
|
|
|
|
data->state.auth_host = strdup(conn->hostname);
|
|
}
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
CURLcode Curl_http_done(struct connectdata *conn)
|
|
{
|
|
struct SessionHandle *data;
|
|
struct HTTP *http;
|
|
|
|
data=conn->data;
|
|
http=conn->proto.http;
|
|
|
|
/* set the proper values (possibly modified on POST) */
|
|
conn->fread = data->set.fread; /* restore */
|
|
conn->fread_in = data->set.in; /* restore */
|
|
|
|
if(http->send_buffer) {
|
|
send_buffer *buff = http->send_buffer;
|
|
|
|
free(buff->buffer);
|
|
free(buff);
|
|
}
|
|
|
|
if(HTTPREQ_POST_FORM == data->set.httpreq) {
|
|
conn->bytecount = http->readbytecount + http->writebytecount;
|
|
|
|
Curl_formclean(http->sendit); /* Now free that whole lot */
|
|
}
|
|
else if(HTTPREQ_PUT == data->set.httpreq)
|
|
conn->bytecount = http->readbytecount + http->writebytecount;
|
|
|
|
if(0 == (http->readbytecount + conn->headerbytecount)) {
|
|
/* nothing was read from the HTTP server, this can't be right
|
|
so we return an error here */
|
|
failf(data, "Empty reply from server");
|
|
return CURLE_GOT_NOTHING;
|
|
}
|
|
|
|
return CURLE_OK;
|
|
}
|
|
|
|
CURLcode Curl_http(struct connectdata *conn)
|
|
{
|
|
struct SessionHandle *data=conn->data;
|
|
char *buf = data->state.buffer; /* this is a short cut to the buffer */
|
|
CURLcode result=CURLE_OK;
|
|
struct HTTP *http;
|
|
struct Cookie *co=NULL; /* no cookies from start */
|
|
char *ppath = conn->ppath; /* three previous function arguments */
|
|
char *host = conn->name;
|
|
const char *te = ""; /* tranfer-encoding */
|
|
char *ptr;
|
|
|
|
if(!conn->proto.http) {
|
|
/* Only allocate this struct if we don't already have it! */
|
|
|
|
http = (struct HTTP *)malloc(sizeof(struct HTTP));
|
|
if(!http)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
memset(http, 0, sizeof(struct HTTP));
|
|
conn->proto.http = http;
|
|
}
|
|
else
|
|
http = conn->proto.http;
|
|
|
|
/* We default to persistant connections */
|
|
conn->bits.close = FALSE;
|
|
|
|
if ( (conn->protocol&(PROT_HTTP|PROT_FTP)) &&
|
|
data->set.upload) {
|
|
data->set.httpreq = HTTPREQ_PUT;
|
|
}
|
|
|
|
/* The User-Agent string has been built in url.c already, because it might
|
|
have been used in the proxy connect, but if we have got a header with
|
|
the user-agent string specified, we erase the previously made string
|
|
here. */
|
|
if(checkheaders(data, "User-Agent:") && conn->allocptr.uagent) {
|
|
free(conn->allocptr.uagent);
|
|
conn->allocptr.uagent=NULL;
|
|
}
|
|
|
|
if((conn->bits.user_passwd) && !checkheaders(data, "Authorization:")) {
|
|
char *authorization;
|
|
|
|
/* To prevent the user+password to get sent to other than the original
|
|
host due to a location-follow, we do some weirdo checks here */
|
|
if(!data->state.this_is_a_follow ||
|
|
!data->state.auth_host ||
|
|
curl_strequal(data->state.auth_host, conn->hostname) ||
|
|
data->set.http_disable_hostname_check_before_authentication) {
|
|
sprintf(data->state.buffer, "%s:%s",
|
|
data->state.user, data->state.passwd);
|
|
if(Curl_base64_encode(data->state.buffer, strlen(data->state.buffer),
|
|
&authorization) >= 0) {
|
|
if(conn->allocptr.userpwd)
|
|
free(conn->allocptr.userpwd);
|
|
conn->allocptr.userpwd = aprintf( "Authorization: Basic %s\015\012",
|
|
authorization);
|
|
free(authorization);
|
|
}
|
|
}
|
|
}
|
|
if((data->change.referer) && !checkheaders(data, "Referer:")) {
|
|
if(conn->allocptr.ref)
|
|
free(conn->allocptr.ref);
|
|
conn->allocptr.ref = aprintf("Referer: %s\015\012", data->change.referer);
|
|
}
|
|
if(data->set.cookie && !checkheaders(data, "Cookie:")) {
|
|
if(conn->allocptr.cookie)
|
|
free(conn->allocptr.cookie);
|
|
conn->allocptr.cookie = aprintf("Cookie: %s\015\012", data->set.cookie);
|
|
}
|
|
|
|
if(!conn->bits.upload_chunky && (data->set.httpreq != HTTPREQ_GET)) {
|
|
/* not a chunky transfer but data is to be sent */
|
|
char *ptr = checkheaders(data, "Transfer-Encoding:");
|
|
if(ptr) {
|
|
/* Some kind of TE is requested, check if 'chunked' is chosen */
|
|
if(Curl_compareheader(ptr, "Transfer-Encoding:", "chunked"))
|
|
/* we have been told explicitly to upload chunky so deal with it! */
|
|
conn->bits.upload_chunky = TRUE;
|
|
}
|
|
}
|
|
|
|
if(conn->bits.upload_chunky) {
|
|
/* RFC2616 section 4.4:
|
|
Messages MUST NOT include both a Content-Length header field and a
|
|
non-identity transfer-coding. If the message does include a non-
|
|
identity transfer-coding, the Content-Length MUST be ignored. */
|
|
|
|
if(!checkheaders(data, "Transfer-Encoding:")) {
|
|
te = "Transfer-Encoding: chunked\r\n";
|
|
}
|
|
else {
|
|
/* The "Transfer-Encoding:" header was already added. */
|
|
te = "";
|
|
}
|
|
}
|
|
|
|
ptr = checkheaders(data, "Host:");
|
|
if(ptr) {
|
|
/* If we have a given custom Host: header, we extract the host name
|
|
in order to possibly use it for cookie reasons later on. */
|
|
char *start = ptr+strlen("Host:");
|
|
char *ptr;
|
|
while(*start && isspace((int)*start ))
|
|
start++;
|
|
ptr = start; /* start host-scanning here */
|
|
|
|
/* scan through the string to find the end (space or colon) */
|
|
while(*ptr && !isspace((int)*ptr) && !(':'==*ptr))
|
|
ptr++;
|
|
|
|
if(ptr != start) {
|
|
int len=ptr-start;
|
|
conn->allocptr.cookiehost = malloc(len+1);
|
|
if(!conn->allocptr.cookiehost)
|
|
return CURLE_OUT_OF_MEMORY;
|
|
memcpy(conn->allocptr.cookiehost, start, len);
|
|
conn->allocptr.cookiehost[len]=0;
|
|
}
|
|
}
|
|
else {
|
|
/* if ptr_host is already set, it is almost OK since we only re-use
|
|
connections to the very same host and port, but when we use a HTTP
|
|
proxy we have a persistant connect and yet we must change the Host:
|
|
header! */
|
|
|
|
if(conn->allocptr.host)
|
|
free(conn->allocptr.host);
|
|
|
|
/* When building Host: headers, we must put the host name within
|
|
[brackets] if the host name is a plain IPv6-address. RFC2732-style. */
|
|
|
|
if(((conn->protocol&PROT_HTTPS) && (conn->remote_port == PORT_HTTPS)) ||
|
|
(!(conn->protocol&PROT_HTTPS) && (conn->remote_port == PORT_HTTP)) )
|
|
/* If (HTTPS on port 443) OR (non-HTTPS on port 80) then don't include
|
|
the port number in the host string */
|
|
conn->allocptr.host = aprintf("Host: %s%s%s\r\n",
|
|
conn->bits.ipv6_ip?"[":"",
|
|
host,
|
|
conn->bits.ipv6_ip?"]":"");
|
|
else
|
|
conn->allocptr.host = aprintf("Host: %s%s%s:%d\r\n",
|
|
conn->bits.ipv6_ip?"[":"",
|
|
host,
|
|
conn->bits.ipv6_ip?"]":"",
|
|
conn->remote_port);
|
|
}
|
|
|
|
if(data->cookies) {
|
|
co = Curl_cookie_getlist(data->cookies,
|
|
conn->allocptr.cookiehost?
|
|
conn->allocptr.cookiehost:host, ppath,
|
|
(bool)(conn->protocol&PROT_HTTPS?TRUE:FALSE));
|
|
}
|
|
|
|
if (data->change.proxy && *data->change.proxy &&
|
|
!data->set.tunnel_thru_httpproxy &&
|
|
!(conn->protocol&PROT_HTTPS)) {
|
|
/* The path sent to the proxy is in fact the entire URL */
|
|
ppath = data->change.url;
|
|
}
|
|
if(HTTPREQ_POST_FORM == data->set.httpreq) {
|
|
/* we must build the whole darned post sequence first, so that we have
|
|
a size of the whole shebang before we start to send it */
|
|
result = Curl_getFormData(&http->sendit, data->set.httppost,
|
|
&http->postsize);
|
|
if(CURLE_OK != result) {
|
|
/* Curl_getFormData() doesn't use failf() */
|
|
failf(data, "failed creating formpost data");
|
|
return result;
|
|
}
|
|
}
|
|
|
|
|
|
if(!checkheaders(data, "Pragma:"))
|
|
http->p_pragma = "Pragma: no-cache\r\n";
|
|
|
|
if(!checkheaders(data, "Accept:"))
|
|
http->p_accept = "Accept: image/gif, image/x-xbitmap, image/jpeg, image/pjpeg, */*\r\n";
|
|
|
|
if(( (HTTPREQ_POST == data->set.httpreq) ||
|
|
(HTTPREQ_POST_FORM == data->set.httpreq) ||
|
|
(HTTPREQ_PUT == data->set.httpreq) ) &&
|
|
conn->resume_from) {
|
|
/**********************************************************************
|
|
* Resuming upload in HTTP means that we PUT or POST and that we have
|
|
* got a resume_from value set. The resume value has already created
|
|
* a Range: header that will be passed along. We need to "fast forward"
|
|
* the file the given number of bytes and decrease the assume upload
|
|
* file size before we continue this venture in the dark lands of HTTP.
|
|
*********************************************************************/
|
|
|
|
if(conn->resume_from < 0 ) {
|
|
/*
|
|
* This is meant to get the size of the present remote-file by itself.
|
|
* We don't support this now. Bail out!
|
|
*/
|
|
conn->resume_from = 0;
|
|
}
|
|
|
|
if(conn->resume_from) {
|
|
/* do we still game? */
|
|
int passed=0;
|
|
|
|
/* Now, let's read off the proper amount of bytes from the
|
|
input. If we knew it was a proper file we could've just
|
|
fseek()ed but we only have a stream here */
|
|
do {
|
|
int readthisamountnow = (conn->resume_from - passed);
|
|
int actuallyread;
|
|
|
|
if(readthisamountnow > BUFSIZE)
|
|
readthisamountnow = BUFSIZE;
|
|
|
|
actuallyread =
|
|
data->set.fread(data->state.buffer, 1, readthisamountnow,
|
|
data->set.in);
|
|
|
|
passed += actuallyread;
|
|
if(actuallyread != readthisamountnow) {
|
|
failf(data, "Could only read %d bytes from the input",
|
|
passed);
|
|
return CURLE_READ_ERROR;
|
|
}
|
|
} while(passed != conn->resume_from); /* loop until done */
|
|
|
|
/* now, decrease the size of the read */
|
|
if(data->set.infilesize>0) {
|
|
data->set.infilesize -= conn->resume_from;
|
|
|
|
if(data->set.infilesize <= 0) {
|
|
failf(data, "File already completely uploaded");
|
|
return CURLE_PARTIAL_FILE;
|
|
}
|
|
}
|
|
/* we've passed, proceed as normal */
|
|
}
|
|
}
|
|
if(conn->bits.use_range) {
|
|
/*
|
|
* A range is selected. We use different headers whether we're downloading
|
|
* or uploading and we always let customized headers override our internal
|
|
* ones if any such are specified.
|
|
*/
|
|
if((data->set.httpreq == HTTPREQ_GET) &&
|
|
!checkheaders(data, "Range:")) {
|
|
/* if a line like this was already allocated, free the previous one */
|
|
if(conn->allocptr.rangeline)
|
|
free(conn->allocptr.rangeline);
|
|
conn->allocptr.rangeline = aprintf("Range: bytes=%s\r\n", conn->range);
|
|
}
|
|
else if((data->set.httpreq != HTTPREQ_GET) &&
|
|
!checkheaders(data, "Content-Range:")) {
|
|
|
|
if(conn->resume_from) {
|
|
/* This is because "resume" was selected */
|
|
long total_expected_size= conn->resume_from + data->set.infilesize;
|
|
conn->allocptr.rangeline = aprintf("Content-Range: bytes %s%ld/%ld\r\n",
|
|
conn->range, total_expected_size-1,
|
|
total_expected_size);
|
|
}
|
|
else {
|
|
/* Range was selected and then we just pass the incoming range and
|
|
append total size */
|
|
conn->allocptr.rangeline = aprintf("Content-Range: bytes %s/%d\r\n",
|
|
conn->range, data->set.infilesize);
|
|
}
|
|
}
|
|
}
|
|
|
|
do {
|
|
/* Use 1.1 unless the use specificly asked for 1.0 */
|
|
const char *httpstring=
|
|
data->set.httpversion==CURL_HTTP_VERSION_1_0?"1.0":"1.1";
|
|
|
|
send_buffer *req_buffer;
|
|
struct curl_slist *headers=data->set.headers;
|
|
|
|
/* initialize a dynamic send-buffer */
|
|
req_buffer = add_buffer_init();
|
|
|
|
/* add the main request stuff */
|
|
add_bufferf(req_buffer,
|
|
"%s " /* GET/HEAD/POST/PUT */
|
|
"%s HTTP/%s\r\n" /* path */
|
|
"%s" /* proxyuserpwd */
|
|
"%s" /* userpwd */
|
|
"%s" /* range */
|
|
"%s" /* user agent */
|
|
"%s" /* cookie */
|
|
"%s" /* host */
|
|
"%s" /* pragma */
|
|
"%s" /* accept */
|
|
"%s" /* accept-encoding */
|
|
"%s" /* referer */
|
|
"%s",/* transfer-encoding */
|
|
|
|
data->set.customrequest?data->set.customrequest:
|
|
(data->set.no_body?"HEAD":
|
|
((HTTPREQ_POST == data->set.httpreq) ||
|
|
(HTTPREQ_POST_FORM == data->set.httpreq))?"POST":
|
|
(HTTPREQ_PUT == data->set.httpreq)?"PUT":"GET"),
|
|
ppath, httpstring,
|
|
(conn->bits.proxy_user_passwd &&
|
|
conn->allocptr.proxyuserpwd)?conn->allocptr.proxyuserpwd:"",
|
|
(conn->bits.user_passwd && conn->allocptr.userpwd)?
|
|
conn->allocptr.userpwd:"",
|
|
(conn->bits.use_range && conn->allocptr.rangeline)?
|
|
conn->allocptr.rangeline:"",
|
|
(data->set.useragent && *data->set.useragent && conn->allocptr.uagent)?
|
|
conn->allocptr.uagent:"",
|
|
(conn->allocptr.cookie?conn->allocptr.cookie:""), /* Cookie: <data> */
|
|
(conn->allocptr.host?conn->allocptr.host:""), /* Host: host */
|
|
http->p_pragma?http->p_pragma:"",
|
|
http->p_accept?http->p_accept:"",
|
|
(data->set.encoding && *data->set.encoding && conn->allocptr.accept_encoding)?
|
|
conn->allocptr.accept_encoding:"", /* 08/28/02 jhrg */
|
|
(data->change.referer && conn->allocptr.ref)?conn->allocptr.ref:"" /* Referer: <data> <CRLF> */,
|
|
te
|
|
);
|
|
|
|
if(co) {
|
|
int count=0;
|
|
struct Cookie *store=co;
|
|
/* now loop through all cookies that matched */
|
|
while(co) {
|
|
if(co->value && strlen(co->value)) {
|
|
if(0 == count) {
|
|
add_bufferf(req_buffer, "Cookie: ");
|
|
}
|
|
add_bufferf(req_buffer,
|
|
"%s%s=%s", count?"; ":"", co->name, co->value);
|
|
count++;
|
|
}
|
|
co = co->next; /* next cookie please */
|
|
}
|
|
if(count) {
|
|
add_buffer(req_buffer, "\r\n", 2);
|
|
}
|
|
Curl_cookie_freelist(store); /* free the cookie list */
|
|
co=NULL;
|
|
}
|
|
|
|
if(data->set.timecondition) {
|
|
struct tm *thistime;
|
|
|
|
/* Phil Karn (Fri, 13 Apr 2001) pointed out that the If-Modified-Since
|
|
* header family should have their times set in GMT as RFC2616 defines:
|
|
* "All HTTP date/time stamps MUST be represented in Greenwich Mean Time
|
|
* (GMT), without exception. For the purposes of HTTP, GMT is exactly
|
|
* equal to UTC (Coordinated Universal Time)." (see page 20 of RFC2616).
|
|
*/
|
|
|
|
#ifdef HAVE_GMTIME_R
|
|
/* thread-safe version */
|
|
struct tm keeptime;
|
|
thistime = (struct tm *)gmtime_r(&data->set.timevalue, &keeptime);
|
|
#else
|
|
thistime = gmtime(&data->set.timevalue);
|
|
#endif
|
|
if(NULL == thistime) {
|
|
failf(data, "localtime() failed!");
|
|
return CURLE_OUT_OF_MEMORY;
|
|
}
|
|
|
|
#ifdef HAVE_STRFTIME
|
|
/* format: "Tue, 15 Nov 1994 12:45:26 GMT" */
|
|
strftime(buf, BUFSIZE-1, "%a, %d %b %Y %H:%M:%S GMT", thistime);
|
|
#else
|
|
/* TODO: Right, we *could* write a replacement here */
|
|
strcpy(buf, "no strftime() support");
|
|
#endif
|
|
switch(data->set.timecondition) {
|
|
case CURL_TIMECOND_IFMODSINCE:
|
|
default:
|
|
add_bufferf(req_buffer,
|
|
"If-Modified-Since: %s\r\n", buf);
|
|
break;
|
|
case CURL_TIMECOND_IFUNMODSINCE:
|
|
add_bufferf(req_buffer,
|
|
"If-Unmodified-Since: %s\r\n", buf);
|
|
break;
|
|
case CURL_TIMECOND_LASTMOD:
|
|
add_bufferf(req_buffer,
|
|
"Last-Modified: %s\r\n", buf);
|
|
break;
|
|
}
|
|
}
|
|
|
|
while(headers) {
|
|
char *ptr = strchr(headers->data, ':');
|
|
if(ptr) {
|
|
/* we require a colon for this to be a true header */
|
|
|
|
ptr++; /* pass the colon */
|
|
while(*ptr && isspace((int)*ptr))
|
|
ptr++;
|
|
|
|
if(*ptr) {
|
|
/* only send this if the contents was non-blank */
|
|
|
|
add_bufferf(req_buffer, "%s\r\n", headers->data);
|
|
}
|
|
}
|
|
headers = headers->next;
|
|
}
|
|
|
|
http->postdata = NULL; /* nothing to post at this point */
|
|
Curl_pgrsSetUploadSize(data, 0); /* upload size is 0 atm */
|
|
|
|
switch(data->set.httpreq) {
|
|
|
|
case HTTPREQ_POST_FORM:
|
|
if(Curl_FormInit(&http->form, http->sendit)) {
|
|
failf(data, "Internal HTTP POST error!");
|
|
return CURLE_HTTP_POST_ERROR;
|
|
}
|
|
|
|
/* set the read function to read from the generated form data */
|
|
conn->fread = (curl_read_callback)Curl_FormReader;
|
|
conn->fread_in = &http->form;
|
|
|
|
http->sending = HTTPSEND_BODY;
|
|
|
|
if(!conn->bits.upload_chunky)
|
|
/* only add Content-Length if not uploading chunked */
|
|
add_bufferf(req_buffer,
|
|
"Content-Length: %d\r\n", http->postsize);
|
|
|
|
if(!checkheaders(data, "Expect:")) {
|
|
/* if not disabled explicitly we add a Expect: 100-continue
|
|
to the headers which actually speeds up post operations (as
|
|
there is one packet coming back from the web server) */
|
|
add_bufferf(req_buffer,
|
|
"Expect: 100-continue\r\n");
|
|
data->set.expect100header = TRUE;
|
|
}
|
|
|
|
if(!checkheaders(data, "Content-Type:")) {
|
|
/* Get Content-Type: line from Curl_FormReadOneLine, which happens
|
|
to always be the first line. We can know this for sure since
|
|
we always build the formpost linked list the same way!
|
|
|
|
The Content-Type header line also contains the MIME boundary
|
|
string etc why disabling this header is likely to not make things
|
|
work, but we support it anyway.
|
|
*/
|
|
char contentType[256];
|
|
int linelength=0;
|
|
linelength = Curl_FormReadOneLine (contentType,
|
|
sizeof(contentType),
|
|
1,
|
|
(FILE *)&http->form);
|
|
if(linelength == -1) {
|
|
failf(data, "Could not get Content-Type header line!");
|
|
return CURLE_HTTP_POST_ERROR;
|
|
}
|
|
add_buffer(req_buffer, contentType, linelength);
|
|
}
|
|
|
|
/* make the request end in a true CRLF */
|
|
add_buffer(req_buffer, "\r\n", 2);
|
|
|
|
/* set upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, http->postsize);
|
|
|
|
/* fire away the whole request to the server */
|
|
result = add_buffer_send(req_buffer, conn->firstsocket, conn,
|
|
&data->info.request_size);
|
|
if(result)
|
|
failf(data, "Failed sending POST request");
|
|
else
|
|
/* setup variables for the upcoming transfer */
|
|
result = Curl_Transfer(conn, conn->firstsocket, -1, TRUE,
|
|
&http->readbytecount,
|
|
conn->firstsocket,
|
|
&http->writebytecount);
|
|
if(result) {
|
|
Curl_formclean(http->sendit); /* free that whole lot */
|
|
return result;
|
|
}
|
|
break;
|
|
|
|
case HTTPREQ_PUT: /* Let's PUT the data to the server! */
|
|
|
|
if((data->set.infilesize>0) && !conn->bits.upload_chunky)
|
|
/* only add Content-Length if not uploading chunked */
|
|
add_bufferf(req_buffer,
|
|
"Content-Length: %d\r\n", /* file size */
|
|
data->set.infilesize );
|
|
|
|
add_bufferf(req_buffer, "\r\n");
|
|
|
|
/* set the upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, data->set.infilesize);
|
|
|
|
/* this sends the buffer and frees all the buffer resources */
|
|
result = add_buffer_send(req_buffer, conn->firstsocket, conn,
|
|
&data->info.request_size);
|
|
if(result)
|
|
failf(data, "Failed sending POST request");
|
|
else
|
|
/* prepare for transfer */
|
|
result = Curl_Transfer(conn, conn->firstsocket, -1, TRUE,
|
|
&http->readbytecount,
|
|
conn->firstsocket,
|
|
&http->writebytecount);
|
|
if(result)
|
|
return result;
|
|
break;
|
|
|
|
case HTTPREQ_POST:
|
|
/* this is the simple POST, using x-www-form-urlencoded style */
|
|
|
|
if(!conn->bits.upload_chunky) {
|
|
/* We only set Content-Length and allow a custom Content-Length if
|
|
we don't upload data chunked, as RFC2616 forbids us to set both
|
|
kinds of headers (Transfer-Encoding: chunked and Content-Length) */
|
|
|
|
if(!checkheaders(data, "Content-Length:"))
|
|
/* we allow replacing this header, although it isn't very wise to
|
|
actually set your own */
|
|
add_bufferf(req_buffer,
|
|
"Content-Length: %d\r\n",
|
|
data->set.postfieldsize?
|
|
data->set.postfieldsize:
|
|
(data->set.postfields?strlen(data->set.postfields):0) );
|
|
}
|
|
|
|
if(!checkheaders(data, "Content-Type:"))
|
|
add_bufferf(req_buffer,
|
|
"Content-Type: application/x-www-form-urlencoded\r\n");
|
|
|
|
add_buffer(req_buffer, "\r\n", 2);
|
|
|
|
/* and here we setup the pointers to the actual data */
|
|
if(data->set.postfields) {
|
|
if(data->set.postfieldsize)
|
|
http->postsize = data->set.postfieldsize;
|
|
else
|
|
http->postsize = strlen(data->set.postfields);
|
|
http->postdata = data->set.postfields;
|
|
|
|
http->sending = HTTPSEND_BODY;
|
|
|
|
conn->fread = (curl_read_callback)readmoredata;
|
|
conn->fread_in = (void *)conn;
|
|
|
|
/* set the upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, http->postsize);
|
|
}
|
|
else
|
|
/* set the upload size to the progress meter */
|
|
Curl_pgrsSetUploadSize(data, data->set.infilesize);
|
|
|
|
/* issue the request, headers-only */
|
|
result = add_buffer_send(req_buffer, conn->firstsocket, conn,
|
|
&data->info.request_size);
|
|
|
|
if(result)
|
|
failf(data, "Failed sending HTTP POST request");
|
|
else
|
|
result =
|
|
Curl_Transfer(conn, conn->firstsocket, -1, TRUE,
|
|
&http->readbytecount,
|
|
conn->firstsocket,
|
|
&http->writebytecount);
|
|
break;
|
|
|
|
default:
|
|
add_buffer(req_buffer, "\r\n", 2);
|
|
|
|
/* issue the request */
|
|
result = add_buffer_send(req_buffer, conn->firstsocket, conn,
|
|
&data->info.request_size);
|
|
|
|
if(result)
|
|
failf(data, "Failed sending HTTP request");
|
|
else
|
|
/* HTTP GET/HEAD download: */
|
|
result = Curl_Transfer(conn, conn->firstsocket, -1, TRUE,
|
|
&http->readbytecount,
|
|
http->postdata?conn->firstsocket:-1,
|
|
http->postdata?&http->writebytecount:NULL);
|
|
}
|
|
if(result)
|
|
return result;
|
|
} while (0); /* this is just a left-over from the multiple document download
|
|
attempts */
|
|
|
|
return CURLE_OK;
|
|
}
|
|
#endif
|