Age | Commit message (Collapse) | Author | Files | Lines |
|
should never have added them in the first place. They don't really
buy anything, and they can hide bugs.
|
|
string and return the port. I cleaned up and added error handling to
the code, but it's basically "alex"'s fix.
(extract_http_url): Rewrote this function to remove all the sscanf()
calls. It's much easier to just split on the path slash (if it's
present) and then strip the user name/password and port from the host
string. Less code, handles more cases!
|
|
this addition follow:
The patch implements a simple reverse proxy (with one funky extra
feature). It has all the regular features: mapping remote servers to local
namespace (ReversePath), disabling forward proxying (ReverseOnly) and HTTP
redirect rewriting (ReverseBaseURL).
The funky feature is this: You map Google to /google/ and the Google front
page opens up fine. Type in stuff and click "Google Search" and you'll get
an error from tinyproxy. Reason for this is that Google's form submits to
"/search" which unfortunately bypasses our /google/ mapping (if they'd
submit to "search" without the slash it would have worked ok). Turn on
ReverseMagic and it starts working....
ReverseMagic "hijacks" one cookie which it sends to the client browser.
This cookie contains the current reverse proxy path mapping (in the above
case /google/) so that even if the site uses absolute links the reverse
proxy still knows where to map the request.
And yes, it works. No, I've never seen this done before - I couldn't find
_any_ working OSS reverse proxies, and the commercial ones I've seen try
to parse the page and fix all links (in the above case changing "/search"
to "/google/search"). The problem with modifying the html is that it might
not be parsable (very common) or it might be encoded so that the proxy
can't read it (mod_gzip or likes).
Hope you like that patch. One caveat - I haven't coded with C in like
three years so my code might be a bit messy.... There shouldn't be any
security problems thou, but you never know. I did all the stuff out of my
memory without reading any RFC's, but I tested everything with Moz, Konq,
IE6, Links and Lynx and they all worked fine.
|
|
so these files needed to be modified to only use the system's
installed regular expression library.
|
|
|
|
cleanly with a C++ compiler. (Tested using GCC 3.3)
|
|
types of upstream configurations correctly. Hopefully, the code is
also a little clearer in it's implementation.
|
|
"ViaProxyName" directive. The "Via" HTTP header is _required_ by the
HTTP spec, so the code has been changed to always send the header.
However, including the proxy's host name could be considered a
security threat, so the "ViaProxyName" directive is used to set the
token sent in the "Via" header. If the directive is not enabled the
proxy's host name will be used.
|
|
call used to build the URL for the upstream proxy. [Patch suggested by
David T. Pierso]
|
|
(upstream_add): Added support to allow ip addresses and networks to be
used when matching an upstream proxy directive.
[Code by Peter da Silva]
|
|
The change was recommended in the C/C++ User Journal magazine.
|
|
convention.
|
|
|
|
server configurable based on the destination host. [Code written by
Peter da Silva]
|
|
standard HTTP port (80 or 443) append the port string to the host
header; otherwise, leave the host string with only the host's domain
name.
Replaced all occurrences of constant 80 and 443 with defines HTTP_PORT
and HTTP_PORT_SSL.
|
|
is used by the transparent proxy code. [Anatole Shaw]
(process_request): Fixed up the transparent proxy code so that
filtering can be done on the whole URL. [Anatole Shaw]
(pull_client_data): Added a bug fix for Internet Explorer (IE). IE
will leave an extra CR and LF after the data in an HTTP POST. The new
code will eat the extra bytes if they're present. Thanks to Yannick
Koehler for finding the bug and offering an explanation as to why it
was happening.
Changed all calls of connptr->remote_content_length to
connptr->content_length.server
|
|
requested server. This fixes a problem when the server is not
listening on the default port, 80.
[Fix suggested by duncan@sapio.co.uk]
|
|
terminating NULL. The va_arg() function requires it to work
properly.
|
|
substitution mechanism. [Steven Young]
|
|
sure the function would compile cleanly regardless of how the
directive was set.
|
|
sure the function's symbol name is stored, even if upstream was not
compiled it. This should keep some compilers from reporting errors.
|
|
comes from the FreeBSD port through James E. Flemer.
|
|
username/password part from the host URI.
(extract_http_url), (extract_ssl_url): Use the new
strip_username_password function to remove any non-host information
from the URI.
|
|
The type was just replaced by "unsigned int" types.
|
|
since it's skipped by the caller before the URL is passed to this
function.
(process_request): Include code to handle proxy FTP requests as
well. This also lead to a bit of a cleanup in the calling conventions
of extract_http_url function. tinyproxy can handle both types of
resources by skipping the leading :// part.
|
|
controlled by the ViaHeader configure directive.
|
|
the data has been sent.
|
|
tinyproxy. There is really no need for this code, since there are
perfectly good programs out there (like rinetd) which are designed for
TCP tunnelling. tinyproxy should be a good HTTP proxy, nothing more,
and nothing less; therefore, the tunnelling code is gone.
|
|
remaining data to the server when the connections are being closed.
It was a one line fix.
|
|
tinyproxy. The additional code is in process_request(), but Petr also had to change around some of the other functions (like process_client_headers and handle_connection.) Note: Right now this code has not been tested, but it is believed to work. To enable run ./configure --enable-transparent-proxy
|
|
newline, we need to add them back on when sending the response header to the client.
|
|
tinyproxy now ignores all the headers after a "double CGI" situation has occurred.
(get_all_headers): Instead of dropping duplicate headers when the "double CGI" situation occurs, tinyproxy will now drop _all_ the headers from the "inner" HTTP response.
|
|
situation is in affect. This is needed to solve a problem with some sites (like cgi.ebay.com) where they are sending two HTTP response lines (and associated headers) with a response.
(get_all_headers): Detect if a "Double CGI" situation has occured.
|
|
was pointed out as being a problem with eBay (cgi.ebay.com)
(process_server_headers): Added code to make skip blank lines before a response line.
|
|
and URLs.
|
|
|
|
|
|
advantage of the fact that the peer's socket information is now stored
inside the connection structure.
|
|
free_request_struct(), so make sure it's pointing to memory.
|
|
of the buffer.h file. This also required the removal of the BUFFER_SIZE
macro, and replace it with the buffer_size() function.
|
|
concept since it blocks waiting for information from the client before any
data from the remote is relayed.
|
|
process_server_headers() functions.
Also, better reporting of errors back to the client.
|
|
recv() call. Thanks to Tom Cross for discovering this bug.
|
|
negative.
|
|
it.
|
|
The remove_connection_headers() function now handles both the Connection
header _and_ the Proxy-Connection header.
|
|
block which had not been allocated.
Also, the "proxy-connection" to the list of headers we don't pass.
|
|
are not passed along.
|
|
headers before sending the HTTP response back. This should be more
standards compliant.
|
|
|