On 15/07/2014 4:25 a.m., Alex Rousskov wrote:
> On 07/12/2014 10:45 PM, Amos Jeffries wrote:
> 
>> +bool
>> +ConnStateData::findProxyProtocolMagic()
>> +{
>> +    // http://www.haproxy.org/download/1.5/doc/proxy-protocol.txt
>> +
>> +    // detect and parse PROXY protocol version 1 header
>> +    if (in.buf.length() > Proxy10magic.length() && 
>> in.buf.startsWith(Proxy10magic)) {
>> +         return parseProxy10();
>> +
>> +        // detect and parse PROXY protocol version 2 header
>> +    } else if (in.buf.length() > Proxy20magic.length() && 
>> in.buf.startsWith(Proxy20magic)) {
>> +        return parseProxy20();
>> +
>> +        // detect and terminate other protocols
>> +    } else if (in.buf.length() >= Proxy20magic.length()) {
>> +        // input other than the PROXY header is a protocol error
>> +        return proxyProtocolError("PROXY protocol error: invalid header");
>> +    }
>> +
> 
> I know you disagree, but the above looks much clearer than the earlier
> code to me. Thank you.
> 
> The "// detect and ..." comments are pretty much not needed now because
> the code is self-documenting! The "else"s are also not needed. Your call
> on whether to remove that redundancy.
> 
> Consider adding
> 
>   // TODO: detect short non-magic prefixes earlier to avoid
>   // waiting for more data which may never come
> 
> but this is not a big deal because most non-malicious clients will not
> send valid non-PROXY requests that is 12 bytes or less, I guess.
> 

Added.

> 
>> +    // XXX: should do this in start(), but SSL/TLS operations begin before 
>> start() is called
> 
> I agree that this should be done in start(). Fortunately, the code this
> XXX refers to does not rely on job protection (AFAICT) so it is not a
> big deal. The XXX is really about the SSL code that makes the move
> difficult.
> 

Where should socket MTU discovery be configured if not at the point
before where the connection actually starts being used?

The constructor where this code block exists is well after accept(), but
also well before socket usage (other than TLS).

> 
>>>> -NAME: follow_x_forwarded_for
>>>> +NAME: proxy_forwarded_access follow_x_forwarded_for
> 
>>> The new name sounds worse than the old one. Hopefully this can be left
>>> as is or renamed to something better after the "proxy-surrogate" issue
>>> is resolved.
> 
>> Is "forwarded_access" obectionable ?
> 
> IMO, it is misleading/awkward. Follow_x_forwarded_for was pretty good.
> We can use something like follow_forwarded_client_info or
> trust_relayed_client_info, but let's wait for the primary naming issue
> to be resolved first. That resolution might help us here.
> 
> 
>>>> +bool
>>>> +ConnStateData::parseProxyProtocolMagic()
>>>
>>> This appears to parse a lot more than just the magic characters. Please
>>> rename to parseProxyProtocolHeader() or similar.
> 
>> The entire PROXY protocol is "magic" connection header. 
> 
> Not really. In this context, "magic" is, roughly, a "rare" string
> constant typically used as a prefix to detect/identify the following
> structure or message. The PROXY protocol header contains both "magical"
> and "regular" parts.
> 
> 
>> +/**
>> + * Test the connection read buffer for PROXY protocol header.
>> + * Version 1 and 2 header currently supported.
>> + */
>> +bool
>> +ConnStateData::findProxyProtocolMagic()
> 
> This method does not just "test". It actually parses. IMO,
> findProxyProtocolMagic() should be called parseProxyProtocolHeader().
> 
> No, it does not matter that the actual non-magic parsing code is in
> other parsing methods. The method description and name should reflect
> what the method does from the caller point of view. The method internals
> are not that important when you are naming and describing the interface.
> 

Okay, okay. Done.

> 
>>>> +    needProxyProtocolHeader_ = xact->squidPort->flags.proxySurrogate;
>>>> +    if (needProxyProtocolHeader_)
>>>> +        proxyProtocolValidateClient(); // will close the connection on 
>>>> failure
>>>
>>> Please do not place things that require job protection in a class
>>> constructor. Calling things like stopSending() and stopReceiving() does
>>> not (or will not) work well when we are just constructing a
>>> ConnStateData object. Nobody may notice (at the right time) that you
>>> "stopped" something because nothing has been started yet.
>>
>> Did not require job protection until converting
>> stopSending()/stopReceiving() into mustStop() calls.
> 
> It only looked that way, but sooner or later that code would break
> because it was essentially stopping the job inside the job's constructor.
> 
> 
>> Also, note that ConnStateData is not a true AsyncJob. It never started
>> with AsynJob::Start() and cleanup is equally nasty as this setup
>> constructor (prior to any changes I am adding).
> 
> Yes, I know that ConnStateData is an AsyncJob that already violates a
> lot of job API principles. That is not your fault, of course. However,
> adding more violations into the job constructor makes things worse
> (somebody would have to rewrite that later).
> 
> 
>> I have done as suggested and created the AsyncJob::Start() functionality
>> for it.
> 
> Thank you. Hopefully, you would be able to keep that change despite the
> extra work it requires.
> 
> 
>> However, this means that the PROXY protocol is no longer capable of
>> being used on https_port. The PROXY protocol header comes before TLS
>> negotiation on the wire, but ConnStateData::start() is only called by
>> our code after SSL/TLS negotiation and SSL-bump operations complete.
> 
> I agree that there is a problem with https_port support, but the "moving
> PROXY ACL check into start() brakes https_port support for PROXY"
> argument does not work for me: I agree that the SSL code in trunk does a
> lot of things between creating the ConnStateData job and calling
> readSomeData(). However, that was true for the original (mk1) patch as
> well! Moreover, the original (mk1) patch could close the connection in
> the ConnStateData constructor and then proceed with negotiating SSL for
> that connection. Thus, I do not think fixing ConnStateData constructor
> is the source of the htts_port support problems. That support was
> already broken in mk1.
> 
> I see two correct ways to address the https_port problem:
> 
> 1. Refuse PROXY support on https_port for now. As any support
> limitation, this is unfortunate, but I think it should be accepted. The
> configuration code should be adjusted to reject Squid configurations
> that use proxy-surrogate with an https_port.

Doing this one. I have no desire for re-writing all the SSL negotiation
logics right now.

> 
> 2. Support PROXY on https_port. As you have mentioned, this probably
> requires moving SSL negotiations inside the ConnStateData job so that
> the job can be Start()ed when it is created instead of passing the job
> as a POD-like structure during SSL negotiations.
> 
> 
>> +<p><em>Known Issue:</em> Due to design issues HTTPS traffic is not yet 
>> accepted
>> +   over this protocol. So use of <em>proxy-surrogate</em> on 
>> <em>https_port</em>
>> +   is not supported.
> 
> If you continue going with #1, please do not blame mysterious "design
> issues" (Squid design? PROXY protocol design? OpenSSL design? Internet
> design?) but simply say that PROXY for https_port is not yet supported.
> There is no design issue here AFAICT. Https_port support just needs more
> development work. This is just an implementation limitation.

They are Squid design issues:

 1) creating the ConnStateData Job before negotiating TLS/SSL using
old-style global functions. But only start()ing it after TLS negotiation.

 2) sharing a socket between ConnStateData read(2) and OpenSSL
openssl_read(2) operations.

Specifically #2 has issues with ConnStateData reading arbitrary amount
of bytes off the socket into its I/O buffer before identifying the end
of PROXY header. So even if we started the ConnStateData early and
paused for TLS/SSL negotiations later, due to AsyncCall delays between
accept() and ConnStateData::start() an unknown amount of TLS/SSL bytes
may be sucked in.

> 
> BTW, when receiving on an https_port, will the PROXY header be
> encrypted? If yes, why not postpone the ACL check until the connection
> is decrypted? And where does the PROXY draft explain/define whether the
> PROXY header should be encrypted?

It is documented as a prefix on the TCP layer payload. So AIUI, the
header is an un-encrypted prefix before the TLS ClientHello.

> 
>> +<p>Squid currently supports receiving HTTP via version 1 or 2 of the 
>> protocol.
> ...
> 
> "receiving HTTP via [PROXY] protocol" sounds awkward to me. The PROXY
> protocol does not envelop or embed the HTTP protocol that follows the
> PROXY header IMO. It just starts a connection with a small header and
> does not have a notion of "PROXY message body".
> 
> If "version 1 or 2" is the PROXY protocol version (rather than the HTTP
> version), then the above is inconsistent with the version 1.5 documented
> later:
> 
>> +proxy-protocol.txt
>> +    Documents Proxy Protocol 1.5, for communicating original client IP
>> +    details between consenting proxies and servers even when
>> +    transparent interception is taking place.
> 
> I think the above ought to say "Documents PROXY Protocol versions 1 and
> 2", just like the draft title.
> 

Fixed.

> 
>> +<p>PROXY protocol provides a simple way for proxies and tunnels of any kind 
>> to
>> +   relay the original client source details ...
> 
>> +proxy-protocol.txt
>> +    Documents Proxy Protocol 1.5, for communicating original client IP
>> +    details ...
> 
> AFAICT, the PROXY protocol supports a lot more than client IP details.
> It also communicates client source port, transport protocol, and the
> destination address details.

Fixed.

> 
> 
>> +<sect1>Support PROXY protocol
> 
> We only support receiving PROXY protocol; Squid does not support being a
> PROXY protocol client, right? Can we be less ambitious in the above
> claim then? Something like "PROXY protocol support (receiving)" or
> "Support for receiving PROXY protocol header" would work better IMO.
> 
> 

Done.

>> +            HTTP message Forwarded header, or
>> +            HTTP message X-Forwarded-For header, or
>> +            PROXY protocol connection header.
> 
>> +    Allowing or Denying the X-Forwarded-For or Forwarded headers to
>> +    be followed to find the original source of a request. Or permitting
>> +    a client proxy to connect using PROXY protocol.
> 
> What happens when the incoming traffic has both an allowed PROXY header
> and an allowed HTTP X-Forwarded-For header? Which takes priority?

 PROXY is about the TCP connection and equivalent to NAT on the local
machine.

 XFF/Forwarded is about the single message where it resides.

When evaluating the security direct TCP is evaluated first, then PROXY,
then XFF entries.

TCP direct IP
  [ PROXY src-IP ]
  XFF last entry
  [ XFF 2nd to last entry ]
  [ ... ]

> What
> if X-Forwarded-For header changes client information in the middle of a
> connection? Please document these cases. I could not find that info in
> the PROXY protocol draft, but perhaps I missed it.

Same thing as if XFF was received on a non-PROXY connection. AFAIK that
changes only the HttpReuqest values for the particular message.

XFF does not change its behaviour in any way due to PROXY protocol
existence. The connection IP is checked for trust then the XFF entries
until one fails. It just happens that the connection IP now comes from
PROXY.

The relation XFF and PROXY have is that the XFF trustworthiness depends
on PROXY being trusted if present. The ACL trust assignment semantics is
identical when handled as diagrammed above. PROXY being a sub-step up
from direct TCP details which also needs to be verified to retain any
trust link between direct TCP client IP and the indirect client IP in XFF.


BTW, this confusion you seem to have between the two is exactly why I am
trying to rename follow_x_forwarded_for - since it does not necessarily
relate to XFF header when evaluating trust of PROXY protocol. And
certainly won't when we upgrade to only supporting Forwarded: header.


> 
> Please adjust the documentation to make it clear whether the features
> are exclusive (either/or) or can co-exist (and/or).
> 
> 
>> +       proxy-surrogate
>> +                    Support for PROXY protocol version 1 or 2 connections.
>> +                    The proxy_forwarded_access is required to whitelist
>> +                    downstream proxies which can be trusted.
>> +
> 
> What happens to the destination information in the PROXY header? Is it
> used in any way? Should we mention how it is used (or the fact that it
> is not used)?

If we can trust the source then we use the PROXY protocol header as per
PROXY protocol. Do we really need to enumerate whole chapters of the
protocol spec in the documentation of its config option?

For example, I dont see anything from RFC 2616 or 7230 documenting
"accel" or anything documenting TLS handshakes for ssl-bump.

If we need detailed long descriptions we have the wiki and/or the PROXY
spec document itself.


> 
> Consider s/downstream/client/.
> 
> 
>> +    /** marks ports receiving PROXY protocol traffic
>> +     *
>> +     * Indicating the following are required:
>> +     *  - PROXY protocol magic header
>> +     *  - src/dst IP retrieved from magic PROXY header
>> +     *  - reverse-proxy traffic prohibited
>> +     *  - intercepted traffic prohibited
>> +     */
>> +    bool proxySurrogate;
> 
> Why is reverse-proxy traffic prohibited? The PROXY protocol draft
> mentions reverse proxies being used, but it is not clear to me whether
> they are only supported as PROXY protocol clients. I do not know all the
> PROXY details, but it seems to me that reverse proxies should be allowed
> as servers (AFAICT, the explicit configuration requirement is the key
> here, not the reverse/forward role of the proxy!).

Hmm. Good point.
I'm removing those prohibitions and replacing with just an implicit
no-spoofing for TPROXY.

> 
>>>> +            debugs(33, 5, "PROXY protocol on connection " << 
>>>> clientConnection);
>>>> +            clientConnection->local = originalDest;
>>>> +            clientConnection->remote = originalClient;
>>>> +            debugs(33, 5, "PROXY upgrade: " << clientConnection);
>>>
>>> We use this kind of address resetting code in many places, right? Please
>>> encapsulate it (together with the debugging) into a
>>> Connection::resetAddrs() or a similar method.
>>
>> Two. PROXY/1.0 and PROXY/2.0 parsers.
> 
> 
> I found more, including:
> 
>> ./log/TcpLogger.cc:    futureConn->remote = remote;
>> ./log/TcpLogger.cc-    futureConn->local.setAnyAddr();

The local setup contains a setAnyAddr() and setIPv4() conditional
optimizations. Using a copy requires adding an otherwise needless local
variable allocation.

> 
>> ./ftp.cc-    conn->local = ftpState->ctrl.conn->local;
>> ./ftp.cc-    conn->local.port(0);
>> ./ftp.cc:    conn->remote = ftpState->ctrl.conn->remote;
>> ./ftp.cc-    conn->remote.port(port);
> 
>> ./ftp.cc-    conn->local = ftpState->ctrl.conn->local;
>> ./ftp.cc-    conn->local.port(0);
>> ./ftp.cc:    conn->remote = ipaddr;
>> ./ftp.cc-    conn->remote.port(port);
> 
>> ./dns_internal.cc-        conn->local = Config.Addrs.udp_incoming;
>> ./dns_internal.cc-
>> ./dns_internal.cc:    conn->remote = nameservers[nsv].S;
> 
> 

Okay. Adding setAddrs(local, remote) to trunk since it is pure scope creep.

> 
>>> * When, in a misconfigured setup, somebody sends a PROXY header to a
>>> regular Squid HTTP port, does the Squid error look obvious/clear enough?
>>> Or will the admin have a hard time understanding why things do not work
>>> in that case?
>>>
>>
>> Trunk will die with a 400 error quoting the PROXY header as the
>> "URL" or buffer content.
> 
> By "die", you mean respond with an HTTP 400 error, right?
> 

Er, yes. And close TCP connection.

> 
>> It seems clear enough to
>> me not to need new code for that type of config error case.
> 
> I disagree because admins do not normally see 400 errors (until users
> start complaining). Most see log messages though. Please note that in a
> multi-port setup, it may not be obvious that something is misconfigured
> because there may be few or no 400 errors when Squid starts serving
> requests.
> 
> There are two kinds of likely misconfigurations:
> 
> 1) Client sends PROXY. Squid is not configured to receive one. We could
> handle this better, but I do not insist on adding code to detect and
> warn about such cases because the required code would be relatively
> complex, and because it would make it even easier to spam cache.log with
> error messages when there is no misconfiguration at all (unless we add
> even more code).
> 
> 2) Client does not send PROXY. Squid is configured to require one. I
> think we should announce such cases in cache.log. There is already code
> to detect this problem. AFAICT, we just need to a logging line.
> 

Added a 33,2 level display of the error message and connection details.
I think higher levels would potentially be too noisy. Terminating
connections on error is a routine part of PROXY.

If we need more complex debugging of this we should probably allocate a
debug section to it. But for now 33 seems fine.

Amos
=== modified file 'doc/release-notes/release-3.5.sgml'
--- doc/release-notes/release-3.5.sgml  2014-07-13 05:28:15 +0000
+++ doc/release-notes/release-3.5.sgml  2014-07-25 12:05:51 +0000
@@ -26,40 +26,41 @@
 <sect1>Known issues
 <p>
 Although this release is deemed good enough for use in many setups, please 
note the existence of 
 <url 
url="http://bugs.squid-cache.org/buglist.cgi?query_format=advanced&amp;product=Squid&amp;bug_status=UNCONFIRMED&amp;bug_status=NEW&amp;bug_status=ASSIGNED&amp;bug_status=REOPENED&amp;version=3.5";
 name="open bugs against Squid-3.5">.
 
 <sect1>Changes since earlier releases of Squid-3.5
 <p>
 The 3.5 change history can be <url 
url="http://www.squid-cache.org/Versions/v3/3.5/changesets/"; name="viewed 
here">.
 
 
 <sect>Major new features since Squid-3.4
 <p>Squid 3.5 represents a new feature release above 3.4.
 
 <p>The most important of these new features are:
 <itemize>
        <item>Support libecap v1.0
        <item>Authentication helper query extensions
        <item>Support named services
        <item>Upgraded squidclient tool
        <item>Helper support for concurrency channels
+       <item>Receive PROXY protocol, Versions 1 & 2
 </itemize>
 
 Most user-facing changes are reflected in squid.conf (see below).
 
 
 <sect1>Support libecap v1.0
 <p>Details at <url url="http://wiki.squid-cache.org/Features/BLAH";>.
 
 <p>The new libecap version allows Squid to better check the version of
   the eCAP adapter being loaded as well as the version of the eCAP library
   being used.
 
 <p>Squid-3.5 can support eCAP adapters built with libecap v1.0,
    but no longer supports adapters built with earlier libecap versions
    due to API changes.
 
 
 <sect1>Authentication helper query extensions
 <p>Details at <url url="http://www.squid-cache.org/Doc/config/auth_param/";>.
 
@@ -146,71 +147,111 @@
    The default is to use X.509 certificate encryption instead.
 
 <p>When performing TLS/SSL server certificates are always verified, the
    results shown at debug level 3. The encrypted type is displayed at debug
    level 2 and the connection is used to send and receive the messages
    regardless of verification results.
 
 
 <sect1>Helper support for concurrency channels
 <p>Helper concurrency greatly reduces the communication lag between Squid
    and its helpers allowing faster transaction speeds even on sequential
    helpers.
 
 <p>The Digest authentication, Store-ID, and URL-rewrite helpers packaged
    with Squid have been updated to support concurrency channels. They will
    auto-detect the <em>channel-ID</em> field and will produce the appropriate
    response format.
    With these helpers concurrency may now be set to 0 or any higher number as 
desired.
 
 
+<sect1>Receive PROXY protocol, Versions 1 & 2
+<p>More info at <url 
url="http://www.haproxy.org/download/1.5/doc/proxy-protocol.txt";>
+
+<p>PROXY protocol provides a simple way for proxies and tunnels of any kind to
+   relay the original client source details without having to alter or 
understand
+   the protocol being relayed on the connection.
+
+<p>Squid currently supports receiving HTTP traffic from a client proxy using 
this protocol.
+   An http_port which has been configured to receive this protocol may only be 
used to
+   receive traffic from client software sending in this protocol.
+   Regular forward-proxy HTTP traffic is not accepted.
+
+<p>The <em>accel</em> and <em>intercept</em> options are still used to 
identify the
+   traffic syntax being delivered by the client proxy.
+
+<p>Squid can be configured by adding an <em>http_port</em>
+   with the <em>proxy-surrogate</em> mode flag. The 
<em>proxy_forwarded_access</em>
+   must also be configured with <em>src</em> ACLs to whitelist proxies which 
are
+   trusted to send correct client details.
+
+<p>Forward-proxy traffic from a client proxy:
+<verbatim>
+ http_port 3128 proxy-surrogate
+ proxy_forwarded_access allow localhost
+</verbatim>
+
+<p>Intercepted traffic from a client proxy or tunnel:
+<verbatim>
+ http_port 3128 intercept proxy-surrogate
+ proxy_forwarded_access allow localhost
+</verbatim>
+
+<p><em>Known Issue:</em>
+   Use of <em>proxy-surrogate</em> on <em>https_port</em> is not supported.
+
+
 <sect>Changes to squid.conf since Squid-3.4
 <p>
 There have been changes to Squid's configuration file since Squid-3.4.
 
 <p>Squid supports reading configuration option parameters from external
    files using the syntax <em>parameters("/path/filename")</em>. For example:
 <verb>
     acl whitelist dstdomain parameters("/etc/squid/whitelist.txt")
 </verb>
 
 <p>The squid.conf macro ${service_name} is added to provide the service name
    of the process parsing the config.
 
 <p>There have also been changes to individual directives in the config file.
 
 This section gives a thorough account of those changes in three categories:
 
 <itemize>
        <item><ref id="newtags" name="New tags">
        <item><ref id="modifiedtags" name="Changes to existing tags">
        <item><ref id="removedtags" name="Removed tags">
 </itemize>
 <p>
 
 <sect1>New tags<label id="newtags">
 <p>
 <descrip>
        <tag>collapsed_forwarding</tag>
        <p>Ported from Squid-2 with no configuration or visible behaviour 
changes.
            Collapsing of requests is performed across SMP workers.
 
+       <tag>proxy_forwarded_access</tag>
+       <p>Renamed from <em>follow_x_forwarded_for</em> and extended to control 
more
+          ways for locating the indirect (original) client IP details.
+
        <tag>send_hit</tag>
        <p>New configuration directive to enable/disable sending cached content
           based on ACL selection. ACL can be based on client request or cached
           response details.
 
        <tag>sslproxy_session_cache_size</tag>
        <p>New directive which sets the cache size to use for TLS/SSL sessions 
cache.
 
        <tag>sslproxy_session_ttl</tag>
        <p>New directive to specify the time in seconds the TLS/SSL session is 
valid.
 
        <tag>store_id_extras</tag>
        <p>New directive to send additional lookup parameters to the configured
           Store-ID helper program. It takes a string which may contain 
logformat %macros.
        <p>The Store-ID helper input format is now:
        <verb>
          [channel-ID] url [extras]
        </verb>
        <p>The default value for extras is: "%&gt;a/%>A %un %>rm myip=%la 
myport=%lp"
 
@@ -259,75 +300,80 @@
        <p>These connections differ from HTTP persistent connections in that 
they
           have not been used for HTTP messaging (and may never be). They may be
           turned into persistent connections after their first use subject to 
the
           same keep-alive critera any HTTP connection is checked for.
 
        <tag>forward_max_tries</tag>
        <p>Default value increased to <em>25 destinations</em> to allow better
           contact and IPv4 failover with domains using long lists of IPv6
           addresses.
 
        <tag>ftp_epsv</tag>
        <p>Converted into an Access List with allow/deny value driven by ACLs
           using Squid standard first line wins matching basis.
        <p>The old values of <em>on</em> and <em>off</em> imply <em>allow 
all</em>
            and <em>deny all</em> respectively and are now deprecated.
           Do not combine use of on/off values with ACL configuration.
 
        <tag>http_port</tag>
        <p><em>protocol=</em> option altered to accept protocol version details.
           Currently supported values are: HTTP, HTTP/1.1, HTTPS, HTTPS/1.1
+       <p><em>New option <em>proxy-surrogate</em> to mark ports receiving PROXY
+          protocol version 1 or 2 traffic.
 
        <tag>https_port</tag>
        <p><em>protocol=</em> option altered to accept protocol version details.
           Currently supported values are: HTTP, HTTP/1.1, HTTPS, HTTPS/1.1
 
        <tag>logformat</tag>
        <p>New format code <em>%credentials</em> to log the client credentials
           token.
        <p>New format code <em>%tS</em> to log transaction start time in
           "seconds.milliseconds" format, similar to the existing access.log
           "current time" field (%ts.%03tu) which logs the corresponding
           transaction finish time.
 
 </descrip>
 
 <sect1>Removed tags<label id="removedtags">
 <p>
 <descrip>
        <tag>cache_dir</tag>
        <p><em>COSS</em> storage type is formally replaced by Rock storage type.
 
        <tag>cache_dns_program</tag>
        <p>DNS external helper interface has been removed. It was no longer
           able to provide high performance service and the internal DNS
           client library with multicast DNS cover all modern use-cases.
 
        <tag>cache_peer</tag>
        <p><em>idle=</em> replaced by <em>standby=</em>.
        <p>NOTE that standby connections are started earlier and available in
           more circumstances than squid-2 idle connections were. They are
           also spread over all IPs of the peer.
 
        <tag>dns_children</tag>
        <p>DNS external helper interface has been removed.
 
+       <tag>follow_x_forwarded_for</tag>
+       <p>Renamed <em>proxy_forwarded_access</em> and extended.
+
 </descrip>
 
 
 <sect>Changes to ./configure options since Squid-3.4
 <p>
 There have been some changes to Squid's build configuration since Squid-3.4.
 
 This section gives an account of those changes in three categories:
 
 <itemize>
        <item><ref id="newoptions" name="New options">
        <item><ref id="modifiedoptions" name="Changes to existing options">
        <item><ref id="removedoptions" name="Removed options">
 </itemize>
 
 
 <sect1>New options<label id="newoptions">
 <p>
 <descrip>
        <tag>BUILDCXX=</tag>

=== modified file 'doc/rfc/1-index.txt'
--- doc/rfc/1-index.txt 2014-06-09 01:38:06 +0000
+++ doc/rfc/1-index.txt 2014-07-25 09:18:15 +0000
@@ -1,40 +1,43 @@
 draft-ietf-radext-digest-auth-06.txt
        RADIUS Extension for Digest Authentication
        A proposed extension to Radius for Digest authentication
        via RADIUS servers.
 
 draft-cooper-webi-wpad-00.txt
 draft-ietf-svrloc-wpad-template-00.txt
        Web Proxy Auto-Discovery Protocol -- WPAD
        documents how MSIE and several other browsers automatically
        find their proxy settings from DHCP and/or DNS
 
 draft-forster-wrec-wccp-v1-00.txt
        WCCP 1.0
 
 draft-wilson-wccp-v2-12-oct-2001.txt
        WCCP 2.0
 
 draft-vinod-carp-v1-03.txt
        Microsoft CARP peering algorithm
 
+proxy-protocol.txt
+       The PROXY protocol, Versions 1 & 2
+
 rfc0959.txt
        FTP
 
 rfc1035.txt
        DNS for IPv4
 
 rfc1157.txt
        A Simple Network Management Protocol (SNMP)
        SNMP v1 Specification. SNMP v2 is documented in several RFCs,
        namely, 1902,1903,1904,1905,1906,1907.
 
 rfc1738.txt
        Uniform Resource Locators (URL)
        (updated by RFC 3986, but not obsoleted)
 
 rfc1902.txt
        Structure of Managament Information (SMI) for SNMPv2
        Management information is viewed as a collection of managed objects,
        the Management Information Base (MIB). MIB modules are
        written using an adapted subset of OSI's Abstract Syntax

=== modified file 'src/Makefile.am'
--- src/Makefile.am     2014-07-23 12:51:55 +0000
+++ src/Makefile.am     2014-07-25 12:05:51 +0000
@@ -1609,40 +1609,41 @@
        acl/libapi.la \
        base/libbase.la \
        libsquid.la \
        ip/libip.la \
        fs/libfs.la \
        comm/libcomm.la \
        eui/libeui.la \
        icmp/libicmp.la icmp/libicmp-core.la \
        log/liblog.la \
        format/libformat.la \
        $(REPL_OBJS) \
        $(DISK_LIBS) \
        $(DISK_OS_LIBS) \
        $(ADAPTATION_LIBS) \
        $(ESI_LIBS) \
        $(SSL_LIBS) \
        anyp/libanyp.la \
        ipc/libipc.la \
        mgr/libmgr.la \
        $(SNMP_LIBS) \
+       parser/libsquid-parser.la \
        $(top_builddir)/lib/libmisccontainers.la \
        $(top_builddir)/lib/libmiscencoding.la \
        $(top_builddir)/lib/libmiscutil.la \
        $(NETTLELIB) \
        $(REGEXLIB) \
        $(SQUID_CPPUNIT_LIBS) \
        $(SQUID_CPPUNIT_LA) \
        $(SSLLIB) \
        $(KRB5LIBS) \
        $(COMPAT_LIB) \
        $(XTRA_LIBS)
 tests_testCacheManager_LDFLAGS = $(LIBADD_DL)
 tests_testCacheManager_DEPENDENCIES = \
        $(REPL_OBJS) \
        $(SQUID_CPPUNIT_LA)
 
 tests_testDiskIO_SOURCES = \
        CacheDigest.h \
        tests/stub_CacheDigest.cc \
        cbdata.cc \
@@ -2037,40 +2038,41 @@
        $(DISKIO_GEN_SOURCE)
 tests_testEvent_LDADD = \
        http/libsquid-http.la \
        ident/libident.la \
        acl/libacls.la \
        acl/libstate.la \
        acl/libapi.la \
        base/libbase.la \
        libsquid.la \
        ip/libip.la \
        fs/libfs.la \
        anyp/libanyp.la \
        icmp/libicmp.la icmp/libicmp-core.la \
        comm/libcomm.la \
        log/liblog.la \
        format/libformat.la \
        $(REPL_OBJS) \
        $(ADAPTATION_LIBS) \
        $(ESI_LIBS) \
        $(SSL_LIBS) \
+       parser/libsquid-parser.la \
        $(top_builddir)/lib/libmisccontainers.la \
        $(top_builddir)/lib/libmiscencoding.la \
        $(top_builddir)/lib/libmiscutil.la \
        $(DISK_LIBS) \
        $(DISK_OS_LIBS) \
        ipc/libipc.la \
        mgr/libmgr.la \
        $(SNMP_LIBS) \
        $(NETTLELIB) \
        $(REGEXLIB) \
        $(SQUID_CPPUNIT_LIBS) \
        $(SQUID_CPPUNIT_LA) \
        $(SSLLIB) \
        $(KRB5LIBS) \
        $(COMPAT_LIB) \
        $(XTRA_LIBS)
 tests_testEvent_LDFLAGS = $(LIBADD_DL)
 tests_testEvent_DEPENDENCIES = \
        $(REPL_OBJS) \
        $(SQUID_CPPUNIT_LA)
@@ -2287,40 +2289,41 @@
        $(DISKIO_GEN_SOURCE)
 tests_testEventLoop_LDADD = \
        http/libsquid-http.la \
        ident/libident.la \
        acl/libacls.la \
        acl/libstate.la \
        acl/libapi.la \
        base/libbase.la \
        libsquid.la \
        ip/libip.la \
        fs/libfs.la \
        anyp/libanyp.la \
        icmp/libicmp.la icmp/libicmp-core.la \
        comm/libcomm.la \
        log/liblog.la \
        format/libformat.la \
        $(REPL_OBJS) \
        $(ADAPTATION_LIBS) \
        $(ESI_LIBS) \
        $(SSL_LIBS) \
+       parser/libsquid-parser.la \
        $(top_builddir)/lib/libmisccontainers.la \
        $(top_builddir)/lib/libmiscencoding.la \
        $(top_builddir)/lib/libmiscutil.la \
        $(DISK_LIBS) \
        $(DISK_OS_LIBS) \
        ipc/libipc.la \
        mgr/libmgr.la \
        $(SNMP_LIBS) \
        $(NETTLELIB) \
        $(REGEXLIB) \
        $(SQUID_CPPUNIT_LIBS) \
        $(SQUID_CPPUNIT_LA) \
        $(SSLLIB) \
        $(KRB5LIBS) \
        $(COMPAT_LIB) \
        $(XTRA_LIBS)
 tests_testEventLoop_LDFLAGS = $(LIBADD_DL)
 tests_testEventLoop_DEPENDENCIES = \
        $(REPL_OBJS) \
        $(SQUID_CPPUNIT_LA)
@@ -2535,40 +2538,41 @@
        acl/libstate.la \
        acl/libapi.la \
        libsquid.la \
        ip/libip.la \
        fs/libfs.la \
        anyp/libanyp.la \
        icmp/libicmp.la icmp/libicmp-core.la \
        comm/libcomm.la \
        log/liblog.la \
        format/libformat.la \
        $(REPL_OBJS) \
        $(DISK_LIBS) \
        $(DISK_OS_LIBS) \
        $(ADAPTATION_LIBS) \
        $(ESI_LIBS) \
        $(SSL_LIBS) \
        ipc/libipc.la \
        base/libbase.la \
        mgr/libmgr.la \
        $(SNMP_LIBS) \
+       parser/libsquid-parser.la \
        $(top_builddir)/lib/libmisccontainers.la \
        $(top_builddir)/lib/libmiscencoding.la \
        $(top_builddir)/lib/libmiscutil.la \
        $(NETTLELIB) \
        $(REGEXLIB) \
        $(SQUID_CPPUNIT_LIBS) \
        $(SQUID_CPPUNIT_LA) \
        $(SSLLIB) \
        $(KRB5LIBS) \
        $(COMPAT_LIB) \
        $(XTRA_LIBS)
 tests_test_http_range_LDFLAGS = $(LIBADD_DL)
 tests_test_http_range_DEPENDENCIES = \
        $(SQUID_CPPUNIT_LA)
 
 tests_testHttpParser_SOURCES = \
        Debug.h \
        HttpParser.cc \
        HttpParser.h \
        MemBuf.cc \
@@ -2825,40 +2829,41 @@
        acl/libacls.la \
        acl/libstate.la \
        acl/libapi.la \
        libsquid.la \
        ip/libip.la \
        fs/libfs.la \
        $(SSL_LIBS) \
        ipc/libipc.la \
        base/libbase.la \
        mgr/libmgr.la \
        anyp/libanyp.la \
        $(SNMP_LIBS) \
        icmp/libicmp.la icmp/libicmp-core.la \
        comm/libcomm.la \
        log/liblog.la \
        format/libformat.la \
        http/libsquid-http.la \
        $(REPL_OBJS) \
        $(ADAPTATION_LIBS) \
        $(ESI_LIBS) \
+       parser/libsquid-parser.la \
        $(top_builddir)/lib/libmisccontainers.la \
        $(top_builddir)/lib/libmiscencoding.la \
        $(top_builddir)/lib/libmiscutil.la \
        $(DISK_OS_LIBS) \
        $(NETTLELIB) \
        $(REGEXLIB) \
        $(SQUID_CPPUNIT_LIBS) \
        $(SQUID_CPPUNIT_LA) \
        $(SSLLIB) \
        $(KRB5LIBS) \
        $(COMPAT_LIB) \
        $(XTRA_LIBS)
 tests_testHttpRequest_LDFLAGS = $(LIBADD_DL)
 tests_testHttpRequest_DEPENDENCIES = \
        $(REPL_OBJS) \
        $(SQUID_CPPUNIT_LA)
 
 ## why so many sources? well httpHeaderTools requites ACLChecklist & friends.
 ## first line - what we are testing.
 tests_testStore_SOURCES= \
@@ -3669,40 +3674,41 @@
        eui/libeui.la \
        acl/libstate.la \
        acl/libapi.la \
        base/libbase.la \
        libsquid.la \
        ip/libip.la \
        fs/libfs.la \
        $(SSL_LIBS) \
        ipc/libipc.la \
        mgr/libmgr.la \
        $(SNMP_LIBS) \
        icmp/libicmp.la icmp/libicmp-core.la \
        comm/libcomm.la \
        log/liblog.la \
        $(DISK_OS_LIBS) \
        format/libformat.la \
        $(REGEXLIB) \
        $(REPL_OBJS) \
        $(ADAPTATION_LIBS) \
        $(ESI_LIBS) \
+       parser/libsquid-parser.la \
        $(top_builddir)/lib/libmisccontainers.la \
        $(top_builddir)/lib/libmiscencoding.la \
        $(top_builddir)/lib/libmiscutil.la \
        $(NETTLELIB) \
        $(COMPAT_LIB) \
        $(SQUID_CPPUNIT_LIBS) \
        $(SQUID_CPPUNIT_LA) \
        $(SSLLIB) \
        $(KRB5LIBS) \
        $(COMPAT_LIB) \
        $(XTRA_LIBS)
 tests_testURL_LDFLAGS = $(LIBADD_DL)
 tests_testURL_DEPENDENCIES = \
        $(REPL_OBJS) \
        $(SQUID_CPPUNIT_LA)
 
 tests_testSBuf_SOURCES= \
        tests/testSBuf.h \
        tests/testSBuf.cc \
        tests/testMain.cc \

=== modified file 'src/anyp/TrafficMode.h'
--- src/anyp/TrafficMode.h      2013-02-04 09:47:50 +0000
+++ src/anyp/TrafficMode.h      2014-07-25 06:12:42 +0000
@@ -8,40 +8,50 @@
  * Set of 'mode' flags defining types of trafic which can be received.
  *
  * Use to determine the processing steps which need to be applied
  * to this traffic under any special circumstances which may apply.
  */
 class TrafficMode
 {
 public:
     TrafficMode() : accelSurrogate(false), natIntercept(false), 
tproxyIntercept(false), tunnelSslBumping(false) {}
     TrafficMode(const TrafficMode &rhs) { operator =(rhs); }
     TrafficMode &operator =(const TrafficMode &rhs) { memcpy(this, &rhs, 
sizeof(TrafficMode)); return *this; }
 
     /** marks HTTP accelerator (reverse/surrogate proxy) traffic
      *
      * Indicating the following are required:
      *  - URL translation from relative to absolute form
      *  - restriction to origin peer relay recommended
      */
     bool accelSurrogate;
 
+    /** marks ports receiving PROXY protocol traffic
+     *
+     * Indicating the following are required:
+     *  - PROXY protocol magic header
+     *  - src/dst IP retrieved from magic PROXY header
+     *  - indirect client IP trust verification is mandatory
+     *  - TLS is not supported
+     */
+    bool proxySurrogate;
+
     /** marks NAT intercepted traffic
      *
      * Indicating the following are required:
      *  - NAT lookups
      *  - URL translation from relative to absolute form
      *  - Same-Origin verification is mandatory
      *  - destination pinning is recommended
      *  - authentication prohibited
      */
     bool natIntercept;
 
     /** marks TPROXY intercepted traffic
      *
      * Indicating the following are required:
      *  - src/dst IP inversion must be performed
      *  - client IP should be spoofed if possible
      *  - URL translation from relative to absolute form
      *  - Same-Origin verification is mandatory
      *  - destination pinning is recommended
      *  - authentication prohibited

=== modified file 'src/cache_cf.cc'
--- src/cache_cf.cc     2014-07-21 14:55:27 +0000
+++ src/cache_cf.cc     2014-07-25 12:05:51 +0000
@@ -3581,45 +3581,53 @@
     } else if (strcmp(token, "transparent") == 0 || strcmp(token, "intercept") 
== 0) {
         if (s->flags.accelSurrogate || s->flags.tproxyIntercept) {
             debugs(3, DBG_CRITICAL, "FATAL: http(s)_port: Intercept mode 
requires its own interception port. It cannot be shared with other modes.");
             self_destruct();
         }
         s->flags.natIntercept = true;
         Ip::Interceptor.StartInterception();
         /* Log information regarding the port modes under interception. */
         debugs(3, DBG_IMPORTANT, "Starting Authentication on port " << s->s);
         debugs(3, DBG_IMPORTANT, "Disabling Authentication on port " << s->s 
<< " (interception enabled)");
     } else if (strcmp(token, "tproxy") == 0) {
         if (s->flags.natIntercept || s->flags.accelSurrogate) {
             debugs(3,DBG_CRITICAL, "FATAL: http(s)_port: TPROXY option 
requires its own interception port. It cannot be shared with other modes.");
             self_destruct();
         }
         s->flags.tproxyIntercept = true;
         Ip::Interceptor.StartTransparency();
         /* Log information regarding the port modes under transparency. */
         debugs(3, DBG_IMPORTANT, "Disabling Authentication on port " << s->s 
<< " (TPROXY enabled)");
 
+        if (s->flags.proxySurrogate) {
+            debugs(3, DBG_IMPORTANT, "Disabling TPROXY Spoofing on port " << 
s->s << " (proxy-surrogate enabled)");
+        }
+
         if (!Ip::Interceptor.ProbeForTproxy(s->s)) {
             debugs(3, DBG_CRITICAL, "FATAL: http(s)_port: TPROXY support in 
the system does not work.");
             self_destruct();
         }
 
+    } else if (strcmp(token, "proxy-surrogate") == 0) {
+        s->flags.proxySurrogate = true;
+        debugs(3, DBG_IMPORTANT, "Disabling TPROXY Spoofing on port " << s->s 
<< " (proxy-surrogate enabled)");
+
     } else if (strncmp(token, "defaultsite=", 12) == 0) {
         if (!s->flags.accelSurrogate) {
             debugs(3, DBG_CRITICAL, "FATAL: http(s)_port: defaultsite option 
requires Acceleration mode flag.");
             self_destruct();
         }
         safe_free(s->defaultsite);
         s->defaultsite = xstrdup(token + 12);
     } else if (strcmp(token, "vhost") == 0) {
         if (!s->flags.accelSurrogate) {
             debugs(3, DBG_CRITICAL, "WARNING: http(s)_port: vhost option is 
deprecated. Use 'accel' mode flag instead.");
         }
         s->flags.accelSurrogate = true;
         s->vhost = true;
     } else if (strcmp(token, "no-vhost") == 0) {
         if (!s->flags.accelSurrogate) {
             debugs(3, DBG_IMPORTANT, "ERROR: http(s)_port: no-vhost option 
requires Acceleration mode flag.");
         }
         s->vhost = false;
     } else if (strcmp(token, "vport") == 0) {
         if (!s->flags.accelSurrogate) {
@@ -3783,84 +3791,91 @@
         self_destruct();
         return;
     }
 
     char *token = ConfigParser::NextToken();
 
     if (!token) {
         self_destruct();
         return;
     }
 
     AnyP::PortCfgPointer s = new AnyP::PortCfg();
     s->setTransport(protocol);
     parsePortSpecification(s, token);
 
     /* parse options ... */
     while ((token = ConfigParser::NextToken())) {
         parse_port_option(s, token);
     }
 
-#if USE_OPENSSL
     if (s->transport.protocol == AnyP::PROTO_HTTPS) {
+#if USE_OPENSSL
         /* ssl-bump on https_port configuration requires either tproxy or 
intercept, and vice versa */
         const bool hijacked = s->flags.isIntercepted();
         if (s->flags.tunnelSslBumping && !hijacked) {
             debugs(3, DBG_CRITICAL, "FATAL: ssl-bump on https_port requires 
tproxy/intercept which is missing.");
             self_destruct();
         }
         if (hijacked && !s->flags.tunnelSslBumping) {
             debugs(3, DBG_CRITICAL, "FATAL: tproxy/intercept on https_port 
requires ssl-bump which is missing.");
             self_destruct();
         }
-    }
 #endif
+        if (s->transport.protocol == AnyP::PROTO_HTTPS) {
+            debugs(3,DBG_CRITICAL, "FATAL: https_port: proxy-surrogate option 
cannot be used on HTTPS ports.");
+            self_destruct();
+        }
+    }
 
     if (Ip::EnableIpv6&IPV6_SPECIAL_SPLITSTACK && s->s.isAnyAddr()) {
         // clone the port options from *s to *(s->next)
         s->next = s->clone();
         s->next->s.setIPv4();
         debugs(3, 3, AnyP::UriScheme(s->transport.protocol).c_str() << "_port: 
clone wildcard address for split-stack: " << s->s << " and " << s->next->s);
     }
 
     while (*head != NULL)
         head = &((*head)->next);
 
     *head = s;
 }
 
 static void
 dump_generic_port(StoreEntry * e, const char *n, const AnyP::PortCfgPointer &s)
 {
     char buf[MAX_IPSTRLEN];
 
     storeAppendPrintf(e, "%s %s",
                       n,
                       s->s.toUrl(buf,MAX_IPSTRLEN));
 
     // MODES and specific sub-options.
     if (s->flags.natIntercept)
         storeAppendPrintf(e, " intercept");
 
     else if (s->flags.tproxyIntercept)
         storeAppendPrintf(e, " tproxy");
 
+    else if (s->flags.proxySurrogate)
+        storeAppendPrintf(e, " proxy-surrogate");
+
     else if (s->flags.accelSurrogate) {
         storeAppendPrintf(e, " accel");
 
         if (s->vhost)
             storeAppendPrintf(e, " vhost");
 
         if (s->vport < 0)
             storeAppendPrintf(e, " vport");
         else if (s->vport > 0)
             storeAppendPrintf(e, " vport=%d", s->vport);
 
         if (s->defaultsite)
             storeAppendPrintf(e, " defaultsite=%s", s->defaultsite);
 
         // TODO: compare against prefix of 'n' instead of assuming http_port
         if (s->transport.protocol != AnyP::PROTO_HTTP)
             storeAppendPrintf(e, " protocol=%s", 
AnyP::UriScheme(s->transport.protocol).c_str());
 
         if (s->allow_direct)
             storeAppendPrintf(e, " allow-direct");

=== modified file 'src/cf.data.pre'
--- src/cf.data.pre     2014-07-21 14:55:27 +0000
+++ src/cf.data.pre     2014-07-25 12:05:51 +0000
@@ -1077,49 +1077,57 @@
 acl localnet src 172.16.0.0/12 # RFC1918 possible internal network
 acl localnet src 192.168.0.0/16        # RFC1918 possible internal network
 acl localnet src fc00::/7       # RFC 4193 local private network range
 acl localnet src fe80::/10      # RFC 4291 link-local (directly plugged) 
machines
 
 acl SSL_ports port 443
 acl Safe_ports port 80         # http
 acl Safe_ports port 21         # ftp
 acl Safe_ports port 443                # https
 acl Safe_ports port 70         # gopher
 acl Safe_ports port 210                # wais
 acl Safe_ports port 1025-65535 # unregistered ports
 acl Safe_ports port 280                # http-mgmt
 acl Safe_ports port 488                # gss-http
 acl Safe_ports port 591                # filemaker
 acl Safe_ports port 777                # multiling http
 acl CONNECT method CONNECT
 NOCOMMENT_END
 DOC_END
 
-NAME: follow_x_forwarded_for
+NAME: proxy_forwarded_access follow_x_forwarded_for
 TYPE: acl_access
-IFDEF: FOLLOW_X_FORWARDED_FOR
 LOC: Config.accessList.followXFF
 DEFAULT_IF_NONE: deny all
-DEFAULT_DOC: X-Forwarded-For header will be ignored.
+DEFAULT_DOC: indirect client IP will not be accepted.
 DOC_START
-       Allowing or Denying the X-Forwarded-For header to be followed to
-       find the original source of a request.
+       Determine which client proxies can be trusted to provide correct
+       information regarding real client IP address.
+
+       The original source details can be relayed in:
+               HTTP message Forwarded header, or
+               HTTP message X-Forwarded-For header, or
+               PROXY protocol connection header.
+
+       Allowing or Denying the X-Forwarded-For or Forwarded headers to
+       be followed to find the original source of a request. Or permitting
+       a client proxy to connect using PROXY protocol.
 
        Requests may pass through a chain of several other proxies
        before reaching us.  The X-Forwarded-For header will contain a
        comma-separated list of the IP addresses in the chain, with the
        rightmost address being the most recent.
 
        If a request reaches us from a source that is allowed by this
        configuration item, then we consult the X-Forwarded-For header
        to see where that host received the request from.  If the
        X-Forwarded-For header contains multiple addresses, we continue
        backtracking until we reach an address for which we are not allowed
        to follow the X-Forwarded-For header, or until we reach the first
        address in the list. For the purpose of ACL used in the
        follow_x_forwarded_for directive the src ACL type always matches
        the address we are testing and srcdomain matches its rDNS.
 
        The end result of this process is an IP address that we will
        refer to as the indirect client address.  This address may
        be treated as the client address for access control, ICAP, delay
        pools and logging, depending on the acl_uses_indirect_client,
@@ -1704,40 +1712,45 @@
                            always      disable always PMTU discovery.
 
                        In many setups of transparently intercepting proxies
                        Path-MTU discovery can not work on traffic towards the
                        clients. This is the case when the intercepting device
                        does not fully track connections and fails to forward
                        ICMP must fragment messages to the cache server. If you
                        have such setup and experience that certain clients
                        sporadically hang or never complete requests set
                        disable-pmtu-discovery option to 'transparent'.
 
           name=        Specifies a internal name for the port. Defaults to
                        the port specification (port or addr:port)
 
           tcpkeepalive[=idle,interval,timeout]
                        Enable TCP keepalive probes of idle connections.
                        In seconds; idle is the initial time before TCP starts
                        probing the connection, interval how often to probe, and
                        timeout the time before giving up.
 
+          proxy-surrogate
+                       Require PROXY protocol version 1 or 2 connections.
+                       The proxy_forwarded_access is required to whitelist
+                       downstream proxies which can be trusted.
+
        If you run Squid on a dual-homed machine with an internal
        and an external interface we recommend you to specify the
        internal address:port in http_port. This way Squid will only be
        visible on the internal address.
 
 NOCOMMENT_START
 
 # Squid normally listens to port 3128
 http_port @DEFAULT_HTTP_PORT@
 NOCOMMENT_END
 DOC_END
 
 NAME: https_port
 IFDEF: USE_OPENSSL
 TYPE: PortCfg
 DEFAULT: none
 LOC: HttpsPortList
 DOC_START
        Usage:  [ip:]port cert=certificate.pem [key=key.pem] [mode] [options...]
 

=== modified file 'src/client_side.cc'
--- src/client_side.cc  2014-07-16 12:10:11 +0000
+++ src/client_side.cc  2014-07-25 12:05:51 +0000
@@ -102,40 +102,41 @@
 #include "fd.h"
 #include "fde.h"
 #include "fqdncache.h"
 #include "FwdState.h"
 #include "globals.h"
 #include "http.h"
 #include "HttpHdrContRange.h"
 #include "HttpHeaderTools.h"
 #include "HttpReply.h"
 #include "HttpRequest.h"
 #include "ident/Config.h"
 #include "ident/Ident.h"
 #include "internal.h"
 #include "ipc/FdNotes.h"
 #include "ipc/StartListening.h"
 #include "log/access_log.h"
 #include "Mem.h"
 #include "MemBuf.h"
 #include "MemObject.h"
 #include "mime_header.h"
+#include "parser/Tokenizer.h"
 #include "profiler/Profiler.h"
 #include "rfc1738.h"
 #include "SquidConfig.h"
 #include "SquidTime.h"
 #include "StatCounters.h"
 #include "StatHist.h"
 #include "Store.h"
 #include "TimeOrTag.h"
 #include "tools.h"
 #include "URL.h"
 
 #if USE_AUTH
 #include "auth/UserRequest.h"
 #endif
 #if USE_DELAY_POOLS
 #include "ClientInfo.h"
 #endif
 #if USE_OPENSSL
 #include "ssl/context_storage.h"
 #include "ssl/gadgets.h"
@@ -2322,40 +2323,42 @@
 
 #if THIS_VIOLATES_HTTP_SPECS_ON_URL_TRANSFORMATION
 
     if ((t = strchr(url, '#')))        /* remove HTML anchors */
         *t = '\0';
 
 #endif
 
     debugs(33,5, HERE << "repare absolute URL from " <<
            (csd->transparent()?"intercept":(csd->port->flags.accelSurrogate ? 
"accel":"")));
     /* Rewrite the URL in transparent or accelerator mode */
     /* NP: there are several cases to traverse here:
      *  - standard mode (forward proxy)
      *  - transparent mode (TPROXY)
      *  - transparent mode with failures
      *  - intercept mode (NAT)
      *  - intercept mode with failures
      *  - accelerator mode (reverse proxy)
      *  - internal URL
      *  - mixed combos of the above with internal URL
+     *  - remote interception with PROXY protocol
+     *  - remote reverse-proxy with PROXY protocol
      */
     if (csd->transparent()) {
         /* intercept or transparent mode, properly working with no failures */
         prepareTransparentURL(csd, http, url, req_hdr);
 
     } else if (internalCheck(url)) {
         /* internal URL mode */
         /* prepend our name & port */
         http->uri = xstrdup(internalLocalUri(NULL, url));
         // We just re-wrote the URL. Must replace the Host: header.
         //  But have not parsed there yet!! flag for local-only handling.
         http->flags.internal = true;
 
     } else if (csd->port->flags.accelSurrogate || csd->switchedToHttps()) {
         /* accelerator mode */
         prepareAcceleratedURL(csd, http, url, req_hdr);
     }
 
     if (!http->uri) {
         /* No special rewrites have been applied above, use the
@@ -2885,67 +2888,312 @@
 bool
 ConnStateData::concurrentRequestQueueFilled() const
 {
     const int existingRequestCount = getConcurrentRequestCount();
 
     // default to the configured pipeline size.
     // add 1 because the head of pipeline is counted in concurrent requests 
and not prefetch queue
     const int concurrentRequestLimit = Config.pipeline_max_prefetch + 1;
 
     // when queue filled already we cant add more.
     if (existingRequestCount >= concurrentRequestLimit) {
         debugs(33, 3, clientConnection << " max concurrent requests reached (" 
<< concurrentRequestLimit << ")");
         debugs(33, 5, clientConnection << " deferring new request until one is 
done");
         return true;
     }
 
     return false;
 }
 
 /**
+ * Perform forwarded_access ACL tests on the client which
+ * connected to PROXY protocol port to see if we trust the
+ * sender enough to accept their PROXY header claim.
+ */
+bool
+ConnStateData::proxyProtocolValidateClient()
+{
+    ACLFilledChecklist ch(Config.accessList.followXFF, NULL, 
clientConnection->rfc931);
+    ch.src_addr = clientConnection->remote;
+    ch.my_addr = clientConnection->local;
+    ch.conn(this);
+
+    if (ch.fastCheck() != ACCESS_ALLOWED)
+        return proxyProtocolError("PROXY client not permitted by ACLs");
+
+    return true;
+}
+
+/**
+ * Perform cleanup on PROXY protocol errors.
+ * If header parsing hits a fatal error terminate the connection,
+ * otherwise wait for more data.
+ */
+bool
+ConnStateData::proxyProtocolError(const char *msg)
+{
+    if (msg) {
+        debugs(33, 2, msg << " from " << clientConnection);
+        mustStop(msg);
+    }
+    return false;
+}
+
+/// magic octet prefix for PROXY protocol version 1
+static const SBuf Proxy10magic("PROXY ", 6);
+
+/// magic octet prefix for PROXY protocol version 2
+static const SBuf 
Proxy20magic("\x0D\x0A\x0D\x0A\x00\x0D\x0A\x51\x55\x49\x54\x0A", 12);
+
+/**
+ * Test the connection read buffer for PROXY protocol header.
+ * Version 1 and 2 header currently supported.
+ */
+bool
+ConnStateData::parseProxyProtocolHeader()
+{
+    // http://www.haproxy.org/download/1.5/doc/proxy-protocol.txt
+
+    // detect and parse PROXY protocol version 1 header
+    if (in.buf.length() > Proxy10magic.length() && 
in.buf.startsWith(Proxy10magic)) {
+         return parseProxy10();
+
+        // detect and parse PROXY protocol version 2 header
+    } else if (in.buf.length() > Proxy20magic.length() && 
in.buf.startsWith(Proxy20magic)) {
+        return parseProxy20();
+
+        // detect and terminate other protocols
+    } else if (in.buf.length() >= Proxy20magic.length()) {
+        // input other than the PROXY header is a protocol error
+        return proxyProtocolError("PROXY protocol error: invalid header");
+    }
+
+    // TODO: detect short non-magic prefixes earlier to avoid
+    // waiting for more data which may never come
+
+    // not enough bytes to parse yet.
+    return false;
+}
+
+/// parse the PROXY/1.0 protocol header from the connection read buffer
+bool
+ConnStateData::parseProxy10()
+{
+    ::Parser::Tokenizer tok(in.buf);
+    tok.skip(Proxy10magic);
+
+    SBuf tcpVersion;
+    if (!tok.prefix(tcpVersion, CharacterSet::ALPHA+CharacterSet::DIGIT))
+        return proxyProtocolError(tok.atEnd()?"PROXY/1.0 error: invalid 
protocol family":NULL);
+
+    if (!tcpVersion.cmp("UNKNOWN")) {
+        // skip to first LF (assumes it is part of CRLF)
+        const SBuf::size_type pos = in.buf.findFirstOf(CharacterSet::LF);
+        if (pos != SBuf::npos) {
+            if (in.buf[pos-1] != '\r')
+                return proxyProtocolError("PROXY/1.0 error: missing CR");
+            // found valid but unusable header
+            in.buf.consume(pos);
+            needProxyProtocolHeader_ = false;
+            return true;
+        }
+        // else, no LF found
+
+        // protocol error only if there are more than 107 bytes prefix header
+        return proxyProtocolError(in.buf.length() > 107? "PROXY/1.0 error: 
missing CRLF":NULL);
+
+    } else if (!tcpVersion.cmp("TCP",3)) {
+
+        // skip SP after protocol version
+        if (!tok.skip(' '))
+            return proxyProtocolError(tok.atEnd()?"PROXY/1.0 error: missing 
SP":NULL);
+
+        SBuf ipa, ipb;
+        int64_t porta, portb;
+        const CharacterSet ipChars =  CharacterSet("IP Address",".:") + 
CharacterSet::HEXDIG;
+
+        // parse  src-IP SP dst-IP SP src-port SP dst-port CRLF
+        if (!tok.prefix(ipa, ipChars) || !tok.skip(' ') ||
+           !tok.prefix(ipb, ipChars) || !tok.skip(' ') ||
+           !tok.int64(porta) || !tok.skip(' ') ||
+           !tok.int64(portb) || !tok.skip('\r') || !tok.skip('\n'))
+            return proxyProtocolError(!tok.atEnd()?"PROXY/1.0 error: invalid 
syntax":NULL);
+
+        in.buf = tok.remaining(); // sync buffers
+        needProxyProtocolHeader_ = false; // found successfully
+
+        // parse IP and port strings
+        Ip::Address originalClient, originalDest;
+
+        if (!originalClient.GetHostByName(ipa.c_str()))
+            return proxyProtocolError("PROXY/1.0 error: invalid src-IP 
address");
+
+        if (!originalDest.GetHostByName(ipb.c_str()))
+            return proxyProtocolError("PROXY/1.0 error: invalid dst-IP 
address");
+
+        if (porta > 0 && porta <= 0xFFFF) // max uint16_t
+            originalClient.port(static_cast<uint16_t>(porta));
+        else
+            return proxyProtocolError("PROXY/1.0 error: invalid src port");
+
+        if (portb > 0 && portb <= 0xFFFF) // max uint16_t
+            originalDest.port(static_cast<uint16_t>(portb));
+        else
+            return proxyProtocolError("PROXY/1.0 error: invalid dst port");
+
+        // we have original client and destination details now
+        // replace the client connection values
+        debugs(33, 5, "PROXY/1.0 protocol on connection " << clientConnection);
+        clientConnection->local = originalDest;
+        clientConnection->remote = originalClient;
+        clientConnection->flags ^= COMM_TRANSPARENT; // prevent TPROXY 
spoofing of this new IP.
+        debugs(33, 5, "PROXY/1.0 upgrade: " << clientConnection);
+
+        // repeat fetch ensuring the new client FQDN can be logged
+        if (Config.onoff.log_fqdn)
+            fqdncache_gethostbyaddr(clientConnection->remote, 
FQDN_LOOKUP_IF_MISS);
+
+        return true;
+    }
+
+    return false;
+}
+
+/// parse the PROXY/2.0 protocol header from the connection read buffer
+bool
+ConnStateData::parseProxy20()
+{
+    if ((in.buf[0] & 0xF0) != 0x20) // version == 2 is mandatory
+        return proxyProtocolError("PROXY/2.0 error: invalid version");
+
+    const char command = (in.buf[0] & 0x0F);
+    if ((command & 0xFE) != 0x00) // values other than 0x0-0x1 are invalid
+        return proxyProtocolError("PROXY/2.0 error: invalid command");
+
+    const char family = (in.buf[1] & 0xF0) >>4;
+    if (family > 0x3) // values other than 0x0-0x3 are invalid
+        return proxyProtocolError("PROXY/2.0 error: invalid family");
+
+    const char proto = (in.buf[1] & 0x0F);
+    if (proto > 0x2) // values other than 0x0-0x2 are invalid
+        return proxyProtocolError("PROXY/2.0 error: invalid protocol type");
+
+    const char *clen = in.buf.rawContent() + Proxy20magic.length() + 2;
+    const uint16_t len = ntohs(*(reinterpret_cast<const uint16_t *>(clen)));
+
+    if (in.buf.length() < Proxy20magic.length() + 4 + len)
+        return false; // need more bytes
+
+    in.buf.consume(Proxy20magic.length() + 4); // 4 being the extra bytes
+    const SBuf extra = in.buf.consume(len);
+    needProxyProtocolHeader_ = false; // found successfully
+
+    // LOCAL connections do nothing with the extras
+    if (command == 0x00/* LOCAL*/)
+        return true;
+
+    typedef union proxy_addr {
+        struct {        /* for TCP/UDP over IPv4, len = 12 */
+            struct in_addr src_addr;
+            struct in_addr dst_addr;
+            uint16_t src_port;
+            uint16_t dst_port;
+        } ipv4_addr;
+        struct {        /* for TCP/UDP over IPv6, len = 36 */
+             struct in6_addr src_addr;
+             struct in6_addr dst_addr;
+             uint16_t src_port;
+             uint16_t dst_port;
+        } ipv6_addr;
+#if NOT_SUPPORTED
+        struct {        /* for AF_UNIX sockets, len = 216 */
+             uint8_t src_addr[108];
+             uint8_t dst_addr[108];
+        } unix_addr;
+#endif
+    } pax;
+
+    const pax *ipu = reinterpret_cast<const pax*>(extra.rawContent());
+
+    // replace the client connection values
+    debugs(33, 5, "PROXY/2.0 protocol on connection " << clientConnection);
+    switch (family)
+    {
+    case 0x1: // IPv4
+        clientConnection->local = ipu->ipv4_addr.dst_addr;
+        clientConnection->local.port(ntohs(ipu->ipv4_addr.dst_port));
+        clientConnection->remote = ipu->ipv4_addr.src_addr;
+        clientConnection->remote.port(ntohs(ipu->ipv4_addr.src_port));
+        clientConnection->flags ^= COMM_TRANSPARENT; // prevent TPROXY 
spoofing of this new IP.
+        break;
+    case 0x2: // IPv6
+        clientConnection->local = ipu->ipv6_addr.dst_addr;
+        clientConnection->local.port(ntohs(ipu->ipv6_addr.dst_port));
+        clientConnection->remote = ipu->ipv6_addr.src_addr;
+        clientConnection->remote.port(ntohs(ipu->ipv6_addr.src_port));
+        clientConnection->flags ^= COMM_TRANSPARENT; // prevent TPROXY 
spoofing of this new IP.
+        break;
+    default: // do nothing
+        break;
+    }
+    debugs(33, 5, "PROXY/2.0 upgrade: " << clientConnection);
+
+    // repeat fetch ensuring the new client FQDN can be logged
+    if (Config.onoff.log_fqdn)
+        fqdncache_gethostbyaddr(clientConnection->remote, FQDN_LOOKUP_IF_MISS);
+
+    return true;
+}
+
+/**
  * Attempt to parse one or more requests from the input buffer.
  * If a request is successfully parsed, even if the next request
  * is only partially parsed, it will return TRUE.
  */
 bool
 ConnStateData::clientParseRequests()
 {
     HttpRequestMethod method;
     bool parsed_req = false;
 
     debugs(33, 5, HERE << clientConnection << ": attempting to parse");
 
     // Loop while we have read bytes that are not needed for producing the body
     // On errors, bodyPipe may become nil, but readMore will be cleared
     while (!in.buf.isEmpty() && !bodyPipe && flags.readMore) {
         connStripBufferWhitespace(this);
 
         /* Don't try to parse if the buffer is empty */
         if (in.buf.isEmpty())
             break;
 
         /* Limit the number of concurrent requests */
         if (concurrentRequestQueueFilled())
             break;
 
         /* Begin the parsing */
         PROF_start(parseHttpRequest);
+
+        // try to parse the PROXY protocol header magic bytes
+        if (needProxyProtocolHeader_ && !parseProxyProtocolHeader())
+            break;
+
         HttpParserInit(&parser_, in.buf.c_str(), in.buf.length());
 
         /* Process request */
         Http::ProtocolVersion http_ver;
         ClientSocketContext *context = parseHttpRequest(this, &parser_, 
&method, &http_ver);
         PROF_stop(parseHttpRequest);
 
         /* partial or incomplete request */
         if (!context) {
             // TODO: why parseHttpRequest can just return parseHttpRequestAbort
             // (which becomes context) but checkHeaderLimits cannot?
             checkHeaderLimits();
             break;
         }
 
         /* status -1 or 1 */
         if (context) {
             debugs(33, 5, HERE << clientConnection << ": parsed a request");
             AsyncCall::Pointer timeoutCall = commCbCall(5, 4, 
"clientLifetimeTimeout",
                                              
CommTimeoutCbPtrFun(clientLifetimeTimeout, context->http));
@@ -3263,114 +3511,130 @@
         sslBumpMode(Ssl::bumpEnd),
         switchedToHttps_(false),
         sslServerBump(NULL),
 #endif
         stoppedSending_(NULL),
         stoppedReceiving_(NULL)
 {
     pinning.host = NULL;
     pinning.port = -1;
     pinning.pinned = false;
     pinning.auth = false;
     pinning.zeroReply = false;
     pinning.peer = NULL;
 
     // store the details required for creating more MasterXaction objects as 
new requests come in
     clientConnection = xact->tcpClient;
     port = xact->squidPort;
     log_addr = xact->tcpClient->remote;
     log_addr.applyMask(Config.Addrs.client_netmask);
 
-    // ensure a buffer is present for this connection
-    in.maybeMakeSpaceAvailable();
-
     if (port->disable_pmtu_discovery != DISABLE_PMTU_OFF &&
             (transparent() || port->disable_pmtu_discovery == 
DISABLE_PMTU_ALWAYS)) {
 #if defined(IP_MTU_DISCOVER) && defined(IP_PMTUDISC_DONT)
         int i = IP_PMTUDISC_DONT;
         if (setsockopt(clientConnection->fd, SOL_IP, IP_MTU_DISCOVER, &i, 
sizeof(i)) < 0)
             debugs(33, 2, "WARNING: Path MTU discovery disabling failed on " 
<< clientConnection << " : " << xstrerror());
 #else
         static bool reported = false;
 
         if (!reported) {
             debugs(33, DBG_IMPORTANT, "NOTICE: Path MTU discovery disabling is 
not supported on your platform.");
             reported = true;
         }
 #endif
     }
+}
+
+void
+ConnStateData::start()
+{
+    // ensure a buffer is present for this connection
+    in.maybeMakeSpaceAvailable();
 
     typedef CommCbMemFunT<ConnStateData, CommCloseCbParams> Dialer;
     AsyncCall::Pointer call = JobCallback(33, 5, Dialer, this, 
ConnStateData::connStateClosed);
     comm_add_close_handler(clientConnection->fd, call);
 
     if (Config.onoff.log_fqdn)
         fqdncache_gethostbyaddr(clientConnection->remote, FQDN_LOOKUP_IF_MISS);
 
 #if USE_IDENT
     if (Ident::TheConfig.identLookup) {
         ACLFilledChecklist identChecklist(Ident::TheConfig.identLookup, NULL, 
NULL);
-        identChecklist.src_addr = xact->tcpClient->remote;
-        identChecklist.my_addr = xact->tcpClient->local;
+        identChecklist.src_addr = clientConnection->remote;
+        identChecklist.my_addr = clientConnection->local;
         if (identChecklist.fastCheck() == ACCESS_ALLOWED)
-            Ident::Start(xact->tcpClient, clientIdentDone, this);
+            Ident::Start(clientConnection, clientIdentDone, this);
     }
 #endif
 
     clientdbEstablished(clientConnection->remote, 1);
 
+    needProxyProtocolHeader_ = port->flags.proxySurrogate;
+    if (needProxyProtocolHeader_) {
+        if (!proxyProtocolValidateClient()) // will close the connection on 
failure
+            return;
+    }
+
+    // prepare any child API state that is needed
+    BodyProducer::start();
+    HttpControlMsgSink::start();
+
+    // if all is well, start reading
     flags.readMore = true;
+    readSomeData();
 }
 
 /** Handle a new connection on HTTP socket. */
 void
 httpAccept(const CommAcceptCbParams &params)
 {
     MasterXaction::Pointer xact = params.xaction;
     AnyP::PortCfgPointer s = xact->squidPort;
 
     // NP: it is possible the port was reconfigured when the call or accept() 
was queued.
 
     if (params.flag != Comm::OK) {
         // Its possible the call was still queued when the client disconnected
         debugs(33, 2, "httpAccept: " << s->listenConn << ": accept failure: " 
<< xstrerr(params.xerrno));
         return;
     }
 
     debugs(33, 4, HERE << params.conn << ": accepted");
     fd_note(params.conn->fd, "client http connect");
 
     if (s->tcp_keepalive.enabled) {
         commSetTcpKeepalive(params.conn->fd, s->tcp_keepalive.idle, 
s->tcp_keepalive.interval, s->tcp_keepalive.timeout);
     }
 
     ++ incoming_sockets_accepted;
 
     // Socket is ready, setup the connection manager to start using it
     ConnStateData *connState = new ConnStateData(xact);
 
     typedef CommCbMemFunT<ConnStateData, CommTimeoutCbParams> TimeoutDialer;
     AsyncCall::Pointer timeoutCall =  JobCallback(33, 5,
                                       TimeoutDialer, connState, 
ConnStateData::requestTimeout);
     commSetConnTimeout(params.conn, Config.Timeout.request, timeoutCall);
 
-    connState->readSomeData();
+    AsyncJob::Start(connState);
 
 #if USE_DELAY_POOLS
     fd_table[params.conn->fd].clientInfo = NULL;
 
     if (Config.onoff.client_db) {
         /* it was said several times that client write limiter does not work 
if client_db is disabled */
 
         ClientDelayPools& pools(Config.ClientDelay.pools);
         ACLFilledChecklist ch(NULL, NULL, NULL);
 
         // TODO: we check early to limit error response bandwith but we
         // should recheck when we can honor delay_pool_uses_indirect
         // TODO: we should also pass the port details for myportname here.
         ch.src_addr = params.conn->remote;
         ch.my_addr = params.conn->local;
 
         for (unsigned int pool = 0; pool < pools.size(); ++pool) {
 
             /* pools require explicit 'allow' to assign a client into them */
             if (pools[pool].access) {
@@ -3524,41 +3788,41 @@
     debugs(83, 3, "clientNegotiateSSL: FD " << fd << " negotiated cipher " <<
            SSL_get_cipher(ssl));
 
     client_cert = SSL_get_peer_certificate(ssl);
 
     if (client_cert != NULL) {
         debugs(83, 3, "clientNegotiateSSL: FD " << fd <<
                " client certificate: subject: " <<
                X509_NAME_oneline(X509_get_subject_name(client_cert), 0, 0));
 
         debugs(83, 3, "clientNegotiateSSL: FD " << fd <<
                " client certificate: issuer: " <<
                X509_NAME_oneline(X509_get_issuer_name(client_cert), 0, 0));
 
         X509_free(client_cert);
     } else {
         debugs(83, 5, "clientNegotiateSSL: FD " << fd <<
                " has no certificate.");
     }
 
-    conn->readSomeData();
+    AsyncJob::Start(conn);
 }
 
 /**
  * If SSL_CTX is given, starts reading the SSL handshake.
  * Otherwise, calls switchToHttps to generate a dynamic SSL_CTX.
  */
 static void
 httpsEstablish(ConnStateData *connState,  SSL_CTX *sslContext, Ssl::BumpMode 
bumpMode)
 {
     SSL *ssl = NULL;
     assert(connState);
     const Comm::ConnectionPointer &details = connState->clientConnection;
 
     if (sslContext && !(ssl = httpsCreate(details, sslContext)))
         return;
 
     typedef CommCbMemFunT<ConnStateData, CommTimeoutCbParams> TimeoutDialer;
     AsyncCall::Pointer timeoutCall = JobCallback(33, 5, TimeoutDialer,
                                      connState, ConnStateData::requestTimeout);
     commSetConnTimeout(details, Config.Timeout.request, timeoutCall);

=== modified file 'src/client_side.h'
--- src/client_side.h   2014-07-14 09:48:47 +0000
+++ src/client_side.h   2014-07-25 12:05:51 +0000
@@ -313,40 +313,41 @@
      \param request   if it is not NULL also checks if the pinning info refers 
to the request client side HttpRequest
      \param CachePeer      if it is not NULL also check if the CachePeer is 
the pinning CachePeer
      \return          The details of the server side connection (may be closed 
if failures were present).
      */
     const Comm::ConnectionPointer validatePinnedConnection(HttpRequest 
*request, const CachePeer *peer);
     /**
      * returts the pinned CachePeer if exists, NULL otherwise
      */
     CachePeer *pinnedPeer() const {return pinning.peer;}
     bool pinnedAuth() const {return pinning.auth;}
 
     // pining related comm callbacks
     void clientPinnedConnectionClosed(const CommCloseCbParams &io);
 
     // comm callbacks
     void clientReadRequest(const CommIoCbParams &io);
     void connStateClosed(const CommCloseCbParams &io);
     void requestTimeout(const CommTimeoutCbParams &params);
 
     // AsyncJob API
+    virtual void start();
     virtual bool doneAll() const { return BodyProducer::doneAll() && false;}
     virtual void swanSong();
 
     /// Changes state so that we close the connection and quit after serving
     /// the client-side-detected error response instead of getting stuck.
     void quitAfterError(HttpRequest *request); // meant to be private
 
     /// The caller assumes responsibility for connection closure detection.
     void stopPinnedConnectionMonitoring();
 
 #if USE_OPENSSL
     /// called by FwdState when it is done bumping the server
     void httpsPeeked(Comm::ConnectionPointer serverConnection);
 
     /// Start to create dynamic SSL_CTX for host or uses static port SSL 
context.
     void getSslContextStart();
     /**
      * Done create dynamic ssl certificate.
      *
      * \param[in] isNew if generated certificate is new, so we need to add 
this certificate to storage.
@@ -382,40 +383,50 @@
 #endif
 
     /* clt_conn_tag=tag annotation access */
     const SBuf &connectionTag() const { return connectionTag_; }
     void connectionTag(const char *aTag) { connectionTag_ = aTag; }
 
 protected:
     void startDechunkingRequest();
     void finishDechunkingRequest(bool withSuccess);
     void abortChunkedRequestBody(const err_type error);
     err_type handleChunkedRequestBody(size_t &putSize);
 
     void startPinnedConnectionMonitoring();
     void clientPinnedConnectionRead(const CommIoCbParams &io);
 
 private:
     int connFinishedWithConn(int size);
     void clientAfterReadingRequests();
     bool concurrentRequestQueueFilled() const;
 
+    /* PROXY protocol functionality */
+    bool proxyProtocolValidateClient();
+    bool parseProxyProtocolHeader();
+    bool parseProxy10();
+    bool parseProxy20();
+    bool proxyProtocolError(const char *reason = NULL);
+
+    /// whether PROXY protocol header is still expected
+    bool needProxyProtocolHeader_;
+
 #if USE_AUTH
     /// some user details that can be used to perform authentication on this 
connection
     Auth::UserRequest::Pointer auth_;
 #endif
 
     HttpParser parser_;
 
     // XXX: CBDATA plays with public/private and leaves the following 
'private' fields all public... :(
 
 #if USE_OPENSSL
     bool switchedToHttps_;
     /// The SSL server host name appears in CONNECT request or the server ip 
address for the intercepted requests
     String sslConnectHostOrIp; ///< The SSL server host name as passed in the 
CONNECT request
     String sslCommonName; ///< CN name for SSL certificate generation
     String sslBumpCertKey; ///< Key to use to store/retrieve generated 
certificate
 
     /// HTTPS server cert. fetching state for bump-ssl-server-first
     Ssl::ServerBump *sslServerBump;
     Ssl::CertSignAlgorithm signAlgorithm; ///< The signing algorithm to use
 #endif

Reply via email to