]> git.ipfire.org Git - thirdparty/squid.git/blob - src/http.cc
Fix assert in StoreEntry::write() when aborted entry is written.
[thirdparty/squid.git] / src / http.cc
1
2 /*
3 * $Id$
4 *
5 * DEBUG: section 11 Hypertext Transfer Protocol (HTTP)
6 * AUTHOR: Harvest Derived
7 *
8 * SQUID Web Proxy Cache http://www.squid-cache.org/
9 * ----------------------------------------------------------
10 *
11 * Squid is the result of efforts by numerous individuals from
12 * the Internet community; see the CONTRIBUTORS file for full
13 * details. Many organizations have provided support for Squid's
14 * development; see the SPONSORS file for full details. Squid is
15 * Copyrighted (C) 2001 by the Regents of the University of
16 * California; see the COPYRIGHT file for full details. Squid
17 * incorporates software developed and/or copyrighted by other
18 * sources; see the CREDITS file for full details.
19 *
20 * This program is free software; you can redistribute it and/or modify
21 * it under the terms of the GNU General Public License as published by
22 * the Free Software Foundation; either version 2 of the License, or
23 * (at your option) any later version.
24 *
25 * This program is distributed in the hope that it will be useful,
26 * but WITHOUT ANY WARRANTY; without even the implied warranty of
27 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
28 * GNU General Public License for more details.
29 *
30 * You should have received a copy of the GNU General Public License
31 * along with this program; if not, write to the Free Software
32 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111, USA.
33 *
34 */
35
36 /*
37 * Anonymizing patch by lutz@as-node.jena.thur.de
38 * have a look into http-anon.c to get more informations.
39 */
40
41 #include "squid.h"
42
43 #include "acl/FilledChecklist.h"
44 #if USE_AUTH
45 #include "auth/UserRequest.h"
46 #endif
47 #include "base/AsyncJobCalls.h"
48 #include "base/TextException.h"
49 #include "base64.h"
50 #include "comm/Write.h"
51 #if USE_DELAY_POOLS
52 #include "DelayPools.h"
53 #endif
54 #include "err_detail_type.h"
55 #include "errorpage.h"
56 #include "http.h"
57 #include "HttpControlMsg.h"
58 #include "HttpHdrContRange.h"
59 #include "HttpHdrSc.h"
60 #include "HttpHdrScTarget.h"
61 #include "HttpReply.h"
62 #include "HttpRequest.h"
63 #include "MemBuf.h"
64 #include "MemObject.h"
65 #include "protos.h"
66 #include "rfc1738.h"
67 #include "SquidTime.h"
68 #include "Store.h"
69
70
71 #define SQUID_ENTER_THROWING_CODE() try {
72 #define SQUID_EXIT_THROWING_CODE(status) \
73 status = true; \
74 } \
75 catch (const std::exception &e) { \
76 debugs (11, 1, "Exception error:" << e.what()); \
77 status = false; \
78 }
79
80 CBDATA_CLASS_INIT(HttpStateData);
81
82 static const char *const crlf = "\r\n";
83
84 static void httpMaybeRemovePublic(StoreEntry *, http_status);
85 static void copyOneHeaderFromClientsideRequestToUpstreamRequest(const HttpHeaderEntry *e, const String strConnection, HttpRequest * request, const HttpRequest * orig_request,
86 HttpHeader * hdr_out, const int we_do_ranges, const http_state_flags);
87
88 HttpStateData::HttpStateData(FwdState *theFwdState) : AsyncJob("HttpStateData"), ServerStateData(theFwdState),
89 lastChunk(0), header_bytes_read(0), reply_bytes_read(0),
90 body_bytes_truncated(0), httpChunkDecoder(NULL)
91 {
92 debugs(11,5,HERE << "HttpStateData " << this << " created");
93 ignoreCacheControl = false;
94 surrogateNoStore = false;
95 fd = fwd->server_fd;
96 readBuf = new MemBuf;
97 readBuf->init(16*1024, 256*1024);
98 orig_request = HTTPMSGLOCK(fwd->request);
99
100 // reset peer response time stats for %<pt
101 orig_request->hier.peer_http_request_sent.tv_sec = 0;
102 orig_request->hier.peer_http_request_sent.tv_usec = 0;
103
104 if (fwd->servers)
105 _peer = fwd->servers->_peer; /* might be NULL */
106
107 if (_peer) {
108 const char *url;
109
110 if (_peer->options.originserver)
111 url = orig_request->urlpath.termedBuf();
112 else
113 url = entry->url();
114
115 HttpRequest * proxy_req = new HttpRequest(orig_request->method,
116 orig_request->protocol, url);
117
118 proxy_req->SetHost(_peer->host);
119
120 proxy_req->port = _peer->http_port;
121
122 proxy_req->flags = orig_request->flags;
123
124 proxy_req->lastmod = orig_request->lastmod;
125
126 proxy_req->flags.proxying = 1;
127
128 HTTPMSGUNLOCK(request);
129
130 request = HTTPMSGLOCK(proxy_req);
131
132 /*
133 * This NEIGHBOR_PROXY_ONLY check probably shouldn't be here.
134 * We might end up getting the object from somewhere else if,
135 * for example, the request to this neighbor fails.
136 */
137 if (_peer->options.proxy_only)
138 entry->releaseRequest();
139
140 #if USE_DELAY_POOLS
141 entry->setNoDelay(_peer->options.no_delay);
142 #endif
143 }
144
145 /*
146 * register the handler to free HTTP state data when the FD closes
147 */
148 typedef CommCbMemFunT<HttpStateData, CommCloseCbParams> Dialer;
149 closeHandler = JobCallback(9, 5,
150 Dialer, this, HttpStateData::httpStateConnClosed);
151 comm_add_close_handler(fd, closeHandler);
152 }
153
154 HttpStateData::~HttpStateData()
155 {
156 /*
157 * don't forget that ~ServerStateData() gets called automatically
158 */
159
160 if (!readBuf->isNull())
161 readBuf->clean();
162
163 delete readBuf;
164
165 if (httpChunkDecoder)
166 delete httpChunkDecoder;
167
168 HTTPMSGUNLOCK(orig_request);
169
170 debugs(11,5, HERE << "HttpStateData " << this << " destroyed; FD " << fd);
171 }
172
173 int
174 HttpStateData::dataDescriptor() const
175 {
176 return fd;
177 }
178 /*
179 static void
180 httpStateFree(int fd, void *data)
181 {
182 HttpStateData *httpState = static_cast<HttpStateData *>(data);
183 debugs(11, 5, "httpStateFree: FD " << fd << ", httpState=" << data);
184 delete httpState;
185 }*/
186
187 void
188 HttpStateData::httpStateConnClosed(const CommCloseCbParams &params)
189 {
190 debugs(11, 5, "httpStateFree: FD " << params.fd << ", httpState=" << params.data);
191 deleteThis("HttpStateData::httpStateConnClosed");
192 }
193
194 int
195 httpCachable(const HttpRequestMethod& method)
196 {
197 /* GET and HEAD are cachable. Others are not. */
198
199 // TODO: replase to HttpRequestMethod::isCachable() ?
200 if (method != METHOD_GET && method != METHOD_HEAD)
201 return 0;
202
203 /* else cachable */
204 return 1;
205 }
206
207 void
208 HttpStateData::httpTimeout(const CommTimeoutCbParams &params)
209 {
210 debugs(11, 4, "httpTimeout: FD " << fd << ": '" << entry->url() << "'" );
211
212 if (entry->store_status == STORE_PENDING) {
213 fwd->fail(errorCon(ERR_READ_TIMEOUT, HTTP_GATEWAY_TIMEOUT, fwd->request));
214 }
215
216 comm_close(fd);
217 }
218
219 static void
220 httpMaybeRemovePublic(StoreEntry * e, http_status status)
221 {
222 int remove = 0;
223 int forbidden = 0;
224 StoreEntry *pe;
225
226 if (!EBIT_TEST(e->flags, KEY_PRIVATE))
227 return;
228
229 switch (status) {
230
231 case HTTP_OK:
232
233 case HTTP_NON_AUTHORITATIVE_INFORMATION:
234
235 case HTTP_MULTIPLE_CHOICES:
236
237 case HTTP_MOVED_PERMANENTLY:
238
239 case HTTP_MOVED_TEMPORARILY:
240
241 case HTTP_GONE:
242
243 case HTTP_NOT_FOUND:
244 remove = 1;
245
246 break;
247
248 case HTTP_FORBIDDEN:
249
250 case HTTP_METHOD_NOT_ALLOWED:
251 forbidden = 1;
252
253 break;
254
255 #if WORK_IN_PROGRESS
256
257 case HTTP_UNAUTHORIZED:
258 forbidden = 1;
259
260 break;
261
262 #endif
263
264 default:
265 #if QUESTIONABLE
266 /*
267 * Any 2xx response should eject previously cached entities...
268 */
269
270 if (status >= 200 && status < 300)
271 remove = 1;
272
273 #endif
274
275 break;
276 }
277
278 if (!remove && !forbidden)
279 return;
280
281 assert(e->mem_obj);
282
283 if (e->mem_obj->request)
284 pe = storeGetPublicByRequest(e->mem_obj->request);
285 else
286 pe = storeGetPublic(e->mem_obj->url, e->mem_obj->method);
287
288 if (pe != NULL) {
289 assert(e != pe);
290 #if USE_HTCP
291 neighborsHtcpClear(e, NULL, e->mem_obj->request, e->mem_obj->method, HTCP_CLR_INVALIDATION);
292 #endif
293 pe->release();
294 }
295
296 /** \par
297 * Also remove any cached HEAD response in case the object has
298 * changed.
299 */
300 if (e->mem_obj->request)
301 pe = storeGetPublicByRequestMethod(e->mem_obj->request, METHOD_HEAD);
302 else
303 pe = storeGetPublic(e->mem_obj->url, METHOD_HEAD);
304
305 if (pe != NULL) {
306 assert(e != pe);
307 #if USE_HTCP
308 neighborsHtcpClear(e, NULL, e->mem_obj->request, HttpRequestMethod(METHOD_HEAD), HTCP_CLR_INVALIDATION);
309 #endif
310 pe->release();
311 }
312 }
313
314 void
315 HttpStateData::processSurrogateControl(HttpReply *reply)
316 {
317 if (request->flags.accelerated && reply->surrogate_control) {
318 HttpHdrScTarget *sctusable = httpHdrScGetMergedTarget(reply->surrogate_control, Config.Accel.surrogate_id);
319
320 if (sctusable) {
321 if (EBIT_TEST(sctusable->mask, SC_NO_STORE) ||
322 (Config.onoff.surrogate_is_remote
323 && EBIT_TEST(sctusable->mask, SC_NO_STORE_REMOTE))) {
324 surrogateNoStore = true;
325 entry->makePrivate();
326 }
327
328 /* The HttpHeader logic cannot tell if the header it's parsing is a reply to an
329 * accelerated request or not...
330 * Still, this is an abstraction breach. - RC
331 */
332 if (sctusable->max_age != -1) {
333 if (sctusable->max_age < sctusable->max_stale)
334 reply->expires = reply->date + sctusable->max_age;
335 else
336 reply->expires = reply->date + sctusable->max_stale;
337
338 /* And update the timestamps */
339 entry->timestampsSet();
340 }
341
342 /* We ignore cache-control directives as per the Surrogate specification */
343 ignoreCacheControl = true;
344
345 httpHdrScTargetDestroy(sctusable);
346 }
347 }
348 }
349
350 int
351 HttpStateData::cacheableReply()
352 {
353 HttpReply const *rep = finalReply();
354 HttpHeader const *hdr = &rep->header;
355 const int cc_mask = (rep->cache_control) ? rep->cache_control->mask : 0;
356 const char *v;
357 #if USE_HTTP_VIOLATIONS
358
359 const refresh_t *R = NULL;
360
361 /* This strange looking define first looks up the refresh pattern
362 * and then checks if the specified flag is set. The main purpose
363 * of this is to simplify the refresh pattern lookup and USE_HTTP_VIOLATIONS
364 * condition
365 */
366 #define REFRESH_OVERRIDE(flag) \
367 ((R = (R ? R : refreshLimits(entry->mem_obj->url))) , \
368 (R && R->flags.flag))
369 #else
370 #define REFRESH_OVERRIDE(flag) 0
371 #endif
372
373 if (surrogateNoStore)
374 return 0;
375
376 // RFC 2616: do not cache replies to responses with no-store CC directive
377 if (request && request->cache_control &&
378 EBIT_TEST(request->cache_control->mask, CC_NO_STORE) &&
379 !REFRESH_OVERRIDE(ignore_no_store))
380 return 0;
381
382 if (!ignoreCacheControl) {
383 if (EBIT_TEST(cc_mask, CC_PRIVATE)) {
384 if (!REFRESH_OVERRIDE(ignore_private))
385 return 0;
386 }
387
388 if (EBIT_TEST(cc_mask, CC_NO_CACHE)) {
389 if (!REFRESH_OVERRIDE(ignore_no_cache))
390 return 0;
391 }
392
393 if (EBIT_TEST(cc_mask, CC_NO_STORE)) {
394 if (!REFRESH_OVERRIDE(ignore_no_store))
395 return 0;
396 }
397 }
398
399 if (request->flags.auth || request->flags.auth_sent) {
400 /*
401 * Responses to requests with authorization may be cached
402 * only if a Cache-Control: public reply header is present.
403 * RFC 2068, sec 14.9.4
404 */
405
406 if (!EBIT_TEST(cc_mask, CC_PUBLIC)) {
407 if (!REFRESH_OVERRIDE(ignore_auth))
408 return 0;
409 }
410 }
411
412 /* Pragma: no-cache in _replies_ is not documented in HTTP,
413 * but servers like "Active Imaging Webcast/2.0" sure do use it */
414 if (hdr->has(HDR_PRAGMA)) {
415 String s = hdr->getList(HDR_PRAGMA);
416 const int no_cache = strListIsMember(&s, "no-cache", ',');
417 s.clean();
418
419 if (no_cache) {
420 if (!REFRESH_OVERRIDE(ignore_no_cache))
421 return 0;
422 }
423 }
424
425 /*
426 * The "multipart/x-mixed-replace" content type is used for
427 * continuous push replies. These are generally dynamic and
428 * probably should not be cachable
429 */
430 if ((v = hdr->getStr(HDR_CONTENT_TYPE)))
431 if (!strncasecmp(v, "multipart/x-mixed-replace", 25))
432 return 0;
433
434 switch (rep->sline.status) {
435 /* Responses that are cacheable */
436
437 case HTTP_OK:
438
439 case HTTP_NON_AUTHORITATIVE_INFORMATION:
440
441 case HTTP_MULTIPLE_CHOICES:
442
443 case HTTP_MOVED_PERMANENTLY:
444
445 case HTTP_GONE:
446 /*
447 * Don't cache objects that need to be refreshed on next request,
448 * unless we know how to refresh it.
449 */
450
451 if (!refreshIsCachable(entry) && !REFRESH_OVERRIDE(store_stale)) {
452 debugs(22, 3, "refreshIsCachable() returned non-cacheable..");
453 return 0;
454 } else
455 return 1;
456
457 /* NOTREACHED */
458 break;
459
460 /* Responses that only are cacheable if the server says so */
461
462 case HTTP_MOVED_TEMPORARILY:
463 case HTTP_TEMPORARY_REDIRECT:
464 if (rep->expires > rep->date && rep->date > 0)
465 return 1;
466 else
467 return 0;
468
469 /* NOTREACHED */
470 break;
471
472 /* Errors can be negatively cached */
473
474 case HTTP_NO_CONTENT:
475
476 case HTTP_USE_PROXY:
477
478 case HTTP_BAD_REQUEST:
479
480 case HTTP_FORBIDDEN:
481
482 case HTTP_NOT_FOUND:
483
484 case HTTP_METHOD_NOT_ALLOWED:
485
486 case HTTP_REQUEST_URI_TOO_LARGE:
487
488 case HTTP_INTERNAL_SERVER_ERROR:
489
490 case HTTP_NOT_IMPLEMENTED:
491
492 case HTTP_BAD_GATEWAY:
493
494 case HTTP_SERVICE_UNAVAILABLE:
495
496 case HTTP_GATEWAY_TIMEOUT:
497 return -1;
498
499 /* NOTREACHED */
500 break;
501
502 /* Some responses can never be cached */
503
504 case HTTP_PARTIAL_CONTENT: /* Not yet supported */
505
506 case HTTP_SEE_OTHER:
507
508 case HTTP_NOT_MODIFIED:
509
510 case HTTP_UNAUTHORIZED:
511
512 case HTTP_PROXY_AUTHENTICATION_REQUIRED:
513
514 case HTTP_INVALID_HEADER: /* Squid header parsing error */
515
516 case HTTP_HEADER_TOO_LARGE:
517
518 case HTTP_PAYMENT_REQUIRED:
519 case HTTP_NOT_ACCEPTABLE:
520 case HTTP_REQUEST_TIMEOUT:
521 case HTTP_CONFLICT:
522 case HTTP_LENGTH_REQUIRED:
523 case HTTP_PRECONDITION_FAILED:
524 case HTTP_REQUEST_ENTITY_TOO_LARGE:
525 case HTTP_UNSUPPORTED_MEDIA_TYPE:
526 case HTTP_UNPROCESSABLE_ENTITY:
527 case HTTP_LOCKED:
528 case HTTP_FAILED_DEPENDENCY:
529 case HTTP_INSUFFICIENT_STORAGE:
530 case HTTP_REQUESTED_RANGE_NOT_SATISFIABLE:
531 case HTTP_EXPECTATION_FAILED:
532
533 return 0;
534
535 default:
536 /* RFC 2616 section 6.1.1: an unrecognized response MUST NOT be cached. */
537 debugs (11, 3, HERE << "Unknown HTTP status code " << rep->sline.status << ". Not cacheable.");
538
539 return 0;
540
541 /* NOTREACHED */
542 break;
543 }
544
545 /* NOTREACHED */
546 }
547
548 /*
549 * For Vary, store the relevant request headers as
550 * virtual headers in the reply
551 * Returns false if the variance cannot be stored
552 */
553 const char *
554 httpMakeVaryMark(HttpRequest * request, HttpReply const * reply)
555 {
556 String vary, hdr;
557 const char *pos = NULL;
558 const char *item;
559 const char *value;
560 int ilen;
561 static String vstr;
562
563 vstr.clean();
564 vary = reply->header.getList(HDR_VARY);
565
566 while (strListGetItem(&vary, ',', &item, &ilen, &pos)) {
567 char *name = (char *)xmalloc(ilen + 1);
568 xstrncpy(name, item, ilen + 1);
569 Tolower(name);
570
571 if (strcmp(name, "*") == 0) {
572 /* Can not handle "Vary: *" withtout ETag support */
573 safe_free(name);
574 vstr.clean();
575 break;
576 }
577
578 strListAdd(&vstr, name, ',');
579 hdr = request->header.getByName(name);
580 safe_free(name);
581 value = hdr.termedBuf();
582
583 if (value) {
584 value = rfc1738_escape_part(value);
585 vstr.append("=\"", 2);
586 vstr.append(value);
587 vstr.append("\"", 1);
588 }
589
590 hdr.clean();
591 }
592
593 vary.clean();
594 #if X_ACCELERATOR_VARY
595
596 pos = NULL;
597 vary = reply->header.getList(HDR_X_ACCELERATOR_VARY);
598
599 while (strListGetItem(&vary, ',', &item, &ilen, &pos)) {
600 char *name = (char *)xmalloc(ilen + 1);
601 xstrncpy(name, item, ilen + 1);
602 Tolower(name);
603 strListAdd(&vstr, name, ',');
604 hdr = request->header.getByName(name);
605 safe_free(name);
606 value = hdr.termedBuf();
607
608 if (value) {
609 value = rfc1738_escape_part(value);
610 vstr.append("=\"", 2);
611 vstr.append(value);
612 vstr.append("\"", 1);
613 }
614
615 hdr.clean();
616 }
617
618 vary.clean();
619 #endif
620
621 debugs(11, 3, "httpMakeVaryMark: " << vstr);
622 return vstr.termedBuf();
623 }
624
625 void
626 HttpStateData::keepaliveAccounting(HttpReply *reply)
627 {
628 if (flags.keepalive)
629 if (_peer)
630 _peer->stats.n_keepalives_sent++;
631
632 if (reply->keep_alive) {
633 if (_peer)
634 _peer->stats.n_keepalives_recv++;
635
636 if (Config.onoff.detect_broken_server_pconns
637 && reply->bodySize(request->method) == -1 && !flags.chunked) {
638 debugs(11, 1, "keepaliveAccounting: Impossible keep-alive header from '" << entry->url() << "'" );
639 // debugs(11, 2, "GOT HTTP REPLY HDR:\n---------\n" << readBuf->content() << "\n----------" );
640 flags.keepalive_broken = 1;
641 }
642 }
643 }
644
645 void
646 HttpStateData::checkDateSkew(HttpReply *reply)
647 {
648 if (reply->date > -1 && !_peer) {
649 int skew = abs((int)(reply->date - squid_curtime));
650
651 if (skew > 86400)
652 debugs(11, 3, "" << request->GetHost() << "'s clock is skewed by " << skew << " seconds!");
653 }
654 }
655
656 /**
657 * This creates the error page itself.. its likely
658 * that the forward ported reply header max size patch
659 * generates non http conformant error pages - in which
660 * case the errors where should be 'BAD_GATEWAY' etc
661 */
662 void
663 HttpStateData::processReplyHeader()
664 {
665 /** Creates a blank header. If this routine is made incremental, this will not do */
666
667 /* NP: all exit points to this function MUST call ctx_exit(ctx) */
668 Ctx ctx = ctx_enter(entry->mem_obj->url);
669
670 debugs(11, 3, "processReplyHeader: key '" << entry->getMD5Text() << "'");
671
672 assert(!flags.headers_parsed);
673
674 if (!readBuf->hasContent()) {
675 ctx_exit(ctx);
676 return;
677 }
678
679 http_status error = HTTP_STATUS_NONE;
680
681 HttpReply *newrep = new HttpReply;
682 const bool parsed = newrep->parse(readBuf, eof, &error);
683
684 if (!parsed && readBuf->contentSize() > 5 && strncmp(readBuf->content(), "HTTP/", 5) != 0 && strncmp(readBuf->content(), "ICY", 3) != 0) {
685 MemBuf *mb;
686 HttpReply *tmprep = new HttpReply;
687 tmprep->setHeaders(HTTP_OK, "Gatewaying", NULL, -1, -1, -1);
688 tmprep->header.putExt("X-Transformed-From", "HTTP/0.9");
689 mb = tmprep->pack();
690 newrep->parse(mb, eof, &error);
691 delete mb;
692 delete tmprep;
693 } else {
694 if (!parsed && error > 0) { // unrecoverable parsing error
695 debugs(11, 3, "processReplyHeader: Non-HTTP-compliant header: '" << readBuf->content() << "'");
696 flags.headers_parsed = 1;
697 newrep->sline.version = HttpVersion(1,1);
698 newrep->sline.status = error;
699 HttpReply *vrep = setVirginReply(newrep);
700 entry->replaceHttpReply(vrep);
701 ctx_exit(ctx);
702 return;
703 }
704
705 if (!parsed) { // need more data
706 assert(!error);
707 assert(!eof);
708 delete newrep;
709 ctx_exit(ctx);
710 return;
711 }
712
713 debugs(11, 9, "GOT HTTP REPLY HDR:\n---------\n" << readBuf->content() << "\n----------");
714
715 header_bytes_read = headersEnd(readBuf->content(), readBuf->contentSize());
716 readBuf->consume(header_bytes_read);
717 }
718
719 newrep->removeStaleWarnings();
720
721 if (newrep->sline.protocol == AnyP::PROTO_HTTP && newrep->sline.status >= 100 && newrep->sline.status < 200) {
722 handle1xx(newrep);
723 ctx_exit(ctx);
724 return;
725 }
726
727 flags.chunked = 0;
728 if (newrep->sline.protocol == AnyP::PROTO_HTTP && newrep->header.chunked()) {
729 flags.chunked = 1;
730 httpChunkDecoder = new ChunkedCodingParser;
731 }
732
733 if (!peerSupportsConnectionPinning())
734 orig_request->flags.connection_auth_disabled = 1;
735
736 HttpReply *vrep = setVirginReply(newrep);
737 flags.headers_parsed = 1;
738
739 keepaliveAccounting(vrep);
740
741 checkDateSkew(vrep);
742
743 processSurrogateControl (vrep);
744
745 /** \todo IF the reply is a 1.0 reply, AND it has a Connection: Header
746 * Parse the header and remove all referenced headers
747 */
748
749 orig_request->hier.peer_reply_status = newrep->sline.status;
750
751 ctx_exit(ctx);
752 }
753
754 /// ignore or start forwarding the 1xx response (a.k.a., control message)
755 void
756 HttpStateData::handle1xx(HttpReply *reply)
757 {
758 HttpMsgPointerT<HttpReply> msg(reply); // will destroy reply if unused
759
760 // one 1xx at a time: we must not be called while waiting for previous 1xx
761 Must(!flags.handling1xx);
762 flags.handling1xx = true;
763
764 if (!orig_request->canHandle1xx()) {
765 debugs(11, 2, HERE << "ignoring client-unsupported 1xx");
766 proceedAfter1xx();
767 return;
768 }
769
770 #if USE_HTTP_VIOLATIONS
771 // check whether the 1xx response forwarding is allowed by squid.conf
772 if (Config.accessList.reply) {
773 ACLFilledChecklist ch(Config.accessList.reply, originalRequest(), NULL);
774 ch.reply = HTTPMSGLOCK(reply);
775 if (!ch.fastCheck()) { // TODO: support slow lookups?
776 debugs(11, 3, HERE << "ignoring denied 1xx");
777 proceedAfter1xx();
778 return;
779 }
780 }
781 #endif // USE_HTTP_VIOLATIONS
782
783 debugs(11, 2, HERE << "forwarding 1xx to client");
784
785 // the Sink will use this to call us back after writing 1xx to the client
786 typedef NullaryMemFunT<HttpStateData> CbDialer;
787 const AsyncCall::Pointer cb = JobCallback(11, 3, CbDialer, this,
788 HttpStateData::proceedAfter1xx);
789 CallJobHere1(11, 4, orig_request->clientConnectionManager, ConnStateData,
790 ConnStateData::sendControlMsg, HttpControlMsg(msg, cb));
791 // If the call is not fired, then the Sink is gone, and HttpStateData
792 // will terminate due to an aborted store entry or another similar error.
793 // If we get stuck, it is not handle1xx fault if we could get stuck
794 // for similar reasons without a 1xx response.
795 }
796
797 /// restores state and resumes processing after 1xx is ignored or forwarded
798 void
799 HttpStateData::proceedAfter1xx()
800 {
801 Must(flags.handling1xx);
802
803 debugs(11, 2, HERE << "consuming " << header_bytes_read <<
804 " header and " << reply_bytes_read << " body bytes read after 1xx");
805 header_bytes_read = 0;
806 reply_bytes_read = 0;
807
808 CallJobHere(11, 3, this, HttpStateData, HttpStateData::processReply);
809 }
810
811
812 /**
813 * returns true if the peer can support connection pinning
814 */
815 bool HttpStateData::peerSupportsConnectionPinning() const
816 {
817 const HttpReply *rep = entry->mem_obj->getReply();
818 const HttpHeader *hdr = &rep->header;
819 bool rc;
820 String header;
821
822 if (!_peer)
823 return true;
824
825 /*If this peer does not support connection pinning (authenticated
826 connections) return false
827 */
828 if (!_peer->connection_auth)
829 return false;
830
831 /*The peer supports connection pinning and the http reply status
832 is not unauthorized, so the related connection can be pinned
833 */
834 if (rep->sline.status != HTTP_UNAUTHORIZED)
835 return true;
836
837 /*The server respond with HTTP_UNAUTHORIZED and the peer configured
838 with "connection-auth=on" we know that the peer supports pinned
839 connections
840 */
841 if (_peer->connection_auth == 1)
842 return true;
843
844 /*At this point peer has configured with "connection-auth=auto"
845 parameter so we need some extra checks to decide if we are going
846 to allow pinned connections or not
847 */
848
849 /*if the peer configured with originserver just allow connection
850 pinning (squid 2.6 behaviour)
851 */
852 if (_peer->options.originserver)
853 return true;
854
855 /*if the connections it is already pinned it is OK*/
856 if (request->flags.pinned)
857 return true;
858
859 /*Allow pinned connections only if the Proxy-support header exists in
860 reply and has in its list the "Session-Based-Authentication"
861 which means that the peer supports connection pinning.
862 */
863 if (!hdr->has(HDR_PROXY_SUPPORT))
864 return false;
865
866 header = hdr->getStrOrList(HDR_PROXY_SUPPORT);
867 /* XXX This ought to be done in a case-insensitive manner */
868 rc = (strstr(header.termedBuf(), "Session-Based-Authentication") != NULL);
869
870 return rc;
871 }
872
873 // Called when we parsed (and possibly adapted) the headers but
874 // had not starting storing (a.k.a., sending) the body yet.
875 void
876 HttpStateData::haveParsedReplyHeaders()
877 {
878 ServerStateData::haveParsedReplyHeaders();
879
880 Ctx ctx = ctx_enter(entry->mem_obj->url);
881 HttpReply *rep = finalReply();
882
883 if (rep->sline.status == HTTP_PARTIAL_CONTENT &&
884 rep->content_range)
885 currentOffset = rep->content_range->spec.offset;
886
887 entry->timestampsSet();
888
889 /* Check if object is cacheable or not based on reply code */
890 debugs(11, 3, "haveParsedReplyHeaders: HTTP CODE: " << rep->sline.status);
891
892 if (neighbors_do_private_keys)
893 httpMaybeRemovePublic(entry, rep->sline.status);
894
895 if (rep->header.has(HDR_VARY)
896 #if X_ACCELERATOR_VARY
897 || rep->header.has(HDR_X_ACCELERATOR_VARY)
898 #endif
899 ) {
900 const char *vary = httpMakeVaryMark(orig_request, rep);
901
902 if (!vary) {
903 entry->makePrivate();
904 if (!fwd->reforwardableStatus(rep->sline.status))
905 EBIT_CLR(entry->flags, ENTRY_FWD_HDR_WAIT);
906 goto no_cache;
907 }
908
909 entry->mem_obj->vary_headers = xstrdup(vary);
910 }
911
912 /*
913 * If its not a reply that we will re-forward, then
914 * allow the client to get it.
915 */
916 if (!fwd->reforwardableStatus(rep->sline.status))
917 EBIT_CLR(entry->flags, ENTRY_FWD_HDR_WAIT);
918
919 switch (cacheableReply()) {
920
921 case 1:
922 entry->makePublic();
923 break;
924
925 case 0:
926 entry->makePrivate();
927 break;
928
929 case -1:
930
931 #if USE_HTTP_VIOLATIONS
932 if (Config.negativeTtl > 0)
933 entry->cacheNegatively();
934 else
935 #endif
936 entry->makePrivate();
937
938 break;
939
940 default:
941 assert(0);
942
943 break;
944 }
945
946 no_cache:
947
948 if (!ignoreCacheControl && rep->cache_control) {
949 if (EBIT_TEST(rep->cache_control->mask, CC_PROXY_REVALIDATE) ||
950 EBIT_TEST(rep->cache_control->mask, CC_MUST_REVALIDATE) ||
951 EBIT_TEST(rep->cache_control->mask, CC_S_MAXAGE))
952 EBIT_SET(entry->flags, ENTRY_REVALIDATE);
953 }
954
955 #if HEADERS_LOG
956 headersLog(1, 0, request->method, rep);
957
958 #endif
959
960 ctx_exit(ctx);
961 }
962
963 HttpStateData::ConnectionStatus
964 HttpStateData::statusIfComplete() const
965 {
966 const HttpReply *rep = virginReply();
967 /** \par
968 * If the reply wants to close the connection, it takes precedence */
969
970 if (httpHeaderHasConnDir(&rep->header, "close"))
971 return COMPLETE_NONPERSISTENT_MSG;
972
973 /** \par
974 * If we didn't send a keep-alive request header, then this
975 * can not be a persistent connection.
976 */
977 if (!flags.keepalive)
978 return COMPLETE_NONPERSISTENT_MSG;
979
980 /** \par
981 * If we haven't sent the whole request then this can not be a persistent
982 * connection.
983 */
984 if (!flags.request_sent) {
985 debugs(11, 2, "statusIfComplete: Request not yet fully sent \"" << RequestMethodStr(orig_request->method) << " " << entry->url() << "\"" );
986 return COMPLETE_NONPERSISTENT_MSG;
987 }
988
989 /** \par
990 * What does the reply have to say about keep-alive?
991 */
992 /**
993 \bug XXX BUG?
994 * If the origin server (HTTP/1.0) does not send a keep-alive
995 * header, but keeps the connection open anyway, what happens?
996 * We'll return here and http.c waits for an EOF before changing
997 * store_status to STORE_OK. Combine this with ENTRY_FWD_HDR_WAIT
998 * and an error status code, and we might have to wait until
999 * the server times out the socket.
1000 */
1001 if (!rep->keep_alive)
1002 return COMPLETE_NONPERSISTENT_MSG;
1003
1004 return COMPLETE_PERSISTENT_MSG;
1005 }
1006
1007 HttpStateData::ConnectionStatus
1008 HttpStateData::persistentConnStatus() const
1009 {
1010 debugs(11, 3, "persistentConnStatus: FD " << fd << " eof=" << eof);
1011 if (eof) // already reached EOF
1012 return COMPLETE_NONPERSISTENT_MSG;
1013
1014 /* If server fd is closing (but we have not been notified yet), stop Comm
1015 I/O to avoid assertions. TODO: Change Comm API to handle callers that
1016 want more I/O after async closing (usually initiated by others). */
1017 // XXX: add canReceive or s/canSend/canTalkToServer/
1018 if (!canSend(fd))
1019 return COMPLETE_NONPERSISTENT_MSG;
1020
1021 /** \par
1022 * In chunked response we do not know the content length but we are absolutely
1023 * sure about the end of response, so we are calling the statusIfComplete to
1024 * decide if we can be persistant
1025 */
1026 if (lastChunk && flags.chunked)
1027 return statusIfComplete();
1028
1029 const HttpReply *vrep = virginReply();
1030 debugs(11, 5, "persistentConnStatus: content_length=" << vrep->content_length);
1031
1032 const int64_t clen = vrep->bodySize(request->method);
1033
1034 debugs(11, 5, "persistentConnStatus: clen=" << clen);
1035
1036 /* If the body size is unknown we must wait for EOF */
1037 if (clen < 0)
1038 return INCOMPLETE_MSG;
1039
1040 /** \par
1041 * If the body size is known, we must wait until we've gotten all of it. */
1042 if (clen > 0) {
1043 // old technique:
1044 // if (entry->mem_obj->endOffset() < vrep->content_length + vrep->hdr_sz)
1045 const int64_t body_bytes_read = reply_bytes_read - header_bytes_read;
1046 debugs(11,5, "persistentConnStatus: body_bytes_read=" <<
1047 body_bytes_read << " content_length=" << vrep->content_length);
1048
1049 if (body_bytes_read < vrep->content_length)
1050 return INCOMPLETE_MSG;
1051
1052 if (body_bytes_truncated > 0) // already read more than needed
1053 return COMPLETE_NONPERSISTENT_MSG; // disable pconns
1054 }
1055
1056 /** \par
1057 * If there is no message body or we got it all, we can be persistent */
1058 return statusIfComplete();
1059 }
1060
1061 /*
1062 * This is the callback after some data has been read from the network
1063 */
1064 /*
1065 void
1066 HttpStateData::ReadReplyWrapper(int fd, char *buf, size_t len, comm_err_t flag, int xerrno, void *data)
1067 {
1068 HttpStateData *httpState = static_cast<HttpStateData *>(data);
1069 assert (fd == httpState->fd);
1070 // assert(buf == readBuf->content());
1071 PROF_start(HttpStateData_readReply);
1072 httpState->readReply(len, flag, xerrno);
1073 PROF_stop(HttpStateData_readReply);
1074 }
1075 */
1076
1077 /* XXX this function is too long! */
1078 void
1079 HttpStateData::readReply(const CommIoCbParams &io)
1080 {
1081 int bin;
1082 int clen;
1083 int len = io.size;
1084
1085 assert(fd == io.fd);
1086
1087 flags.do_next_read = 0;
1088
1089 debugs(11, 5, "httpReadReply: FD " << fd << ": len " << len << ".");
1090
1091 // Bail out early on COMM_ERR_CLOSING - close handlers will tidy up for us
1092 if (io.flag == COMM_ERR_CLOSING) {
1093 debugs(11, 3, "http socket closing");
1094 return;
1095 }
1096
1097 if (EBIT_TEST(entry->flags, ENTRY_ABORTED)) {
1098 maybeReadVirginBody();
1099 return;
1100 }
1101
1102 // handle I/O errors
1103 if (io.flag != COMM_OK || len < 0) {
1104 debugs(11, 2, "httpReadReply: FD " << fd << ": read failure: " << xstrerror() << ".");
1105
1106 if (ignoreErrno(io.xerrno)) {
1107 flags.do_next_read = 1;
1108 } else {
1109 ErrorState *err;
1110 err = errorCon(ERR_READ_ERROR, HTTP_BAD_GATEWAY, fwd->request);
1111 err->xerrno = io.xerrno;
1112 fwd->fail(err);
1113 flags.do_next_read = 0;
1114 comm_close(fd);
1115 }
1116
1117 return;
1118 }
1119
1120 // update I/O stats
1121 if (len > 0) {
1122 readBuf->appended(len);
1123 reply_bytes_read += len;
1124 #if USE_DELAY_POOLS
1125 DelayId delayId = entry->mem_obj->mostBytesAllowed();
1126 delayId.bytesIn(len);
1127 #endif
1128
1129 kb_incr(&statCounter.server.all.kbytes_in, len);
1130 kb_incr(&statCounter.server.http.kbytes_in, len);
1131 IOStats.Http.reads++;
1132
1133 for (clen = len - 1, bin = 0; clen; bin++)
1134 clen >>= 1;
1135
1136 IOStats.Http.read_hist[bin]++;
1137
1138 // update peer response time stats (%<pt)
1139 const timeval &sent = orig_request->hier.peer_http_request_sent;
1140 orig_request->hier.peer_response_time =
1141 sent.tv_sec ? tvSubMsec(sent, current_time) : -1;
1142 }
1143
1144 /** \par
1145 * Here the RFC says we should ignore whitespace between replies, but we can't as
1146 * doing so breaks HTTP/0.9 replies beginning with witespace, and in addition
1147 * the response splitting countermeasures is extremely likely to trigger on this,
1148 * not allowing connection reuse in the first place.
1149 */
1150 #if DONT_DO_THIS
1151 if (!flags.headers_parsed && len > 0 && fd_table[fd].uses > 1) {
1152 /* Skip whitespace between replies */
1153
1154 while (len > 0 && xisspace(*buf))
1155 memmove(buf, buf + 1, len--);
1156
1157 if (len == 0) {
1158 /* Continue to read... */
1159 /* Timeout NOT increased. This whitespace was from previous reply */
1160 flags.do_next_read = 1;
1161 maybeReadVirginBody();
1162 return;
1163 }
1164 }
1165
1166 #endif
1167
1168 if (len == 0) { // reached EOF?
1169 eof = 1;
1170 flags.do_next_read = 0;
1171
1172 /* Bug 2879: Replies may terminate with \r\n then EOF instead of \r\n\r\n
1173 * Ensure here that we have at minimum two \r\n when EOF is seen.
1174 * TODO: Add eof parameter to headersEnd() and move this hack there.
1175 */
1176 if (readBuf->contentSize() && !flags.headers_parsed) {
1177 /*
1178 * Yes Henrik, there is a point to doing this. When we
1179 * called httpProcessReplyHeader() before, we didn't find
1180 * the end of headers, but now we are definately at EOF, so
1181 * we want to process the reply headers.
1182 */
1183 /* Fake an "end-of-headers" to work around such broken servers */
1184 readBuf->append("\r\n", 2);
1185 }
1186 }
1187
1188 processReply();
1189 }
1190
1191 /// processes the already read and buffered response data, possibly after
1192 /// waiting for asynchronous 1xx control message processing
1193 void
1194 HttpStateData::processReply()
1195 {
1196
1197 if (flags.handling1xx) { // we came back after handling a 1xx response
1198 debugs(11, 5, HERE << "done with 1xx handling");
1199 flags.handling1xx = false;
1200 Must(!flags.headers_parsed);
1201 }
1202
1203 if (!flags.headers_parsed) { // have not parsed headers yet?
1204 PROF_start(HttpStateData_processReplyHeader);
1205 processReplyHeader();
1206 PROF_stop(HttpStateData_processReplyHeader);
1207
1208 if (!continueAfterParsingHeader()) // parsing error or need more data
1209 return; // TODO: send errors to ICAP
1210
1211 adaptOrFinalizeReply();
1212 // adaptOrFinalizeReply() may abort or otherwise "close" the entry
1213 if (!entry->isAccepting())
1214 return;
1215 }
1216
1217 // kick more reads if needed and/or process the response body, if any
1218 PROF_start(HttpStateData_processReplyBody);
1219 processReplyBody(); // may call serverComplete()
1220 PROF_stop(HttpStateData_processReplyBody);
1221 }
1222
1223 /**
1224 \retval true if we can continue with processing the body or doing ICAP.
1225 */
1226 bool
1227 HttpStateData::continueAfterParsingHeader()
1228 {
1229 if (flags.handling1xx) {
1230 debugs(11, 5, HERE << "wait for 1xx handling");
1231 Must(!flags.headers_parsed);
1232 return false;
1233 }
1234
1235 if (!flags.headers_parsed && !eof) {
1236 debugs(11, 9, HERE << "needs more at " << readBuf->contentSize());
1237 flags.do_next_read = 1;
1238 /** \retval false If we have not finished parsing the headers and may get more data.
1239 * Schedules more reads to retrieve the missing data.
1240 */
1241 maybeReadVirginBody(); // schedules all kinds of reads; TODO: rename
1242 return false;
1243 }
1244
1245 /** If we are done with parsing, check for errors */
1246
1247 err_type error = ERR_NONE;
1248
1249 if (flags.headers_parsed) { // parsed headers, possibly with errors
1250 // check for header parsing errors
1251 if (HttpReply *vrep = virginReply()) {
1252 const http_status s = vrep->sline.status;
1253 const HttpVersion &v = vrep->sline.version;
1254 if (s == HTTP_INVALID_HEADER && v != HttpVersion(0,9)) {
1255 debugs(11, DBG_IMPORTANT, "WARNING: HTTP: Invalid Response: Bad header encountered from " << entry->url() << " AKA " << orig_request->GetHost() << orig_request->urlpath.termedBuf() );
1256 error = ERR_INVALID_RESP;
1257 } else if (s == HTTP_HEADER_TOO_LARGE) {
1258 fwd->dontRetry(true);
1259 error = ERR_TOO_BIG;
1260 } else {
1261 return true; // done parsing, got reply, and no error
1262 }
1263 } else {
1264 // parsed headers but got no reply
1265 debugs(11, DBG_IMPORTANT, "WARNING: HTTP: Invalid Response: No reply at all for " << entry->url() << " AKA " << orig_request->GetHost() << orig_request->urlpath.termedBuf() );
1266 error = ERR_INVALID_RESP;
1267 }
1268 } else {
1269 assert(eof);
1270 if (readBuf->hasContent()) {
1271 error = ERR_INVALID_RESP;
1272 debugs(11, DBG_IMPORTANT, "WARNING: HTTP: Invalid Response: Headers did not parse at all for " << entry->url() << " AKA " << orig_request->GetHost() << orig_request->urlpath.termedBuf() );
1273 } else {
1274 error = ERR_ZERO_SIZE_OBJECT;
1275 debugs(11, (orig_request->flags.accelerated?DBG_IMPORTANT:2), "WARNING: HTTP: Invalid Response: No object data received for " <<
1276 entry->url() << " AKA " << orig_request->GetHost() << orig_request->urlpath.termedBuf() );
1277 }
1278 }
1279
1280 assert(error != ERR_NONE);
1281 entry->reset();
1282 fwd->fail(errorCon(error, HTTP_BAD_GATEWAY, fwd->request));
1283 flags.do_next_read = 0;
1284 comm_close(fd);
1285 return false; // quit on error
1286 }
1287
1288 /** truncate what we read if we read too much so that writeReplyBody()
1289 writes no more than what we should have read */
1290 void
1291 HttpStateData::truncateVirginBody()
1292 {
1293 assert(flags.headers_parsed);
1294
1295 HttpReply *vrep = virginReply();
1296 int64_t clen = -1;
1297 if (!vrep->expectingBody(request->method, clen) || clen < 0)
1298 return; // no body or a body of unknown size, including chunked
1299
1300 const int64_t body_bytes_read = reply_bytes_read - header_bytes_read;
1301 if (body_bytes_read - body_bytes_truncated <= clen)
1302 return; // we did not read too much or already took care of the extras
1303
1304 if (const int64_t extras = body_bytes_read - body_bytes_truncated - clen) {
1305 // server sent more that the advertised content length
1306 debugs(11,5, HERE << "body_bytes_read=" << body_bytes_read <<
1307 " clen=" << clen << '/' << vrep->content_length <<
1308 " body_bytes_truncated=" << body_bytes_truncated << '+' << extras);
1309
1310 readBuf->truncate(extras);
1311 body_bytes_truncated += extras;
1312 }
1313 }
1314
1315 /**
1316 * Call this when there is data from the origin server
1317 * which should be sent to either StoreEntry, or to ICAP...
1318 */
1319 void
1320 HttpStateData::writeReplyBody()
1321 {
1322 truncateVirginBody(); // if needed
1323 const char *data = readBuf->content();
1324 int len = readBuf->contentSize();
1325 addVirginReplyBody(data, len);
1326 readBuf->consume(len);
1327 }
1328
1329 bool
1330 HttpStateData::decodeAndWriteReplyBody()
1331 {
1332 const char *data = NULL;
1333 int len;
1334 bool wasThereAnException = false;
1335 assert(flags.chunked);
1336 assert(httpChunkDecoder);
1337 SQUID_ENTER_THROWING_CODE();
1338 MemBuf decodedData;
1339 decodedData.init();
1340 const bool doneParsing = httpChunkDecoder->parse(readBuf,&decodedData);
1341 len = decodedData.contentSize();
1342 data=decodedData.content();
1343 addVirginReplyBody(data, len);
1344 if (doneParsing) {
1345 lastChunk = 1;
1346 flags.do_next_read = 0;
1347 }
1348 SQUID_EXIT_THROWING_CODE(wasThereAnException);
1349 return wasThereAnException;
1350 }
1351
1352 /**
1353 * processReplyBody has two purposes:
1354 * 1 - take the reply body data, if any, and put it into either
1355 * the StoreEntry, or give it over to ICAP.
1356 * 2 - see if we made it to the end of the response (persistent
1357 * connections and such)
1358 */
1359 void
1360 HttpStateData::processReplyBody()
1361 {
1362 AsyncCall::Pointer call;
1363 Ip::Address client_addr;
1364 bool ispinned = false;
1365
1366 if (!flags.headers_parsed) {
1367 flags.do_next_read = 1;
1368 maybeReadVirginBody();
1369 return;
1370 }
1371
1372 #if USE_ADAPTATION
1373 debugs(11,5, HERE << "adaptationAccessCheckPending=" << adaptationAccessCheckPending);
1374 if (adaptationAccessCheckPending)
1375 return;
1376
1377 #endif
1378
1379 /*
1380 * At this point the reply headers have been parsed and consumed.
1381 * That means header content has been removed from readBuf and
1382 * it contains only body data.
1383 */
1384 if (flags.chunked) {
1385 if (!decodeAndWriteReplyBody()) {
1386 flags.do_next_read = 0;
1387 serverComplete();
1388 return;
1389 }
1390 } else
1391 writeReplyBody();
1392
1393 if (EBIT_TEST(entry->flags, ENTRY_ABORTED)) {
1394 /*
1395 * The above writeReplyBody() call could ABORT this entry,
1396 * in that case, the server FD should already be closed.
1397 * there's nothing for us to do.
1398 */
1399 (void) 0;
1400 } else
1401 switch (persistentConnStatus()) {
1402 case INCOMPLETE_MSG:
1403 debugs(11, 5, "processReplyBody: INCOMPLETE_MSG");
1404 /* Wait for more data or EOF condition */
1405 if (flags.keepalive_broken) {
1406 call = NULL;
1407 commSetTimeout(fd, 10, call);
1408 } else {
1409 call = NULL;
1410 commSetTimeout(fd, Config.Timeout.read, call);
1411 }
1412
1413 flags.do_next_read = 1;
1414 break;
1415
1416 case COMPLETE_PERSISTENT_MSG:
1417 debugs(11, 5, "processReplyBody: COMPLETE_PERSISTENT_MSG");
1418 /* yes we have to clear all these! */
1419 call = NULL;
1420 commSetTimeout(fd, -1, call);
1421 flags.do_next_read = 0;
1422
1423 comm_remove_close_handler(fd, closeHandler);
1424 closeHandler = NULL;
1425 fwd->unregister(fd);
1426
1427 if (orig_request->flags.spoof_client_ip)
1428 client_addr = orig_request->client_addr;
1429
1430
1431 if (request->flags.pinned) {
1432 ispinned = true;
1433 } else if (request->flags.connection_auth && request->flags.auth_sent) {
1434 ispinned = true;
1435 }
1436
1437 if (orig_request->pinnedConnection() && ispinned) {
1438 orig_request->pinnedConnection()->pinConnection(fd, orig_request, _peer,
1439 (request->flags.connection_auth != 0));
1440 } else {
1441 fwd->pconnPush(fd, _peer, request, orig_request->GetHost(), client_addr);
1442 }
1443
1444 fd = -1;
1445
1446 serverComplete();
1447 return;
1448
1449 case COMPLETE_NONPERSISTENT_MSG:
1450 debugs(11, 5, "processReplyBody: COMPLETE_NONPERSISTENT_MSG");
1451 serverComplete();
1452 return;
1453 }
1454
1455 maybeReadVirginBody();
1456 }
1457
1458 void
1459 HttpStateData::maybeReadVirginBody()
1460 {
1461 // we may need to grow the buffer if headers do not fit
1462 const int minRead = flags.headers_parsed ? 0 :1024;
1463 const int read_size = replyBodySpace(*readBuf, minRead);
1464
1465 debugs(11,9, HERE << (flags.do_next_read ? "may" : "wont") <<
1466 " read up to " << read_size << " bytes from FD " << fd);
1467
1468 /*
1469 * why <2? Because delayAwareRead() won't actually read if
1470 * you ask it to read 1 byte. The delayed read request
1471 * just gets re-queued until the client side drains, then
1472 * the I/O thread hangs. Better to not register any read
1473 * handler until we get a notification from someone that
1474 * its okay to read again.
1475 */
1476 if (read_size < 2)
1477 return;
1478
1479 if (flags.do_next_read) {
1480 flags.do_next_read = 0;
1481 typedef CommCbMemFunT<HttpStateData, CommIoCbParams> Dialer;
1482 entry->delayAwareRead(fd, readBuf->space(read_size), read_size,
1483 JobCallback(11, 5, Dialer, this, HttpStateData::readReply));
1484 }
1485 }
1486
1487 /// called after writing the very last request byte (body, last-chunk, etc)
1488 void
1489 HttpStateData::wroteLast(const CommIoCbParams &io)
1490 {
1491 debugs(11, 5, HERE << "FD " << fd << ": size " << io.size << ": errflag " << io.flag << ".");
1492 #if URL_CHECKSUM_DEBUG
1493
1494 entry->mem_obj->checkUrlChecksum();
1495 #endif
1496
1497 if (io.size > 0) {
1498 fd_bytes(fd, io.size, FD_WRITE);
1499 kb_incr(&statCounter.server.all.kbytes_out, io.size);
1500 kb_incr(&statCounter.server.http.kbytes_out, io.size);
1501 }
1502
1503 if (io.flag == COMM_ERR_CLOSING)
1504 return;
1505
1506 if (io.flag) {
1507 ErrorState *err;
1508 err = errorCon(ERR_WRITE_ERROR, HTTP_BAD_GATEWAY, fwd->request);
1509 err->xerrno = io.xerrno;
1510 fwd->fail(err);
1511 comm_close(fd);
1512 return;
1513 }
1514
1515 sendComplete();
1516 }
1517
1518 /// successfully wrote the entire request (including body, last-chunk, etc.)
1519 void
1520 HttpStateData::sendComplete()
1521 {
1522 /*
1523 * Set the read timeout here because it hasn't been set yet.
1524 * We only set the read timeout after the request has been
1525 * fully written to the server-side. If we start the timeout
1526 * after connection establishment, then we are likely to hit
1527 * the timeout for POST/PUT requests that have very large
1528 * request bodies.
1529 */
1530 typedef CommCbMemFunT<HttpStateData, CommTimeoutCbParams> TimeoutDialer;
1531 AsyncCall::Pointer timeoutCall = JobCallback(11, 5,
1532 TimeoutDialer, this, HttpStateData::httpTimeout);
1533
1534 commSetTimeout(fd, Config.Timeout.read, timeoutCall);
1535
1536 flags.request_sent = 1;
1537
1538 orig_request->hier.peer_http_request_sent = current_time;
1539 }
1540
1541 // Close the HTTP server connection. Used by serverComplete().
1542 void
1543 HttpStateData::closeServer()
1544 {
1545 debugs(11,5, HERE << "closing HTTP server FD " << fd << " this " << this);
1546
1547 if (fd >= 0) {
1548 fwd->unregister(fd);
1549 comm_remove_close_handler(fd, closeHandler);
1550 closeHandler = NULL;
1551 comm_close(fd);
1552 fd = -1;
1553 }
1554 }
1555
1556 bool
1557 HttpStateData::doneWithServer() const
1558 {
1559 return fd < 0;
1560 }
1561
1562
1563 /*
1564 * Fixup authentication request headers for special cases
1565 */
1566 static void
1567 httpFixupAuthentication(HttpRequest * request, HttpRequest * orig_request, const HttpHeader * hdr_in, HttpHeader * hdr_out, http_state_flags flags)
1568 {
1569 http_hdr_type header = flags.originpeer ? HDR_AUTHORIZATION : HDR_PROXY_AUTHORIZATION;
1570
1571 /* Nothing to do unless we are forwarding to a peer */
1572 if (!request->flags.proxying)
1573 return;
1574
1575 /* Needs to be explicitly enabled */
1576 if (!orig_request->peer_login)
1577 return;
1578
1579 /* Maybe already dealt with? */
1580 if (hdr_out->has(header))
1581 return;
1582
1583 /* Nothing to do here for PASSTHRU */
1584 if (strcmp(orig_request->peer_login, "PASSTHRU") == 0)
1585 return;
1586
1587 /* PROXYPASS is a special case, single-signon to servers with the proxy password (basic only) */
1588 if (flags.originpeer && strcmp(orig_request->peer_login, "PROXYPASS") == 0 && hdr_in->has(HDR_PROXY_AUTHORIZATION)) {
1589 const char *auth = hdr_in->getStr(HDR_PROXY_AUTHORIZATION);
1590
1591 if (auth && strncasecmp(auth, "basic ", 6) == 0) {
1592 hdr_out->putStr(header, auth);
1593 return;
1594 }
1595 }
1596
1597 /* Special mode to pass the username to the upstream cache */
1598 if (*orig_request->peer_login == '*') {
1599 char loginbuf[256];
1600 const char *username = "-";
1601
1602 if (orig_request->extacl_user.size())
1603 username = orig_request->extacl_user.termedBuf();
1604 #if USE_AUTH
1605 else if (orig_request->auth_user_request != NULL)
1606 username = orig_request->auth_user_request->username();
1607 #endif
1608
1609 snprintf(loginbuf, sizeof(loginbuf), "%s%s", username, orig_request->peer_login + 1);
1610
1611 httpHeaderPutStrf(hdr_out, header, "Basic %s",
1612 old_base64_encode(loginbuf));
1613 return;
1614 }
1615
1616 /* external_acl provided credentials */
1617 if (orig_request->extacl_user.size() && orig_request->extacl_passwd.size() &&
1618 (strcmp(orig_request->peer_login, "PASS") == 0 ||
1619 strcmp(orig_request->peer_login, "PROXYPASS") == 0)) {
1620 char loginbuf[256];
1621 snprintf(loginbuf, sizeof(loginbuf), SQUIDSTRINGPH ":" SQUIDSTRINGPH,
1622 SQUIDSTRINGPRINT(orig_request->extacl_user),
1623 SQUIDSTRINGPRINT(orig_request->extacl_passwd));
1624 httpHeaderPutStrf(hdr_out, header, "Basic %s",
1625 old_base64_encode(loginbuf));
1626 return;
1627 }
1628
1629 /* Kerberos login to peer */
1630 #if HAVE_AUTH_MODULE_NEGOTIATE && HAVE_KRB5 && HAVE_GSSAPI
1631 if (strncmp(orig_request->peer_login, "NEGOTIATE",strlen("NEGOTIATE")) == 0) {
1632 char *Token=NULL;
1633 char *PrincipalName=NULL,*p;
1634 if ((p=strchr(orig_request->peer_login,':')) != NULL ) {
1635 PrincipalName=++p;
1636 }
1637 Token = peer_proxy_negotiate_auth(PrincipalName,request->peer_host);
1638 if (Token) {
1639 httpHeaderPutStrf(hdr_out, HDR_PROXY_AUTHORIZATION, "Negotiate %s",Token);
1640 }
1641 return;
1642 }
1643 #endif /* HAVE_KRB5 && HAVE_GSSAPI */
1644
1645 httpHeaderPutStrf(hdr_out, header, "Basic %s",
1646 old_base64_encode(orig_request->peer_login));
1647 return;
1648 }
1649
1650 /*
1651 * build request headers and append them to a given MemBuf
1652 * used by buildRequestPrefix()
1653 * note: initialised the HttpHeader, the caller is responsible for Clean()-ing
1654 */
1655 void
1656 HttpStateData::httpBuildRequestHeader(HttpRequest * request,
1657 HttpRequest * orig_request,
1658 StoreEntry * entry,
1659 HttpHeader * hdr_out,
1660 const http_state_flags flags)
1661 {
1662 /* building buffer for complex strings */
1663 #define BBUF_SZ (MAX_URL+32)
1664 LOCAL_ARRAY(char, bbuf, BBUF_SZ);
1665 LOCAL_ARRAY(char, ntoabuf, MAX_IPSTRLEN);
1666 const HttpHeader *hdr_in = &orig_request->header;
1667 const HttpHeaderEntry *e = NULL;
1668 HttpHeaderPos pos = HttpHeaderInitPos;
1669 assert (hdr_out->owner == hoRequest);
1670
1671 /* append our IMS header */
1672 if (request->lastmod > -1)
1673 hdr_out->putTime(HDR_IF_MODIFIED_SINCE, request->lastmod);
1674
1675 bool we_do_ranges = decideIfWeDoRanges (orig_request);
1676
1677 String strConnection (hdr_in->getList(HDR_CONNECTION));
1678
1679 while ((e = hdr_in->getEntry(&pos)))
1680 copyOneHeaderFromClientsideRequestToUpstreamRequest(e, strConnection, request, orig_request, hdr_out, we_do_ranges, flags);
1681
1682 /* Abstraction break: We should interpret multipart/byterange responses
1683 * into offset-length data, and this works around our inability to do so.
1684 */
1685 if (!we_do_ranges && orig_request->multipartRangeRequest()) {
1686 /* don't cache the result */
1687 orig_request->flags.cachable = 0;
1688 /* pretend it's not a range request */
1689 delete orig_request->range;
1690 orig_request->range = NULL;
1691 orig_request->flags.range = 0;
1692 }
1693
1694 /* append Via */
1695 if (Config.onoff.via) {
1696 String strVia;
1697 strVia = hdr_in->getList(HDR_VIA);
1698 snprintf(bbuf, BBUF_SZ, "%d.%d %s",
1699 orig_request->http_ver.major,
1700 orig_request->http_ver.minor, ThisCache);
1701 strListAdd(&strVia, bbuf, ',');
1702 hdr_out->putStr(HDR_VIA, strVia.termedBuf());
1703 strVia.clean();
1704 }
1705
1706 if (orig_request->flags.accelerated) {
1707 /* Append Surrogate-Capabilities */
1708 String strSurrogate(hdr_in->getList(HDR_SURROGATE_CAPABILITY));
1709 #if USE_SQUID_ESI
1710 snprintf(bbuf, BBUF_SZ, "%s=\"Surrogate/1.0 ESI/1.0\"", Config.Accel.surrogate_id);
1711 #else
1712 snprintf(bbuf, BBUF_SZ, "%s=\"Surrogate/1.0\"", Config.Accel.surrogate_id);
1713 #endif
1714 strListAdd(&strSurrogate, bbuf, ',');
1715 hdr_out->putStr(HDR_SURROGATE_CAPABILITY, strSurrogate.termedBuf());
1716 }
1717
1718 /** \pre Handle X-Forwarded-For */
1719 if (strcmp(opt_forwarded_for, "delete") != 0) {
1720
1721 String strFwd = hdr_in->getList(HDR_X_FORWARDED_FOR);
1722
1723 if (strFwd.size() > 65536/2) {
1724 // There is probably a forwarding loop with Via detection disabled.
1725 // If we do nothing, String will assert on overflow soon.
1726 // TODO: Terminate all transactions with huge XFF?
1727 strFwd = "error";
1728
1729 static int warnedCount = 0;
1730 if (warnedCount++ < 100) {
1731 const char *url = entry ? entry->url() : urlCanonical(orig_request);
1732 debugs(11, 1, "Warning: likely forwarding loop with " << url);
1733 }
1734 }
1735
1736 if (strcmp(opt_forwarded_for, "on") == 0) {
1737 /** If set to ON - append client IP or 'unknown'. */
1738 if ( orig_request->client_addr.IsNoAddr() )
1739 strListAdd(&strFwd, "unknown", ',');
1740 else
1741 strListAdd(&strFwd, orig_request->client_addr.NtoA(ntoabuf, MAX_IPSTRLEN), ',');
1742 } else if (strcmp(opt_forwarded_for, "off") == 0) {
1743 /** If set to OFF - append 'unknown'. */
1744 strListAdd(&strFwd, "unknown", ',');
1745 } else if (strcmp(opt_forwarded_for, "transparent") == 0) {
1746 /** If set to TRANSPARENT - pass through unchanged. */
1747 } else if (strcmp(opt_forwarded_for, "truncate") == 0) {
1748 /** If set to TRUNCATE - drop existing list and replace with client IP or 'unknown'. */
1749 if ( orig_request->client_addr.IsNoAddr() )
1750 strFwd = "unknown";
1751 else
1752 strFwd = orig_request->client_addr.NtoA(ntoabuf, MAX_IPSTRLEN);
1753 }
1754 if (strFwd.size() > 0)
1755 hdr_out->putStr(HDR_X_FORWARDED_FOR, strFwd.termedBuf());
1756 }
1757 /** If set to DELETE - do not copy through. */
1758
1759 /* append Host if not there already */
1760 if (!hdr_out->has(HDR_HOST)) {
1761 if (orig_request->peer_domain) {
1762 hdr_out->putStr(HDR_HOST, orig_request->peer_domain);
1763 } else if (orig_request->port == urlDefaultPort(orig_request->protocol)) {
1764 /* use port# only if not default */
1765 hdr_out->putStr(HDR_HOST, orig_request->GetHost());
1766 } else {
1767 httpHeaderPutStrf(hdr_out, HDR_HOST, "%s:%d",
1768 orig_request->GetHost(),
1769 (int) orig_request->port);
1770 }
1771 }
1772
1773 /* append Authorization if known in URL, not in header and going direct */
1774 if (!hdr_out->has(HDR_AUTHORIZATION)) {
1775 if (!request->flags.proxying && *request->login) {
1776 httpHeaderPutStrf(hdr_out, HDR_AUTHORIZATION, "Basic %s",
1777 old_base64_encode(request->login));
1778 }
1779 }
1780
1781 /* Fixup (Proxy-)Authorization special cases. Plain relaying dealt with above */
1782 httpFixupAuthentication(request, orig_request, hdr_in, hdr_out, flags);
1783
1784 /* append Cache-Control, add max-age if not there already */
1785 {
1786 HttpHdrCc *cc = hdr_in->getCc();
1787
1788 if (!cc)
1789 cc = httpHdrCcCreate();
1790
1791 #if 0 /* see bug 2330 */
1792 /* Set no-cache if determined needed but not found */
1793 if (orig_request->flags.nocache)
1794 EBIT_SET(cc->mask, CC_NO_CACHE);
1795 #endif
1796
1797 /* Add max-age only without no-cache */
1798 if (!EBIT_TEST(cc->mask, CC_MAX_AGE) && !EBIT_TEST(cc->mask, CC_NO_CACHE)) {
1799 const char *url =
1800 entry ? entry->url() : urlCanonical(orig_request);
1801 httpHdrCcSetMaxAge(cc, getMaxAge(url));
1802
1803 if (request->urlpath.size())
1804 assert(strstr(url, request->urlpath.termedBuf()));
1805 }
1806
1807 /* Enforce sibling relations */
1808 if (flags.only_if_cached)
1809 EBIT_SET(cc->mask, CC_ONLY_IF_CACHED);
1810
1811 hdr_out->putCc(cc);
1812
1813 httpHdrCcDestroy(cc);
1814 }
1815
1816 /* maybe append Connection: keep-alive */
1817 if (flags.keepalive) {
1818 hdr_out->putStr(HDR_CONNECTION, "keep-alive");
1819 }
1820
1821 /* append Front-End-Https */
1822 if (flags.front_end_https) {
1823 if (flags.front_end_https == 1 || request->protocol == AnyP::PROTO_HTTPS)
1824 hdr_out->putStr(HDR_FRONT_END_HTTPS, "On");
1825 }
1826
1827 if (flags.chunked_request) {
1828 // Do not just copy the original value so that if the client-side
1829 // starts decode other encodings, this code may remain valid.
1830 hdr_out->putStr(HDR_TRANSFER_ENCODING, "chunked");
1831 }
1832
1833 /* Now mangle the headers. */
1834 if (Config2.onoff.mangle_request_headers)
1835 httpHdrMangleList(hdr_out, request, ROR_REQUEST);
1836
1837 strConnection.clean();
1838 }
1839
1840 /**
1841 * Decides whether a particular header may be cloned from the received Clients request
1842 * to our outgoing fetch request.
1843 */
1844 void
1845 copyOneHeaderFromClientsideRequestToUpstreamRequest(const HttpHeaderEntry *e, const String strConnection, HttpRequest * request, const HttpRequest * orig_request, HttpHeader * hdr_out, const int we_do_ranges, const http_state_flags flags)
1846 {
1847 debugs(11, 5, "httpBuildRequestHeader: " << e->name << ": " << e->value );
1848
1849 switch (e->id) {
1850
1851 /** \par RFC 2616 sect 13.5.1 - Hop-by-Hop headers which Squid should not pass on. */
1852
1853 case HDR_PROXY_AUTHORIZATION:
1854 /** \par Proxy-Authorization:
1855 * Only pass on proxy authentication to peers for which
1856 * authentication forwarding is explicitly enabled
1857 */
1858 if (!flags.originpeer && flags.proxying && orig_request->peer_login &&
1859 (strcmp(orig_request->peer_login, "PASS") == 0 ||
1860 strcmp(orig_request->peer_login, "PROXYPASS") == 0 ||
1861 strcmp(orig_request->peer_login, "PASSTHRU") == 0)) {
1862 hdr_out->addEntry(e->clone());
1863 }
1864 break;
1865
1866 /** \par RFC 2616 sect 13.5.1 - Hop-by-Hop headers which Squid does not pass on. */
1867
1868 case HDR_CONNECTION: /** \par Connection: */
1869 case HDR_TE: /** \par TE: */
1870 case HDR_KEEP_ALIVE: /** \par Keep-Alive: */
1871 case HDR_PROXY_AUTHENTICATE: /** \par Proxy-Authenticate: */
1872 case HDR_TRAILER: /** \par Trailer: */
1873 case HDR_UPGRADE: /** \par Upgrade: */
1874 case HDR_TRANSFER_ENCODING: /** \par Transfer-Encoding: */
1875 break;
1876
1877
1878 /** \par OTHER headers I haven't bothered to track down yet. */
1879
1880 case HDR_AUTHORIZATION:
1881 /** \par WWW-Authorization:
1882 * Pass on WWW authentication */
1883
1884 if (!flags.originpeer) {
1885 hdr_out->addEntry(e->clone());
1886 } else {
1887 /** \note In accelerators, only forward authentication if enabled
1888 * (see also httpFixupAuthentication for special cases)
1889 */
1890 if (orig_request->peer_login &&
1891 (strcmp(orig_request->peer_login, "PASS") == 0 ||
1892 strcmp(orig_request->peer_login, "PASSTHRU") == 0 ||
1893 strcmp(orig_request->peer_login, "PROXYPASS") == 0)) {
1894 hdr_out->addEntry(e->clone());
1895 }
1896 }
1897
1898 break;
1899
1900 case HDR_HOST:
1901 /** \par Host:
1902 * Normally Squid rewrites the Host: header.
1903 * However, there is one case when we don't: If the URL
1904 * went through our redirector and the admin configured
1905 * 'redir_rewrites_host' to be off.
1906 */
1907 if (orig_request->peer_domain)
1908 hdr_out->putStr(HDR_HOST, orig_request->peer_domain);
1909 else if (request->flags.redirected && !Config.onoff.redir_rewrites_host)
1910 hdr_out->addEntry(e->clone());
1911 else {
1912 /* use port# only if not default */
1913
1914 if (orig_request->port == urlDefaultPort(orig_request->protocol)) {
1915 hdr_out->putStr(HDR_HOST, orig_request->GetHost());
1916 } else {
1917 httpHeaderPutStrf(hdr_out, HDR_HOST, "%s:%d",
1918 orig_request->GetHost(),
1919 (int) orig_request->port);
1920 }
1921 }
1922
1923 break;
1924
1925 case HDR_IF_MODIFIED_SINCE:
1926 /** \par If-Modified-Since:
1927 * append unless we added our own;
1928 * \note at most one client's ims header can pass through */
1929
1930 if (!hdr_out->has(HDR_IF_MODIFIED_SINCE))
1931 hdr_out->addEntry(e->clone());
1932
1933 break;
1934
1935 case HDR_MAX_FORWARDS:
1936 /** \par Max-Forwards:
1937 * pass only on TRACE or OPTIONS requests */
1938 if (orig_request->method == METHOD_TRACE || orig_request->method == METHOD_OPTIONS) {
1939 const int64_t hops = e->getInt64();
1940
1941 if (hops > 0)
1942 hdr_out->putInt64(HDR_MAX_FORWARDS, hops - 1);
1943 }
1944
1945 break;
1946
1947 case HDR_VIA:
1948 /** \par Via:
1949 * If Via is disabled then forward any received header as-is.
1950 * Otherwise leave for explicit updated addition later. */
1951
1952 if (!Config.onoff.via)
1953 hdr_out->addEntry(e->clone());
1954
1955 break;
1956
1957 case HDR_RANGE:
1958
1959 case HDR_IF_RANGE:
1960
1961 case HDR_REQUEST_RANGE:
1962 /** \par Range:, If-Range:, Request-Range:
1963 * Only pass if we accept ranges */
1964 if (!we_do_ranges)
1965 hdr_out->addEntry(e->clone());
1966
1967 break;
1968
1969 case HDR_PROXY_CONNECTION: // SHOULD ignore. But doing so breaks things.
1970 break;
1971
1972 case HDR_CONTENT_LENGTH:
1973 // pass through unless we chunk; also, keeping this away from default
1974 // prevents request smuggling via Connection: Content-Length tricks
1975 if (!flags.chunked_request)
1976 hdr_out->addEntry(e->clone());
1977 break;
1978
1979 case HDR_X_FORWARDED_FOR:
1980
1981 case HDR_CACHE_CONTROL:
1982 /** \par X-Forwarded-For:, Cache-Control:
1983 * handled specially by Squid, so leave off for now.
1984 * append these after the loop if needed */
1985 break;
1986
1987 case HDR_FRONT_END_HTTPS:
1988 /** \par Front-End-Https:
1989 * Pass thru only if peer is configured with front-end-https */
1990 if (!flags.front_end_https)
1991 hdr_out->addEntry(e->clone());
1992
1993 break;
1994
1995 default:
1996 /** \par default.
1997 * pass on all other header fields
1998 * which are NOT listed by the special Connection: header. */
1999
2000 if (strConnection.size()>0 && strListIsMember(&strConnection, e->name.termedBuf(), ',')) {
2001 debugs(11, 2, "'" << e->name << "' header cropped by Connection: definition");
2002 return;
2003 }
2004
2005 hdr_out->addEntry(e->clone());
2006 }
2007 }
2008
2009 bool
2010 HttpStateData::decideIfWeDoRanges (HttpRequest * orig_request)
2011 {
2012 bool result = true;
2013 /* decide if we want to do Ranges ourselves
2014 * and fetch the whole object now)
2015 * We want to handle Ranges ourselves iff
2016 * - we can actually parse client Range specs
2017 * - the specs are expected to be simple enough (e.g. no out-of-order ranges)
2018 * - reply will be cachable
2019 * (If the reply will be uncachable we have to throw it away after
2020 * serving this request, so it is better to forward ranges to
2021 * the server and fetch only the requested content)
2022 */
2023
2024 int64_t roffLimit = orig_request->getRangeOffsetLimit();
2025
2026 if (NULL == orig_request->range || !orig_request->flags.cachable
2027 || orig_request->range->offsetLimitExceeded(roffLimit) || orig_request->flags.connection_auth)
2028 result = false;
2029
2030 debugs(11, 8, "decideIfWeDoRanges: range specs: " <<
2031 orig_request->range << ", cachable: " <<
2032 orig_request->flags.cachable << "; we_do_ranges: " << result);
2033
2034 return result;
2035 }
2036
2037 /* build request prefix and append it to a given MemBuf;
2038 * return the length of the prefix */
2039 mb_size_t
2040 HttpStateData::buildRequestPrefix(HttpRequest * aRequest,
2041 HttpRequest * original_request,
2042 StoreEntry * sentry,
2043 MemBuf * mb)
2044 {
2045 const int offset = mb->size;
2046 HttpVersion httpver(1,1);
2047 mb->Printf("%s %s HTTP/%d.%d\r\n",
2048 RequestMethodStr(aRequest->method),
2049 aRequest->urlpath.size() ? aRequest->urlpath.termedBuf() : "/",
2050 httpver.major,httpver.minor);
2051 /* build and pack headers */
2052 {
2053 HttpHeader hdr(hoRequest);
2054 Packer p;
2055 httpBuildRequestHeader(aRequest, original_request, sentry, &hdr, flags);
2056
2057 if (aRequest->flags.pinned && aRequest->flags.connection_auth)
2058 aRequest->flags.auth_sent = 1;
2059 else if (hdr.has(HDR_AUTHORIZATION))
2060 aRequest->flags.auth_sent = 1;
2061
2062 packerToMemInit(&p, mb);
2063 hdr.packInto(&p);
2064 hdr.clean();
2065 packerClean(&p);
2066 }
2067 /* append header terminator */
2068 mb->append(crlf, 2);
2069 return mb->size - offset;
2070 }
2071
2072 /* This will be called when connect completes. Write request. */
2073 bool
2074 HttpStateData::sendRequest()
2075 {
2076 MemBuf mb;
2077
2078 debugs(11, 5, "httpSendRequest: FD " << fd << ", request " << request << ", this " << this << ".");
2079
2080 if (!canSend(fd)) {
2081 debugs(11,3, HERE << "cannot send request to closing FD " << fd);
2082 assert(closeHandler != NULL);
2083 return false;
2084 }
2085
2086 typedef CommCbMemFunT<HttpStateData, CommTimeoutCbParams> TimeoutDialer;
2087 AsyncCall::Pointer timeoutCall = JobCallback(11, 5,
2088 TimeoutDialer, this, HttpStateData::httpTimeout);
2089 commSetTimeout(fd, Config.Timeout.lifetime, timeoutCall);
2090 flags.do_next_read = 1;
2091 maybeReadVirginBody();
2092
2093 if (orig_request->body_pipe != NULL) {
2094 if (!startRequestBodyFlow()) // register to receive body data
2095 return false;
2096 typedef CommCbMemFunT<HttpStateData, CommIoCbParams> Dialer;
2097 requestSender = JobCallback(11,5,
2098 Dialer, this, HttpStateData::sentRequestBody);
2099
2100 Must(!flags.chunked_request);
2101 // use chunked encoding if we do not know the length
2102 if (orig_request->content_length < 0)
2103 flags.chunked_request = 1;
2104 } else {
2105 assert(!requestBodySource);
2106 typedef CommCbMemFunT<HttpStateData, CommIoCbParams> Dialer;
2107 requestSender = JobCallback(11,5,
2108 Dialer, this, HttpStateData::wroteLast);
2109 }
2110
2111 if (_peer != NULL) {
2112 if (_peer->options.originserver) {
2113 flags.proxying = 0;
2114 flags.originpeer = 1;
2115 } else {
2116 flags.proxying = 1;
2117 flags.originpeer = 0;
2118 }
2119 } else {
2120 flags.proxying = 0;
2121 flags.originpeer = 0;
2122 }
2123
2124 /*
2125 * Is keep-alive okay for all request methods?
2126 */
2127 if (orig_request->flags.must_keepalive)
2128 flags.keepalive = 1;
2129 else if (!Config.onoff.server_pconns)
2130 flags.keepalive = 0;
2131 else if (_peer == NULL)
2132 flags.keepalive = 1;
2133 else if (_peer->stats.n_keepalives_sent < 10)
2134 flags.keepalive = 1;
2135 else if ((double) _peer->stats.n_keepalives_recv /
2136 (double) _peer->stats.n_keepalives_sent > 0.50)
2137 flags.keepalive = 1;
2138
2139 if (_peer) {
2140 if (neighborType(_peer, request) == PEER_SIBLING &&
2141 !_peer->options.allow_miss)
2142 flags.only_if_cached = 1;
2143
2144 flags.front_end_https = _peer->front_end_https;
2145 }
2146
2147 mb.init();
2148 request->peer_host=_peer?_peer->host:NULL;
2149 buildRequestPrefix(request, orig_request, entry, &mb);
2150 debugs(11, 6, "httpSendRequest: FD " << fd << ":\n" << mb.buf);
2151 Comm::Write(fd, &mb, requestSender);
2152
2153 return true;
2154 }
2155
2156 bool
2157 HttpStateData::getMoreRequestBody(MemBuf &buf)
2158 {
2159 // parent's implementation can handle the no-encoding case
2160 if (!flags.chunked_request)
2161 return ServerStateData::getMoreRequestBody(buf);
2162
2163 MemBuf raw;
2164
2165 Must(requestBodySource != NULL);
2166 if (!requestBodySource->getMoreData(raw))
2167 return false; // no request body bytes to chunk yet
2168
2169 // optimization: pre-allocate buffer size that should be enough
2170 const mb_size_t rawDataSize = raw.contentSize();
2171 // we may need to send: hex-chunk-size CRLF raw-data CRLF last-chunk
2172 buf.init(16 + 2 + rawDataSize + 2 + 5, raw.max_capacity);
2173
2174 buf.Printf("%x\r\n", static_cast<unsigned int>(rawDataSize));
2175 buf.append(raw.content(), rawDataSize);
2176 buf.Printf("\r\n");
2177
2178 Must(rawDataSize > 0); // we did not accidently created last-chunk above
2179
2180 // Do not send last-chunk unless we successfully received everything
2181 if (receivedWholeRequestBody) {
2182 Must(!flags.sentLastChunk);
2183 flags.sentLastChunk = true;
2184 buf.append("0\r\n\r\n", 5);
2185 }
2186
2187 return true;
2188 }
2189
2190 void
2191 httpStart(FwdState *fwd)
2192 {
2193 debugs(11, 3, "httpStart: \"" << RequestMethodStr(fwd->request->method) << " " << fwd->entry->url() << "\"" );
2194 HttpStateData *httpState = new HttpStateData(fwd);
2195
2196 if (!httpState->sendRequest()) {
2197 debugs(11, 3, "httpStart: aborted");
2198 delete httpState;
2199 return;
2200 }
2201
2202 statCounter.server.all.requests++;
2203 statCounter.server.http.requests++;
2204
2205 /*
2206 * We used to set the read timeout here, but not any more.
2207 * Now its set in httpSendComplete() after the full request,
2208 * including request body, has been written to the server.
2209 */
2210 }
2211
2212 /// if broken posts are enabled for the request, try to fix and return true
2213 bool
2214 HttpStateData::finishingBrokenPost()
2215 {
2216 #if USE_HTTP_VIOLATIONS
2217 if (!Config.accessList.brokenPosts) {
2218 debugs(11, 5, HERE << "No brokenPosts list");
2219 return false;
2220 }
2221
2222 ACLFilledChecklist ch(Config.accessList.brokenPosts, originalRequest(), NULL);
2223 if (!ch.fastCheck()) {
2224 debugs(11, 5, HERE << "didn't match brokenPosts");
2225 return false;
2226 }
2227
2228 if (!canSend(fd)) {
2229 debugs(11,2, HERE << "ignoring broken POST for closing FD " << fd);
2230 assert(closeHandler != NULL);
2231 return true; // prevent caller from proceeding as if nothing happened
2232 }
2233
2234 debugs(11, 2, "finishingBrokenPost: fixing broken POST");
2235 typedef CommCbMemFunT<HttpStateData, CommIoCbParams> Dialer;
2236 requestSender = JobCallback(11,5,
2237 Dialer, this, HttpStateData::wroteLast);
2238 Comm::Write(fd, "\r\n", 2, requestSender, NULL);
2239 return true;
2240 #else
2241 return false;
2242 #endif /* USE_HTTP_VIOLATIONS */
2243 }
2244
2245 /// if needed, write last-chunk to end the request body and return true
2246 bool
2247 HttpStateData::finishingChunkedRequest()
2248 {
2249 if (flags.sentLastChunk) {
2250 debugs(11, 5, HERE << "already sent last-chunk");
2251 return false;
2252 }
2253
2254 Must(receivedWholeRequestBody); // or we should not be sending last-chunk
2255 flags.sentLastChunk = true;
2256
2257 typedef CommCbMemFunT<HttpStateData, CommIoCbParams> Dialer;
2258 requestSender = JobCallback(11,5,
2259 Dialer, this, HttpStateData::wroteLast);
2260 Comm::Write(fd, "0\r\n\r\n", 5, requestSender, NULL);
2261 return true;
2262 }
2263
2264 void
2265 HttpStateData::doneSendingRequestBody()
2266 {
2267 ServerStateData::doneSendingRequestBody();
2268 debugs(11,5, HERE << "doneSendingRequestBody: FD " << fd);
2269
2270 // do we need to write something after the last body byte?
2271 if (flags.chunked_request && finishingChunkedRequest())
2272 return;
2273 if (!flags.chunked_request && finishingBrokenPost())
2274 return;
2275
2276 sendComplete();
2277 }
2278
2279 // more origin request body data is available
2280 void
2281 HttpStateData::handleMoreRequestBodyAvailable()
2282 {
2283 if (eof || fd < 0) {
2284 // XXX: we should check this condition in other callbacks then!
2285 // TODO: Check whether this can actually happen: We should unsubscribe
2286 // as a body consumer when the above condition(s) are detected.
2287 debugs(11, 1, HERE << "Transaction aborted while reading HTTP body");
2288 return;
2289 }
2290
2291 assert(requestBodySource != NULL);
2292
2293 if (requestBodySource->buf().hasContent()) {
2294 // XXX: why does not this trigger a debug message on every request?
2295
2296 if (flags.headers_parsed && !flags.abuse_detected) {
2297 flags.abuse_detected = 1;
2298 debugs(11, 1, "http handleMoreRequestBodyAvailable: Likely proxy abuse detected '" << orig_request->client_addr << "' -> '" << entry->url() << "'" );
2299
2300 if (virginReply()->sline.status == HTTP_INVALID_HEADER) {
2301 comm_close(fd);
2302 return;
2303 }
2304 }
2305 }
2306
2307 HttpStateData::handleMoreRequestBodyAvailable();
2308 }
2309
2310 // premature end of the request body
2311 void
2312 HttpStateData::handleRequestBodyProducerAborted()
2313 {
2314 ServerStateData::handleRequestBodyProducerAborted();
2315 if (entry->isEmpty()) {
2316 debugs(11, 3, "request body aborted: FD " << fd);
2317 ErrorState *err;
2318 // We usually get here when ICAP REQMOD aborts during body processing.
2319 // We might also get here if client-side aborts, but then our response
2320 // should not matter because either client-side will provide its own or
2321 // there will be no response at all (e.g., if the the client has left).
2322 err = errorCon(ERR_ICAP_FAILURE, HTTP_INTERNAL_SERVER_ERROR, fwd->request);
2323 err->xerrno = ERR_DETAIL_SRV_REQMOD_REQ_BODY;
2324 fwd->fail(err);
2325 }
2326
2327 abortTransaction("request body producer aborted");
2328 }
2329
2330 // called when we wrote request headers(!) or a part of the body
2331 void
2332 HttpStateData::sentRequestBody(const CommIoCbParams &io)
2333 {
2334 if (io.size > 0)
2335 kb_incr(&statCounter.server.http.kbytes_out, io.size);
2336
2337 ServerStateData::sentRequestBody(io);
2338 }
2339
2340 // Quickly abort the transaction
2341 // TODO: destruction should be sufficient as the destructor should cleanup,
2342 // including canceling close handlers
2343 void
2344 HttpStateData::abortTransaction(const char *reason)
2345 {
2346 debugs(11,5, HERE << "aborting transaction for " << reason <<
2347 "; FD " << fd << ", this " << this);
2348
2349 if (fd >= 0) {
2350 comm_close(fd);
2351 return;
2352 }
2353
2354 fwd->handleUnregisteredServerEnd();
2355 deleteThis("HttpStateData::abortTransaction");
2356 }
2357
2358 HttpRequest *
2359 HttpStateData::originalRequest()
2360 {
2361 return orig_request;
2362 }