]> git.saurik.com Git - apt.git/blob - methods/server.cc
ignore AllowMem parameter in cache generation
[apt.git] / methods / server.cc
1 // -*- mode: cpp; mode: fold -*-
2 // Description /*{{{*/
3 /* ######################################################################
4
5 HTTP and HTTPS share a lot of common code and these classes are
6 exactly the dumping ground for this common code
7
8 ##################################################################### */
9 /*}}}*/
10 // Include Files /*{{{*/
11 #include <config.h>
12
13 #include <apt-pkg/acquire-method.h>
14 #include <apt-pkg/configuration.h>
15 #include <apt-pkg/error.h>
16 #include <apt-pkg/fileutl.h>
17 #include <apt-pkg/strutl.h>
18
19 #include <ctype.h>
20 #include <signal.h>
21 #include <stdio.h>
22 #include <stdlib.h>
23 #include <sys/stat.h>
24 #include <sys/time.h>
25 #include <time.h>
26 #include <unistd.h>
27 #include <iostream>
28 #include <limits>
29 #include <map>
30 #include <string>
31 #include <vector>
32
33 #include "server.h"
34
35 #include <apti18n.h>
36 /*}}}*/
37 using namespace std;
38
39 string ServerMethod::FailFile;
40 int ServerMethod::FailFd = -1;
41 time_t ServerMethod::FailTime = 0;
42
43 // ServerState::RunHeaders - Get the headers before the data /*{{{*/
44 // ---------------------------------------------------------------------
45 /* Returns 0 if things are OK, 1 if an IO error occurred and 2 if a header
46 parse error occurred */
47 ServerState::RunHeadersResult ServerState::RunHeaders(FileFd * const File,
48 const std::string &Uri)
49 {
50 State = Header;
51
52 Owner->Status(_("Waiting for headers"));
53
54 Major = 0;
55 Minor = 0;
56 Result = 0;
57 TotalFileSize = 0;
58 JunkSize = 0;
59 StartPos = 0;
60 Encoding = Closes;
61 HaveContent = false;
62 time(&Date);
63
64 do
65 {
66 string Data;
67 if (ReadHeaderLines(Data) == false)
68 continue;
69
70 if (Owner->Debug == true)
71 clog << "Answer for: " << Uri << endl << Data;
72
73 for (string::const_iterator I = Data.begin(); I < Data.end(); ++I)
74 {
75 string::const_iterator J = I;
76 for (; J != Data.end() && *J != '\n' && *J != '\r'; ++J);
77 if (HeaderLine(string(I,J)) == false)
78 return RUN_HEADERS_PARSE_ERROR;
79 I = J;
80 }
81
82 // 100 Continue is a Nop...
83 if (Result == 100)
84 continue;
85
86 // Tidy up the connection persistence state.
87 if (Encoding == Closes && HaveContent == true)
88 Persistent = false;
89
90 return RUN_HEADERS_OK;
91 }
92 while (LoadNextResponse(false, File) == true);
93
94 return RUN_HEADERS_IO_ERROR;
95 }
96 /*}}}*/
97 // ServerState::HeaderLine - Process a header line /*{{{*/
98 // ---------------------------------------------------------------------
99 /* */
100 bool ServerState::HeaderLine(string Line)
101 {
102 if (Line.empty() == true)
103 return true;
104
105 string::size_type Pos = Line.find(' ');
106 if (Pos == string::npos || Pos+1 > Line.length())
107 {
108 // Blah, some servers use "connection:closes", evil.
109 Pos = Line.find(':');
110 if (Pos == string::npos || Pos + 2 > Line.length())
111 return _error->Error(_("Bad header line"));
112 Pos++;
113 }
114
115 // Parse off any trailing spaces between the : and the next word.
116 string::size_type Pos2 = Pos;
117 while (Pos2 < Line.length() && isspace(Line[Pos2]) != 0)
118 Pos2++;
119
120 string Tag = string(Line,0,Pos);
121 string Val = string(Line,Pos2);
122
123 if (stringcasecmp(Tag.c_str(),Tag.c_str()+4,"HTTP") == 0)
124 {
125 // Evil servers return no version
126 if (Line[4] == '/')
127 {
128 int const elements = sscanf(Line.c_str(),"HTTP/%3u.%3u %3u%359[^\n]",&Major,&Minor,&Result,Code);
129 if (elements == 3)
130 {
131 Code[0] = '\0';
132 if (Owner != NULL && Owner->Debug == true)
133 clog << "HTTP server doesn't give Reason-Phrase for " << Result << std::endl;
134 }
135 else if (elements != 4)
136 return _error->Error(_("The HTTP server sent an invalid reply header"));
137 }
138 else
139 {
140 Major = 0;
141 Minor = 9;
142 if (sscanf(Line.c_str(),"HTTP %3u%359[^\n]",&Result,Code) != 2)
143 return _error->Error(_("The HTTP server sent an invalid reply header"));
144 }
145
146 /* Check the HTTP response header to get the default persistence
147 state. */
148 if (Major < 1)
149 Persistent = false;
150 else
151 {
152 if (Major == 1 && Minor == 0)
153 Persistent = false;
154 else
155 Persistent = true;
156 }
157
158 return true;
159 }
160
161 if (stringcasecmp(Tag,"Content-Length:") == 0)
162 {
163 if (Encoding == Closes)
164 Encoding = Stream;
165 HaveContent = true;
166
167 unsigned long long * DownloadSizePtr = &DownloadSize;
168 if (Result == 416)
169 DownloadSizePtr = &JunkSize;
170
171 *DownloadSizePtr = strtoull(Val.c_str(), NULL, 10);
172 if (*DownloadSizePtr >= std::numeric_limits<unsigned long long>::max())
173 return _error->Errno("HeaderLine", _("The HTTP server sent an invalid Content-Length header"));
174 else if (*DownloadSizePtr == 0)
175 HaveContent = false;
176
177 // On partial content (206) the Content-Length less than the real
178 // size, so do not set it here but leave that to the Content-Range
179 // header instead
180 if(Result != 206 && TotalFileSize == 0)
181 TotalFileSize = DownloadSize;
182
183 return true;
184 }
185
186 if (stringcasecmp(Tag,"Content-Type:") == 0)
187 {
188 HaveContent = true;
189 return true;
190 }
191
192 if (stringcasecmp(Tag,"Content-Range:") == 0)
193 {
194 HaveContent = true;
195
196 // §14.16 says 'byte-range-resp-spec' should be a '*' in case of 416
197 if (Result == 416 && sscanf(Val.c_str(), "bytes */%llu",&TotalFileSize) == 1)
198 ; // we got the expected filesize which is all we wanted
199 else if (sscanf(Val.c_str(),"bytes %llu-%*u/%llu",&StartPos,&TotalFileSize) != 2)
200 return _error->Error(_("The HTTP server sent an invalid Content-Range header"));
201 if ((unsigned long long)StartPos > TotalFileSize)
202 return _error->Error(_("This HTTP server has broken range support"));
203
204 // figure out what we will download
205 DownloadSize = TotalFileSize - StartPos;
206 return true;
207 }
208
209 if (stringcasecmp(Tag,"Transfer-Encoding:") == 0)
210 {
211 HaveContent = true;
212 if (stringcasecmp(Val,"chunked") == 0)
213 Encoding = Chunked;
214 return true;
215 }
216
217 if (stringcasecmp(Tag,"Connection:") == 0)
218 {
219 if (stringcasecmp(Val,"close") == 0)
220 Persistent = false;
221 if (stringcasecmp(Val,"keep-alive") == 0)
222 Persistent = true;
223 return true;
224 }
225
226 if (stringcasecmp(Tag,"Last-Modified:") == 0)
227 {
228 if (RFC1123StrToTime(Val.c_str(), Date) == false)
229 return _error->Error(_("Unknown date format"));
230 return true;
231 }
232
233 if (stringcasecmp(Tag,"Location:") == 0)
234 {
235 Location = Val;
236 return true;
237 }
238
239 return true;
240 }
241 /*}}}*/
242 // ServerState::ServerState - Constructor /*{{{*/
243 ServerState::ServerState(URI Srv, ServerMethod *Owner) : ServerName(Srv), TimeOut(120), Owner(Owner)
244 {
245 Reset();
246 }
247 /*}}}*/
248 bool ServerState::AddPartialFileToHashes(FileFd &File) /*{{{*/
249 {
250 File.Truncate(StartPos);
251 return GetHashes()->AddFD(File, StartPos);
252 }
253 /*}}}*/
254
255 bool ServerMethod::Configuration(string Message) /*{{{*/
256 {
257 if (pkgAcqMethod::Configuration(Message) == false)
258 return false;
259
260 DropPrivsOrDie();
261
262 return true;
263 }
264 /*}}}*/
265
266 // ServerMethod::DealWithHeaders - Handle the retrieved header data /*{{{*/
267 // ---------------------------------------------------------------------
268 /* We look at the header data we got back from the server and decide what
269 to do. Returns DealWithHeadersResult (see http.h for details).
270 */
271 ServerMethod::DealWithHeadersResult
272 ServerMethod::DealWithHeaders(FetchResult &Res)
273 {
274 // Not Modified
275 if (Server->Result == 304)
276 {
277 unlink(Queue->DestFile.c_str());
278 Res.IMSHit = true;
279 Res.LastModified = Queue->LastModified;
280 return IMS_HIT;
281 }
282
283 /* Redirect
284 *
285 * Note that it is only OK for us to treat all redirection the same
286 * because we *always* use GET, not other HTTP methods. There are
287 * three redirection codes for which it is not appropriate that we
288 * redirect. Pass on those codes so the error handling kicks in.
289 */
290 if (AllowRedirect
291 && (Server->Result > 300 && Server->Result < 400)
292 && (Server->Result != 300 // Multiple Choices
293 && Server->Result != 304 // Not Modified
294 && Server->Result != 306)) // (Not part of HTTP/1.1, reserved)
295 {
296 if (Server->Location.empty() == true);
297 else if (Server->Location[0] == '/' && Queue->Uri.empty() == false)
298 {
299 URI Uri = Queue->Uri;
300 if (Uri.Host.empty() == false)
301 NextURI = URI::SiteOnly(Uri);
302 else
303 NextURI.clear();
304 NextURI.append(DeQuoteString(Server->Location));
305 return TRY_AGAIN_OR_REDIRECT;
306 }
307 else
308 {
309 NextURI = DeQuoteString(Server->Location);
310 URI tmpURI = NextURI;
311 URI Uri = Queue->Uri;
312 // same protocol redirects are okay
313 if (tmpURI.Access == Uri.Access)
314 return TRY_AGAIN_OR_REDIRECT;
315 // as well as http to https
316 else if (Uri.Access == "http" && tmpURI.Access == "https")
317 return TRY_AGAIN_OR_REDIRECT;
318 }
319 /* else pass through for error message */
320 }
321 // retry after an invalid range response without partial data
322 else if (Server->Result == 416)
323 {
324 struct stat SBuf;
325 if (stat(Queue->DestFile.c_str(),&SBuf) >= 0 && SBuf.st_size > 0)
326 {
327 bool partialHit = false;
328 if (Queue->ExpectedHashes.usable() == true)
329 {
330 Hashes resultHashes(Queue->ExpectedHashes);
331 FileFd file(Queue->DestFile, FileFd::ReadOnly);
332 Server->TotalFileSize = file.FileSize();
333 Server->Date = file.ModificationTime();
334 resultHashes.AddFD(file);
335 HashStringList const hashList = resultHashes.GetHashStringList();
336 partialHit = (Queue->ExpectedHashes == hashList);
337 }
338 else if ((unsigned long long)SBuf.st_size == Server->TotalFileSize)
339 partialHit = true;
340 if (partialHit == true)
341 {
342 // the file is completely downloaded, but was not moved
343 if (Server->HaveContent == true)
344 {
345 // Send to error page to dev/null
346 FileFd DevNull("/dev/null",FileFd::WriteExists);
347 Server->RunData(&DevNull);
348 }
349 Server->HaveContent = false;
350 Server->StartPos = Server->TotalFileSize;
351 Server->Result = 200;
352 }
353 else if (unlink(Queue->DestFile.c_str()) == 0)
354 {
355 NextURI = Queue->Uri;
356 return TRY_AGAIN_OR_REDIRECT;
357 }
358 }
359 }
360
361 /* We have a reply we dont handle. This should indicate a perm server
362 failure */
363 if (Server->Result < 200 || Server->Result >= 300)
364 {
365 std::string err;
366 strprintf(err, "HttpError%u", Server->Result);
367 SetFailReason(err);
368 _error->Error("%u %s", Server->Result, Server->Code);
369 if (Server->HaveContent == true)
370 return ERROR_WITH_CONTENT_PAGE;
371 return ERROR_UNRECOVERABLE;
372 }
373
374 // This is some sort of 2xx 'data follows' reply
375 Res.LastModified = Server->Date;
376 Res.Size = Server->TotalFileSize;
377
378 // Open the file
379 delete File;
380 File = new FileFd(Queue->DestFile,FileFd::WriteAny);
381 if (_error->PendingError() == true)
382 return ERROR_NOT_FROM_SERVER;
383
384 FailFile = Queue->DestFile;
385 FailFile.c_str(); // Make sure we dont do a malloc in the signal handler
386 FailFd = File->Fd();
387 FailTime = Server->Date;
388
389 if (Server->InitHashes(Queue->ExpectedHashes) == false || Server->AddPartialFileToHashes(*File) == false)
390 {
391 _error->Errno("read",_("Problem hashing file"));
392 return ERROR_NOT_FROM_SERVER;
393 }
394 if (Server->StartPos > 0)
395 Res.ResumePoint = Server->StartPos;
396
397 SetNonBlock(File->Fd(),true);
398 return FILE_IS_OPEN;
399 }
400 /*}}}*/
401 // ServerMethod::SigTerm - Handle a fatal signal /*{{{*/
402 // ---------------------------------------------------------------------
403 /* This closes and timestamps the open file. This is necessary to get
404 resume behavoir on user abort */
405 void ServerMethod::SigTerm(int)
406 {
407 if (FailFd == -1)
408 _exit(100);
409
410 struct timeval times[2];
411 times[0].tv_sec = FailTime;
412 times[1].tv_sec = FailTime;
413 times[0].tv_usec = times[1].tv_usec = 0;
414 utimes(FailFile.c_str(), times);
415 close(FailFd);
416
417 _exit(100);
418 }
419 /*}}}*/
420 // ServerMethod::Fetch - Fetch an item /*{{{*/
421 // ---------------------------------------------------------------------
422 /* This adds an item to the pipeline. We keep the pipeline at a fixed
423 depth. */
424 bool ServerMethod::Fetch(FetchItem *)
425 {
426 if (Server == 0)
427 return true;
428
429 // Queue the requests
430 int Depth = -1;
431 for (FetchItem *I = Queue; I != 0 && Depth < (signed)PipelineDepth;
432 I = I->Next, Depth++)
433 {
434 if (Depth >= 0)
435 {
436 // If pipelining is disabled, we only queue 1 request
437 if (Server->Pipeline == false)
438 break;
439 // if we have no hashes, do at most one such request
440 // as we can't fixup pipeling misbehaviors otherwise
441 else if (I->ExpectedHashes.usable() == false)
442 break;
443 }
444
445 // Make sure we stick with the same server
446 if (Server->Comp(I->Uri) == false)
447 break;
448 if (QueueBack == I)
449 {
450 QueueBack = I->Next;
451 SendReq(I);
452 continue;
453 }
454 }
455
456 return true;
457 }
458 /*}}}*/
459 // ServerMethod::Loop - Main loop /*{{{*/
460 int ServerMethod::Loop()
461 {
462 typedef vector<string> StringVector;
463 typedef vector<string>::iterator StringVectorIterator;
464 map<string, StringVector> Redirected;
465
466 signal(SIGTERM,SigTerm);
467 signal(SIGINT,SigTerm);
468
469 Server = 0;
470
471 int FailCounter = 0;
472 while (1)
473 {
474 // We have no commands, wait for some to arrive
475 if (Queue == 0)
476 {
477 if (WaitFd(STDIN_FILENO) == false)
478 return 0;
479 }
480
481 /* Run messages, we can accept 0 (no message) if we didn't
482 do a WaitFd above.. Otherwise the FD is closed. */
483 int Result = Run(true);
484 if (Result != -1 && (Result != 0 || Queue == 0))
485 {
486 if(FailReason.empty() == false ||
487 _config->FindB("Acquire::http::DependOnSTDIN", true) == true)
488 return 100;
489 else
490 return 0;
491 }
492
493 if (Queue == 0)
494 continue;
495
496 // Connect to the server
497 if (Server == 0 || Server->Comp(Queue->Uri) == false)
498 {
499 delete Server;
500 Server = CreateServerState(Queue->Uri);
501 }
502 /* If the server has explicitly said this is the last connection
503 then we pre-emptively shut down the pipeline and tear down
504 the connection. This will speed up HTTP/1.0 servers a tad
505 since we don't have to wait for the close sequence to
506 complete */
507 if (Server->Persistent == false)
508 Server->Close();
509
510 // Reset the pipeline
511 if (Server->IsOpen() == false)
512 QueueBack = Queue;
513
514 // Connnect to the host
515 if (Server->Open() == false)
516 {
517 Fail(true);
518 delete Server;
519 Server = 0;
520 continue;
521 }
522
523 // Fill the pipeline.
524 Fetch(0);
525
526 // Fetch the next URL header data from the server.
527 switch (Server->RunHeaders(File, Queue->Uri))
528 {
529 case ServerState::RUN_HEADERS_OK:
530 break;
531
532 // The header data is bad
533 case ServerState::RUN_HEADERS_PARSE_ERROR:
534 {
535 _error->Error(_("Bad header data"));
536 Fail(true);
537 RotateDNS();
538 continue;
539 }
540
541 // The server closed a connection during the header get..
542 default:
543 case ServerState::RUN_HEADERS_IO_ERROR:
544 {
545 FailCounter++;
546 _error->Discard();
547 Server->Close();
548 Server->Pipeline = false;
549
550 if (FailCounter >= 2)
551 {
552 Fail(_("Connection failed"),true);
553 FailCounter = 0;
554 }
555
556 RotateDNS();
557 continue;
558 }
559 };
560
561 // Decide what to do.
562 FetchResult Res;
563 Res.Filename = Queue->DestFile;
564 switch (DealWithHeaders(Res))
565 {
566 // Ok, the file is Open
567 case FILE_IS_OPEN:
568 {
569 URIStart(Res);
570
571 // Run the data
572 bool Result = true;
573
574 // ensure we don't fetch too much
575 // we could do "Server->MaximumSize = Queue->MaximumSize" here
576 // but that would break the clever pipeline messup detection
577 // so instead we use the size of the biggest item in the queue
578 Server->MaximumSize = FindMaximumObjectSizeInQueue();
579
580 if (Server->HaveContent)
581 Result = Server->RunData(File);
582
583 /* If the server is sending back sizeless responses then fill in
584 the size now */
585 if (Res.Size == 0)
586 Res.Size = File->Size();
587
588 // Close the file, destroy the FD object and timestamp it
589 FailFd = -1;
590 delete File;
591 File = 0;
592
593 // Timestamp
594 struct timeval times[2];
595 times[0].tv_sec = times[1].tv_sec = Server->Date;
596 times[0].tv_usec = times[1].tv_usec = 0;
597 utimes(Queue->DestFile.c_str(), times);
598
599 // Send status to APT
600 if (Result == true)
601 {
602 Hashes * const resultHashes = Server->GetHashes();
603 HashStringList const hashList = resultHashes->GetHashStringList();
604 if (PipelineDepth != 0 && Queue->ExpectedHashes.usable() == true && Queue->ExpectedHashes != hashList)
605 {
606 // we did not get the expected hash… mhhh:
607 // could it be that server/proxy messed up pipelining?
608 FetchItem * BeforeI = Queue;
609 for (FetchItem *I = Queue->Next; I != 0 && I != QueueBack; I = I->Next)
610 {
611 if (I->ExpectedHashes.usable() == true && I->ExpectedHashes == hashList)
612 {
613 // yes, he did! Disable pipelining and rewrite queue
614 if (Server->Pipeline == true)
615 {
616 // FIXME: fake a warning message as we have no proper way of communicating here
617 std::string out;
618 strprintf(out, _("Automatically disabled %s due to incorrect response from server/proxy. (man 5 apt.conf)"), "Acquire::http::PipelineDepth");
619 std::cerr << "W: " << out << std::endl;
620 Server->Pipeline = false;
621 // we keep the PipelineDepth value so that the rest of the queue can be fixed up as well
622 }
623 Rename(Res.Filename, I->DestFile);
624 Res.Filename = I->DestFile;
625 BeforeI->Next = I->Next;
626 I->Next = Queue;
627 Queue = I;
628 break;
629 }
630 BeforeI = I;
631 }
632 }
633 Res.TakeHashes(*resultHashes);
634 URIDone(Res);
635 }
636 else
637 {
638 if (Server->IsOpen() == false)
639 {
640 FailCounter++;
641 _error->Discard();
642 Server->Close();
643
644 if (FailCounter >= 2)
645 {
646 Fail(_("Connection failed"),true);
647 FailCounter = 0;
648 }
649
650 QueueBack = Queue;
651 }
652 else
653 {
654 Server->Close();
655 Fail(true);
656 }
657 }
658 break;
659 }
660
661 // IMS hit
662 case IMS_HIT:
663 {
664 URIDone(Res);
665 break;
666 }
667
668 // Hard server error, not found or something
669 case ERROR_UNRECOVERABLE:
670 {
671 Fail();
672 break;
673 }
674
675 // Hard internal error, kill the connection and fail
676 case ERROR_NOT_FROM_SERVER:
677 {
678 delete File;
679 File = 0;
680
681 Fail();
682 RotateDNS();
683 Server->Close();
684 break;
685 }
686
687 // We need to flush the data, the header is like a 404 w/ error text
688 case ERROR_WITH_CONTENT_PAGE:
689 {
690 Fail();
691
692 // Send to content to dev/null
693 File = new FileFd("/dev/null",FileFd::WriteExists);
694 Server->RunData(File);
695 delete File;
696 File = 0;
697 break;
698 }
699
700 // Try again with a new URL
701 case TRY_AGAIN_OR_REDIRECT:
702 {
703 // Clear rest of response if there is content
704 if (Server->HaveContent)
705 {
706 File = new FileFd("/dev/null",FileFd::WriteExists);
707 Server->RunData(File);
708 delete File;
709 File = 0;
710 }
711
712 /* Detect redirect loops. No more redirects are allowed
713 after the same URI is seen twice in a queue item. */
714 StringVector &R = Redirected[Queue->DestFile];
715 bool StopRedirects = false;
716 if (R.empty() == true)
717 R.push_back(Queue->Uri);
718 else if (R[0] == "STOP" || R.size() > 10)
719 StopRedirects = true;
720 else
721 {
722 for (StringVectorIterator I = R.begin(); I != R.end(); ++I)
723 if (Queue->Uri == *I)
724 {
725 R[0] = "STOP";
726 break;
727 }
728
729 R.push_back(Queue->Uri);
730 }
731
732 if (StopRedirects == false)
733 Redirect(NextURI);
734 else
735 Fail();
736
737 break;
738 }
739
740 default:
741 Fail(_("Internal error"));
742 break;
743 }
744
745 FailCounter = 0;
746 }
747
748 return 0;
749 }
750 /*}}}*/
751 /*{{{*/
752 unsigned long long
753 ServerMethod::FindMaximumObjectSizeInQueue() const
754 {
755 unsigned long long MaxSizeInQueue = 0;
756 for (FetchItem *I = Queue; I != 0 && I != QueueBack; I = I->Next)
757 MaxSizeInQueue = std::max(MaxSizeInQueue, I->MaximumSize);
758 return MaxSizeInQueue;
759 }
760 /*}}}*/