1 // -*- mode: cpp; mode: fold -*-
3 /* ######################################################################
5 HTTP and HTTPS share a lot of common code and these classes are
6 exactly the dumping ground for this common code
8 ##################################################################### */
10 // Include Files /*{{{*/
13 #include <apt-pkg/acquire-method.h>
14 #include <apt-pkg/configuration.h>
15 #include <apt-pkg/error.h>
16 #include <apt-pkg/fileutl.h>
17 #include <apt-pkg/strutl.h>
39 string
ServerMethod::FailFile
;
40 int ServerMethod::FailFd
= -1;
41 time_t ServerMethod::FailTime
= 0;
43 // ServerState::RunHeaders - Get the headers before the data /*{{{*/
44 // ---------------------------------------------------------------------
45 /* Returns 0 if things are OK, 1 if an IO error occurred and 2 if a header
46 parse error occurred */
47 ServerState::RunHeadersResult
ServerState::RunHeaders(FileFd
* const File
,
48 const std::string
&Uri
)
52 Owner
->Status(_("Waiting for headers"));
67 if (ReadHeaderLines(Data
) == false)
70 if (Owner
->Debug
== true)
71 clog
<< "Answer for: " << Uri
<< endl
<< Data
;
73 for (string::const_iterator I
= Data
.begin(); I
< Data
.end(); ++I
)
75 string::const_iterator J
= I
;
76 for (; J
!= Data
.end() && *J
!= '\n' && *J
!= '\r'; ++J
);
77 if (HeaderLine(string(I
,J
)) == false)
78 return RUN_HEADERS_PARSE_ERROR
;
82 // 100 Continue is a Nop...
86 // Tidy up the connection persistence state.
87 if (Encoding
== Closes
&& HaveContent
== true)
90 return RUN_HEADERS_OK
;
92 while (LoadNextResponse(false, File
) == true);
94 return RUN_HEADERS_IO_ERROR
;
97 // ServerState::HeaderLine - Process a header line /*{{{*/
98 // ---------------------------------------------------------------------
100 bool ServerState::HeaderLine(string Line
)
102 if (Line
.empty() == true)
105 string::size_type Pos
= Line
.find(' ');
106 if (Pos
== string::npos
|| Pos
+1 > Line
.length())
108 // Blah, some servers use "connection:closes", evil.
109 Pos
= Line
.find(':');
110 if (Pos
== string::npos
|| Pos
+ 2 > Line
.length())
111 return _error
->Error(_("Bad header line"));
115 // Parse off any trailing spaces between the : and the next word.
116 string::size_type Pos2
= Pos
;
117 while (Pos2
< Line
.length() && isspace_ascii(Line
[Pos2
]) != 0)
120 string Tag
= string(Line
,0,Pos
);
121 string Val
= string(Line
,Pos2
);
123 if (stringcasecmp(Tag
.c_str(),Tag
.c_str()+4,"HTTP") == 0)
125 // Evil servers return no version
128 int const elements
= sscanf(Line
.c_str(),"HTTP/%3u.%3u %3u%359[^\n]",&Major
,&Minor
,&Result
,Code
);
132 if (Owner
!= NULL
&& Owner
->Debug
== true)
133 clog
<< "HTTP server doesn't give Reason-Phrase for " << Result
<< std::endl
;
135 else if (elements
!= 4)
136 return _error
->Error(_("The HTTP server sent an invalid reply header"));
142 if (sscanf(Line
.c_str(),"HTTP %3u%359[^\n]",&Result
,Code
) != 2)
143 return _error
->Error(_("The HTTP server sent an invalid reply header"));
146 /* Check the HTTP response header to get the default persistence
152 if (Major
== 1 && Minor
== 0)
167 if (stringcasecmp(Tag
,"Content-Length:") == 0)
169 if (Encoding
== Closes
)
173 unsigned long long * DownloadSizePtr
= &DownloadSize
;
175 DownloadSizePtr
= &JunkSize
;
177 *DownloadSizePtr
= strtoull(Val
.c_str(), NULL
, 10);
178 if (*DownloadSizePtr
>= std::numeric_limits
<unsigned long long>::max())
179 return _error
->Errno("HeaderLine", _("The HTTP server sent an invalid Content-Length header"));
180 else if (*DownloadSizePtr
== 0)
183 // On partial content (206) the Content-Length less than the real
184 // size, so do not set it here but leave that to the Content-Range
186 if(Result
!= 206 && TotalFileSize
== 0)
187 TotalFileSize
= DownloadSize
;
192 if (stringcasecmp(Tag
,"Content-Type:") == 0)
198 if (stringcasecmp(Tag
,"Content-Range:") == 0)
202 // §14.16 says 'byte-range-resp-spec' should be a '*' in case of 416
203 if (Result
== 416 && sscanf(Val
.c_str(), "bytes */%llu",&TotalFileSize
) == 1)
204 ; // we got the expected filesize which is all we wanted
205 else if (sscanf(Val
.c_str(),"bytes %llu-%*u/%llu",&StartPos
,&TotalFileSize
) != 2)
206 return _error
->Error(_("The HTTP server sent an invalid Content-Range header"));
207 if ((unsigned long long)StartPos
> TotalFileSize
)
208 return _error
->Error(_("This HTTP server has broken range support"));
210 // figure out what we will download
211 DownloadSize
= TotalFileSize
- StartPos
;
215 if (stringcasecmp(Tag
,"Transfer-Encoding:") == 0)
218 if (stringcasecmp(Val
,"chunked") == 0)
223 if (stringcasecmp(Tag
,"Connection:") == 0)
225 if (stringcasecmp(Val
,"close") == 0)
227 if (stringcasecmp(Val
,"keep-alive") == 0)
232 if (stringcasecmp(Tag
,"Last-Modified:") == 0)
234 if (RFC1123StrToTime(Val
.c_str(), Date
) == false)
235 return _error
->Error(_("Unknown date format"));
239 if (stringcasecmp(Tag
,"Location:") == 0)
248 // ServerState::ServerState - Constructor /*{{{*/
249 ServerState::ServerState(URI Srv
, ServerMethod
*Owner
) :
250 DownloadSize(0), ServerName(Srv
), TimeOut(120), Owner(Owner
)
255 bool ServerState::AddPartialFileToHashes(FileFd
&File
) /*{{{*/
257 File
.Truncate(StartPos
);
258 return GetHashes()->AddFD(File
, StartPos
);
262 // ServerMethod::DealWithHeaders - Handle the retrieved header data /*{{{*/
263 // ---------------------------------------------------------------------
264 /* We look at the header data we got back from the server and decide what
265 to do. Returns DealWithHeadersResult (see http.h for details).
267 ServerMethod::DealWithHeadersResult
268 ServerMethod::DealWithHeaders(FetchResult
&Res
)
271 if (Server
->Result
== 304)
273 RemoveFile("server", Queue
->DestFile
);
275 Res
.LastModified
= Queue
->LastModified
;
281 * Note that it is only OK for us to treat all redirection the same
282 * because we *always* use GET, not other HTTP methods. There are
283 * three redirection codes for which it is not appropriate that we
284 * redirect. Pass on those codes so the error handling kicks in.
287 && (Server
->Result
> 300 && Server
->Result
< 400)
288 && (Server
->Result
!= 300 // Multiple Choices
289 && Server
->Result
!= 304 // Not Modified
290 && Server
->Result
!= 306)) // (Not part of HTTP/1.1, reserved)
292 if (Server
->Location
.empty() == true);
293 else if (Server
->Location
[0] == '/' && Queue
->Uri
.empty() == false)
295 URI Uri
= Queue
->Uri
;
296 if (Uri
.Host
.empty() == false)
297 NextURI
= URI::SiteOnly(Uri
);
300 NextURI
.append(DeQuoteString(Server
->Location
));
301 return TRY_AGAIN_OR_REDIRECT
;
305 NextURI
= DeQuoteString(Server
->Location
);
306 URI tmpURI
= NextURI
;
307 URI Uri
= Queue
->Uri
;
308 // same protocol redirects are okay
309 if (tmpURI
.Access
== Uri
.Access
)
310 return TRY_AGAIN_OR_REDIRECT
;
311 // as well as http to https
312 else if (Uri
.Access
== "http" && tmpURI
.Access
== "https")
313 return TRY_AGAIN_OR_REDIRECT
;
315 /* else pass through for error message */
317 // retry after an invalid range response without partial data
318 else if (Server
->Result
== 416)
321 if (stat(Queue
->DestFile
.c_str(),&SBuf
) >= 0 && SBuf
.st_size
> 0)
323 bool partialHit
= false;
324 if (Queue
->ExpectedHashes
.usable() == true)
326 Hashes
resultHashes(Queue
->ExpectedHashes
);
327 FileFd
file(Queue
->DestFile
, FileFd::ReadOnly
);
328 Server
->TotalFileSize
= file
.FileSize();
329 Server
->Date
= file
.ModificationTime();
330 resultHashes
.AddFD(file
);
331 HashStringList
const hashList
= resultHashes
.GetHashStringList();
332 partialHit
= (Queue
->ExpectedHashes
== hashList
);
334 else if ((unsigned long long)SBuf
.st_size
== Server
->TotalFileSize
)
336 if (partialHit
== true)
338 // the file is completely downloaded, but was not moved
339 if (Server
->HaveContent
== true)
341 // Send to error page to dev/null
342 FileFd
DevNull("/dev/null",FileFd::WriteExists
);
343 Server
->RunData(&DevNull
);
345 Server
->HaveContent
= false;
346 Server
->StartPos
= Server
->TotalFileSize
;
347 Server
->Result
= 200;
349 else if (RemoveFile("server", Queue
->DestFile
))
351 NextURI
= Queue
->Uri
;
352 return TRY_AGAIN_OR_REDIRECT
;
357 /* We have a reply we don't handle. This should indicate a perm server
359 if (Server
->Result
< 200 || Server
->Result
>= 300)
362 strprintf(err
, "HttpError%u", Server
->Result
);
364 _error
->Error("%u %s", Server
->Result
, Server
->Code
);
365 if (Server
->HaveContent
== true)
366 return ERROR_WITH_CONTENT_PAGE
;
367 return ERROR_UNRECOVERABLE
;
370 // This is some sort of 2xx 'data follows' reply
371 Res
.LastModified
= Server
->Date
;
372 Res
.Size
= Server
->TotalFileSize
;
376 File
= new FileFd(Queue
->DestFile
,FileFd::WriteAny
);
377 if (_error
->PendingError() == true)
378 return ERROR_NOT_FROM_SERVER
;
380 FailFile
= Queue
->DestFile
;
381 FailFile
.c_str(); // Make sure we don't do a malloc in the signal handler
383 FailTime
= Server
->Date
;
385 if (Server
->InitHashes(Queue
->ExpectedHashes
) == false || Server
->AddPartialFileToHashes(*File
) == false)
387 _error
->Errno("read",_("Problem hashing file"));
388 return ERROR_NOT_FROM_SERVER
;
390 if (Server
->StartPos
> 0)
391 Res
.ResumePoint
= Server
->StartPos
;
393 SetNonBlock(File
->Fd(),true);
397 // ServerMethod::SigTerm - Handle a fatal signal /*{{{*/
398 // ---------------------------------------------------------------------
399 /* This closes and timestamps the open file. This is necessary to get
400 resume behavoir on user abort */
401 void ServerMethod::SigTerm(int)
406 struct timeval times
[2];
407 times
[0].tv_sec
= FailTime
;
408 times
[1].tv_sec
= FailTime
;
409 times
[0].tv_usec
= times
[1].tv_usec
= 0;
410 utimes(FailFile
.c_str(), times
);
416 // ServerMethod::Fetch - Fetch an item /*{{{*/
417 // ---------------------------------------------------------------------
418 /* This adds an item to the pipeline. We keep the pipeline at a fixed
420 bool ServerMethod::Fetch(FetchItem
*)
422 if (Server
== nullptr || QueueBack
== nullptr)
425 // If pipelining is disabled, we only queue 1 request
426 auto const AllowedDepth
= Server
->Pipeline
? PipelineDepth
: 0;
427 // how deep is our pipeline currently?
428 decltype(PipelineDepth
) CurrentDepth
= 0;
429 for (FetchItem
const *I
= Queue
; I
!= QueueBack
; I
= I
->Next
)
433 // Make sure we stick with the same server
434 if (Server
->Comp(QueueBack
->Uri
) == false)
437 bool const UsableHashes
= QueueBack
->ExpectedHashes
.usable();
438 // if we have no hashes, do at most one such request
439 // as we can't fixup pipeling misbehaviors otherwise
440 if (CurrentDepth
!= 0 && UsableHashes
== false)
443 if (UsableHashes
&& FileExists(QueueBack
->DestFile
))
445 FileFd
partial(QueueBack
->DestFile
, FileFd::ReadOnly
);
446 Hashes
wehave(QueueBack
->ExpectedHashes
);
447 if (QueueBack
->ExpectedHashes
.FileSize() == partial
.FileSize())
449 if (wehave
.AddFD(partial
) &&
450 wehave
.GetHashStringList() == QueueBack
->ExpectedHashes
)
453 Res
.Filename
= QueueBack
->DestFile
;
454 Res
.ResumePoint
= QueueBack
->ExpectedHashes
.FileSize();
456 // move item to the start of the queue as URIDone will
457 // always dequeued the first item in the queue
458 if (Queue
!= QueueBack
)
460 FetchItem
*Prev
= Queue
;
461 for (; Prev
->Next
!= QueueBack
; Prev
= Prev
->Next
)
462 /* look for the previous queue item */;
463 Prev
->Next
= QueueBack
->Next
;
464 QueueBack
->Next
= Queue
;
466 QueueBack
= Prev
->Next
;
468 Res
.TakeHashes(wehave
);
473 RemoveFile("Fetch-Partial", QueueBack
->DestFile
);
476 auto const Tmp
= QueueBack
;
477 QueueBack
= QueueBack
->Next
;
480 } while (CurrentDepth
<= AllowedDepth
&& QueueBack
!= nullptr);
485 // ServerMethod::Loop - Main loop /*{{{*/
486 int ServerMethod::Loop()
488 typedef vector
<string
> StringVector
;
489 typedef vector
<string
>::iterator StringVectorIterator
;
490 map
<string
, StringVector
> Redirected
;
492 signal(SIGTERM
,SigTerm
);
493 signal(SIGINT
,SigTerm
);
500 // We have no commands, wait for some to arrive
503 if (WaitFd(STDIN_FILENO
) == false)
507 /* Run messages, we can accept 0 (no message) if we didn't
508 do a WaitFd above.. Otherwise the FD is closed. */
509 int Result
= Run(true);
510 if (Result
!= -1 && (Result
!= 0 || Queue
== 0))
512 if(FailReason
.empty() == false ||
513 _config
->FindB("Acquire::http::DependOnSTDIN", true) == true)
522 // Connect to the server
523 if (Server
== 0 || Server
->Comp(Queue
->Uri
) == false)
524 Server
= CreateServerState(Queue
->Uri
);
526 /* If the server has explicitly said this is the last connection
527 then we pre-emptively shut down the pipeline and tear down
528 the connection. This will speed up HTTP/1.0 servers a tad
529 since we don't have to wait for the close sequence to
531 if (Server
->Persistent
== false)
534 // Reset the pipeline
535 if (Server
->IsOpen() == false)
538 // Connnect to the host
539 if (Server
->Open() == false)
546 // Fill the pipeline.
549 // Fetch the next URL header data from the server.
550 switch (Server
->RunHeaders(File
, Queue
->Uri
))
552 case ServerState::RUN_HEADERS_OK
:
555 // The header data is bad
556 case ServerState::RUN_HEADERS_PARSE_ERROR
:
558 _error
->Error(_("Bad header data"));
564 // The server closed a connection during the header get..
566 case ServerState::RUN_HEADERS_IO_ERROR
:
571 Server
->Pipeline
= false;
572 Server
->PipelineAllowed
= false;
574 if (FailCounter
>= 2)
576 Fail(_("Connection failed"),true);
585 // Decide what to do.
587 Res
.Filename
= Queue
->DestFile
;
588 switch (DealWithHeaders(Res
))
590 // Ok, the file is Open
598 // ensure we don't fetch too much
599 // we could do "Server->MaximumSize = Queue->MaximumSize" here
600 // but that would break the clever pipeline messup detection
601 // so instead we use the size of the biggest item in the queue
602 Server
->MaximumSize
= FindMaximumObjectSizeInQueue();
604 if (Server
->HaveContent
)
605 Result
= Server
->RunData(File
);
607 /* If the server is sending back sizeless responses then fill in
610 Res
.Size
= File
->Size();
612 // Close the file, destroy the FD object and timestamp it
618 struct timeval times
[2];
619 times
[0].tv_sec
= times
[1].tv_sec
= Server
->Date
;
620 times
[0].tv_usec
= times
[1].tv_usec
= 0;
621 utimes(Queue
->DestFile
.c_str(), times
);
623 // Send status to APT
626 Hashes
* const resultHashes
= Server
->GetHashes();
627 HashStringList
const hashList
= resultHashes
->GetHashStringList();
628 if (PipelineDepth
!= 0 && Queue
->ExpectedHashes
.usable() == true && Queue
->ExpectedHashes
!= hashList
)
630 // we did not get the expected hash… mhhh:
631 // could it be that server/proxy messed up pipelining?
632 FetchItem
* BeforeI
= Queue
;
633 for (FetchItem
*I
= Queue
->Next
; I
!= 0 && I
!= QueueBack
; I
= I
->Next
)
635 if (I
->ExpectedHashes
.usable() == true && I
->ExpectedHashes
== hashList
)
637 // yes, he did! Disable pipelining and rewrite queue
638 if (Server
->Pipeline
== true)
640 // FIXME: fake a warning message as we have no proper way of communicating here
642 strprintf(out
, _("Automatically disabled %s due to incorrect response from server/proxy. (man 5 apt.conf)"), "Acquire::http::PipelineDepth");
643 std::cerr
<< "W: " << out
<< std::endl
;
644 Server
->Pipeline
= false;
645 Server
->PipelineAllowed
= false;
646 // we keep the PipelineDepth value so that the rest of the queue can be fixed up as well
648 Rename(Res
.Filename
, I
->DestFile
);
649 Res
.Filename
= I
->DestFile
;
650 BeforeI
->Next
= I
->Next
;
658 Res
.TakeHashes(*resultHashes
);
663 if (Server
->IsOpen() == false)
669 if (FailCounter
>= 2)
671 Fail(_("Connection failed"),true);
693 // Hard server error, not found or something
694 case ERROR_UNRECOVERABLE
:
700 // Hard internal error, kill the connection and fail
701 case ERROR_NOT_FROM_SERVER
:
712 // We need to flush the data, the header is like a 404 w/ error text
713 case ERROR_WITH_CONTENT_PAGE
:
717 // Send to content to dev/null
718 File
= new FileFd("/dev/null",FileFd::WriteExists
);
719 Server
->RunData(File
);
725 // Try again with a new URL
726 case TRY_AGAIN_OR_REDIRECT
:
728 // Clear rest of response if there is content
729 if (Server
->HaveContent
)
731 File
= new FileFd("/dev/null",FileFd::WriteExists
);
732 Server
->RunData(File
);
737 /* Detect redirect loops. No more redirects are allowed
738 after the same URI is seen twice in a queue item. */
739 StringVector
&R
= Redirected
[Queue
->DestFile
];
740 bool StopRedirects
= false;
741 if (R
.empty() == true)
742 R
.push_back(Queue
->Uri
);
743 else if (R
[0] == "STOP" || R
.size() > 10)
744 StopRedirects
= true;
747 for (StringVectorIterator I
= R
.begin(); I
!= R
.end(); ++I
)
748 if (Queue
->Uri
== *I
)
754 R
.push_back(Queue
->Uri
);
757 if (StopRedirects
== false)
766 Fail(_("Internal error"));
776 unsigned long long ServerMethod::FindMaximumObjectSizeInQueue() const /*{{{*/
778 unsigned long long MaxSizeInQueue
= 0;
779 for (FetchItem
*I
= Queue
; I
!= 0 && I
!= QueueBack
; I
= I
->Next
)
780 MaxSizeInQueue
= std::max(MaxSizeInQueue
, I
->MaximumSize
);
781 return MaxSizeInQueue
;
784 ServerMethod::ServerMethod(char const * const Binary
, char const * const Ver
,unsigned long const Flags
) :/*{{{*/
785 aptMethod(Binary
, Ver
, Flags
), Server(nullptr), File(NULL
), PipelineDepth(10),
786 AllowRedirect(false), Debug(false)