//-*- mode: cpp; mode: fold -*- // Description /*{{{*/ // $Id: http.cc,v 1.59 2004/05/08 19:42:35 mdz Exp $ /* ###################################################################### HTTPS Acquire Method - This is the HTTPS acquire method for APT. It uses libcurl ##################################################################### */ /*}}}*/ // Include Files /*{{{*/ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "https.h" #include /*}}}*/ using namespace std; struct APT_HIDDEN CURLUserPointer { HttpsMethod * const https; HttpsMethod::FetchResult * const Res; HttpsMethod::FetchItem const * const Itm; CURLUserPointer(HttpsMethod * const https, HttpsMethod::FetchResult * const Res, HttpsMethod::FetchItem const * const Itm) : https(https), Res(Res), Itm(Itm) {} }; size_t HttpsMethod::parse_header(void *buffer, size_t size, size_t nmemb, void *userp) { size_t len = size * nmemb; CURLUserPointer *me = static_cast(userp); std::string line((char*) buffer, len); for (--len; len > 0; --len) if (isspace_ascii(line[len]) == 0) { ++len; break; } line.erase(len); if (line.empty() == true) { if (me->https->Server->Result != 416 && me->https->Server->StartPos != 0) ; else if (me->https->Server->Result == 416) { bool partialHit = false; if (me->Itm->ExpectedHashes.usable() == true) { Hashes resultHashes(me->Itm->ExpectedHashes); FileFd file(me->Itm->DestFile, FileFd::ReadOnly); me->https->Server->TotalFileSize = file.FileSize(); me->https->Server->Date = file.ModificationTime(); resultHashes.AddFD(file); HashStringList const hashList = resultHashes.GetHashStringList(); partialHit = (me->Itm->ExpectedHashes == hashList); } else if (me->https->Server->Result == 416 && me->https->Server->TotalFileSize == me->https->File->FileSize()) partialHit = true; if (partialHit == true) { me->https->Server->Result = 200; me->https->Server->StartPos = me->https->Server->TotalFileSize; // the actual size is not important for https as curl will deal with it // by itself and e.g. doesn't bother us with transport-encoding… me->https->Server->JunkSize = std::numeric_limits::max(); } else me->https->Server->StartPos = 0; } else me->https->Server->StartPos = 0; me->Res->LastModified = me->https->Server->Date; me->Res->Size = me->https->Server->TotalFileSize; me->Res->ResumePoint = me->https->Server->StartPos; // we expect valid data, so tell our caller we get the file now if (me->https->Server->Result >= 200 && me->https->Server->Result < 300) { if (me->https->Server->JunkSize == 0 && me->Res->Size != 0 && me->Res->Size > me->Res->ResumePoint) me->https->URIStart(*me->Res); if (me->https->Server->AddPartialFileToHashes(*(me->https->File)) == false) return 0; } } else if (me->https->Server->HeaderLine(line) == false) return 0; return size*nmemb; } size_t HttpsMethod::write_data(void *buffer, size_t size, size_t nmemb, void *userp) { HttpsMethod *me = static_cast(userp); size_t buffer_size = size * nmemb; // we don't need to count the junk here, just drop anything we get as // we don't always know how long it would be, e.g. in chunked encoding. if (me->Server->JunkSize != 0) return buffer_size; if(me->File->Write(buffer, buffer_size) != true) return 0; if(me->Queue->MaximumSize > 0) { unsigned long long const TotalWritten = me->File->Tell(); if (TotalWritten > me->Queue->MaximumSize) { me->SetFailReason("MaximumSizeExceeded"); _error->Error("Writing more data than expected (%llu > %llu)", TotalWritten, me->Queue->MaximumSize); return 0; } } if (me->Server->GetHashes()->Add((unsigned char const * const)buffer, buffer_size) == false) return 0; return buffer_size; } // HttpsServerState::HttpsServerState - Constructor /*{{{*/ HttpsServerState::HttpsServerState(URI Srv,HttpsMethod * Owner) : ServerState(Srv, Owner), Hash(NULL) { TimeOut = _config->FindI("Acquire::https::Timeout",TimeOut); Reset(); } /*}}}*/ bool HttpsServerState::InitHashes(HashStringList const &ExpectedHashes) /*{{{*/ { delete Hash; Hash = new Hashes(ExpectedHashes); return true; } /*}}}*/ APT_PURE Hashes * HttpsServerState::GetHashes() /*{{{*/ { return Hash; } /*}}}*/ void HttpsMethod::SetupProxy() /*{{{*/ { URI ServerName = Queue->Uri; // Determine the proxy setting AutoDetectProxy(ServerName); // Curl should never read proxy settings from the environment, as // we determine which proxy to use. Do this for consistency among // methods and prevent an environment variable overriding a // no-proxy ("DIRECT") setting in apt.conf. curl_easy_setopt(curl, CURLOPT_PROXY, ""); // Determine the proxy setting - try https first, fallback to http and use env at last string UseProxy = _config->Find("Acquire::https::Proxy::" + ServerName.Host, _config->Find("Acquire::http::Proxy::" + ServerName.Host).c_str()); if (UseProxy.empty() == true) UseProxy = _config->Find("Acquire::https::Proxy", _config->Find("Acquire::http::Proxy").c_str()); // User want to use NO proxy, so nothing to setup if (UseProxy == "DIRECT") return; // Parse no_proxy, a comma (,) separated list of domains we don't want to use // a proxy for so we stop right here if it is in the list if (getenv("no_proxy") != 0 && CheckDomainList(ServerName.Host,getenv("no_proxy")) == true) return; if (UseProxy.empty() == true) { const char* result = getenv("https_proxy"); // FIXME: Fall back to http_proxy is to remain compatible with // existing setups and behaviour of apt.conf. This should be // deprecated in the future (including apt.conf). Most other // programs do not fall back to http proxy settings and neither // should Apt. if (result == NULL) result = getenv("http_proxy"); UseProxy = result == NULL ? "" : result; } // Determine what host and port to use based on the proxy settings if (UseProxy.empty() == false) { Proxy = UseProxy; if (Proxy.Port != 1) curl_easy_setopt(curl, CURLOPT_PROXYPORT, Proxy.Port); curl_easy_setopt(curl, CURLOPT_PROXY, Proxy.Host.c_str()); if (Proxy.User.empty() == false || Proxy.Password.empty() == false) { curl_easy_setopt(curl, CURLOPT_PROXYUSERNAME, Proxy.User.c_str()); curl_easy_setopt(curl, CURLOPT_PROXYPASSWORD, Proxy.Password.c_str()); } } } /*}}}*/ // HttpsMethod::Fetch - Fetch an item /*{{{*/ // --------------------------------------------------------------------- /* This adds an item to the pipeline. We keep the pipeline at a fixed depth. */ bool HttpsMethod::Fetch(FetchItem *Itm) { struct stat SBuf; struct curl_slist *headers=NULL; char curl_errorstr[CURL_ERROR_SIZE]; URI Uri = Itm->Uri; string remotehost = Uri.Host; // TODO: // - http::Pipeline-Depth // - error checking/reporting // - more debug options? (CURLOPT_DEBUGFUNCTION?) curl_easy_reset(curl); SetupProxy(); maybe_add_auth (Uri, _config->FindFile("Dir::Etc::netrc")); if(Uri.User.empty() || Uri.Password.empty()) { auto const netrcparts = _config->FindDir("Dir::Etc::netrcparts"); if (not netrcparts.empty()) { _error->PushToStack(); for (auto const &netrc : GetListOfFilesInDir(netrcparts, "conf", true, true)) { maybe_add_auth (Uri, netrc); if (Uri.User.empty() == false || Uri.Password.empty() == false) break; } _error->RevertToStack(); } } // The "+" is encoded as a workaround for a amazon S3 bug // see LP bugs #1003633 and #1086997. (taken from http method) Uri.Path = QuoteString(Uri.Path, "+~ "); FetchResult Res; CURLUserPointer userp(this, &Res, Itm); // callbacks curl_easy_setopt(curl, CURLOPT_URL, static_cast(Uri).c_str()); curl_easy_setopt(curl, CURLOPT_HEADERFUNCTION, parse_header); curl_easy_setopt(curl, CURLOPT_WRITEHEADER, &userp); curl_easy_setopt(curl, CURLOPT_WRITEFUNCTION, write_data); curl_easy_setopt(curl, CURLOPT_WRITEDATA, this); // options curl_easy_setopt(curl, CURLOPT_NOPROGRESS, true); curl_easy_setopt(curl, CURLOPT_FILETIME, true); // only allow curl to handle https, not the other stuff it supports curl_easy_setopt(curl, CURLOPT_PROTOCOLS, CURLPROTO_HTTPS); curl_easy_setopt(curl, CURLOPT_REDIR_PROTOCOLS, CURLPROTO_HTTPS); // SSL parameters are set by default to the common (non mirror-specific) value // if available (or a default one) and gets overload by mirror-specific ones. // File containing the list of trusted CA. string cainfo = _config->Find("Acquire::https::CaInfo",""); string knob = "Acquire::https::"+remotehost+"::CaInfo"; cainfo = _config->Find(knob.c_str(),cainfo.c_str()); if(cainfo.empty() == false) curl_easy_setopt(curl, CURLOPT_CAINFO,cainfo.c_str()); // Check server certificate against previous CA list ... bool peer_verify = _config->FindB("Acquire::https::Verify-Peer",true); knob = "Acquire::https::" + remotehost + "::Verify-Peer"; peer_verify = _config->FindB(knob.c_str(), peer_verify); curl_easy_setopt(curl, CURLOPT_SSL_VERIFYPEER, peer_verify); // ... and hostname against cert CN or subjectAltName bool verify = _config->FindB("Acquire::https::Verify-Host",true); knob = "Acquire::https::"+remotehost+"::Verify-Host"; verify = _config->FindB(knob.c_str(),verify); int const default_verify = (verify == true) ? 2 : 0; curl_easy_setopt(curl, CURLOPT_SSL_VERIFYHOST, default_verify); // Also enforce issuer of server certificate using its cert string issuercert = _config->Find("Acquire::https::IssuerCert",""); knob = "Acquire::https::"+remotehost+"::IssuerCert"; issuercert = _config->Find(knob.c_str(),issuercert.c_str()); if(issuercert.empty() == false) curl_easy_setopt(curl, CURLOPT_ISSUERCERT,issuercert.c_str()); // For client authentication, certificate file ... string pem = _config->Find("Acquire::https::SslCert",""); knob = "Acquire::https::"+remotehost+"::SslCert"; pem = _config->Find(knob.c_str(),pem.c_str()); if(pem.empty() == false) curl_easy_setopt(curl, CURLOPT_SSLCERT, pem.c_str()); // ... and associated key. string key = _config->Find("Acquire::https::SslKey",""); knob = "Acquire::https::"+remotehost+"::SslKey"; key = _config->Find(knob.c_str(),key.c_str()); if(key.empty() == false) curl_easy_setopt(curl, CURLOPT_SSLKEY, key.c_str()); // Allow forcing SSL version to SSLv3 or TLSv1 (SSLv2 is not // supported by GnuTLS). long final_version = CURL_SSLVERSION_DEFAULT; string sslversion = _config->Find("Acquire::https::SslForceVersion",""); knob = "Acquire::https::"+remotehost+"::SslForceVersion"; sslversion = _config->Find(knob.c_str(),sslversion.c_str()); if(sslversion == "TLSv1") final_version = CURL_SSLVERSION_TLSv1; else if(sslversion == "SSLv3") final_version = CURL_SSLVERSION_SSLv3; curl_easy_setopt(curl, CURLOPT_SSLVERSION, final_version); // CRL file string crlfile = _config->Find("Acquire::https::CrlFile",""); knob = "Acquire::https::"+remotehost+"::CrlFile"; crlfile = _config->Find(knob.c_str(),crlfile.c_str()); if(crlfile.empty() == false) curl_easy_setopt(curl, CURLOPT_CRLFILE, crlfile.c_str()); // cache-control if(_config->FindB("Acquire::https::No-Cache", _config->FindB("Acquire::http::No-Cache",false)) == false) { // cache enabled if (_config->FindB("Acquire::https::No-Store", _config->FindB("Acquire::http::No-Store",false)) == true) headers = curl_slist_append(headers,"Cache-Control: no-store"); stringstream ss; ioprintf(ss, "Cache-Control: max-age=%u", _config->FindI("Acquire::https::Max-Age", _config->FindI("Acquire::http::Max-Age",0))); headers = curl_slist_append(headers, ss.str().c_str()); } else { // cache disabled by user headers = curl_slist_append(headers, "Cache-Control: no-cache"); headers = curl_slist_append(headers, "Pragma: no-cache"); } curl_easy_setopt(curl, CURLOPT_HTTPHEADER, headers); // speed limit int const dlLimit = _config->FindI("Acquire::https::Dl-Limit", _config->FindI("Acquire::http::Dl-Limit",0))*1024; if (dlLimit > 0) curl_easy_setopt(curl, CURLOPT_MAX_RECV_SPEED_LARGE, dlLimit); // set header curl_easy_setopt(curl, CURLOPT_USERAGENT, _config->Find("Acquire::https::User-Agent", _config->Find("Acquire::http::User-Agent", "Debian APT-CURL/1.0 (" PACKAGE_VERSION ")").c_str()).c_str()); // set timeout int const timeout = _config->FindI("Acquire::https::Timeout", _config->FindI("Acquire::http::Timeout",120)); curl_easy_setopt(curl, CURLOPT_CONNECTTIMEOUT, timeout); //set really low lowspeed timeout (see #497983) curl_easy_setopt(curl, CURLOPT_LOW_SPEED_LIMIT, DL_MIN_SPEED); curl_easy_setopt(curl, CURLOPT_LOW_SPEED_TIME, timeout); // set redirect options and default to 10 redirects curl_easy_setopt(curl, CURLOPT_FOLLOWLOCATION, AllowRedirect); curl_easy_setopt(curl, CURLOPT_MAXREDIRS, 10); if(_config->FindB("Acquire::ForceIPv4", false) == true) curl_easy_setopt(curl, CURLOPT_IPRESOLVE, CURL_IPRESOLVE_V4); else if(_config->FindB("Acquire::ForceIPv6", false) == true) curl_easy_setopt(curl, CURLOPT_IPRESOLVE, CURL_IPRESOLVE_V6); // debug if (Debug == true) curl_easy_setopt(curl, CURLOPT_VERBOSE, true); // error handling curl_errorstr[0] = '\0'; curl_easy_setopt(curl, CURLOPT_ERRORBUFFER, curl_errorstr); // If we ask for uncompressed files servers might respond with content- // negotiation which lets us end up with compressed files we do not support, // see 657029, 657560 and co, so if we have no extension on the request // ask for text only. As a sidenote: If there is nothing to negotate servers // seem to be nice and ignore it. if (_config->FindB("Acquire::https::SendAccept", _config->FindB("Acquire::http::SendAccept", true)) == true) { size_t const filepos = Itm->Uri.find_last_of('/'); string const file = Itm->Uri.substr(filepos + 1); if (flExtension(file) == file) headers = curl_slist_append(headers, "Accept: text/*"); } // if we have the file send an if-range query with a range header if (stat(Itm->DestFile.c_str(),&SBuf) >= 0 && SBuf.st_size > 0) { std::string Buf; strprintf(Buf, "Range: bytes=%lli-", (long long) SBuf.st_size); headers = curl_slist_append(headers, Buf.c_str()); strprintf(Buf, "If-Range: %s", TimeRFC1123(SBuf.st_mtime).c_str()); headers = curl_slist_append(headers, Buf.c_str()); } else if(Itm->LastModified > 0) { curl_easy_setopt(curl, CURLOPT_TIMECONDITION, CURL_TIMECOND_IFMODSINCE); curl_easy_setopt(curl, CURLOPT_TIMEVALUE, Itm->LastModified); } // go for it - if the file exists, append on it File = new FileFd(Itm->DestFile, FileFd::WriteAny); Server = CreateServerState(Itm->Uri); if (Server->InitHashes(Itm->ExpectedHashes) == false) return false; // keep apt updated Res.Filename = Itm->DestFile; // get it! CURLcode success = curl_easy_perform(curl); // If the server returns 200 OK but the If-Modified-Since condition is not // met, CURLINFO_CONDITION_UNMET will be set to 1 long curl_condition_unmet = 0; curl_easy_getinfo(curl, CURLINFO_CONDITION_UNMET, &curl_condition_unmet); File->Close(); curl_slist_free_all(headers); // cleanup if (success != CURLE_OK) { #pragma GCC diagnostic push #pragma GCC diagnostic ignored "-Wswitch" switch (success) { case CURLE_COULDNT_RESOLVE_PROXY: case CURLE_COULDNT_RESOLVE_HOST: SetFailReason("ResolveFailure"); break; case CURLE_COULDNT_CONNECT: SetFailReason("ConnectionRefused"); break; case CURLE_OPERATION_TIMEDOUT: SetFailReason("Timeout"); break; } #pragma GCC diagnostic pop // only take curls technical errors if we haven't our own // (e.g. for the maximum size limit we have and curls can be confusing) if (_error->PendingError() == false) _error->Error("%s", curl_errorstr); else _error->Warning("curl: %s", curl_errorstr); return false; } // server says file not modified if (Server->Result == 304 || curl_condition_unmet == 1) { RemoveFile("https", File->Name()); Res.IMSHit = true; Res.LastModified = Itm->LastModified; Res.Size = 0; URIDone(Res); return true; } Res.IMSHit = false; if (Server->Result != 200 && // OK Server->Result != 206 && // Partial Server->Result != 416) // invalid Range { char err[255]; snprintf(err, sizeof(err) - 1, "HttpError%i", Server->Result); SetFailReason(err); _error->Error("%i %s", Server->Result, Server->Code); // unlink, no need keep 401/404 page content in partial/ RemoveFile("https", File->Name()); return false; } // invalid range-request if (Server->Result == 416) { RemoveFile("https", File->Name()); delete File; Redirect(Itm->Uri); return true; } struct stat resultStat; if (unlikely(stat(File->Name().c_str(), &resultStat) != 0)) { _error->Errno("stat", "Unable to access file %s", File->Name().c_str()); return false; } Res.Size = resultStat.st_size; // Timestamp curl_easy_getinfo(curl, CURLINFO_FILETIME, &Res.LastModified); if (Res.LastModified != -1) { struct timeval times[2]; times[0].tv_sec = Res.LastModified; times[1].tv_sec = Res.LastModified; times[0].tv_usec = times[1].tv_usec = 0; utimes(File->Name().c_str(), times); } else Res.LastModified = resultStat.st_mtime; // take hashes Res.TakeHashes(*(Server->GetHashes())); // keep apt updated URIDone(Res); // cleanup delete File; return true; } /*}}}*/ // HttpsMethod::Configuration - Handle a configuration message /*{{{*/ bool HttpsMethod::Configuration(string Message) { if (ServerMethod::Configuration(Message) == false) return false; AllowRedirect = _config->FindB("Acquire::https::AllowRedirect", _config->FindB("Acquire::http::AllowRedirect", true)); Debug = _config->FindB("Debug::Acquire::https",false); return true; } /*}}}*/ std::unique_ptr HttpsMethod::CreateServerState(URI const &uri)/*{{{*/ { return std::unique_ptr(new HttpsServerState(uri, this)); } /*}}}*/ int main() { setlocale(LC_ALL, ""); HttpsMethod Mth; curl_global_init(CURL_GLOBAL_SSL) ; return Mth.Run(); }