// -*- mode: cpp; mode: fold -*- // Description /*{{{*/ // $Id: apt-ftparchive.cc,v 1.8.2.3 2004/01/02 22:01:48 mdz Exp $ /* ###################################################################### apt-ftparchive - Efficient work-alike for dpkg-scanpackages Let contents be disabled from the conf ##################################################################### */ /*}}}*/ // Include Files /*{{{*/ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "cachedb.h" #include "override.h" #include "apt-ftparchive.h" #include "multicompress.h" #include "writer.h" #include /*}}}*/ using namespace std; ostream c0out(0); ostream c1out(0); ostream c2out(0); ofstream devnull("/dev/null"); unsigned Quiet = 0; // struct PackageMap - List of all package files in the config file /*{{{*/ // --------------------------------------------------------------------- /* */ struct PackageMap { // General Stuff string BaseDir; string InternalPrefix; string FLFile; string PkgExt; string SrcExt; // Stuff for the Package File string PkgFile; string BinCacheDB; string SrcCacheDB; string BinOverride; string ExtraOverride; // We generate for this given arch string Arch; // Stuff for the Source File string SrcFile; string SrcOverride; string SrcExtraOverride; // Translation master file bool LongDesc; TranslationWriter *TransWriter; // Contents string Contents; string ContentsHead; // Random things string Tag; string PkgCompress; string CntCompress; string SrcCompress; string PathPrefix; unsigned int DeLinkLimit; mode_t Permissions; bool ContentsDone; bool PkgDone; bool SrcDone; time_t ContentsMTime; struct ContentsCompare : public binary_function { inline bool operator() (const PackageMap &x,const PackageMap &y) {return x.ContentsMTime < y.ContentsMTime;}; }; struct DBCompare : public binary_function { inline bool operator() (const PackageMap &x,const PackageMap &y) {return x.BinCacheDB < y.BinCacheDB;}; }; struct SrcDBCompare : public binary_function { inline bool operator() (const PackageMap &x,const PackageMap &y) {return x.SrcCacheDB < y.SrcCacheDB;}; }; void GetGeneral(Configuration &Setup,Configuration &Block); bool GenPackages(Configuration &Setup,struct CacheDB::Stats &Stats); bool GenSources(Configuration &Setup,struct CacheDB::Stats &Stats); bool GenContents(Configuration &Setup, vector::iterator Begin, vector::iterator End, unsigned long &Left); PackageMap() : LongDesc(true), TransWriter(NULL), DeLinkLimit(0), Permissions(1), ContentsDone(false), PkgDone(false), SrcDone(false), ContentsMTime(0) {}; }; /*}}}*/ // PackageMap::GetGeneral - Common per-section definitions /*{{{*/ // --------------------------------------------------------------------- /* */ void PackageMap::GetGeneral(Configuration &Setup,Configuration &Block) { PathPrefix = Block.Find("PathPrefix"); if (Block.FindB("External-Links",true) == false) DeLinkLimit = Setup.FindI("Default::DeLinkLimit",UINT_MAX); else DeLinkLimit = 0; PkgCompress = Block.Find("Packages::Compress", Setup.Find("Default::Packages::Compress",". gzip").c_str()); CntCompress = Block.Find("Contents::Compress", Setup.Find("Default::Contents::Compress",". gzip").c_str()); SrcCompress = Block.Find("Sources::Compress", Setup.Find("Default::Sources::Compress",". gzip").c_str()); SrcExt = Block.Find("Sources::Extensions", Setup.Find("Default::Sources::Extensions",".dsc").c_str()); PkgExt = Block.Find("Packages::Extensions", Setup.Find("Default::Packages::Extensions",".deb").c_str()); Permissions = Setup.FindI("Default::FileMode",0644); if (FLFile.empty() == false) FLFile = flCombine(Setup.Find("Dir::FileListDir"),FLFile); if (Contents == " ") Contents= string(); } /*}}}*/ // PackageMap::GenPackages - Actually generate a Package file /*{{{*/ // --------------------------------------------------------------------- /* This generates the Package File described by this object. */ bool PackageMap::GenPackages(Configuration &Setup,struct CacheDB::Stats &Stats) { if (PkgFile.empty() == true) return true; string ArchiveDir = Setup.FindDir("Dir::ArchiveDir"); string OverrideDir = Setup.FindDir("Dir::OverrideDir"); string CacheDir = Setup.FindDir("Dir::CacheDir"); struct timeval StartTime; gettimeofday(&StartTime,0); PkgDone = true; // Create a package writer object. PackagesWriter Packages(flCombine(CacheDir,BinCacheDB), flCombine(OverrideDir,BinOverride), flCombine(OverrideDir,ExtraOverride), Arch); if (PkgExt.empty() == false && Packages.SetExts(PkgExt) == false) return _error->Error(_("Package extension list is too long")); if (_error->PendingError() == true) return _error->Error(_("Error processing directory %s"),BaseDir.c_str()); Packages.PathPrefix = PathPrefix; Packages.DirStrip = ArchiveDir; Packages.InternalPrefix = flCombine(ArchiveDir,InternalPrefix); Packages.TransWriter = TransWriter; Packages.LongDescription = LongDesc; Packages.Stats.DeLinkBytes = Stats.DeLinkBytes; Packages.DeLinkLimit = DeLinkLimit; // Create a compressor object MultiCompress Comp(flCombine(ArchiveDir,PkgFile), PkgCompress,Permissions); Packages.Output = Comp.Input; if (_error->PendingError() == true) return _error->Error(_("Error processing directory %s"),BaseDir.c_str()); c0out << ' ' << BaseDir << ":" << flush; // Do recursive directory searching if (FLFile.empty() == true) { if (Packages.RecursiveScan(flCombine(ArchiveDir,BaseDir)) == false) return false; } else { if (Packages.LoadFileList(ArchiveDir,FLFile) == false) return false; } Packages.Output = 0; // Just in case // Finish compressing unsigned long long Size; if (Comp.Finalize(Size) == false) { c0out << endl; return _error->Error(_("Error processing directory %s"),BaseDir.c_str()); } if (Size != 0) c0out << " New " << SizeToStr(Size) << "B "; else c0out << ' '; struct timeval NewTime; gettimeofday(&NewTime,0); double Delta = NewTime.tv_sec - StartTime.tv_sec + (NewTime.tv_usec - StartTime.tv_usec)/1000000.0; c0out << Packages.Stats.Packages << " files " << /* SizeToStr(Packages.Stats.MD5Bytes) << "B/" << */ SizeToStr(Packages.Stats.Bytes) << "B " << TimeToStr((long)Delta) << endl; if(_config->FindB("APT::FTPArchive::ShowCacheMisses", false) == true) c0out << " Misses in Cache: " << Packages.Stats.Misses<< endl; Stats.Add(Packages.Stats); Stats.DeLinkBytes = Packages.Stats.DeLinkBytes; return !_error->PendingError(); } /*}}}*/ // PackageMap::GenSources - Actually generate a Source file /*{{{*/ // --------------------------------------------------------------------- /* This generates the Sources File described by this object. */ bool PackageMap::GenSources(Configuration &Setup,struct CacheDB::Stats &Stats) { if (SrcFile.empty() == true) return true; string ArchiveDir = Setup.FindDir("Dir::ArchiveDir"); string OverrideDir = Setup.FindDir("Dir::OverrideDir"); string CacheDir = Setup.FindDir("Dir::CacheDir"); struct timeval StartTime; gettimeofday(&StartTime,0); SrcDone = true; // Create a package writer object. SourcesWriter Sources(flCombine(CacheDir, SrcCacheDB), flCombine(OverrideDir,BinOverride), flCombine(OverrideDir,SrcOverride), flCombine(OverrideDir,SrcExtraOverride)); if (SrcExt.empty() == false && Sources.SetExts(SrcExt) == false) return _error->Error(_("Source extension list is too long")); if (_error->PendingError() == true) return _error->Error(_("Error processing directory %s"),BaseDir.c_str()); Sources.PathPrefix = PathPrefix; Sources.DirStrip = ArchiveDir; Sources.InternalPrefix = flCombine(ArchiveDir,InternalPrefix); Sources.DeLinkLimit = DeLinkLimit; Sources.Stats.DeLinkBytes = Stats.DeLinkBytes; // Create a compressor object MultiCompress Comp(flCombine(ArchiveDir,SrcFile), SrcCompress,Permissions); Sources.Output = Comp.Input; if (_error->PendingError() == true) return _error->Error(_("Error processing directory %s"),BaseDir.c_str()); c0out << ' ' << BaseDir << ":" << flush; // Do recursive directory searching if (FLFile.empty() == true) { if (Sources.RecursiveScan(flCombine(ArchiveDir,BaseDir))== false) return false; } else { if (Sources.LoadFileList(ArchiveDir,FLFile) == false) return false; } Sources.Output = 0; // Just in case // Finish compressing unsigned long long Size; if (Comp.Finalize(Size) == false) { c0out << endl; return _error->Error(_("Error processing directory %s"),BaseDir.c_str()); } if (Size != 0) c0out << " New " << SizeToStr(Size) << "B "; else c0out << ' '; struct timeval NewTime; gettimeofday(&NewTime,0); double Delta = NewTime.tv_sec - StartTime.tv_sec + (NewTime.tv_usec - StartTime.tv_usec)/1000000.0; c0out << Sources.Stats.Packages << " pkgs in " << TimeToStr((long)Delta) << endl; if(_config->FindB("APT::FTPArchive::ShowCacheMisses", false) == true) c0out << " Misses in Cache: " << Sources.Stats.Misses << endl; Stats.Add(Sources.Stats); Stats.DeLinkBytes = Sources.Stats.DeLinkBytes; return !_error->PendingError(); } /*}}}*/ // PackageMap::GenContents - Actually generate a Contents file /*{{{*/ // --------------------------------------------------------------------- /* This generates the contents file partially described by this object. It searches the given iterator range for other package files that map into this contents file and includes their data as well when building. */ bool PackageMap::GenContents(Configuration &Setup, vector::iterator Begin, vector::iterator End, unsigned long &Left) { if (Contents.empty() == true) return true; if (Left == 0) return true; string ArchiveDir = Setup.FindDir("Dir::ArchiveDir"); string CacheDir = Setup.FindDir("Dir::CacheDir"); string OverrideDir = Setup.FindDir("Dir::OverrideDir"); struct timeval StartTime; gettimeofday(&StartTime,0); // Create a package writer object. ContentsWriter Contents("", Arch); if (PkgExt.empty() == false && Contents.SetExts(PkgExt) == false) return _error->Error(_("Package extension list is too long")); if (_error->PendingError() == true) return false; MultiCompress Comp(flCombine(ArchiveDir,this->Contents), CntCompress,Permissions); Comp.UpdateMTime = Setup.FindI("Default::ContentsAge",10)*24*60*60; Contents.Output = Comp.Input; if (_error->PendingError() == true) return false; // Write the header out. if (ContentsHead.empty() == false) { FileFd Head(flCombine(OverrideDir,ContentsHead),FileFd::ReadOnly); if (_error->PendingError() == true) return false; unsigned long long Size = Head.Size(); unsigned char Buf[4096]; while (Size != 0) { unsigned long long ToRead = Size; if (Size > sizeof(Buf)) ToRead = sizeof(Buf); if (Head.Read(Buf,ToRead) == false) return false; if (fwrite(Buf,1,ToRead,Comp.Input) != ToRead) return _error->Errno("fwrite",_("Error writing header to contents file")); Size -= ToRead; } } /* Go over all the package file records and parse all the package files associated with this contents file into one great big honking memory structure, then dump the sorted version */ c0out << ' ' << this->Contents << ":" << flush; for (vector::iterator I = Begin; I != End; ++I) { if (I->Contents != this->Contents) continue; Contents.Prefix = ArchiveDir; Contents.ReadyDB(flCombine(CacheDir,I->BinCacheDB)); Contents.ReadFromPkgs(flCombine(ArchiveDir,I->PkgFile), I->PkgCompress); I->ContentsDone = true; } Contents.Finish(); // Finish compressing unsigned long long Size; if (Comp.Finalize(Size) == false || _error->PendingError() == true) { c0out << endl; return _error->Error(_("Error processing contents %s"), this->Contents.c_str()); } if (Size != 0) { c0out << " New " << SizeToStr(Size) << "B "; if (Left > Size) Left -= Size; else Left = 0; } else c0out << ' '; struct timeval NewTime; gettimeofday(&NewTime,0); double Delta = NewTime.tv_sec - StartTime.tv_sec + (NewTime.tv_usec - StartTime.tv_usec)/1000000.0; if(_config->FindB("APT::FTPArchive::ShowCacheMisses", false) == true) c0out << " Misses in Cache: " << Contents.Stats.Misses<< endl; c0out << Contents.Stats.Packages << " files " << SizeToStr(Contents.Stats.Bytes) << "B " << TimeToStr((long)Delta) << endl; return true; } /*}}}*/ // LoadTree - Load a 'tree' section from the Generate Config /*{{{*/ // --------------------------------------------------------------------- /* This populates the PkgList with all the possible permutations of the section/arch lists. */ static void LoadTree(vector &PkgList,Configuration &Setup) { // Load the defaults string DDir = Setup.Find("TreeDefault::Directory", "$(DIST)/$(SECTION)/binary-$(ARCH)/"); string DSDir = Setup.Find("TreeDefault::SrcDirectory", "$(DIST)/$(SECTION)/source/"); string DPkg = Setup.Find("TreeDefault::Packages", "$(DIST)/$(SECTION)/binary-$(ARCH)/Packages"); string DTrans = Setup.Find("TreeDefault::Translation", "$(DIST)/$(SECTION)/i18n/Translation-en"); string DIPrfx = Setup.Find("TreeDefault::InternalPrefix", "$(DIST)/$(SECTION)/"); string DContents = Setup.Find("TreeDefault::Contents", "$(DIST)/$(SECTION)/Contents-$(ARCH)"); string DContentsH = Setup.Find("TreeDefault::Contents::Header",""); string DBCache = Setup.Find("TreeDefault::BinCacheDB", "packages-$(ARCH).db"); string SrcDBCache = Setup.Find("TreeDefault::SrcCacheDB", "sources-$(SECTION).db"); string DSources = Setup.Find("TreeDefault::Sources", "$(DIST)/$(SECTION)/source/Sources"); string DFLFile = Setup.Find("TreeDefault::FileList", ""); string DSFLFile = Setup.Find("TreeDefault::SourceFileList", ""); mode_t const Permissions = Setup.FindI("Default::FileMode",0644); bool const LongDescription = Setup.FindB("Default::LongDescription", _config->FindB("APT::FTPArchive::LongDescription", true)); string const TranslationCompress = Setup.Find("Default::Translation::Compress",". gzip").c_str(); // Process 'tree' type sections const Configuration::Item *Top = Setup.Tree("tree"); for (Top = (Top == 0?0:Top->Child); Top != 0;) { Configuration Block(Top); string Dist = Top->Tag; // Parse the sections string Tmp = Block.Find("Sections"); const char *Sections = Tmp.c_str(); string Section; while (ParseQuoteWord(Sections,Section) == true) { string Arch; struct SubstVar const Vars[] = {{"$(DIST)",&Dist}, {"$(SECTION)",&Section}, {"$(ARCH)",&Arch}, {NULL, NULL}}; mode_t const Perms = Block.FindI("FileMode", Permissions); bool const LongDesc = Block.FindB("LongDescription", LongDescription); TranslationWriter *TransWriter; if (DTrans.empty() == false && LongDesc == false) { string const TranslationFile = flCombine(Setup.FindDir("Dir::ArchiveDir"), SubstVar(Block.Find("Translation", DTrans.c_str()), Vars)); string const TransCompress = Block.Find("Translation::Compress", TranslationCompress); TransWriter = new TranslationWriter(TranslationFile, TransCompress, Perms); } else TransWriter = NULL; string const Tmp2 = Block.Find("Architectures"); const char *Archs = Tmp2.c_str(); while (ParseQuoteWord(Archs,Arch) == true) { PackageMap Itm; Itm.Permissions = Perms; Itm.BinOverride = SubstVar(Block.Find("BinOverride"),Vars); Itm.InternalPrefix = SubstVar(Block.Find("InternalPrefix",DIPrfx.c_str()),Vars); if (stringcasecmp(Arch,"source") == 0) { Itm.SrcOverride = SubstVar(Block.Find("SrcOverride"),Vars); Itm.BaseDir = SubstVar(Block.Find("SrcDirectory",DSDir.c_str()),Vars); Itm.SrcFile = SubstVar(Block.Find("Sources",DSources.c_str()),Vars); Itm.Tag = SubstVar("$(DIST)/$(SECTION)/source",Vars); Itm.FLFile = SubstVar(Block.Find("SourceFileList",DSFLFile.c_str()),Vars); Itm.SrcExtraOverride = SubstVar(Block.Find("SrcExtraOverride"),Vars); Itm.SrcCacheDB = SubstVar(Block.Find("SrcCacheDB",SrcDBCache.c_str()),Vars); } else { Itm.BinCacheDB = SubstVar(Block.Find("BinCacheDB",DBCache.c_str()),Vars); Itm.BaseDir = SubstVar(Block.Find("Directory",DDir.c_str()),Vars); Itm.PkgFile = SubstVar(Block.Find("Packages",DPkg.c_str()),Vars); Itm.Tag = SubstVar("$(DIST)/$(SECTION)/$(ARCH)",Vars); Itm.Arch = Arch; Itm.LongDesc = LongDesc; if (TransWriter != NULL) { TransWriter->IncreaseRefCounter(); Itm.TransWriter = TransWriter; } Itm.Contents = SubstVar(Block.Find("Contents",DContents.c_str()),Vars); Itm.ContentsHead = SubstVar(Block.Find("Contents::Header",DContentsH.c_str()),Vars); Itm.FLFile = SubstVar(Block.Find("FileList",DFLFile.c_str()),Vars); Itm.ExtraOverride = SubstVar(Block.Find("ExtraOverride"),Vars); } Itm.GetGeneral(Setup,Block); PkgList.push_back(Itm); } // we didn't use this TransWriter, so we can release it if (TransWriter != NULL && TransWriter->GetRefCounter() == 0) delete TransWriter; } Top = Top->Next; } } /*}}}*/ // LoadBinDir - Load a 'bindirectory' section from the Generate Config /*{{{*/ // --------------------------------------------------------------------- /* */ static void LoadBinDir(vector &PkgList,Configuration &Setup) { mode_t const Permissions = Setup.FindI("Default::FileMode",0644); // Process 'bindirectory' type sections const Configuration::Item *Top = Setup.Tree("bindirectory"); for (Top = (Top == 0?0:Top->Child); Top != 0;) { Configuration Block(Top); PackageMap Itm; Itm.PkgFile = Block.Find("Packages"); Itm.SrcFile = Block.Find("Sources"); Itm.BinCacheDB = Block.Find("BinCacheDB"); Itm.SrcCacheDB = Block.Find("SrcCacheDB"); Itm.BinOverride = Block.Find("BinOverride"); Itm.ExtraOverride = Block.Find("ExtraOverride"); Itm.SrcExtraOverride = Block.Find("SrcExtraOverride"); Itm.SrcOverride = Block.Find("SrcOverride"); Itm.BaseDir = Top->Tag; Itm.FLFile = Block.Find("FileList"); Itm.InternalPrefix = Block.Find("InternalPrefix",Top->Tag.c_str()); Itm.Contents = Block.Find("Contents"); Itm.ContentsHead = Block.Find("Contents::Header"); Itm.Permissions = Block.FindI("FileMode", Permissions); Itm.GetGeneral(Setup,Block); PkgList.push_back(Itm); Top = Top->Next; } } /*}}}*/ // ShowHelp - Show the help text /*{{{*/ // --------------------------------------------------------------------- /* */ static bool ShowHelp(CommandLine &) { ioprintf(cout,_("%s %s for %s compiled on %s %s\n"),PACKAGE,PACKAGE_VERSION, COMMON_ARCH,__DATE__,__TIME__); if (_config->FindB("version") == true) return true; cout << _("Usage: apt-ftparchive [options] command\n" "Commands: packages binarypath [overridefile [pathprefix]]\n" " sources srcpath [overridefile [pathprefix]]\n" " contents path\n" " release path\n" " generate config [groups]\n" " clean config\n" "\n" "apt-ftparchive generates index files for Debian archives. It supports\n" "many styles of generation from fully automated to functional replacements\n" "for dpkg-scanpackages and dpkg-scansources\n" "\n" "apt-ftparchive generates Package files from a tree of .debs. The\n" "Package file contains the contents of all the control fields from\n" "each package as well as the MD5 hash and filesize. An override file\n" "is supported to force the value of Priority and Section.\n" "\n" "Similarly apt-ftparchive generates Sources files from a tree of .dscs.\n" "The --source-override option can be used to specify a src override file\n" "\n" "The 'packages' and 'sources' command should be run in the root of the\n" "tree. BinaryPath should point to the base of the recursive search and \n" "override file should contain the override flags. Pathprefix is\n" "appended to the filename fields if present. Example usage from the \n" "Debian archive:\n" " apt-ftparchive packages dists/potato/main/binary-i386/ > \\\n" " dists/potato/main/binary-i386/Packages\n" "\n" "Options:\n" " -h This help text\n" " --md5 Control MD5 generation\n" " -s=? Source override file\n" " -q Quiet\n" " -d=? Select the optional caching database\n" " --no-delink Enable delinking debug mode\n" " --contents Control contents file generation\n" " -c=? Read this configuration file\n" " -o=? Set an arbitrary configuration option") << endl; return true; } /*}}}*/ // SimpleGenPackages - Generate a Packages file for a directory tree /*{{{*/ // --------------------------------------------------------------------- /* This emulates dpkg-scanpackages's command line interface. 'mostly' */ static bool SimpleGenPackages(CommandLine &CmdL) { if (CmdL.FileSize() < 2) return ShowHelp(CmdL); string Override; if (CmdL.FileSize() >= 3) Override = CmdL.FileList[2]; // Create a package writer object. PackagesWriter Packages(_config->Find("APT::FTPArchive::DB"), Override, "", _config->Find("APT::FTPArchive::Architecture")); if (_error->PendingError() == true) return false; if (CmdL.FileSize() >= 4) Packages.PathPrefix = CmdL.FileList[3]; // Do recursive directory searching if (Packages.RecursiveScan(CmdL.FileList[1]) == false) return false; // Give some stats if asked for if(_config->FindB("APT::FTPArchive::ShowCacheMisses", false) == true) c0out << " Misses in Cache: " << Packages.Stats.Misses<< endl; return true; } /*}}}*/ // SimpleGenContents - Generate a Contents listing /*{{{*/ // --------------------------------------------------------------------- /* */ static bool SimpleGenContents(CommandLine &CmdL) { if (CmdL.FileSize() < 2) return ShowHelp(CmdL); // Create a package writer object. ContentsWriter Contents(_config->Find("APT::FTPArchive::DB"), _config->Find("APT::FTPArchive::Architecture")); if (_error->PendingError() == true) return false; // Do recursive directory searching if (Contents.RecursiveScan(CmdL.FileList[1]) == false) return false; Contents.Finish(); return true; } /*}}}*/ // SimpleGenSources - Generate a Sources file for a directory tree /*{{{*/ // --------------------------------------------------------------------- /* This emulates dpkg-scanpackages's command line interface. 'mostly' */ static bool SimpleGenSources(CommandLine &CmdL) { if (CmdL.FileSize() < 2) return ShowHelp(CmdL); string Override; if (CmdL.FileSize() >= 3) Override = CmdL.FileList[2]; string SOverride; if (Override.empty() == false) SOverride = Override + ".src"; SOverride = _config->Find("APT::FTPArchive::SourceOverride", SOverride.c_str()); // Create a package writer object. SourcesWriter Sources(_config->Find("APT::FTPArchive::DB"),Override,SOverride); if (_error->PendingError() == true) return false; if (CmdL.FileSize() >= 4) Sources.PathPrefix = CmdL.FileList[3]; // Do recursive directory searching if (Sources.RecursiveScan(CmdL.FileList[1]) == false) return false; // Give some stats if asked for if(_config->FindB("APT::FTPArchive::ShowCacheMisses", false) == true) c0out << " Misses in Cache: " << Sources.Stats.Misses<< endl; return true; } /*}}}*/ // SimpleGenRelease - Generate a Release file for a directory tree /*{{{*/ // --------------------------------------------------------------------- static bool SimpleGenRelease(CommandLine &CmdL) { if (CmdL.FileSize() < 2) return ShowHelp(CmdL); string Dir = CmdL.FileList[1]; ReleaseWriter Release(""); Release.DirStrip = Dir; if (_error->PendingError() == true) return false; if (Release.RecursiveScan(Dir) == false) return false; Release.Finish(); return true; } /*}}}*/ // DoGeneratePackagesAndSources - Helper for Generate /*{{{*/ // --------------------------------------------------------------------- static bool DoGeneratePackagesAndSources(Configuration &Setup, vector &PkgList, struct CacheDB::Stats &SrcStats, struct CacheDB::Stats &Stats, CommandLine &CmdL) { if (CmdL.FileSize() <= 2) { for (vector::iterator I = PkgList.begin(); I != PkgList.end(); ++I) if (I->GenPackages(Setup,Stats) == false) _error->DumpErrors(); for (vector::iterator I = PkgList.begin(); I != PkgList.end(); ++I) if (I->GenSources(Setup,SrcStats) == false) _error->DumpErrors(); } else { // Make a choice list out of the package list.. RxChoiceList *List = new RxChoiceList[2*PkgList.size()+1]; RxChoiceList *End = List; for (vector::iterator I = PkgList.begin(); I != PkgList.end(); ++I) { End->UserData = &(*I); End->Str = I->BaseDir.c_str(); End++; End->UserData = &(*I); End->Str = I->Tag.c_str(); End++; } End->Str = 0; // Regex it if (RegexChoice(List,CmdL.FileList + 2,CmdL.FileList + CmdL.FileSize()) == 0) { delete [] List; return _error->Error(_("No selections matched")); } _error->DumpErrors(); // Do the generation for Packages for (End = List; End->Str != 0; End++) { if (End->Hit == false) continue; PackageMap *I = (PackageMap *)End->UserData; if (I->PkgDone == true) continue; if (I->GenPackages(Setup,Stats) == false) _error->DumpErrors(); } // Do the generation for Sources for (End = List; End->Str != 0; End++) { if (End->Hit == false) continue; PackageMap *I = (PackageMap *)End->UserData; if (I->SrcDone == true) continue; if (I->GenSources(Setup,SrcStats) == false) _error->DumpErrors(); } delete [] List; } // close the Translation master files for (vector::reverse_iterator I = PkgList.rbegin(); I != PkgList.rend(); ++I) if (I->TransWriter != NULL && I->TransWriter->DecreaseRefCounter() == 0) delete I->TransWriter; return true; } /*}}}*/ // DoGenerateContents - Helper for Generate to generate the Contents /*{{{*/ // --------------------------------------------------------------------- static bool DoGenerateContents(Configuration &Setup, vector &PkgList, CommandLine &CmdL) { c1out << "Packages done, Starting contents." << endl; // Sort the contents file list by date string ArchiveDir = Setup.FindDir("Dir::ArchiveDir"); for (vector::iterator I = PkgList.begin(); I != PkgList.end(); ++I) { struct stat A; if (MultiCompress::GetStat(flCombine(ArchiveDir,I->Contents), I->CntCompress,A) == false) time(&I->ContentsMTime); else I->ContentsMTime = A.st_mtime; } stable_sort(PkgList.begin(),PkgList.end(),PackageMap::ContentsCompare()); /* Now for Contents.. The process here is to do a make-like dependency check. Each contents file is verified to be newer than the package files that describe the debs it indexes. Since the package files contain hashes of the .debs this means they have not changed either so the contents must be up to date. */ unsigned long MaxContentsChange = Setup.FindI("Default::MaxContentsChange",UINT_MAX)*1024; for (vector::iterator I = PkgList.begin(); I != PkgList.end(); ++I) { // This record is not relevant if (I->ContentsDone == true || I->Contents.empty() == true) continue; // Do not do everything if the user specified sections. if (CmdL.FileSize() > 2 && I->PkgDone == false) continue; struct stat A,B; if (MultiCompress::GetStat(flCombine(ArchiveDir,I->Contents),I->CntCompress,A) == true) { if (MultiCompress::GetStat(flCombine(ArchiveDir,I->PkgFile),I->PkgCompress,B) == false) { _error->Warning(_("Some files are missing in the package file group `%s'"),I->PkgFile.c_str()); continue; } if (A.st_mtime > B.st_mtime) continue; } if (I->GenContents(Setup,PkgList.begin(),PkgList.end(), MaxContentsChange) == false) _error->DumpErrors(); // Hit the limit? if (MaxContentsChange == 0) { c1out << "Hit contents update byte limit" << endl; break; } } return true; } /*}}}*/ // Generate - Full generate, using a config file /*{{{*/ // --------------------------------------------------------------------- /* */ static bool Generate(CommandLine &CmdL) { struct CacheDB::Stats SrcStats; if (CmdL.FileSize() < 2) return ShowHelp(CmdL); struct timeval StartTime; gettimeofday(&StartTime,0); struct CacheDB::Stats Stats; // Read the configuration file. Configuration Setup; if (ReadConfigFile(Setup,CmdL.FileList[1],true) == false) return false; vector PkgList; LoadTree(PkgList,Setup); LoadBinDir(PkgList,Setup); // Sort by cache DB to improve IO locality. stable_sort(PkgList.begin(),PkgList.end(),PackageMap::DBCompare()); stable_sort(PkgList.begin(),PkgList.end(),PackageMap::SrcDBCompare()); // Generate packages if (_config->FindB("APT::FTPArchive::ContentsOnly", false) == false) { if(DoGeneratePackagesAndSources(Setup, PkgList, SrcStats, Stats, CmdL) == false) return false; } else { c1out << "Skipping Packages/Sources generation" << endl; } // do Contents if needed if (_config->FindB("APT::FTPArchive::Contents", true) == true) if (DoGenerateContents(Setup, PkgList, CmdL) == false) return false; struct timeval NewTime; gettimeofday(&NewTime,0); double Delta = NewTime.tv_sec - StartTime.tv_sec + (NewTime.tv_usec - StartTime.tv_usec)/1000000.0; c1out << "Done. " << SizeToStr(Stats.Bytes) << "B in " << Stats.Packages << " archives. Took " << TimeToStr((long)Delta) << endl; return true; } /*}}}*/ // Clean - Clean out the databases /*{{{*/ // --------------------------------------------------------------------- /* */ static bool Clean(CommandLine &CmdL) { if (CmdL.FileSize() != 2) return ShowHelp(CmdL); // Read the configuration file. Configuration Setup; if (ReadConfigFile(Setup,CmdL.FileList[1],true) == false) return false; vector PkgList; LoadTree(PkgList,Setup); LoadBinDir(PkgList,Setup); // Sort by cache DB to improve IO locality. stable_sort(PkgList.begin(),PkgList.end(),PackageMap::DBCompare()); stable_sort(PkgList.begin(),PkgList.end(),PackageMap::SrcDBCompare()); string CacheDir = Setup.FindDir("Dir::CacheDir"); for (vector::iterator I = PkgList.begin(); I != PkgList.end(); ) { if(I->BinCacheDB != "") c0out << I->BinCacheDB << endl; if(I->SrcCacheDB != "") c0out << I->SrcCacheDB << endl; CacheDB DB(flCombine(CacheDir,I->BinCacheDB)); CacheDB DB_SRC(flCombine(CacheDir,I->SrcCacheDB)); if (DB.Clean() == false) _error->DumpErrors(); if (DB_SRC.Clean() == false) _error->DumpErrors(); string CacheDB = I->BinCacheDB; string SrcCacheDB = I->SrcCacheDB; while(I != PkgList.end() && I->BinCacheDB == CacheDB && I->SrcCacheDB == SrcCacheDB) ++I; } return true; } /*}}}*/ int main(int argc, const char *argv[]) { setlocale(LC_ALL, ""); CommandLine::Args Args[] = { {'h',"help","help",0}, {0,"md5","APT::FTPArchive::MD5",0}, {0,"sha1","APT::FTPArchive::SHA1",0}, {0,"sha256","APT::FTPArchive::SHA256",0}, {'v',"version","version",0}, {'d',"db","APT::FTPArchive::DB",CommandLine::HasArg}, {'s',"source-override","APT::FTPArchive::SourceOverride",CommandLine::HasArg}, {'q',"quiet","quiet",CommandLine::IntLevel}, {'q',"silent","quiet",CommandLine::IntLevel}, {0,"delink","APT::FTPArchive::DeLinkAct",0}, {0,"readonly","APT::FTPArchive::ReadOnlyDB",0}, {0,"contents","APT::FTPArchive::Contents",0}, {'a',"arch","APT::FTPArchive::Architecture",CommandLine::HasArg}, {'c',"config-file",0,CommandLine::ConfigFile}, {'o',"option",0,CommandLine::ArbItem}, {0,0,0,0}}; CommandLine::Dispatch Cmds[] = {{"packages",&SimpleGenPackages}, {"contents",&SimpleGenContents}, {"sources",&SimpleGenSources}, {"release",&SimpleGenRelease}, {"generate",&Generate}, {"clean",&Clean}, {"help",&ShowHelp}, {0,0}}; // Parse the command line and initialize the package library CommandLine CmdL(Args,_config); ParseCommandLine(CmdL, Cmds, Args, &_config, NULL, argc, argv, ShowHelp); // Setup the output streams c0out.rdbuf(clog.rdbuf()); c1out.rdbuf(clog.rdbuf()); c2out.rdbuf(clog.rdbuf()); Quiet = _config->FindI("quiet",0); if (Quiet > 0) c0out.rdbuf(devnull.rdbuf()); if (Quiet > 1) c1out.rdbuf(devnull.rdbuf()); // Match the operation CmdL.DispatchArg(Cmds); if (_error->empty() == false) { bool Errors = _error->PendingError(); _error->DumpErrors(); return Errors == true?100:0; } return 0; }