// -*- mode: cpp; mode: fold -*-
// Description /*{{{*/
-// $Id: pkgcachegen.cc,v 1.49 2001/05/27 05:55:27 jgg Exp $
+// $Id: pkgcachegen.cc,v 1.53.2.1 2003/12/24 23:09:17 mdz Exp $
/* ######################################################################
Package Cache Generator - Generator for the cache structure.
#include <apt-pkg/sptr.h>
#include <apt-pkg/pkgsystem.h>
+#include <apt-pkg/tagfile.h>
+
#include <apti18n.h>
#include <vector>
// ---------------------------------------------------------------------
/* We set the diry flag and make sure that is written to the disk */
pkgCacheGenerator::pkgCacheGenerator(DynamicMMap *pMap,OpProgress *Prog) :
- Map(*pMap), Cache(pMap,false), Progress(Prog)
+ Map(*pMap), Cache(pMap,false), Progress(Prog),
+ FoundFileDeps(0)
{
CurrentFile = 0;
memset(UniqHash,0,sizeof(UniqHash));
pkgCache::PkgIterator Pkg;
if (NewPackage(Pkg,PackageName) == false)
- return _error->Error(_("Error occured while processing %s (NewPackage)"),PackageName.c_str());
+ return _error->Error(_("Error occurred while processing %s (NewPackage)"),PackageName.c_str());
Counter++;
if (Counter % 100 == 0 && Progress != 0)
Progress->Progress(List.Offset());
if (Version.empty() == true)
{
if (List.UsePackage(Pkg,pkgCache::VerIterator(Cache)) == false)
- return _error->Error(_("Error occured while processing %s (UsePackage1)"),PackageName.c_str());
+ return _error->Error(_("Error occurred while processing %s (UsePackage1)"),
+ PackageName.c_str());
continue;
}
if (Res == 0 && Ver->Hash == Hash)
{
if (List.UsePackage(Pkg,Ver) == false)
- return _error->Error(_("Error occured while processing %s (UsePackage2)"),PackageName.c_str());
+ return _error->Error(_("Error occurred while processing %s (UsePackage2)"),
+ PackageName.c_str());
if (NewFileVer(Ver,List) == false)
- return _error->Error(_("Error occured while processing %s (NewFileVer1)"),PackageName.c_str());
+ return _error->Error(_("Error occurred while processing %s (NewFileVer1)"),
+ PackageName.c_str());
// Read only a single record and return
if (OutVer != 0)
{
*OutVer = Ver;
+ FoundFileDeps |= List.HasFileDeps();
return true;
}
Ver->ParentPkg = Pkg.Index();
Ver->Hash = Hash;
if (List.NewVersion(Ver) == false)
- return _error->Error(_("Error occured while processing %s (NewVersion1)"),PackageName.c_str());
+ return _error->Error(_("Error occurred while processing %s (NewVersion1)"),
+ PackageName.c_str());
if (List.UsePackage(Pkg,Ver) == false)
- return _error->Error(_("Error occured while processing %s (UsePackage3)"),PackageName.c_str());
+ return _error->Error(_("Error occurred while processing %s (UsePackage3)"),
+ PackageName.c_str());
if (NewFileVer(Ver,List) == false)
- return _error->Error(_("Error occured while processing %s (NewVersion2)"),PackageName.c_str());
+ return _error->Error(_("Error occurred while processing %s (NewVersion2)"),
+ PackageName.c_str());
// Read only a single record and return
if (OutVer != 0)
{
*OutVer = Ver;
+ FoundFileDeps |= List.HasFileDeps();
return true;
}
}
+ FoundFileDeps |= List.HasFileDeps();
+
+ if (Cache.HeaderP->PackageCount >= (1ULL<<sizeof(Cache.PkgP->ID)*8)-1)
+ return _error->Error(_("Wow, you exceeded the number of package "
+ "names this APT is capable of."));
+ if (Cache.HeaderP->VersionCount >= (1ULL<<(sizeof(Cache.VerP->ID)*8))-1)
+ return _error->Error(_("Wow, you exceeded the number of versions "
+ "this APT is capable of."));
+ if (Cache.HeaderP->DependsCount >= (1ULL<<(sizeof(Cache.DepP->ID)*8))-1ULL)
+ return _error->Error(_("Wow, you exceeded the number of dependencies "
+ "this APT is capable of."));
+ return true;
+}
+ /*}}}*/
+// CacheGenerator::MergeFileProvides - Merge file provides /*{{{*/
+// ---------------------------------------------------------------------
+/* If we found any file depends while parsing the main list we need to
+ resolve them. Since it is undesired to load the entire list of files
+ into the cache as virtual packages we do a two stage effort. MergeList
+ identifies the file depends and this creates Provdies for them by
+ re-parsing all the indexs. */
+bool pkgCacheGenerator::MergeFileProvides(ListParser &List)
+{
+ List.Owner = this;
+
+ unsigned int Counter = 0;
+ while (List.Step() == true)
+ {
+ string PackageName = List.Package();
+ if (PackageName.empty() == true)
+ return false;
+ string Version = List.Version();
+ if (Version.empty() == true)
+ continue;
+
+ pkgCache::PkgIterator Pkg = Cache.FindPkg(PackageName);
+ if (Pkg.end() == true)
+ return _error->Error(_("Error occurred while processing %s (FindPkg)"),
+ PackageName.c_str());
+ Counter++;
+ if (Counter % 100 == 0 && Progress != 0)
+ Progress->Progress(List.Offset());
+
+ unsigned long Hash = List.VersionHash();
+ pkgCache::VerIterator Ver = Pkg.VersionList();
+ for (; Ver.end() == false; Ver++)
+ {
+ if (Ver->Hash == Hash && Version.c_str() == Ver.VerStr())
+ {
+ if (List.CollectFileProvides(Cache,Ver) == false)
+ return _error->Error(_("Error occurred while processing %s (CollectFileProvides)"),PackageName.c_str());
+ break;
+ }
+ }
+
+ if (Ver.end() == true)
+ _error->Warning(_("Package %s %s was not found while processing file dependencies"),PackageName.c_str(),Version.c_str());
+ }
+
return true;
}
/*}}}*/
// CacheGenerator::NewPackage - Add a new package /*{{{*/
// ---------------------------------------------------------------------
/* This creates a new package structure and adds it to the hash table */
-bool pkgCacheGenerator::NewPackage(pkgCache::PkgIterator &Pkg,string Name)
+bool pkgCacheGenerator::NewPackage(pkgCache::PkgIterator &Pkg,const string &Name)
{
Pkg = Cache.FindPkg(Name);
if (Pkg.end() == false)
// ---------------------------------------------------------------------
/* This puts a version structure in the linked list */
unsigned long pkgCacheGenerator::NewVersion(pkgCache::VerIterator &Ver,
- string VerStr,
+ const string &VerStr,
unsigned long Next)
{
// Get a structure
/* This creates a dependency element in the tree. It is linked to the
version and to the package that it is pointing to. */
bool pkgCacheGenerator::ListParser::NewDepends(pkgCache::VerIterator Ver,
- string PackageName,
- string Version,
+ const string &PackageName,
+ const string &Version,
unsigned int Op,
unsigned int Type)
{
OldDepLast = &D->NextDepends;
OldDepVer = Ver;
}
+
+ // Is it a file dependency?
+ if (PackageName[0] == '/')
+ FoundFileDeps = true;
Dep->NextDepends = *OldDepLast;
*OldDepLast = Dep.Index();
// ---------------------------------------------------------------------
/* */
bool pkgCacheGenerator::ListParser::NewProvides(pkgCache::VerIterator Ver,
- string PackageName,
- string Version)
+ const string &PackageName,
+ const string &Version)
{
pkgCache &Cache = Owner->Cache;
// ---------------------------------------------------------------------
/* This is used to select which file is to be associated with all newly
added versions. The caller is responsible for setting the IMS fields. */
-bool pkgCacheGenerator::SelectFile(string File,string Site,
+bool pkgCacheGenerator::SelectFile(const string &File,const string &Site,
const pkgIndexFile &Index,
unsigned long Flags)
{
/* This just verifies that each file in the list of index files exists,
has matching attributes with the cache and the cache does not have
any extra files. */
-static bool CheckValidity(string CacheFile, FileIterator Start,
+static bool CheckValidity(const string &CacheFile, FileIterator Start,
FileIterator End,MMap **OutMap = 0)
{
// No file, certainly invalid
if ((*Start)->Exists() == false)
{
+#if 0 // mvo: we no longer give a message here (Default Sources spec)
_error->WarningE("stat",_("Couldn't stat source package list %s"),
(*Start)->Describe().c_str());
+#endif
continue;
}
unsigned long &CurrentSize,unsigned long TotalSize,
FileIterator Start, FileIterator End)
{
- for (; Start != End; Start++)
+ FileIterator I;
+ for (I = Start; I != End; I++)
{
- if ((*Start)->HasPackages() == false)
+ if ((*I)->HasPackages() == false)
continue;
- if ((*Start)->Exists() == false)
+ if ((*I)->Exists() == false)
continue;
- if ((*Start)->FindInCache(Gen.GetCache()).end() == false)
+ if ((*I)->FindInCache(Gen.GetCache()).end() == false)
{
_error->Warning("Duplicate sources.list entry %s",
- (*Start)->Describe().c_str());
+ (*I)->Describe().c_str());
continue;
}
- unsigned long Size = (*Start)->Size();
- Progress.OverallProgress(CurrentSize,TotalSize,Size,_("Reading Package Lists"));
+ unsigned long Size = (*I)->Size();
+ Progress.OverallProgress(CurrentSize,TotalSize,Size,_("Reading package lists"));
CurrentSize += Size;
- if ((*Start)->Merge(Gen,Progress) == false)
+ if ((*I)->Merge(Gen,Progress) == false)
return false;
}
+
+ if (Gen.HasFileDeps() == true)
+ {
+ Progress.Done();
+ TotalSize = ComputeSize(Start, End);
+ CurrentSize = 0;
+ for (I = Start; I != End; I++)
+ {
+ unsigned long Size = (*I)->Size();
+ Progress.OverallProgress(CurrentSize,TotalSize,Size,_("Collecting File Provides"));
+ CurrentSize += Size;
+ if ((*I)->MergeFileProvides(Gen,Progress) == false)
+ return false;
+ }
+ }
return true;
}
bool pkgMakeStatusCache(pkgSourceList &List,OpProgress &Progress,
MMap **OutMap,bool AllowMem)
{
- unsigned long MapSize = _config->FindI("APT::Cache-Limit",6*1024*1024);
+ unsigned long MapSize = _config->FindI("APT::Cache-Limit",12*1024*1024);
+
+ vector<pkgIndexFile *> Files;
+ for (vector<metaIndex *>::const_iterator i = List.begin();
+ i != List.end();
+ i++)
+ {
+ vector <pkgIndexFile *> *Indexes = (*i)->GetIndexFiles();
+ for (vector<pkgIndexFile *>::const_iterator j = Indexes->begin();
+ j != Indexes->end();
+ j++)
+ Files.push_back (*j);
+ }
- vector<pkgIndexFile *> Files(List.begin(),List.end());
unsigned long EndOfSource = Files.size();
if (_system->AddStatusFiles(Files) == false)
return false;
if (Writeable == false && AllowMem == false && CacheFile.empty() == false)
return _error->Error(_("Unable to write to %s"),flNotFile(CacheFile).c_str());
- Progress.OverallProgress(0,1,1,_("Reading Package Lists"));
+ Progress.OverallProgress(0,1,1,_("Reading package lists"));
// Cache is OK, Fin.
if (CheckValidity(CacheFile,Files.begin(),Files.end(),OutMap) == true)
{
- Progress.OverallProgress(1,1,1,_("Reading Package Lists"));
+ Progress.OverallProgress(1,1,1,_("Reading package lists"));
return true;
}
{
unlink(CacheFile.c_str());
CacheF = new FileFd(CacheFile,FileFd::WriteEmpty);
+ fchmod(CacheF->Fd(),0644);
Map = new DynamicMMap(*CacheF,MMap::Public,MapSize);
if (_error->PendingError() == true)
return false;
FileFd SCacheF(SrcCacheFile,FileFd::WriteEmpty);
if (_error->PendingError() == true)
return false;
+
+ fchmod(SCacheF.Fd(),0644);
+
// Write out the main data
if (SCacheF.Write(Map->Data(),Map->Size()) == false)
return _error->Error(_("IO Error saving source cache"));
if (SCacheF.Seek(0) == false ||
SCacheF.Write(Map->Data(),sizeof(*Gen.GetCache().HeaderP)) == false)
return _error->Error(_("IO Error saving source cache"));
- SCacheF.Sync();
Gen.GetCache().HeaderP->Dirty = true;
+ SCacheF.Sync();
}
// Build the status cache
/* */
bool pkgMakeOnlyStatusCache(OpProgress &Progress,DynamicMMap **OutMap)
{
- unsigned long MapSize = _config->FindI("APT::Cache-Limit",4*1024*1024);
+ unsigned long MapSize = _config->FindI("APT::Cache-Limit",8*1024*1024);
vector<pkgIndexFile *> Files;
unsigned long EndOfSource = Files.size();
if (_system->AddStatusFiles(Files) == false)
TotalSize = ComputeSize(Files.begin()+EndOfSource,Files.end());
// Build the status cache
- Progress.OverallProgress(0,1,1,_("Reading Package Lists"));
+ Progress.OverallProgress(0,1,1,_("Reading package lists"));
pkgCacheGenerator Gen(Map.Get(),&Progress);
if (_error->PendingError() == true)
return false;