]>
git.saurik.com Git - apt.git/blob - apt-pkg/pkgcachegen.cc
1 // -*- mode: cpp; mode: fold -*-
3 // $Id: pkgcachegen.cc,v 1.53.2.1 2003/12/24 23:09:17 mdz Exp $
4 /* ######################################################################
6 Package Cache Generator - Generator for the cache structure.
8 This builds the cache structure from the abstract package list parser.
10 ##################################################################### */
12 // Include Files /*{{{*/
15 #include <apt-pkg/pkgcachegen.h>
16 #include <apt-pkg/error.h>
17 #include <apt-pkg/version.h>
18 #include <apt-pkg/progress.h>
19 #include <apt-pkg/sourcelist.h>
20 #include <apt-pkg/configuration.h>
21 #include <apt-pkg/aptconfiguration.h>
22 #include <apt-pkg/strutl.h>
23 #include <apt-pkg/sptr.h>
24 #include <apt-pkg/pkgsystem.h>
25 #include <apt-pkg/macros.h>
26 #include <apt-pkg/tagfile.h>
27 #include <apt-pkg/metaindex.h>
28 #include <apt-pkg/fileutl.h>
38 typedef std::vector
<pkgIndexFile
*>::iterator FileIterator
;
39 template <typename Iter
> std::vector
<Iter
*> pkgCacheGenerator::Dynamic
<Iter
>::toReMap
;
41 static bool IsDuplicateDescription(pkgCache::DescIterator Desc
,
42 MD5SumValue
const &CurMd5
, std::string
const &CurLang
);
46 // CacheGenerator::pkgCacheGenerator - Constructor /*{{{*/
47 // ---------------------------------------------------------------------
48 /* We set the dirty flag and make sure that is written to the disk */
49 pkgCacheGenerator::pkgCacheGenerator(DynamicMMap
*pMap
,OpProgress
*Prog
) :
50 Map(*pMap
), Cache(pMap
,false), Progress(Prog
),
54 memset(UniqHash
,0,sizeof(UniqHash
));
56 if (_error
->PendingError() == true)
61 // Setup the map interface..
62 Cache
.HeaderP
= (pkgCache::Header
*)Map
.Data();
63 if (Map
.RawAllocate(sizeof(pkgCache::Header
)) == 0 && _error
->PendingError() == true)
66 Map
.UsePools(*Cache
.HeaderP
->Pools
,sizeof(Cache
.HeaderP
->Pools
)/sizeof(Cache
.HeaderP
->Pools
[0]));
69 *Cache
.HeaderP
= pkgCache::Header();
70 map_ptrloc
const idxVerSysName
= WriteStringInMap(_system
->VS
->Label
);
71 Cache
.HeaderP
->VerSysName
= idxVerSysName
;
72 // this pointer is set in ReMap, but we need it now for WriteUniqString
73 Cache
.StringItemP
= (pkgCache::StringItem
*)Map
.Data();
74 map_ptrloc
const idxArchitecture
= WriteUniqString(_config
->Find("APT::Architecture"));
75 Cache
.HeaderP
->Architecture
= idxArchitecture
;
76 if (unlikely(idxVerSysName
== 0 || idxArchitecture
== 0))
82 // Map directly from the existing file
84 Map
.UsePools(*Cache
.HeaderP
->Pools
,sizeof(Cache
.HeaderP
->Pools
)/sizeof(Cache
.HeaderP
->Pools
[0]));
85 if (Cache
.VS
!= _system
->VS
)
87 _error
->Error(_("Cache has an incompatible versioning system"));
92 Cache
.HeaderP
->Dirty
= true;
93 Map
.Sync(0,sizeof(pkgCache::Header
));
96 // CacheGenerator::~pkgCacheGenerator - Destructor /*{{{*/
97 // ---------------------------------------------------------------------
98 /* We sync the data then unset the dirty flag in two steps so as to
99 advoid a problem during a crash */
100 pkgCacheGenerator::~pkgCacheGenerator()
102 if (_error
->PendingError() == true)
104 if (Map
.Sync() == false)
107 Cache
.HeaderP
->Dirty
= false;
108 Cache
.HeaderP
->CacheFileSize
= Map
.Size();
109 Map
.Sync(0,sizeof(pkgCache::Header
));
112 void pkgCacheGenerator::ReMap(void const * const oldMap
, void const * const newMap
) {/*{{{*/
113 if (oldMap
== newMap
)
116 if (_config
->FindB("Debug::pkgCacheGen", false))
117 std::clog
<< "Remaping from " << oldMap
<< " to " << newMap
<< std::endl
;
121 CurrentFile
+= (pkgCache::PackageFile
*) newMap
- (pkgCache::PackageFile
*) oldMap
;
123 for (size_t i
= 0; i
< _count(UniqHash
); ++i
)
124 if (UniqHash
[i
] != 0)
125 UniqHash
[i
] += (pkgCache::StringItem
*) newMap
- (pkgCache::StringItem
*) oldMap
;
127 for (std::vector
<pkgCache::GrpIterator
*>::const_iterator i
= Dynamic
<pkgCache::GrpIterator
>::toReMap
.begin();
128 i
!= Dynamic
<pkgCache::GrpIterator
>::toReMap
.end(); ++i
)
129 (*i
)->ReMap(oldMap
, newMap
);
130 for (std::vector
<pkgCache::PkgIterator
*>::const_iterator i
= Dynamic
<pkgCache::PkgIterator
>::toReMap
.begin();
131 i
!= Dynamic
<pkgCache::PkgIterator
>::toReMap
.end(); ++i
)
132 (*i
)->ReMap(oldMap
, newMap
);
133 for (std::vector
<pkgCache::VerIterator
*>::const_iterator i
= Dynamic
<pkgCache::VerIterator
>::toReMap
.begin();
134 i
!= Dynamic
<pkgCache::VerIterator
>::toReMap
.end(); ++i
)
135 (*i
)->ReMap(oldMap
, newMap
);
136 for (std::vector
<pkgCache::DepIterator
*>::const_iterator i
= Dynamic
<pkgCache::DepIterator
>::toReMap
.begin();
137 i
!= Dynamic
<pkgCache::DepIterator
>::toReMap
.end(); ++i
)
138 (*i
)->ReMap(oldMap
, newMap
);
139 for (std::vector
<pkgCache::DescIterator
*>::const_iterator i
= Dynamic
<pkgCache::DescIterator
>::toReMap
.begin();
140 i
!= Dynamic
<pkgCache::DescIterator
>::toReMap
.end(); ++i
)
141 (*i
)->ReMap(oldMap
, newMap
);
142 for (std::vector
<pkgCache::PrvIterator
*>::const_iterator i
= Dynamic
<pkgCache::PrvIterator
>::toReMap
.begin();
143 i
!= Dynamic
<pkgCache::PrvIterator
>::toReMap
.end(); ++i
)
144 (*i
)->ReMap(oldMap
, newMap
);
145 for (std::vector
<pkgCache::PkgFileIterator
*>::const_iterator i
= Dynamic
<pkgCache::PkgFileIterator
>::toReMap
.begin();
146 i
!= Dynamic
<pkgCache::PkgFileIterator
>::toReMap
.end(); ++i
)
147 (*i
)->ReMap(oldMap
, newMap
);
149 // CacheGenerator::WriteStringInMap /*{{{*/
150 map_ptrloc
pkgCacheGenerator::WriteStringInMap(const char *String
,
151 const unsigned long &Len
) {
152 void const * const oldMap
= Map
.Data();
153 map_ptrloc
const index
= Map
.WriteString(String
, Len
);
155 ReMap(oldMap
, Map
.Data());
159 // CacheGenerator::WriteStringInMap /*{{{*/
160 map_ptrloc
pkgCacheGenerator::WriteStringInMap(const char *String
) {
161 void const * const oldMap
= Map
.Data();
162 map_ptrloc
const index
= Map
.WriteString(String
);
164 ReMap(oldMap
, Map
.Data());
168 map_ptrloc
pkgCacheGenerator::AllocateInMap(const unsigned long &size
) {/*{{{*/
169 void const * const oldMap
= Map
.Data();
170 map_ptrloc
const index
= Map
.Allocate(size
);
172 ReMap(oldMap
, Map
.Data());
176 // CacheGenerator::MergeList - Merge the package list /*{{{*/
177 // ---------------------------------------------------------------------
178 /* This provides the generation of the entries in the cache. Each loop
179 goes through a single package record from the underlying parse engine. */
180 bool pkgCacheGenerator::MergeList(ListParser
&List
,
181 pkgCache::VerIterator
*OutVer
)
185 unsigned int Counter
= 0;
186 while (List
.Step() == true)
188 string
const PackageName
= List
.Package();
189 if (PackageName
.empty() == true)
193 if (Counter
% 100 == 0 && Progress
!= 0)
194 Progress
->Progress(List
.Offset());
196 string Arch
= List
.Architecture();
197 string
const Version
= List
.Version();
198 if (Version
.empty() == true && Arch
.empty() == true)
200 // package descriptions
201 if (MergeListGroup(List
, PackageName
) == false)
206 if (Arch
.empty() == true)
208 // use the pseudo arch 'none' for arch-less packages
210 /* We might built a SingleArchCache here, which we don't want to blow up
211 just for these :none packages to a proper MultiArchCache, so just ensure
212 that we have always a native package structure first for SingleArch */
213 pkgCache::PkgIterator NP
;
214 Dynamic
<pkgCache::PkgIterator
> DynPkg(NP
);
215 if (NewPackage(NP
, PackageName
, _config
->Find("APT::Architecture")) == false)
216 // TRANSLATOR: The first placeholder is a package name,
217 // the other two should be copied verbatim as they include debug info
218 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
219 PackageName
.c_str(), "NewPackage", 0);
222 // Get a pointer to the package structure
223 pkgCache::PkgIterator Pkg
;
224 Dynamic
<pkgCache::PkgIterator
> DynPkg(Pkg
);
225 if (NewPackage(Pkg
, PackageName
, Arch
) == false)
226 // TRANSLATOR: The first placeholder is a package name,
227 // the other two should be copied verbatim as they include debug info
228 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
229 PackageName
.c_str(), "NewPackage", 1);
232 if (Version
.empty() == true)
234 if (MergeListPackage(List
, Pkg
) == false)
239 if (MergeListVersion(List
, Pkg
, Version
, OutVer
) == false)
245 FoundFileDeps
|= List
.HasFileDeps();
250 if (Cache
.HeaderP
->PackageCount
>= (1ULL<<sizeof(Cache
.PkgP
->ID
)*8)-1)
251 return _error
->Error(_("Wow, you exceeded the number of package "
252 "names this APT is capable of."));
253 if (Cache
.HeaderP
->VersionCount
>= (1ULL<<(sizeof(Cache
.VerP
->ID
)*8))-1)
254 return _error
->Error(_("Wow, you exceeded the number of versions "
255 "this APT is capable of."));
256 if (Cache
.HeaderP
->DescriptionCount
>= (1ULL<<(sizeof(Cache
.DescP
->ID
)*8))-1)
257 return _error
->Error(_("Wow, you exceeded the number of descriptions "
258 "this APT is capable of."));
259 if (Cache
.HeaderP
->DependsCount
>= (1ULL<<(sizeof(Cache
.DepP
->ID
)*8))-1ULL)
260 return _error
->Error(_("Wow, you exceeded the number of dependencies "
261 "this APT is capable of."));
263 FoundFileDeps
|= List
.HasFileDeps();
266 // CacheGenerator::MergeListGroup /*{{{*/
267 bool pkgCacheGenerator::MergeListGroup(ListParser
&List
, std::string
const &GrpName
)
269 pkgCache::GrpIterator Grp
= Cache
.FindGrp(GrpName
);
270 // a group has no data on it's own, only packages have it but these
271 // stanzas like this come from Translation- files to add descriptions,
272 // but without a version we don't need a description for it…
273 if (Grp
.end() == true)
275 Dynamic
<pkgCache::GrpIterator
> DynGrp(Grp
);
277 pkgCache::PkgIterator Pkg
;
278 Dynamic
<pkgCache::PkgIterator
> DynPkg(Pkg
);
279 for (Pkg
= Grp
.PackageList(); Pkg
.end() == false; Pkg
= Grp
.NextPkg(Pkg
))
280 if (MergeListPackage(List
, Pkg
) == false)
286 // CacheGenerator::MergeListPackage /*{{{*/
287 bool pkgCacheGenerator::MergeListPackage(ListParser
&List
, pkgCache::PkgIterator
&Pkg
)
289 // we first process the package, then the descriptions
290 // (for deb this package processing is in fact a no-op)
291 pkgCache::VerIterator
Ver(Cache
);
292 Dynamic
<pkgCache::VerIterator
> DynVer(Ver
);
293 if (List
.UsePackage(Pkg
, Ver
) == false)
294 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
295 Pkg
.Name(), "UsePackage", 1);
297 // Find the right version to write the description
298 MD5SumValue CurMd5
= List
.Description_md5();
299 if (CurMd5
.Value().empty() == true || List
.Description().empty() == true)
301 std::string CurLang
= List
.DescriptionLanguage();
303 for (Ver
= Pkg
.VersionList(); Ver
.end() == false; ++Ver
)
305 pkgCache::DescIterator VerDesc
= Ver
.DescriptionList();
307 // a version can only have one md5 describing it
308 if (VerDesc
.end() == true || MD5SumValue(VerDesc
.md5()) != CurMd5
)
311 // don't add a new description if we have one for the given
313 if (IsDuplicateDescription(VerDesc
, CurMd5
, CurLang
) == true)
316 pkgCache::DescIterator Desc
;
317 Dynamic
<pkgCache::DescIterator
> DynDesc(Desc
);
319 map_ptrloc
const descindex
= NewDescription(Desc
, CurLang
, CurMd5
, VerDesc
->md5sum
);
320 if (unlikely(descindex
== 0 && _error
->PendingError()))
321 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
322 Pkg
.Name(), "NewDescription", 1);
324 Desc
->ParentPkg
= Pkg
.Index();
326 // we add at the end, so that the start is constant as we need
327 // that to be able to efficiently share these lists
328 VerDesc
= Ver
.DescriptionList(); // old value might be invalid after ReMap
329 for (;VerDesc
.end() == false && VerDesc
->NextDesc
!= 0; ++VerDesc
);
330 map_ptrloc
* const LastNextDesc
= (VerDesc
.end() == true) ? &Ver
->DescriptionList
: &VerDesc
->NextDesc
;
331 *LastNextDesc
= descindex
;
333 if (NewFileDesc(Desc
,List
) == false)
334 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
335 Pkg
.Name(), "NewFileDesc", 1);
337 // we can stop here as all "same" versions will share the description
344 // CacheGenerator::MergeListVersion /*{{{*/
345 bool pkgCacheGenerator::MergeListVersion(ListParser
&List
, pkgCache::PkgIterator
&Pkg
,
346 std::string
const &Version
, pkgCache::VerIterator
* &OutVer
)
348 pkgCache::VerIterator Ver
= Pkg
.VersionList();
349 Dynamic
<pkgCache::VerIterator
> DynVer(Ver
);
350 map_ptrloc
*LastVer
= &Pkg
->VersionList
;
351 void const * oldMap
= Map
.Data();
353 unsigned long const Hash
= List
.VersionHash();
354 if (Ver
.end() == false)
356 /* We know the list is sorted so we use that fact in the search.
357 Insertion of new versions is done with correct sorting */
359 for (; Ver
.end() == false; LastVer
= &Ver
->NextVer
, Ver
++)
361 Res
= Cache
.VS
->CmpVersion(Version
,Ver
.VerStr());
362 // Version is higher as current version - insert here
365 // Versionstrings are equal - is hash also equal?
366 if (Res
== 0 && Ver
->Hash
== Hash
)
368 // proceed with the next till we have either the right
369 // or we found another version (which will be lower)
372 /* We already have a version for this item, record that we saw it */
373 if (Res
== 0 && Ver
.end() == false && Ver
->Hash
== Hash
)
375 if (List
.UsePackage(Pkg
,Ver
) == false)
376 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
377 Pkg
.Name(), "UsePackage", 2);
379 if (NewFileVer(Ver
,List
) == false)
380 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
381 Pkg
.Name(), "NewFileVer", 1);
383 // Read only a single record and return
395 map_ptrloc
const verindex
= NewVersion(Ver
, Version
, Pkg
.Index(), Hash
, *LastVer
);
396 if (verindex
== 0 && _error
->PendingError())
397 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
398 Pkg
.Name(), "NewVersion", 1);
400 if (oldMap
!= Map
.Data())
401 LastVer
+= (map_ptrloc
*) Map
.Data() - (map_ptrloc
*) oldMap
;
404 if (unlikely(List
.NewVersion(Ver
) == false))
405 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
406 Pkg
.Name(), "NewVersion", 2);
408 if (unlikely(List
.UsePackage(Pkg
,Ver
) == false))
409 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
410 Pkg
.Name(), "UsePackage", 3);
412 if (unlikely(NewFileVer(Ver
,List
) == false))
413 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
414 Pkg
.Name(), "NewFileVer", 2);
416 pkgCache::GrpIterator Grp
= Pkg
.Group();
417 Dynamic
<pkgCache::GrpIterator
> DynGrp(Grp
);
419 /* If it is the first version of this package we need to add implicit
420 Multi-Arch dependencies to all other package versions in the group now -
421 otherwise we just add them for this new version */
422 if (Pkg
.VersionList()->NextVer
== 0)
424 pkgCache::PkgIterator P
= Grp
.PackageList();
425 Dynamic
<pkgCache::PkgIterator
> DynP(P
);
426 for (; P
.end() != true; P
= Grp
.NextPkg(P
))
428 if (P
->ID
== Pkg
->ID
)
430 pkgCache::VerIterator V
= P
.VersionList();
431 Dynamic
<pkgCache::VerIterator
> DynV(V
);
432 for (; V
.end() != true; ++V
)
433 if (unlikely(AddImplicitDepends(V
, Pkg
) == false))
434 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
435 Pkg
.Name(), "AddImplicitDepends", 1);
437 /* :none packages are packages without an architecture. They are forbidden by
438 debian-policy, so usually they will only be in (old) dpkg status files -
439 and dpkg will complain about them - and are pretty rare. We therefore do
440 usually not create conflicts while the parent is created, but only if a :none
441 package (= the target) appears. This creates incorrect dependencies on :none
442 for architecture-specific dependencies on the package we copy from, but we
443 will ignore this bug as architecture-specific dependencies are only allowed
444 in jessie and until then the :none packages should be extinct (hopefully).
445 In other words: This should work long enough to allow graceful removal of
446 these packages, it is not supposed to allow users to keep using them … */
447 if (strcmp(Pkg
.Arch(), "none") == 0)
449 pkgCache::PkgIterator M
= Grp
.FindPreferredPkg();
450 if (M
.end() == false && Pkg
!= M
)
452 pkgCache::DepIterator D
= M
.RevDependsList();
453 Dynamic
<pkgCache::DepIterator
> DynD(D
);
454 for (; D
.end() == false; ++D
)
456 if ((D
->Type
!= pkgCache::Dep::Conflicts
&&
457 D
->Type
!= pkgCache::Dep::DpkgBreaks
&&
458 D
->Type
!= pkgCache::Dep::Replaces
) ||
459 D
.ParentPkg().Group() == Grp
)
462 map_ptrloc
*OldDepLast
= NULL
;
463 pkgCache::VerIterator ConVersion
= D
.ParentVer();
464 Dynamic
<pkgCache::VerIterator
> DynV(ConVersion
);
465 // duplicate the Conflicts/Breaks/Replaces for :none arch
466 NewDepends(Pkg
, ConVersion
, D
->Version
,
467 D
->CompareOp
, D
->Type
, OldDepLast
);
472 if (unlikely(AddImplicitDepends(Grp
, Pkg
, Ver
) == false))
473 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
474 Pkg
.Name(), "AddImplicitDepends", 2);
476 // Read only a single record and return
483 /* Record the Description (it is not translated) */
484 MD5SumValue CurMd5
= List
.Description_md5();
485 if (CurMd5
.Value().empty() == true || List
.Description().empty() == true)
487 std::string CurLang
= List
.DescriptionLanguage();
489 /* Before we add a new description we first search in the group for
490 a version with a description of the same MD5 - if so we reuse this
491 description group instead of creating our own for this version */
492 for (pkgCache::PkgIterator P
= Grp
.PackageList();
493 P
.end() == false; P
= Grp
.NextPkg(P
))
495 for (pkgCache::VerIterator V
= P
.VersionList();
496 V
.end() == false; ++V
)
498 if (IsDuplicateDescription(V
.DescriptionList(), CurMd5
, "") == false)
500 Ver
->DescriptionList
= V
->DescriptionList
;
505 // We haven't found reusable descriptions, so add the first description
506 pkgCache::DescIterator Desc
= Ver
.DescriptionList();
507 Dynamic
<pkgCache::DescIterator
> DynDesc(Desc
);
509 map_ptrloc
const descindex
= NewDescription(Desc
, CurLang
, CurMd5
, 0);
510 if (unlikely(descindex
== 0 && _error
->PendingError()))
511 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
512 Pkg
.Name(), "NewDescription", 2);
514 Desc
->ParentPkg
= Pkg
.Index();
515 Ver
->DescriptionList
= descindex
;
517 if (NewFileDesc(Desc
,List
) == false)
518 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
519 Pkg
.Name(), "NewFileDesc", 2);
525 // CacheGenerator::MergeFileProvides - Merge file provides /*{{{*/
526 // ---------------------------------------------------------------------
527 /* If we found any file depends while parsing the main list we need to
528 resolve them. Since it is undesired to load the entire list of files
529 into the cache as virtual packages we do a two stage effort. MergeList
530 identifies the file depends and this creates Provdies for them by
531 re-parsing all the indexs. */
532 bool pkgCacheGenerator::MergeFileProvides(ListParser
&List
)
536 unsigned int Counter
= 0;
537 while (List
.Step() == true)
539 string PackageName
= List
.Package();
540 if (PackageName
.empty() == true)
542 string Version
= List
.Version();
543 if (Version
.empty() == true)
546 pkgCache::PkgIterator Pkg
= Cache
.FindPkg(PackageName
);
547 Dynamic
<pkgCache::PkgIterator
> DynPkg(Pkg
);
548 if (Pkg
.end() == true)
549 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
550 PackageName
.c_str(), "FindPkg", 1);
552 if (Counter
% 100 == 0 && Progress
!= 0)
553 Progress
->Progress(List
.Offset());
555 unsigned long Hash
= List
.VersionHash();
556 pkgCache::VerIterator Ver
= Pkg
.VersionList();
557 Dynamic
<pkgCache::VerIterator
> DynVer(Ver
);
558 for (; Ver
.end() == false; ++Ver
)
560 if (Ver
->Hash
== Hash
&& Version
== Ver
.VerStr())
562 if (List
.CollectFileProvides(Cache
,Ver
) == false)
563 return _error
->Error(_("Error occurred while processing %s (%s%d)"),
564 PackageName
.c_str(), "CollectFileProvides", 1);
569 if (Ver
.end() == true)
570 _error
->Warning(_("Package %s %s was not found while processing file dependencies"),PackageName
.c_str(),Version
.c_str());
576 // CacheGenerator::NewGroup - Add a new group /*{{{*/
577 // ---------------------------------------------------------------------
578 /* This creates a new group structure and adds it to the hash table */
579 bool pkgCacheGenerator::NewGroup(pkgCache::GrpIterator
&Grp
, const string
&Name
)
581 Grp
= Cache
.FindGrp(Name
);
582 if (Grp
.end() == false)
586 map_ptrloc
const Group
= AllocateInMap(sizeof(pkgCache::Group
));
587 if (unlikely(Group
== 0))
590 Grp
= pkgCache::GrpIterator(Cache
, Cache
.GrpP
+ Group
);
591 map_ptrloc
const idxName
= WriteStringInMap(Name
);
592 if (unlikely(idxName
== 0))
596 // Insert it into the hash table
597 unsigned long const Hash
= Cache
.Hash(Name
);
598 map_ptrloc
*insertAt
= &Cache
.HeaderP
->GrpHashTable
[Hash
];
599 while (*insertAt
!= 0 && strcasecmp(Name
.c_str(), Cache
.StrP
+ (Cache
.GrpP
+ *insertAt
)->Name
) > 0)
600 insertAt
= &(Cache
.GrpP
+ *insertAt
)->Next
;
601 Grp
->Next
= *insertAt
;
604 Grp
->ID
= Cache
.HeaderP
->GroupCount
++;
608 // CacheGenerator::NewPackage - Add a new package /*{{{*/
609 // ---------------------------------------------------------------------
610 /* This creates a new package structure and adds it to the hash table */
611 bool pkgCacheGenerator::NewPackage(pkgCache::PkgIterator
&Pkg
,const string
&Name
,
612 const string
&Arch
) {
613 pkgCache::GrpIterator Grp
;
614 Dynamic
<pkgCache::GrpIterator
> DynGrp(Grp
);
615 if (unlikely(NewGroup(Grp
, Name
) == false))
618 Pkg
= Grp
.FindPkg(Arch
);
619 if (Pkg
.end() == false)
623 map_ptrloc
const Package
= AllocateInMap(sizeof(pkgCache::Package
));
624 if (unlikely(Package
== 0))
626 Pkg
= pkgCache::PkgIterator(Cache
,Cache
.PkgP
+ Package
);
628 // Insert the package into our package list
629 if (Grp
->FirstPackage
== 0) // the group is new
631 Grp
->FirstPackage
= Package
;
632 // Insert it into the hash table
633 unsigned long const Hash
= Cache
.Hash(Name
);
634 map_ptrloc
*insertAt
= &Cache
.HeaderP
->PkgHashTable
[Hash
];
635 while (*insertAt
!= 0 && strcasecmp(Name
.c_str(), Cache
.StrP
+ (Cache
.PkgP
+ *insertAt
)->Name
) > 0)
636 insertAt
= &(Cache
.PkgP
+ *insertAt
)->NextPackage
;
637 Pkg
->NextPackage
= *insertAt
;
640 else // Group the Packages together
642 // this package is the new last package
643 pkgCache::PkgIterator
LastPkg(Cache
, Cache
.PkgP
+ Grp
->LastPackage
);
644 Pkg
->NextPackage
= LastPkg
->NextPackage
;
645 LastPkg
->NextPackage
= Package
;
647 Grp
->LastPackage
= Package
;
649 // Set the name, arch and the ID
650 Pkg
->Name
= Grp
->Name
;
651 Pkg
->Group
= Grp
.Index();
652 // all is mapped to the native architecture
653 map_ptrloc
const idxArch
= (Arch
== "all") ? Cache
.HeaderP
->Architecture
: WriteUniqString(Arch
.c_str());
654 if (unlikely(idxArch
== 0))
657 Pkg
->ID
= Cache
.HeaderP
->PackageCount
++;
662 // CacheGenerator::AddImplicitDepends /*{{{*/
663 bool pkgCacheGenerator::AddImplicitDepends(pkgCache::GrpIterator
&G
,
664 pkgCache::PkgIterator
&P
,
665 pkgCache::VerIterator
&V
)
667 // copy P.Arch() into a string here as a cache remap
668 // in NewDepends() later may alter the pointer location
669 string Arch
= P
.Arch() == NULL
? "" : P
.Arch();
670 map_ptrloc
*OldDepLast
= NULL
;
671 /* MultiArch handling introduces a lot of implicit Dependencies:
672 - MultiArch: same → Co-Installable if they have the same version
673 - All others conflict with all other group members */
674 bool const coInstall
= ((V
->MultiArch
& pkgCache::Version::Same
) == pkgCache::Version::Same
);
675 pkgCache::PkgIterator D
= G
.PackageList();
676 Dynamic
<pkgCache::PkgIterator
> DynD(D
);
677 map_ptrloc
const VerStrIdx
= V
->VerStr
;
678 for (; D
.end() != true; D
= G
.NextPkg(D
))
680 if (Arch
== D
.Arch() || D
->VersionList
== 0)
682 /* We allow only one installed arch at the time
683 per group, therefore each group member conflicts
684 with all other group members */
685 if (coInstall
== true)
687 // Replaces: ${self}:other ( << ${binary:Version})
688 NewDepends(D
, V
, VerStrIdx
,
689 pkgCache::Dep::Less
, pkgCache::Dep::Replaces
,
691 // Breaks: ${self}:other (!= ${binary:Version})
692 NewDepends(D
, V
, VerStrIdx
,
693 pkgCache::Dep::NotEquals
, pkgCache::Dep::DpkgBreaks
,
696 // Conflicts: ${self}:other
698 pkgCache::Dep::NoOp
, pkgCache::Dep::Conflicts
,
704 bool pkgCacheGenerator::AddImplicitDepends(pkgCache::VerIterator
&V
,
705 pkgCache::PkgIterator
&D
)
707 /* MultiArch handling introduces a lot of implicit Dependencies:
708 - MultiArch: same → Co-Installable if they have the same version
709 - All others conflict with all other group members */
710 map_ptrloc
*OldDepLast
= NULL
;
711 bool const coInstall
= ((V
->MultiArch
& pkgCache::Version::Same
) == pkgCache::Version::Same
);
712 if (coInstall
== true)
714 map_ptrloc
const VerStrIdx
= V
->VerStr
;
715 // Replaces: ${self}:other ( << ${binary:Version})
716 NewDepends(D
, V
, VerStrIdx
,
717 pkgCache::Dep::Less
, pkgCache::Dep::Replaces
,
719 // Breaks: ${self}:other (!= ${binary:Version})
720 NewDepends(D
, V
, VerStrIdx
,
721 pkgCache::Dep::NotEquals
, pkgCache::Dep::DpkgBreaks
,
724 // Conflicts: ${self}:other
726 pkgCache::Dep::NoOp
, pkgCache::Dep::Conflicts
,
733 // CacheGenerator::NewFileVer - Create a new File<->Version association /*{{{*/
734 // ---------------------------------------------------------------------
736 bool pkgCacheGenerator::NewFileVer(pkgCache::VerIterator
&Ver
,
739 if (CurrentFile
== 0)
743 map_ptrloc
const VerFile
= AllocateInMap(sizeof(pkgCache::VerFile
));
747 pkgCache::VerFileIterator
VF(Cache
,Cache
.VerFileP
+ VerFile
);
748 VF
->File
= CurrentFile
- Cache
.PkgFileP
;
750 // Link it to the end of the list
751 map_ptrloc
*Last
= &Ver
->FileList
;
752 for (pkgCache::VerFileIterator V
= Ver
.FileList(); V
.end() == false; ++V
)
754 VF
->NextFile
= *Last
;
757 VF
->Offset
= List
.Offset();
758 VF
->Size
= List
.Size();
759 if (Cache
.HeaderP
->MaxVerFileSize
< VF
->Size
)
760 Cache
.HeaderP
->MaxVerFileSize
= VF
->Size
;
761 Cache
.HeaderP
->VerFileCount
++;
766 // CacheGenerator::NewVersion - Create a new Version /*{{{*/
767 // ---------------------------------------------------------------------
768 /* This puts a version structure in the linked list */
769 unsigned long pkgCacheGenerator::NewVersion(pkgCache::VerIterator
&Ver
,
770 const string
&VerStr
,
771 map_ptrloc
const ParentPkg
,
772 unsigned long const Hash
,
776 map_ptrloc
const Version
= AllocateInMap(sizeof(pkgCache::Version
));
781 Ver
= pkgCache::VerIterator(Cache
,Cache
.VerP
+ Version
);
782 //Dynamic<pkgCache::VerIterator> DynV(Ver); // caller MergeListVersion already takes care of it
784 Ver
->ParentPkg
= ParentPkg
;
786 Ver
->ID
= Cache
.HeaderP
->VersionCount
++;
788 // try to find the version string in the group for reuse
789 pkgCache::PkgIterator Pkg
= Ver
.ParentPkg();
790 pkgCache::GrpIterator Grp
= Pkg
.Group();
791 if (Pkg
.end() == false && Grp
.end() == false)
793 for (pkgCache::PkgIterator P
= Grp
.PackageList(); P
.end() == false; P
= Grp
.NextPkg(P
))
797 for (pkgCache::VerIterator V
= P
.VersionList(); V
.end() == false; ++V
)
799 int const cmp
= strcmp(V
.VerStr(), VerStr
.c_str());
802 Ver
->VerStr
= V
->VerStr
;
810 // haven't found the version string, so create
811 map_ptrloc
const idxVerStr
= WriteStringInMap(VerStr
);
812 if (unlikely(idxVerStr
== 0))
814 Ver
->VerStr
= idxVerStr
;
818 // CacheGenerator::NewFileDesc - Create a new File<->Desc association /*{{{*/
819 // ---------------------------------------------------------------------
821 bool pkgCacheGenerator::NewFileDesc(pkgCache::DescIterator
&Desc
,
824 if (CurrentFile
== 0)
828 map_ptrloc
const DescFile
= AllocateInMap(sizeof(pkgCache::DescFile
));
832 pkgCache::DescFileIterator
DF(Cache
,Cache
.DescFileP
+ DescFile
);
833 DF
->File
= CurrentFile
- Cache
.PkgFileP
;
835 // Link it to the end of the list
836 map_ptrloc
*Last
= &Desc
->FileList
;
837 for (pkgCache::DescFileIterator D
= Desc
.FileList(); D
.end() == false; ++D
)
840 DF
->NextFile
= *Last
;
843 DF
->Offset
= List
.Offset();
844 DF
->Size
= List
.Size();
845 if (Cache
.HeaderP
->MaxDescFileSize
< DF
->Size
)
846 Cache
.HeaderP
->MaxDescFileSize
= DF
->Size
;
847 Cache
.HeaderP
->DescFileCount
++;
852 // CacheGenerator::NewDescription - Create a new Description /*{{{*/
853 // ---------------------------------------------------------------------
854 /* This puts a description structure in the linked list */
855 map_ptrloc
pkgCacheGenerator::NewDescription(pkgCache::DescIterator
&Desc
,
857 const MD5SumValue
&md5sum
,
858 map_ptrloc idxmd5str
)
861 map_ptrloc
const Description
= AllocateInMap(sizeof(pkgCache::Description
));
862 if (Description
== 0)
866 Desc
= pkgCache::DescIterator(Cache
,Cache
.DescP
+ Description
);
867 Desc
->ID
= Cache
.HeaderP
->DescriptionCount
++;
868 map_ptrloc
const idxlanguage_code
= WriteUniqString(Lang
);
869 if (unlikely(idxlanguage_code
== 0))
871 Desc
->language_code
= idxlanguage_code
;
874 Desc
->md5sum
= idxmd5str
;
877 map_ptrloc
const idxmd5sum
= WriteStringInMap(md5sum
.Value());
878 if (unlikely(idxmd5sum
== 0))
880 Desc
->md5sum
= idxmd5sum
;
886 // CacheGenerator::NewDepends - Create a dependency element /*{{{*/
887 // ---------------------------------------------------------------------
888 /* This creates a dependency element in the tree. It is linked to the
889 version and to the package that it is pointing to. */
890 bool pkgCacheGenerator::NewDepends(pkgCache::PkgIterator
&Pkg
,
891 pkgCache::VerIterator
&Ver
,
892 string
const &Version
,
893 unsigned int const &Op
,
894 unsigned int const &Type
,
895 map_ptrloc
* &OldDepLast
)
897 map_ptrloc index
= 0;
898 if (Version
.empty() == false)
900 int const CmpOp
= Op
& 0x0F;
901 // =-deps are used (79:1) for lockstep on same-source packages (e.g. data-packages)
902 if (CmpOp
== pkgCache::Dep::Equals
&& strcmp(Version
.c_str(), Ver
.VerStr()) == 0)
907 void const * const oldMap
= Map
.Data();
908 index
= WriteStringInMap(Version
);
909 if (unlikely(index
== 0))
911 if (OldDepLast
!= 0 && oldMap
!= Map
.Data())
912 OldDepLast
+= (map_ptrloc
*) Map
.Data() - (map_ptrloc
*) oldMap
;
915 return NewDepends(Pkg
, Ver
, index
, Op
, Type
, OldDepLast
);
917 bool pkgCacheGenerator::NewDepends(pkgCache::PkgIterator
&Pkg
,
918 pkgCache::VerIterator
&Ver
,
919 map_ptrloc
const Version
,
920 unsigned int const &Op
,
921 unsigned int const &Type
,
922 map_ptrloc
* &OldDepLast
)
924 void const * const oldMap
= Map
.Data();
926 map_ptrloc
const Dependency
= AllocateInMap(sizeof(pkgCache::Dependency
));
927 if (unlikely(Dependency
== 0))
931 pkgCache::DepIterator
Dep(Cache
,Cache
.DepP
+ Dependency
);
932 Dynamic
<pkgCache::DepIterator
> DynDep(Dep
);
933 Dep
->ParentVer
= Ver
.Index();
936 Dep
->Version
= Version
;
937 Dep
->ID
= Cache
.HeaderP
->DependsCount
++;
939 // Link it to the package
940 Dep
->Package
= Pkg
.Index();
941 Dep
->NextRevDepends
= Pkg
->RevDepends
;
942 Pkg
->RevDepends
= Dep
.Index();
944 // Do we know where to link the Dependency to?
945 if (OldDepLast
== NULL
)
947 OldDepLast
= &Ver
->DependsList
;
948 for (pkgCache::DepIterator D
= Ver
.DependsList(); D
.end() == false; ++D
)
949 OldDepLast
= &D
->NextDepends
;
950 } else if (oldMap
!= Map
.Data())
951 OldDepLast
+= (map_ptrloc
*) Map
.Data() - (map_ptrloc
*) oldMap
;
953 Dep
->NextDepends
= *OldDepLast
;
954 *OldDepLast
= Dep
.Index();
955 OldDepLast
= &Dep
->NextDepends
;
960 // ListParser::NewDepends - Create the environment for a new dependency /*{{{*/
961 // ---------------------------------------------------------------------
962 /* This creates a Group and the Package to link this dependency to if
963 needed and handles also the caching of the old endpoint */
964 bool pkgCacheGenerator::ListParser::NewDepends(pkgCache::VerIterator
&Ver
,
965 const string
&PackageName
,
967 const string
&Version
,
971 pkgCache::GrpIterator Grp
;
972 Dynamic
<pkgCache::GrpIterator
> DynGrp(Grp
);
973 if (unlikely(Owner
->NewGroup(Grp
, PackageName
) == false))
976 // Locate the target package
977 pkgCache::PkgIterator Pkg
= Grp
.FindPkg(Arch
);
978 // we don't create 'none' packages and their dependencies if we can avoid it …
979 if (Pkg
.end() == true && Arch
== "none" && strcmp(Ver
.ParentPkg().Arch(), "none") != 0)
981 Dynamic
<pkgCache::PkgIterator
> DynPkg(Pkg
);
982 if (Pkg
.end() == true) {
983 if (unlikely(Owner
->NewPackage(Pkg
, PackageName
, Arch
) == false))
987 // Is it a file dependency?
988 if (unlikely(PackageName
[0] == '/'))
989 FoundFileDeps
= true;
991 /* Caching the old end point speeds up generation substantially */
992 if (OldDepVer
!= Ver
) {
997 return Owner
->NewDepends(Pkg
, Ver
, Version
, Op
, Type
, OldDepLast
);
1000 // ListParser::NewProvides - Create a Provides element /*{{{*/
1001 // ---------------------------------------------------------------------
1003 bool pkgCacheGenerator::ListParser::NewProvides(pkgCache::VerIterator
&Ver
,
1004 const string
&PkgName
,
1005 const string
&PkgArch
,
1006 const string
&Version
)
1008 pkgCache
&Cache
= Owner
->Cache
;
1010 // We do not add self referencing provides
1011 if (Ver
.ParentPkg().Name() == PkgName
&& (PkgArch
== Ver
.ParentPkg().Arch() ||
1012 (PkgArch
== "all" && strcmp((Cache
.StrP
+ Cache
.HeaderP
->Architecture
), Ver
.ParentPkg().Arch()) == 0)))
1016 map_ptrloc
const Provides
= Owner
->AllocateInMap(sizeof(pkgCache::Provides
));
1017 if (unlikely(Provides
== 0))
1019 Cache
.HeaderP
->ProvidesCount
++;
1022 pkgCache::PrvIterator
Prv(Cache
,Cache
.ProvideP
+ Provides
,Cache
.PkgP
);
1023 Dynamic
<pkgCache::PrvIterator
> DynPrv(Prv
);
1024 Prv
->Version
= Ver
.Index();
1025 Prv
->NextPkgProv
= Ver
->ProvidesList
;
1026 Ver
->ProvidesList
= Prv
.Index();
1027 if (Version
.empty() == false) {
1028 map_ptrloc
const idxProvideVersion
= WriteString(Version
);
1029 Prv
->ProvideVersion
= idxProvideVersion
;
1030 if (unlikely(idxProvideVersion
== 0))
1034 // Locate the target package
1035 pkgCache::PkgIterator Pkg
;
1036 Dynamic
<pkgCache::PkgIterator
> DynPkg(Pkg
);
1037 if (unlikely(Owner
->NewPackage(Pkg
,PkgName
, PkgArch
) == false))
1040 // Link it to the package
1041 Prv
->ParentPkg
= Pkg
.Index();
1042 Prv
->NextProvides
= Pkg
->ProvidesList
;
1043 Pkg
->ProvidesList
= Prv
.Index();
1048 // CacheGenerator::SelectFile - Select the current file being parsed /*{{{*/
1049 // ---------------------------------------------------------------------
1050 /* This is used to select which file is to be associated with all newly
1051 added versions. The caller is responsible for setting the IMS fields. */
1052 bool pkgCacheGenerator::SelectFile(const string
&File
,const string
&Site
,
1053 const pkgIndexFile
&Index
,
1054 unsigned long Flags
)
1056 // Get some space for the structure
1057 map_ptrloc
const idxFile
= AllocateInMap(sizeof(*CurrentFile
));
1058 if (unlikely(idxFile
== 0))
1060 CurrentFile
= Cache
.PkgFileP
+ idxFile
;
1063 map_ptrloc
const idxFileName
= WriteStringInMap(File
);
1064 map_ptrloc
const idxSite
= WriteUniqString(Site
);
1065 if (unlikely(idxFileName
== 0 || idxSite
== 0))
1067 CurrentFile
->FileName
= idxFileName
;
1068 CurrentFile
->Site
= idxSite
;
1069 CurrentFile
->NextFile
= Cache
.HeaderP
->FileList
;
1070 CurrentFile
->Flags
= Flags
;
1071 CurrentFile
->ID
= Cache
.HeaderP
->PackageFileCount
;
1072 map_ptrloc
const idxIndexType
= WriteUniqString(Index
.GetType()->Label
);
1073 if (unlikely(idxIndexType
== 0))
1075 CurrentFile
->IndexType
= idxIndexType
;
1077 Cache
.HeaderP
->FileList
= CurrentFile
- Cache
.PkgFileP
;
1078 Cache
.HeaderP
->PackageFileCount
++;
1081 Progress
->SubProgress(Index
.Size());
1085 // CacheGenerator::WriteUniqueString - Insert a unique string /*{{{*/
1086 // ---------------------------------------------------------------------
1087 /* This is used to create handles to strings. Given the same text it
1088 always returns the same number */
1089 unsigned long pkgCacheGenerator::WriteUniqString(const char *S
,
1092 /* We use a very small transient hash table here, this speeds up generation
1093 by a fair amount on slower machines */
1094 pkgCache::StringItem
*&Bucket
= UniqHash
[(S
[0]*5 + S
[1]) % _count(UniqHash
)];
1096 stringcmp(S
,S
+Size
,Cache
.StrP
+ Bucket
->String
) == 0)
1097 return Bucket
->String
;
1099 // Search for an insertion point
1100 pkgCache::StringItem
*I
= Cache
.StringItemP
+ Cache
.HeaderP
->StringList
;
1102 map_ptrloc
*Last
= &Cache
.HeaderP
->StringList
;
1103 for (; I
!= Cache
.StringItemP
; Last
= &I
->NextItem
,
1104 I
= Cache
.StringItemP
+ I
->NextItem
)
1106 Res
= stringcmp(S
,S
+Size
,Cache
.StrP
+ I
->String
);
1119 void const * const oldMap
= Map
.Data();
1120 map_ptrloc
const Item
= AllocateInMap(sizeof(pkgCache::StringItem
));
1124 map_ptrloc
const idxString
= WriteStringInMap(S
,Size
);
1125 if (unlikely(idxString
== 0))
1127 if (oldMap
!= Map
.Data()) {
1128 Last
+= (map_ptrloc
*) Map
.Data() - (map_ptrloc
*) oldMap
;
1129 I
+= (pkgCache::StringItem
*) Map
.Data() - (pkgCache::StringItem
*) oldMap
;
1133 // Fill in the structure
1134 pkgCache::StringItem
*ItemP
= Cache
.StringItemP
+ Item
;
1135 ItemP
->NextItem
= I
- Cache
.StringItemP
;
1136 ItemP
->String
= idxString
;
1139 return ItemP
->String
;
1142 // CheckValidity - Check that a cache is up-to-date /*{{{*/
1143 // ---------------------------------------------------------------------
1144 /* This just verifies that each file in the list of index files exists,
1145 has matching attributes with the cache and the cache does not have
1147 static bool CheckValidity(const string
&CacheFile
,
1148 pkgSourceList
&List
,
1153 bool const Debug
= _config
->FindB("Debug::pkgCacheGen", false);
1154 // No file, certainly invalid
1155 if (CacheFile
.empty() == true || FileExists(CacheFile
) == false)
1158 std::clog
<< "CacheFile doesn't exist" << std::endl
;
1162 if (List
.GetLastModifiedTime() > GetModificationTime(CacheFile
))
1165 std::clog
<< "sources.list is newer than the cache" << std::endl
;
1170 FileFd
CacheF(CacheFile
,FileFd::ReadOnly
);
1171 SPtr
<MMap
> Map
= new MMap(CacheF
,0);
1172 pkgCache
Cache(Map
);
1173 if (_error
->PendingError() == true || Map
->Size() == 0)
1176 std::clog
<< "Errors are pending or Map is empty()" << std::endl
;
1181 /* Now we check every index file, see if it is in the cache,
1182 verify the IMS data and check that it is on the disk too.. */
1183 SPtrArray
<bool> Visited
= new bool[Cache
.HeaderP
->PackageFileCount
];
1184 memset(Visited
,0,sizeof(*Visited
)*Cache
.HeaderP
->PackageFileCount
);
1185 for (; Start
!= End
; ++Start
)
1188 std::clog
<< "Checking PkgFile " << (*Start
)->Describe() << ": ";
1189 if ((*Start
)->HasPackages() == false)
1192 std::clog
<< "Has NO packages" << std::endl
;
1196 if ((*Start
)->Exists() == false)
1198 #if 0 // mvo: we no longer give a message here (Default Sources spec)
1199 _error
->WarningE("stat",_("Couldn't stat source package list %s"),
1200 (*Start
)->Describe().c_str());
1203 std::clog
<< "file doesn't exist" << std::endl
;
1207 // FindInCache is also expected to do an IMS check.
1208 pkgCache::PkgFileIterator File
= (*Start
)->FindInCache(Cache
);
1209 if (File
.end() == true)
1212 std::clog
<< "FindInCache returned end-Pointer" << std::endl
;
1216 Visited
[File
->ID
] = true;
1218 std::clog
<< "with ID " << File
->ID
<< " is valid" << std::endl
;
1221 for (unsigned I
= 0; I
!= Cache
.HeaderP
->PackageFileCount
; I
++)
1222 if (Visited
[I
] == false)
1225 std::clog
<< "File with ID" << I
<< " wasn't visited" << std::endl
;
1229 if (_error
->PendingError() == true)
1233 std::clog
<< "Validity failed because of pending errors:" << std::endl
;
1234 _error
->DumpErrors();
1241 *OutMap
= Map
.UnGuard();
1245 // ComputeSize - Compute the total size of a bunch of files /*{{{*/
1246 // ---------------------------------------------------------------------
1247 /* Size is kind of an abstract notion that is only used for the progress
1249 static unsigned long ComputeSize(FileIterator Start
,FileIterator End
)
1251 unsigned long TotalSize
= 0;
1252 for (; Start
!= End
; ++Start
)
1254 if ((*Start
)->HasPackages() == false)
1256 TotalSize
+= (*Start
)->Size();
1261 // BuildCache - Merge the list of index files into the cache /*{{{*/
1262 // ---------------------------------------------------------------------
1264 static bool BuildCache(pkgCacheGenerator
&Gen
,
1265 OpProgress
*Progress
,
1266 unsigned long &CurrentSize
,unsigned long TotalSize
,
1267 FileIterator Start
, FileIterator End
)
1270 for (I
= Start
; I
!= End
; ++I
)
1272 if ((*I
)->HasPackages() == false)
1275 if ((*I
)->Exists() == false)
1278 if ((*I
)->FindInCache(Gen
.GetCache()).end() == false)
1280 _error
->Warning("Duplicate sources.list entry %s",
1281 (*I
)->Describe().c_str());
1285 unsigned long Size
= (*I
)->Size();
1286 if (Progress
!= NULL
)
1287 Progress
->OverallProgress(CurrentSize
,TotalSize
,Size
,_("Reading package lists"));
1288 CurrentSize
+= Size
;
1290 if ((*I
)->Merge(Gen
,Progress
) == false)
1294 if (Gen
.HasFileDeps() == true)
1296 if (Progress
!= NULL
)
1298 TotalSize
= ComputeSize(Start
, End
);
1300 for (I
= Start
; I
!= End
; ++I
)
1302 unsigned long Size
= (*I
)->Size();
1303 if (Progress
!= NULL
)
1304 Progress
->OverallProgress(CurrentSize
,TotalSize
,Size
,_("Collecting File Provides"));
1305 CurrentSize
+= Size
;
1306 if ((*I
)->MergeFileProvides(Gen
,Progress
) == false)
1314 // CacheGenerator::CreateDynamicMMap - load an mmap with configuration options /*{{{*/
1315 DynamicMMap
* pkgCacheGenerator::CreateDynamicMMap(FileFd
*CacheF
, unsigned long Flags
) {
1316 unsigned long const MapStart
= _config
->FindI("APT::Cache-Start", 24*1024*1024);
1317 unsigned long const MapGrow
= _config
->FindI("APT::Cache-Grow", 1*1024*1024);
1318 unsigned long const MapLimit
= _config
->FindI("APT::Cache-Limit", 0);
1319 Flags
|= MMap::Moveable
;
1320 if (_config
->FindB("APT::Cache-Fallback", false) == true)
1321 Flags
|= MMap::Fallback
;
1323 return new DynamicMMap(*CacheF
, Flags
, MapStart
, MapGrow
, MapLimit
);
1325 return new DynamicMMap(Flags
, MapStart
, MapGrow
, MapLimit
);
1328 // CacheGenerator::MakeStatusCache - Construct the status cache /*{{{*/
1329 // ---------------------------------------------------------------------
1330 /* This makes sure that the status cache (the cache that has all
1331 index files from the sources list and all local ones) is ready
1332 to be mmaped. If OutMap is not zero then a MMap object representing
1333 the cache will be stored there. This is pretty much mandetory if you
1334 are using AllowMem. AllowMem lets the function be run as non-root
1335 where it builds the cache 'fast' into a memory buffer. */
1336 __deprecated
bool pkgMakeStatusCache(pkgSourceList
&List
,OpProgress
&Progress
,
1337 MMap
**OutMap
, bool AllowMem
)
1338 { return pkgCacheGenerator::MakeStatusCache(List
, &Progress
, OutMap
, AllowMem
); }
1339 bool pkgCacheGenerator::MakeStatusCache(pkgSourceList
&List
,OpProgress
*Progress
,
1340 MMap
**OutMap
,bool AllowMem
)
1342 bool const Debug
= _config
->FindB("Debug::pkgCacheGen", false);
1344 std::vector
<pkgIndexFile
*> Files
;
1345 for (std::vector
<metaIndex
*>::const_iterator i
= List
.begin();
1349 std::vector
<pkgIndexFile
*> *Indexes
= (*i
)->GetIndexFiles();
1350 for (std::vector
<pkgIndexFile
*>::const_iterator j
= Indexes
->begin();
1351 j
!= Indexes
->end();
1353 Files
.push_back (*j
);
1356 unsigned long const EndOfSource
= Files
.size();
1357 if (_system
->AddStatusFiles(Files
) == false)
1360 // Decide if we can write to the files..
1361 string
const CacheFile
= _config
->FindFile("Dir::Cache::pkgcache");
1362 string
const SrcCacheFile
= _config
->FindFile("Dir::Cache::srcpkgcache");
1364 // ensure the cache directory exists
1365 if (CacheFile
.empty() == false || SrcCacheFile
.empty() == false)
1367 string dir
= _config
->FindDir("Dir::Cache");
1368 size_t const len
= dir
.size();
1369 if (len
> 5 && dir
.find("/apt/", len
- 6, 5) == len
- 5)
1370 dir
= dir
.substr(0, len
- 5);
1371 if (CacheFile
.empty() == false)
1372 CreateDirectory(dir
, flNotFile(CacheFile
));
1373 if (SrcCacheFile
.empty() == false)
1374 CreateDirectory(dir
, flNotFile(SrcCacheFile
));
1377 // Decide if we can write to the cache
1378 bool Writeable
= false;
1379 if (CacheFile
.empty() == false)
1380 Writeable
= access(flNotFile(CacheFile
).c_str(),W_OK
) == 0;
1382 if (SrcCacheFile
.empty() == false)
1383 Writeable
= access(flNotFile(SrcCacheFile
).c_str(),W_OK
) == 0;
1385 std::clog
<< "Do we have write-access to the cache files? " << (Writeable
? "YES" : "NO") << std::endl
;
1387 if (Writeable
== false && AllowMem
== false && CacheFile
.empty() == false)
1388 return _error
->Error(_("Unable to write to %s"),flNotFile(CacheFile
).c_str());
1390 if (Progress
!= NULL
)
1391 Progress
->OverallProgress(0,1,1,_("Reading package lists"));
1393 // Cache is OK, Fin.
1394 if (CheckValidity(CacheFile
, List
, Files
.begin(),Files
.end(),OutMap
) == true)
1396 if (Progress
!= NULL
)
1397 Progress
->OverallProgress(1,1,1,_("Reading package lists"));
1399 std::clog
<< "pkgcache.bin is valid - no need to build anything" << std::endl
;
1402 else if (Debug
== true)
1403 std::clog
<< "pkgcache.bin is NOT valid" << std::endl
;
1405 /* At this point we know we need to reconstruct the package cache,
1407 SPtr
<FileFd
> CacheF
;
1408 SPtr
<DynamicMMap
> Map
;
1409 if (Writeable
== true && CacheFile
.empty() == false)
1411 _error
->PushToStack();
1412 unlink(CacheFile
.c_str());
1413 CacheF
= new FileFd(CacheFile
,FileFd::WriteAtomic
);
1414 fchmod(CacheF
->Fd(),0644);
1415 Map
= CreateDynamicMMap(CacheF
, MMap::Public
);
1416 if (_error
->PendingError() == true)
1418 delete CacheF
.UnGuard();
1419 delete Map
.UnGuard();
1421 std::clog
<< "Open filebased MMap FAILED" << std::endl
;
1423 if (AllowMem
== false)
1425 _error
->MergeWithStack();
1428 _error
->RevertToStack();
1432 _error
->MergeWithStack();
1434 std::clog
<< "Open filebased MMap" << std::endl
;
1437 if (Writeable
== false || CacheFile
.empty() == true)
1439 // Just build it in memory..
1440 Map
= CreateDynamicMMap(NULL
);
1442 std::clog
<< "Open memory Map (not filebased)" << std::endl
;
1445 // Lets try the source cache.
1446 unsigned long CurrentSize
= 0;
1447 unsigned long TotalSize
= 0;
1448 if (CheckValidity(SrcCacheFile
, List
, Files
.begin(),
1449 Files
.begin()+EndOfSource
) == true)
1452 std::clog
<< "srcpkgcache.bin is valid - populate MMap with it." << std::endl
;
1453 // Preload the map with the source cache
1454 FileFd
SCacheF(SrcCacheFile
,FileFd::ReadOnly
);
1455 unsigned long const alloc
= Map
->RawAllocate(SCacheF
.Size());
1456 if ((alloc
== 0 && _error
->PendingError())
1457 || SCacheF
.Read((unsigned char *)Map
->Data() + alloc
,
1458 SCacheF
.Size()) == false)
1461 TotalSize
= ComputeSize(Files
.begin()+EndOfSource
,Files
.end());
1463 // Build the status cache
1464 pkgCacheGenerator
Gen(Map
.Get(),Progress
);
1465 if (_error
->PendingError() == true)
1467 if (BuildCache(Gen
,Progress
,CurrentSize
,TotalSize
,
1468 Files
.begin()+EndOfSource
,Files
.end()) == false)
1474 std::clog
<< "srcpkgcache.bin is NOT valid - rebuild" << std::endl
;
1475 TotalSize
= ComputeSize(Files
.begin(),Files
.end());
1477 // Build the source cache
1478 pkgCacheGenerator
Gen(Map
.Get(),Progress
);
1479 if (_error
->PendingError() == true)
1481 if (BuildCache(Gen
,Progress
,CurrentSize
,TotalSize
,
1482 Files
.begin(),Files
.begin()+EndOfSource
) == false)
1486 if (Writeable
== true && SrcCacheFile
.empty() == false)
1488 FileFd
SCacheF(SrcCacheFile
,FileFd::WriteAtomic
);
1489 if (_error
->PendingError() == true)
1492 fchmod(SCacheF
.Fd(),0644);
1494 // Write out the main data
1495 if (SCacheF
.Write(Map
->Data(),Map
->Size()) == false)
1496 return _error
->Error(_("IO Error saving source cache"));
1499 // Write out the proper header
1500 Gen
.GetCache().HeaderP
->Dirty
= false;
1501 if (SCacheF
.Seek(0) == false ||
1502 SCacheF
.Write(Map
->Data(),sizeof(*Gen
.GetCache().HeaderP
)) == false)
1503 return _error
->Error(_("IO Error saving source cache"));
1504 Gen
.GetCache().HeaderP
->Dirty
= true;
1508 // Build the status cache
1509 if (BuildCache(Gen
,Progress
,CurrentSize
,TotalSize
,
1510 Files
.begin()+EndOfSource
,Files
.end()) == false)
1514 std::clog
<< "Caches are ready for shipping" << std::endl
;
1516 if (_error
->PendingError() == true)
1522 delete Map
.UnGuard();
1523 *OutMap
= new MMap(*CacheF
,0);
1527 *OutMap
= Map
.UnGuard();
1534 // CacheGenerator::MakeOnlyStatusCache - Build only a status files cache/*{{{*/
1535 // ---------------------------------------------------------------------
1537 __deprecated
bool pkgMakeOnlyStatusCache(OpProgress
&Progress
,DynamicMMap
**OutMap
)
1538 { return pkgCacheGenerator::MakeOnlyStatusCache(&Progress
, OutMap
); }
1539 bool pkgCacheGenerator::MakeOnlyStatusCache(OpProgress
*Progress
,DynamicMMap
**OutMap
)
1541 std::vector
<pkgIndexFile
*> Files
;
1542 unsigned long EndOfSource
= Files
.size();
1543 if (_system
->AddStatusFiles(Files
) == false)
1546 SPtr
<DynamicMMap
> Map
= CreateDynamicMMap(NULL
);
1547 unsigned long CurrentSize
= 0;
1548 unsigned long TotalSize
= 0;
1550 TotalSize
= ComputeSize(Files
.begin()+EndOfSource
,Files
.end());
1552 // Build the status cache
1553 if (Progress
!= NULL
)
1554 Progress
->OverallProgress(0,1,1,_("Reading package lists"));
1555 pkgCacheGenerator
Gen(Map
.Get(),Progress
);
1556 if (_error
->PendingError() == true)
1558 if (BuildCache(Gen
,Progress
,CurrentSize
,TotalSize
,
1559 Files
.begin()+EndOfSource
,Files
.end()) == false)
1562 if (_error
->PendingError() == true)
1564 *OutMap
= Map
.UnGuard();
1569 // IsDuplicateDescription /*{{{*/
1570 static bool IsDuplicateDescription(pkgCache::DescIterator Desc
,
1571 MD5SumValue
const &CurMd5
, std::string
const &CurLang
)
1573 // Descriptions in the same link-list have all the same md5
1574 if (Desc
.end() == true || MD5SumValue(Desc
.md5()) != CurMd5
)
1576 for (; Desc
.end() == false; ++Desc
)
1577 if (Desc
.LanguageCode() == CurLang
)
1582 // CacheGenerator::FinishCache /*{{{*/
1583 bool pkgCacheGenerator::FinishCache(OpProgress
*Progress
)