#!/usr/bin/python -tt # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Library General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # Copyright 2005 Duke University import os import os.path import rpm import re import types import errno import time import sre_constants import glob import logging import logging.config from ConfigParser import ParsingError, ConfigParser import Errors import rpmsack import rpmUtils.updates import rpmUtils.arch import rpmUtils.transaction import comps import config import repos import misc from parser import ConfigPreProcessor import transactioninfo import urlgrabber from urlgrabber.grabber import URLGrabError from packageSack import ListPackageSack import depsolve import plugins import logginglevels import yumRepo from packages import parsePackages, YumAvailablePackage, YumLocalPackage, YumInstalledPackage from constants import * __version__ = '3.0.6' class YumBase(depsolve.Depsolve): """This is a primary structure and base class. It houses the objects and methods needed to perform most things in yum. It is almost an abstract class in that you will need to add your own class above it for most real use.""" def __init__(self): depsolve.Depsolve.__init__(self) self.tsInfo = None self.rpmdb = None self.up = None self.comps = None self.pkgSack = None self.logger = logging.getLogger("yum.YumBase") self.verbose_logger = logging.getLogger("yum.verbose.YumBase") self.repos = repos.RepoStorage() # class of repositories # Start with plugins disabled self.disablePlugins() self.localPackages = [] # for local package handling def _transactionDataFactory(self): """Factory method returning TransactionData object""" if self.conf.enable_group_conditionals: return transactioninfo.ConditionalTransactionData() return transactioninfo.TransactionData() def doGenericSetup(self, cache=0): """do a default setup for all the normal/necessary yum components, really just a shorthand for testing""" self.doConfigSetup(init_plugins=False) self.conf.cache = cache self.doTsSetup() self.doRpmDBSetup() self.doRepoSetup() self.doSackSetup() def doConfigSetup(self, fn='/etc/yum/yum.conf', root='/', init_plugins=True, plugin_types=(plugins.TYPE_CORE,), optparser=None, debuglevel=None, errorlevel=None): ''' Parse and load Yum's configuration files and call hooks initialise plugins and logging. @param fn: Path to main configuration file to parse (yum.conf). @param root: Filesystem root to use. @param init_plugins: If False, plugins will not be loaded here. If True, plugins will be loaded if the "plugins" option is enabled in the configuration file. @param plugin_types: As per doPluginSetup() @param optparser: As per doPluginSetup() @param debuglevel: Debug level to use for logging. If None, the debug level will be read from the configuration file. @param errorlevel: Error level to use for logging. If None, the debug level will be read from the configuration file. ''' # TODO: Remove this block when we no longer support configs outside # of /etc/yum/ if fn == '/etc/yum/yum.conf' and not os.path.exists(fn): # Try the old default fn = '/etc/yum.conf' startupconf = config.readStartupConfig(fn, root) if debuglevel != None: startupconf.debuglevel = debuglevel if errorlevel != None: startupconf.errorlevel = errorlevel self.doLoggingSetup(startupconf.debuglevel, startupconf.errorlevel) if init_plugins and startupconf.plugins: self.doPluginSetup(optparser, plugin_types, startupconf.pluginpath, startupconf.pluginconfpath) self.conf = config.readMainConfig(startupconf) self.yumvar = self.conf.yumvar self.getReposFromConfig() # who are we: self.conf.uid = os.geteuid() self.doFileLogSetup(self.conf.uid, self.conf.logfile) self.plugins.run('init') def doLoggingSetup(self, debuglevel, errorlevel): ''' Perform logging related setup. @param debuglevel: Debug logging level to use. @param errorlevel: Error logging level to use. ''' logginglevels.doLoggingSetup(debuglevel, errorlevel) def doFileLogSetup(self, uid, logfile): logginglevels.setFileLog(uid, logfile) def getReposFromConfig(self): """read in repositories from config main and .repo files""" #FIXME this method could be a simpler reposlist = [] # Check yum.conf for repositories for section in self.conf.cfg.sections(): # All sections except [main] are repositories if section == 'main': continue try: thisrepo = self.readRepoConfig(self.conf.cfg, section) except (Errors.RepoError, Errors.ConfigError), e: self.logger.warning(e) else: reposlist.append(thisrepo) # Read .repo files from directories specified by the reposdir option # (typically /etc/yum.repos.d and /etc/yum/repos.d) parser = ConfigParser() for reposdir in self.conf.reposdir: if os.path.exists(self.conf.installroot+'/'+reposdir): reposdir = self.conf.installroot + '/' + reposdir if os.path.isdir(reposdir): for repofn in glob.glob('%s/*.repo' % reposdir): confpp_obj = ConfigPreProcessor(repofn, vars=self.yumvar) try: parser.readfp(confpp_obj) except ParsingError, e: msg = str(e) raise Errors.ConfigError, msg # Check sections in the .repo files that were just slurped up for section in parser.sections(): try: thisrepo = self.readRepoConfig(parser, section) except (Errors.RepoError, Errors.ConfigError), e: self.logger.warning(e) else: reposlist.append(thisrepo) # Got our list of repo objects, add them to the repos collection for thisrepo in reposlist: try: self.repos.add(thisrepo) except Errors.RepoError, e: self.logger.warning(e) continue def readRepoConfig(self, parser, section): '''Parse an INI file section for a repository. @param parser: ConfParser or similar to read INI file values from. @param section: INI file section to read. @return: YumRepository instance. ''' repo = yumRepo.YumRepository(section) repo.populate(parser, section, self.conf) # Ensure that the repo name is set if not repo.name: repo.name = section self.logger.error('Repository %r is missing name in configuration, ' 'using id' % section) # Set attributes not from the config file repo.basecachedir = self.conf.cachedir repo.yumvar.update(self.conf.yumvar) repo.cfg = parser return repo def disablePlugins(self): '''Disable yum plugins ''' self.plugins = plugins.DummyYumPlugins() def doPluginSetup(self, optparser=None, plugin_types=None, searchpath=None, confpath=None): '''Initialise and enable yum plugins. Note: doConfigSetup() will initialise plugins if instructed to. Only call this method directly if not calling doConfigSetup() or calling doConfigSetup(init_plugins=False). @param optparser: The OptionParser instance for this run (optional) @param plugin_types: A sequence specifying the types of plugins to load. This should be sequnce containing one or more of the yum.plugins.TYPE_... constants. If None (the default), all plugins will be loaded. @param searchpath: A list of directories to look in for plugins. A default will be used if no value is specified. @param confpath: A list of directories to look in for plugin configuration files. A default will be used if no value is specified. ''' if isinstance(plugins, plugins.YumPlugins): raise RuntimeError("plugins already initialised") self.plugins = plugins.YumPlugins(self, searchpath, optparser, plugin_types, confpath) def doTsSetup(self): """setup all the transaction set storage items we'll need This can't happen in __init__ b/c we don't know our installroot yet""" if self.tsInfo != None and self.ts != None: return if not self.conf.installroot: raise Errors.YumBaseError, 'Setting up TransactionSets before config class is up' self.tsInfo = self._transactionDataFactory() self.initActionTs() def doRpmDBSetup(self): """sets up a holder object for important information from the rpmdb""" if self.rpmdb is None: self.verbose_logger.debug('Reading Local RPMDB') self.rpmdb = rpmsack.RPMDBPackageSack(root=self.conf.installroot) def closeRpmDB(self): """closes down the instances of the rpmdb we have wangling around""" self.rpmdb = None self.ts = None self.up = None if self.comps != None: self.comps.compiled = False def doRepoSetup(self, thisrepo=None): """grabs the repomd.xml for each enabled repository and sets up the basics of the repository""" self.plugins.run('prereposetup') if thisrepo is None: repos = self.repos.listEnabled() else: repos = self.repos.findRepos(thisrepo) if len(repos) < 1: self.logger.critical('No Repositories Available to Set Up') num = 1 for repo in repos: repo.setup(self.conf.cache) num += 1 if self.repos.callback and len(repos) > 0: self.repos.callback.progressbar(num, len(repos), repo.id) self.plugins.run('postreposetup') def doSackSetup(self, archlist=None, thisrepo=None): """populates the package sacks for information from our repositories, takes optional archlist for archs to include""" if self.pkgSack and thisrepo is None: self.verbose_logger.log(logginglevels.DEBUG_4, 'skipping reposetup, pkgsack exists') return if thisrepo is None: repos = self.repos.listEnabled() else: repos = self.repos.findRepos(thisrepo) self.verbose_logger.debug('Setting up Package Sacks') if not archlist: archlist = rpmUtils.arch.getArchList() archdict = {} for arch in archlist: archdict[arch] = 1 self.repos.getPackageSack().setCompatArchs(archdict) self.repos.populateSack(which=repos) self.pkgSack = self.repos.getPackageSack() self.excludePackages() self.pkgSack.excludeArchs(archlist) for repo in repos: self.excludePackages(repo) self.includePackages(repo) self.plugins.run('exclude') self.pkgSack.buildIndexes() def doUpdateSetup(self): """setups up the update object in the base class and fills out the updates, obsoletes and others lists""" if self.up != None: return self.verbose_logger.debug('Building updates object') #FIXME - add checks for the other pkglists to see if we should # raise an error if self.pkgSack is None: self.doRepoSetup() self.doSackSetup() self.up = rpmUtils.updates.Updates(self.rpmdb.simplePkgList(), self.pkgSack.simplePkgList()) if self.conf.debuglevel >= 6: self.up.debug = 1 if self.conf.obsoletes: self.up.rawobsoletes = self.pkgSack.returnObsoletes(newest=True) self.up.exactarch = self.conf.exactarch self.up.exactarchlist = self.conf.exactarchlist self.up.doUpdates() if self.conf.obsoletes: self.up.doObsoletes() self.up.condenseUpdates() def doGroupSetup(self): """create the groups object that will store the comps metadata finds the repos with groups, gets their comps data and merge it into the group object""" self.verbose_logger.debug('Getting group metadata') reposWithGroups = [] for repo in self.repos.listGroupsEnabled(): self.verbose_logger.debug('repos.listGroupsEnabled is %s' % repo) if repo.groups_added: # already added the groups from this repo reposWithGroups.append(repo) continue if not repo.ready(): raise Errors.RepoError, "Repository '%s' not yet setup" % repo try: groupremote = repo.getGroupLocation() except Errors.RepoMDError, e: pass else: reposWithGroups.append(repo) # now we know which repos actually have groups files. overwrite = self.conf.overwrite_groups if self.comps is None: self.comps = comps.Comps(overwrite_groups = overwrite) for repo in reposWithGroups: # if repo == 'base' : # nodefault = True # else : # nodefault = False if repo.groups_added: # already added the groups from this repo continue self.verbose_logger.debug('Adding group file from repository: %s', repo) groupfile = repo.getGroups() self.verbose_logger.debug('groupfile %s for repo : %s', (groupfile,repo)) try: self.comps.add(groupfile) except Errors.GroupsError, e: self.logger.critical('Failed to add groups file for repository: %s' % repo) else: repo.groups_added = True if self.comps.compscount == 0: raise Errors.GroupsError, 'No Groups Available in any repository' self.doRpmDBSetup() pkglist = self.rpmdb.simplePkgList() self.comps.compile(pkglist) def doSackFilelistPopulate(self): """convenience function to populate the repos with the filelist metadata it also is simply to only emit a log if anything actually gets populated""" necessary = False for repo in self.repos.listEnabled(): if 'filelists' in repo.sack.added[repo]: continue else: necessary = True if necessary: msg = 'Importing additional filelist information' self.verbose_logger.log(logginglevels.INFO_2, msg) self.repos.populateSack(with='filelists') def buildTransaction(self): """go through the packages in the transaction set, find them in the packageSack or rpmdb, and pack up the ts accordingly""" self.plugins.run('preresolve') (rescode, restring) = self.resolveDeps() self.plugins.run('postresolve', rescode=rescode, restring=restring) if self.tsInfo.changed: (rescode, restring) = self.resolveDeps() return rescode, restring def runTransaction(self, cb): """takes an rpm callback object, performs the transaction""" self.plugins.run('pretrans') errors = self.ts.run(cb.callback, '') if errors: raise Errors.YumBaseError, errors if not self.conf.keepcache: self.cleanUsedHeadersPackages() self.plugins.run('posttrans') def excludePackages(self, repo=None): """removes packages from packageSacks based on global exclude lists, command line excludes and per-repository excludes, takes optional repo object to use.""" # if not repo: then assume global excludes, only # if repo: then do only that repos' packages and excludes if not repo: # global only excludelist = self.conf.exclude repoid = None else: excludelist = repo.getExcludePkgList() repoid = repo.id if len(excludelist) == 0: return if not repo: self.verbose_logger.log(logginglevels.INFO_2, 'Excluding Packages in global exclude list') else: self.verbose_logger.log(logginglevels.INFO_2, 'Excluding Packages from %s', repo.name) exactmatch, matched, unmatched = \ parsePackages(self.pkgSack.returnPackages(repoid), excludelist, casematch=1) for po in exactmatch + matched: self.verbose_logger.debug('Excluding %s', po) self.pkgSack.delPackage(po) self.verbose_logger.log(logginglevels.INFO_2, 'Finished') def includePackages(self, repo): """removes packages from packageSacks based on list of packages, to include. takes repoid as a mandatory argument.""" includelist = repo.getIncludePkgList() if len(includelist) == 0: return pkglist = self.pkgSack.returnPackages(repo.id) exactmatch, matched, unmatched = \ parsePackages(pkglist, includelist, casematch=1) self.verbose_logger.log(logginglevels.INFO_2, 'Reducing %s to included packages only', repo.name) rmlist = [] for po in pkglist: if po in exactmatch + matched: self.verbose_logger.debug('Keeping included package %s', po) continue else: rmlist.append(po) for po in rmlist: self.verbose_logger.debug('Removing unmatched package %s', po) self.pkgSack.delPackage(po) self.verbose_logger.log(logginglevels.INFO_2, 'Finished') def doLock(self, lockfile = YUM_PID_FILE): """perform the yum locking, raise yum-based exceptions, not OSErrors""" # if we're not root then we don't lock - just return nicely if self.conf.uid != 0: return root = self.conf.installroot lockfile = root + '/' + lockfile # lock in the chroot lockfile = os.path.normpath(lockfile) # get rid of silly preceding extra / mypid=str(os.getpid()) while not self._lock(lockfile, mypid, 0644): fd = open(lockfile, 'r') try: oldpid = int(fd.readline()) except ValueError: # bogus data in the pid file. Throw away. self._unlock(lockfile) else: if oldpid == os.getpid(): # if we own the lock, we're fine return try: os.kill(oldpid, 0) except OSError, e: if e[0] == errno.ESRCH: # The pid doesn't exist self._unlock(lockfile) else: # Whoa. What the heck happened? msg = 'Unable to check if PID %s is active' % oldpid raise Errors.LockError(1, msg) else: # Another copy seems to be running. msg = 'Existing lock %s: another copy is running. Aborting.' % lockfile raise Errors.LockError(0, msg) def doUnlock(self, lockfile = YUM_PID_FILE): """do the unlock for yum""" # if we're not root then we don't lock - just return nicely if self.conf.uid != 0: return root = self.conf.installroot lockfile = root + '/' + lockfile # lock in the chroot self._unlock(lockfile) def _lock(self, filename, contents='', mode=0777): lockdir = os.path.dirname(filename) try: if not os.path.exists(lockdir): os.makedirs(lockdir, mode=0755) fd = os.open(filename, os.O_EXCL|os.O_CREAT|os.O_WRONLY, mode) except OSError, msg: if not msg.errno == errno.EEXIST: raise msg return 0 else: os.write(fd, contents) os.close(fd) return 1 def _unlock(self, filename): try: os.unlink(filename) except OSError, msg: pass def verifyPkg(self, fo, po, raiseError): """verifies the package is what we expect it to be raiseError = defaults to 0 - if 1 then will raise a URLGrabError if the file does not check out. otherwise it returns false for a failure, true for success""" if type(fo) is types.InstanceType: fo = fo.filename if not po.verifyLocalPkg(): if raiseError: raise URLGrabError(-1, 'Package does not match intended download') else: return False ylp = YumLocalPackage(self.rpmdb.readOnlyTS(), fo) if ylp.pkgtup != po.pkgtup: if raiseError: raise URLGrabError(-1, 'Package does not match intended download') else: return False return True def verifyChecksum(self, fo, checksumType, csum): """Verify the checksum of the file versus the provided checksum""" try: filesum = misc.checksum(checksumType, fo) except Errors.MiscError, e: raise URLGrabError(-3, 'Could not perform checksum') if filesum != csum: raise URLGrabError(-1, 'Package does not match checksum') return 0 def downloadPkgs(self, pkglist, callback=None): """download list of package objects handed to you, output based on callback, raise yum.Errors.YumBaseError on problems""" errors = {} def adderror(po, msg): errors.setdefault(po, []).append(msg) self.plugins.run('predownload', pkglist=pkglist) repo_cached = False remote_pkgs = [] for po in pkglist: if hasattr(po, 'pkgtype') and po.pkgtype == 'local': continue local = po.localPkg() if os.path.exists(local): cursize = os.stat(local)[6] totsize = long(po.size) if not po.verifyLocalPkg(): if po.repo.cache: repo_cached = True adderror(po, 'package fails checksum but caching is ' 'enabled for %s' % po.repo.id) if cursize >= totsize: # otherwise keep it around for regetting os.unlink(local) else: self.verbose_logger.debug("using local copy of %s" %(po,)) continue remote_pkgs.append(po) # caching is enabled and the package # just failed to check out there's no # way to save this, report the error and return if (self.conf.cache or repo_cached) and errors: return errors i = 0 for po in remote_pkgs: i += 1 checkfunc = (self.verifyPkg, (po, 1), {}) dirstat = os.statvfs(po.repo.pkgdir) if (dirstat.f_bavail * dirstat.f_bsize) <= long(po.size): adderror(po, 'Insufficient space in download directory %s ' 'to download' % po.repo.pkgdir) continue try: text = '(%s/%s): %s' % (i, len(remote_pkgs), os.path.basename(po.returnSimple('relativepath'))) mylocal = po.repo.getPackage(po, checkfunc=checkfunc, text=text, cache=po.repo.http_caching != 'none', ) except Errors.RepoError, e: adderror(po, str(e)) else: po.localpath = mylocal if errors.has_key(po): del errors[po] self.plugins.run('postdownload', pkglist=pkglist, errors=errors) return errors def verifyHeader(self, fo, po, raiseError): """check the header out via it's naevr, internally""" if type(fo) is types.InstanceType: fo = fo.filename try: hlist = rpm.readHeaderListFromFile(fo) hdr = hlist[0] except (rpm.error, IndexError): if raiseError: raise URLGrabError(-1, 'Header is not complete.') else: return 0 yip = YumInstalledPackage(hdr) # we're using YumInstalledPackage b/c # it takes headers if yip.pkgtup != po.pkgtup: if raiseError: raise URLGrabError(-1, 'Header does not match intended download') else: return 0 return 1 def downloadHeader(self, po): """download a header from a package object. output based on callback, raise yum.Errors.YumBaseError on problems""" if hasattr(po, 'pkgtype') and po.pkgtype == 'local': return errors = {} local = po.localHdr() repo = self.repos.getRepo(po.repoid) if os.path.exists(local): try: result = self.verifyHeader(local, po, raiseError=1) except URLGrabError, e: # might add a check for length of file - if it is < # required doing a reget try: os.unlink(local) except OSError, e: pass else: po.hdrpath = local return else: if self.conf.cache: raise Errors.RepoError, \ 'Header not in local cache and caching-only mode enabled. Cannot download %s' % po.hdrpath if self.dsCallback: self.dsCallback.downloadHeader(po.name) try: checkfunc = (self.verifyHeader, (po, 1), {}) hdrpath = repo.getHeader(po, checkfunc=checkfunc, cache=repo.http_caching != 'none', ) except Errors.RepoError, e: saved_repo_error = e try: os.unlink(local) except OSError, e: raise Errors.RepoError, saved_repo_error else: raise else: po.hdrpath = hdrpath return def sigCheckPkg(self, po): '''Take a package object and attempt to verify GPG signature if required Returns (result, error_string) where result is 0 - GPG signature verifies ok or verification is not required 1 - GPG verification failed but installation of the right GPG key might help 2 - Fatal GPG verifcation error, give up ''' if hasattr(po, 'pkgtype') and po.pkgtype == 'local': check = self.conf.gpgcheck hasgpgkey = 0 else: repo = self.repos.getRepo(po.repoid) check = repo.gpgcheck hasgpgkey = not not repo.gpgkey if check: ts = self.rpmdb.readOnlyTS() sigresult = rpmUtils.miscutils.checkSig(ts, po.localPkg()) localfn = os.path.basename(po.localPkg()) if sigresult == 0: result = 0 msg = '' elif sigresult == 1: if hasgpgkey: result = 1 else: result = 2 msg = 'Public key for %s is not installed' % localfn elif sigresult == 2: result = 2 msg = 'Problem opening package %s' % localfn elif sigresult == 3: if hasgpgkey: result = 1 else: result = 2 result = 1 msg = 'Public key for %s is not trusted' % localfn elif sigresult == 4: result = 2 msg = 'Package %s is not signed' % localfn else: result =0 msg = '' return result, msg def cleanUsedHeadersPackages(self): filelist = [] for txmbr in self.tsInfo: if txmbr.po.state not in TS_INSTALL_STATES: continue if txmbr.po.repoid == "installed": continue if not self.repos.repos.has_key(txmbr.po.repoid): continue # make sure it's not a local file repo = self.repos.repos[txmbr.po.repoid] local = False for u in repo.baseurl: if u.startswith("file:"): local = True break if local: filelist.extend([txmbr.po.localHdr()]) else: filelist.extend([txmbr.po.localPkg(), txmbr.po.localHdr()]) # now remove them for fn in filelist: try: os.unlink(fn) except OSError, e: self.logger.warning('Cannot remove %s', fn) continue else: self.verbose_logger.log(logginglevels.DEBUG_4, '%s removed', fn) def cleanHeaders(self): filelist = [] ext = 'hdr' removed = 0 for repo in self.repos.listEnabled(): repo.dirSetup() path = repo.hdrdir filelist = misc.getFileList(path, ext, filelist) for hdr in filelist: try: os.unlink(hdr) except OSError, e: self.logger.critical('Cannot remove header %s', hdr) continue else: self.verbose_logger.log(logginglevels.DEBUG_4, 'Header %s removed', hdr) removed+=1 msg = '%d headers removed' % removed return 0, [msg] def cleanPackages(self): filelist = [] ext = 'rpm' removed = 0 for repo in self.repos.listEnabled(): repo.dirSetup() path = repo.pkgdir filelist = misc.getFileList(path, ext, filelist) for pkg in filelist: try: os.unlink(pkg) except OSError, e: self.logger.critical('Cannot remove package %s', pkg) continue else: self.verbose_logger.log(logginglevels.DEBUG_4, 'Package %s removed', pkg) removed+=1 msg = '%d packages removed' % removed return 0, [msg] def cleanSqlite(self): filelist = [] ext = 'sqlite' removed = 0 for repo in self.repos.listEnabled(): repo.dirSetup() path = repo.cachedir filelist = misc.getFileList(path, ext, filelist) for item in filelist: try: os.unlink(item) except OSError, e: self.logger.critical('Cannot remove sqlite cache file %s', item) continue else: self.verbose_logger.log(logginglevels.DEBUG_4, 'Cache file %s removed', item) removed+=1 msg = '%d cache files removed' % removed return 0, [msg] def cleanMetadata(self): filelist = [] exts = ['xml.gz', 'xml', 'cachecookie'] removed = 0 for ext in exts: for repo in self.repos.listEnabled(): repo.dirSetup() path = repo.cachedir filelist = misc.getFileList(path, ext, filelist) for item in filelist: try: os.unlink(item) except OSError, e: self.logger.critical('Cannot remove metadata file %s', item) continue else: self.verbose_logger.log(logginglevels.DEBUG_4, 'metadata file %s removed', item) removed+=1 msg = '%d metadata files removed' % removed return 0, [msg] def doPackageLists(self, pkgnarrow='all'): """generates lists of packages, un-reduced, based on pkgnarrow option""" ygh = misc.GenericHolder() installed = [] available = [] updates = [] obsoletes = [] obsoletesTuples = [] recent = [] extras = [] # list all packages - those installed and available, don't 'think about it' if pkgnarrow == 'all': self.doRepoSetup() self.doRpmDBSetup() inst = self.rpmdb.simplePkgList() for po in self.rpmdb: installed.append(po) if self.conf.showdupesfromrepos: avail = self.pkgSack.returnPackages() else: avail = self.pkgSack.returnNewestByNameArch() for pkg in avail: pkgtup = (pkg.name, pkg.arch, pkg.epoch, pkg.version, pkg.release) if pkgtup not in inst: available.append(pkg) # produce the updates list of tuples elif pkgnarrow == 'updates': self.doRepoSetup() self.doRpmDBSetup() self.doUpdateSetup() for (n,a,e,v,r) in self.up.getUpdatesList(): matches = self.pkgSack.searchNevra(name=n, arch=a, epoch=e, ver=v, rel=r) if len(matches) > 1: updates.append(matches[0]) self.verbose_logger.log(logginglevels.DEBUG_1, 'More than one identical match in sack for %s', matches[0]) elif len(matches) == 1: updates.append(matches[0]) else: self.verbose_logger.log(logginglevels.DEBUG_1, 'Nothing matches %s.%s %s:%s-%s from update', n,a,e,v,r) # installed only elif pkgnarrow == 'installed': self.doRpmDBSetup() for po in self.rpmdb: installed.append(po) # available in a repository elif pkgnarrow == 'available': self.doRepoSetup() self.doRpmDBSetup() inst = self.rpmdb.simplePkgList() if self.conf.showdupesfromrepos: avail = self.pkgSack.returnPackages() else: avail = self.pkgSack.returnNewestByNameArch() for pkg in avail: pkgtup = (pkg.name, pkg.arch, pkg.epoch, pkg.version, pkg.release) if pkgtup not in inst: available.append(pkg) # not in a repo but installed elif pkgnarrow == 'extras': # we must compare the installed set versus the repo set # anything installed but not in a repo is an extra self.doRepoSetup() self.doRpmDBSetup() avail = self.pkgSack.simplePkgList() for po in self.rpmdb: if po.pkgtup not in avail: extras.append(po) # obsoleting packages (and what they obsolete) elif pkgnarrow == 'obsoletes': self.doRepoSetup() self.doRpmDBSetup() self.conf.obsoletes = 1 self.doUpdateSetup() for (pkgtup, instTup) in self.up.getObsoletesTuples(): (n,a,e,v,r) = pkgtup pkgs = self.pkgSack.searchNevra(name=n, arch=a, ver=v, rel=r, epoch=e) instpo = self.rpmdb.searchPkgTuple(instTup)[0] # the first one for po in pkgs: obsoletes.append(po) obsoletesTuples.append((po, instpo)) # packages recently added to the repositories elif pkgnarrow == 'recent': now = time.time() recentlimit = now-(self.conf.recent*86400) ftimehash = {} self.doRepoSetup() if self.conf.showdupesfromrepos: avail = self.pkgSack.returnPackages() else: avail = self.pkgSack.returnNewestByNameArch() for po in avail: ftime = int(po.returnSimple('filetime')) if ftime > recentlimit: if not ftimehash.has_key(ftime): ftimehash[ftime] = [po] else: ftimehash[ftime].append(po) for sometime in ftimehash.keys(): for po in ftimehash[sometime]: recent.append(po) ygh.installed = installed ygh.available = available ygh.updates = updates ygh.obsoletes = obsoletes ygh.obsoletesTuples = obsoletesTuples ygh.recent = recent ygh.extras = extras return ygh def findDeps(self, pkgs): """Return the dependencies for a given package object list, as well possible solutions for those dependencies. Returns the deps as a dict of dicts: packageobject = [reqs] = [list of satisfying pkgs]""" results = {} self.doRepoSetup() for pkg in pkgs: results[pkg] = {} reqs = pkg.requires reqs.sort() pkgresults = results[pkg] # shorthand so we don't have to do the # double bracket thing for req in reqs: (r,f,v) = req if r.startswith('rpmlib('): continue satisfiers = [] for po in self.whatProvides(r, f, v): satisfiers.append(po) pkgresults[req] = satisfiers return results def searchGenerator(self, fields, criteria): """Generator method to lighten memory load for some searches. This is the preferred search function to use.""" self.doRepoSetup() self.doRpmDBSetup() for string in criteria: restring = misc.refineSearchPattern(string) try: crit_re = re.compile(restring, flags=re.I) except sre_constants.error, e: raise Errors.MiscError, \ 'Search Expression: %s is an invalid Regular Expression.\n' % string for sack in self.pkgSack, self.rpmdb: for po in sack: tmpvalues = [] for field in fields: value = po.returnSimple(field) if value and crit_re.search(value): tmpvalues.append(value) if len(tmpvalues) > 0: yield (po, tmpvalues) def searchPackages(self, fields, criteria, callback=None): """Search specified fields for matches to criteria optional callback specified to print out results as you go. Callback is a simple function of: callback(po, matched values list). It will just return a dict of dict[po]=matched values list""" matches = {} match_gen = self.searchGenerator(fields, criteria) for (po, matched_strings) in match_gen: if callback: callback(po, matched_strings) if not matches.has_key(po): matches[po] = [] matches[po].extend(matched_strings) return matches def searchPackageProvides(self, args, callback=None): self.doRepoSetup() matches = {} # search deps the simple way first for arg in args: self.verbose_logger.log(logginglevels.DEBUG_1, 'searching the simple way') pkgs = self.returnPackagesByDep(arg) for po in pkgs: if callback: callback(po, [arg]) matches[po] = [arg] # search pkgSack - fully populate the worthwhile metadata to search # if it even vaguely matches self.verbose_logger.log(logginglevels.DEBUG_1, 'fully populating the necessary data') for arg in args: matched = 0 globs = ['.*bin\/.*', '.*\/etc\/.*', '^\/usr\/lib\/sendmail$'] for glob in globs: globc = re.compile(glob) if globc.match(arg): matched = 1 if not matched: self.doSackFilelistPopulate() for arg in args: # assume we have to search every package, unless we can refine the search set where = self.pkgSack # this is annoying. If the user doesn't use any glob or regex-like # or regexes then we can use the where 'like' search in sqlite # if they do use globs or regexes then we can't b/c the string # will no longer have much meaning to use it for matches if hasattr(self.pkgSack, 'searchAll'): if not re.match('.*[\*,\[,\],\{,\},\?,\+,\%].*', arg): self.verbose_logger.log(logginglevels.DEBUG_1, 'Using the like search') where = self.pkgSack.searchAll(arg, query_type='like') self.verbose_logger.log(logginglevels.DEBUG_1, 'Searching %d packages', len(where)) self.verbose_logger.log(logginglevels.DEBUG_1, 'refining the search expression of %s', arg) restring = misc.refineSearchPattern(arg) self.verbose_logger.log(logginglevels.DEBUG_1, 'refined search: %s', restring) try: arg_re = re.compile(restring, flags=re.I) except sre_constants.error, e: raise Errors.MiscError, \ 'Search Expression: %s is an invalid Regular Expression.\n' % arg for po in where: self.verbose_logger.log(logginglevels.DEBUG_2, 'searching package %s', po) tmpvalues = [] self.verbose_logger.log(logginglevels.DEBUG_2, 'searching in file entries') for filetype in po.returnFileTypes(): for fn in po.returnFileEntries(ftype=filetype): if arg_re.search(fn): tmpvalues.append(fn) self.verbose_logger.log(logginglevels.DEBUG_2, 'searching in provides entries') for (p_name, p_flag, (p_e, p_v, p_r)) in po.provides: if arg_re.search(p_name): prov = misc.prco_tuple_to_string((p_name, p_flag, (p_e, p_v, p_r))) tmpvalues.append(prov) if len(tmpvalues) > 0: if callback: callback(po, tmpvalues) matches[po] = tmpvalues self.doRpmDBSetup() # installed rpms, too taglist = ['filelist', 'dirnames', 'provides_names'] arg_re = [] for arg in args: restring = misc.refineSearchPattern(arg) try: reg = re.compile(restring, flags=re.I) except sre_constants.error, e: raise Errors.MiscError, \ 'Search Expression: %s is an invalid Regular Expression.\n' % arg for po in self.rpmdb: tmpvalues = [] searchlist = [] for tag in taglist: tagdata = po.returnSimple(tag) if tagdata is None: continue if type(tagdata) is types.ListType: searchlist.extend(tagdata) else: searchlist.append(tagdata) for item in searchlist: if reg.search(item): tmpvalues.append(item) del searchlist if len(tmpvalues) > 0: if callback: callback(po, tmpvalues) matches[po] = tmpvalues return matches def doGroupLists(self, uservisible=0): """returns two lists of groups, installed groups and available groups optional 'uservisible' bool to tell it whether or not to return only groups marked as uservisible""" installed = [] available = [] for grp in self.comps.groups: if grp.installed: if uservisible: if grp.user_visible: installed.append(grp) else: installed.append(grp) else: if uservisible: if grp.user_visible: available.append(grp) else: available.append(grp) return installed, available def groupRemove(self, grpid): """mark all the packages in this group to be removed""" txmbrs_used = [] self.doGroupSetup() thisgroup = self.comps.return_group(grpid) if not thisgroup: raise Errors.GroupsError, "No Group named %s exists" % grpid thisgroup.toremove = True pkgs = thisgroup.packages for pkg in thisgroup.packages: txmbrs = self.remove(name=pkg) txmbrs_used.extend(txmbrs) for txmbr in txmbrs: txmbr.groups.append(thisgroup.groupid) return txmbrs_used def groupUnremove(self, grpid): """unmark any packages in the group from being removed""" self.doGroupSetup() thisgroup = self.comps.return_group(grpid) if not thisgroup: raise Errors.GroupsError, "No Group named %s exists" % grpid thisgroup.toremove = False pkgs = thisgroup.packages for pkg in thisgroup.packages: for txmbr in self.tsInfo: if txmbr.po.name == pkg and txmbr.po.state in TS_INSTALL_STATES: try: txmbr.groups.remove(grpid) except ValueError: self.verbose_logger.log(logginglevels.DEBUG_1, "package %s was not marked in group %s", txmbr.po, grpid) continue # if there aren't any other groups mentioned then remove the pkg if len(txmbr.groups) == 0: self.tsInfo.remove(txmbr.po.pkgtup) def selectGroup(self, grpid): """mark all the packages in the group to be installed returns a list of transaction members it added to the transaction set""" txmbrs_used = [] if not self.comps: self.doGroupSetup() if not self.comps.has_group(grpid): raise Errors.GroupsError, "No Group named %s exists" % grpid thisgroup = self.comps.return_group(grpid) if not thisgroup: raise Errors.GroupsError, "No Group named %s exists" % grpid if thisgroup.selected: return txmbrs_used thisgroup.selected = True pkgs = [] if 'mandatory' in self.conf.group_package_types: pkgs.extend(thisgroup.mandatory_packages.keys()) if 'default' in self.conf.group_package_types: pkgs.extend(thisgroup.default_packages.keys()) if 'optional' in self.conf.group_package_types: pkgs.extend(thisgroup.optional_packages.keys()) for pkg in pkgs: self.verbose_logger.log(logginglevels.DEBUG_2, 'Adding package %s from group %s', pkg, thisgroup.groupid) try: txmbrs = self.install(name = pkg) except Errors.InstallError, e: self.verbose_logger.debug('No package named %s available to be installed', pkg) else: txmbrs_used.extend(txmbrs) for txmbr in txmbrs: txmbr.groups.append(thisgroup.groupid) if self.conf.enable_group_conditionals: for condreq, cond in thisgroup.conditional_packages.iteritems(): if self._isPackageInstalled(cond): try: txmbrs = self.install(name = condreq) except Errors.InstallError: # we don't care if the package doesn't exist continue txmbrs_used.extend(txmbrs) for txmbr in txmbrs: txmbr.groups.append(thisgroup.groupid) continue # Otherwise we hook into tsInfo.add pkgs = self.pkgSack.searchNevra(name=condreq) if pkgs: pkgs = self.bestPackagesFromList(pkgs) if self.tsInfo.conditionals.has_key(cond): self.tsInfo.conditionals[cond].extend(pkgs) else: self.tsInfo.conditionals[cond] = pkgs return txmbrs_used def deselectGroup(self, grpid): """de-mark all the packages in the group for install""" if not self.comps: self.doGroupSetup() if not self.comps.has_group(grpid): raise Errors.GroupsError, "No Group named %s exists" % grpid thisgroup = self.comps.return_group(grpid) if not thisgroup: raise Errors.GroupsError, "No Group named %s exists" % grpid thisgroup.selected = False for pkgname in thisgroup.packages: for txmbr in self.tsInfo: if txmbr.po.name == pkgname and txmbr.po.state in TS_INSTALL_STATES: try: txmbr.groups.remove(grpid) except ValueError: self.verbose_logger.log(logginglevels.DEBUG_1, "package %s was not marked in group %s", txmbr.po, grpid) continue # if there aren't any other groups mentioned then remove the pkg if len(txmbr.groups) == 0: self.tsInfo.remove(txmbr.po.pkgtup) def getPackageObject(self, pkgtup): """retrieves a packageObject from a pkgtuple - if we need to pick and choose which one is best we better call out to some method from here to pick the best pkgobj if there are more than one response - right now it's more rudimentary.""" # look it up in the self.localPackages first: for po in self.localPackages: if po.pkgtup == pkgtup: return po pkgs = self.pkgSack.searchPkgTuple(pkgtup) if len(pkgs) == 0: raise Errors.DepError, 'Package tuple %s could not be found in packagesack' % str(pkgtup) return None if len(pkgs) > 1: # boy it'd be nice to do something smarter here FIXME result = pkgs[0] else: result = pkgs[0] # which should be the only # this is where we could do something to figure out which repository # is the best one to pull from return result def getInstalledPackageObject(self, pkgtup): """returns a YumInstallPackage object for the pkgtup specified""" #FIXME - this should probably emit a deprecation warning telling # people to just use the command below po = self.rpmdb.searchPkgTuple(pkgtup)[0] # take the first one return po def gpgKeyCheck(self): """checks for the presence of gpg keys in the rpmdb returns 0 if no keys returns 1 if keys""" gpgkeyschecked = self.conf.cachedir + '/.gpgkeyschecked.yum' if os.path.exists(gpgkeyschecked): return 1 myts = rpmUtils.transaction.initReadOnlyTransaction(root=self.conf.installroot) myts.pushVSFlags(~(rpm._RPMVSF_NOSIGNATURES|rpm._RPMVSF_NODIGESTS)) idx = myts.dbMatch('name', 'gpg-pubkey') keys = idx.count() del idx del myts if keys == 0: return 0 else: mydir = os.path.dirname(gpgkeyschecked) if not os.path.exists(mydir): os.makedirs(mydir) fo = open(gpgkeyschecked, 'w') fo.close() del fo return 1 def returnPackagesByDep(self, depstring): """Pass in a generic [build]require string and this function will pass back the packages it finds providing that dep.""" results = [] self.doRepoSetup() # parse the string out # either it is 'dep (some operator) e:v-r' # or /file/dep # or packagename depname = depstring depflags = None depver = None if depstring[0] != '/': # not a file dep - look at it for being versioned if re.search('[>=<]', depstring): # versioned try: depname, flagsymbol, depver = depstring.split() except ValueError, e: raise Errors.YumBaseError, 'Invalid versioned dependency string, try quoting it.' if not SYMBOLFLAGS.has_key(flagsymbol): raise Errors.YumBaseError, 'Invalid version flag' depflags = SYMBOLFLAGS[flagsymbol] sack = self.whatProvides(depname, depflags, depver) results = sack.returnPackages() return results def returnPackageByDep(self, depstring): """Pass in a generic [build]require string and this function will pass back the best(or first) package it finds providing that dep.""" try: pkglist = self.returnPackagesByDep(depstring) except Errors.YumBaseError: raise Errors.YumBaseError, 'No Package found for %s' % depstring result = self._bestPackageFromList(pkglist) if result is None: raise Errors.YumBaseError, 'No Package found for %s' % depstring return result def returnInstalledPackagesByDep(self, depstring): """Pass in a generic [build]require string and this function will pass back the installed packages it finds providing that dep.""" results = [] self.doRpmDBSetup() # parse the string out # either it is 'dep (some operator) e:v-r' # or /file/dep # or packagename depname = depstring depflags = None depver = None if depstring[0] != '/': # not a file dep - look at it for being versioned if re.search('[>=<]', depstring): # versioned try: depname, flagsymbol, depver = depstring.split() except ValueError: raise Errors.YumBaseError, 'Invalid versioned dependency string, try quoting it.' if not SYMBOLFLAGS.has_key(flagsymbol): raise Errors.YumBaseError, 'Invalid version flag' depflags = SYMBOLFLAGS[flagsymbol] pkglist = self.rpmdb.whatProvides(depname, depflags, depver) for pkgtup in pkglist: results.append(self.getInstalledPackageObject(pkgtup)) return results def _bestPackageFromList(self, pkglist): """take list of package objects and return the best package object. If the list is empty, return None. Note: this is not aware of multilib so make sure you're only passing it packages of a single arch group.""" if len(pkglist) == 0: return None if len(pkglist) == 1: return pkglist[0] mysack = ListPackageSack() mysack.addList(pkglist) bestlist = mysack.returnNewestByNameArch() # get rid of all lesser vers best = bestlist[0] for pkg in bestlist[1:]: if len(pkg.name) < len(best.name): # shortest name silliness best = pkg continue elif len(pkg.name) > len(best.name): continue # compare arch arch = rpmUtils.arch.getBestArchFromList([pkg.arch, best.arch]) if arch == pkg.arch: best = pkg continue return best def bestPackagesFromList(self, pkglist, arch=None): """Takes a list of packages, returns the best packages. This function is multilib aware so that it will not compare multilib to singlelib packages""" returnlist = [] compatArchList = rpmUtils.arch.getArchList(arch) multiLib = [] singleLib = [] noarch = [] for po in pkglist: if po.arch not in compatArchList: continue elif po.arch in ("noarch"): noarch.append(po) elif rpmUtils.arch.isMultiLibArch(arch=po.arch): multiLib.append(po) else: singleLib.append(po) # we now have three lists. find the best package(s) of each multi = self._bestPackageFromList(multiLib) single = self._bestPackageFromList(singleLib) no = self._bestPackageFromList(noarch) # now, to figure out which arches we actually want # if there aren't noarch packages, it's easy. multi + single if no is None: if multi: returnlist.append(multi) if single: returnlist.append(single) # if there's a noarch and it's newer than the multilib, we want # just the noarch. otherwise, we want multi + single elif multi: best = self._bestPackageFromList([multi,no]) if best.arch == "noarch": returnlist.append(no) else: if multi: returnlist.append(multi) if single: returnlist.append(single) # similar for the non-multilib case elif single: best = self._bestPackageFromList([single,no]) if best.arch == "noarch": returnlist.append(no) else: returnlist.append(single) # if there's not a multi or single lib, then we want the noarch else: returnlist.append(no) return returnlist def install(self, po=None, **kwargs): """try to mark for install the item specified. Uses provided package object, if available. If not it uses the kwargs and gets the best packages from the keyword options provided returns the list of txmbr of the items it installs """ self.doRepoSetup() self.doSackSetup() self.doRpmDBSetup() self.doUpdateSetup() pkgs = [] if po: if isinstance(po, YumAvailablePackage) or isinstance(po, YumLocalPackage): pkgs.append(po) else: raise Errors.InstallError, 'Package Object was not a package object instance' else: if not kwargs.keys(): raise Errors.InstallError, 'Nothing specified to install' if kwargs.has_key('pattern'): exactmatch, matched, unmatched = \ parsePackages(self.pkgSack.returnPackages(),[kwargs['pattern']] , casematch=1) pkgs.extend(exactmatch) pkgs.extend(matched) else: nevra_dict = self._nevra_kwarg_parse(kwargs) pkgs = self.pkgSack.searchNevra(name=nevra_dict['name'], epoch=nevra_dict['epoch'], arch=nevra_dict['arch'], ver=nevra_dict['version'], rel=nevra_dict['release']) if pkgs: pkgSack = ListPackageSack(pkgs) pkgs = pkgSack.returnNewestByName() del(pkgSack) lst = [] lst.extend(self.bestPackagesFromList(pkgs)) pkgs = lst if len(pkgs) == 0: #FIXME - this is where we could check to see if it already installed # for returning better errors raise Errors.InstallError, 'No package(s) available to install' # FIXME - lots more checking here # - install instead of erase # - better error handling/reporting tx_return = [] for po in pkgs: if self.tsInfo.exists(pkgtup=po.pkgtup): self.verbose_logger.log(logginglevels.DEBUG_1, 'Package: %s - already in transaction set', po) tx_return.extend(self.tsInfo.getMembers(pkgtup=po.pkgtup)) continue # make sure this shouldn't be passed to update: if self.up.updating_dict.has_key(po.pkgtup): txmbrs = self.update(po=po) tx_return.extend(txmbrs) continue # make sure it's not already installed if self.rpmdb.installed(name=po.name, arch=po.arch, epoch=po.epoch, rel=po.rel, ver=po.ver): self.logger.warning('Package %s already installed and latest version', po) continue # make sure we're not installing a package which is obsoleted by something # else in the repo thispkgobsdict = self.up.checkForObsolete([po.pkgtup]) if thispkgobsdict.has_key(po.pkgtup): obsoleting = thispkgobsdict[po.pkgtup][0] obsoleting_pkg = self.getPackageObject(obsoleting) self.install(po=obsoleting_pkg) continue txmbr = self.tsInfo.addInstall(po) tx_return.append(txmbr) return tx_return def update(self, po=None, **kwargs): """try to mark for update the item(s) specified. po is a package object - if that is there, mark it for update, if possible else use **kwargs to match the package needing update if nothing is specified at all then attempt to update everything returns the list of txmbr of the items it marked for update""" # do updates list # do obsoletes list # check for args - if no po nor kwargs, do them all # if po, do it, ignore all else # if no po do kwargs # uninstalled pkgs called for update get returned with errors in a list, maybe? self.doRepoSetup() self.doSackSetup() self.doRpmDBSetup() self.doUpdateSetup() updates = self.up.getUpdatesTuples() if self.conf.obsoletes: obsoletes = self.up.getObsoletesTuples(newest=1) else: obsoletes = [] tx_return = [] if not po and not kwargs.keys(): # update everything (the easy case) self.verbose_logger.log(logginglevels.DEBUG_2, 'Updating Everything') for (obsoleting, installed) in obsoletes: obsoleting_pkg = self.getPackageObject(obsoleting) installed_pkg = self.rpmdb.searchPkgTuple(installed)[0] txmbr = self.tsInfo.addObsoleting(obsoleting_pkg, installed_pkg) self.tsInfo.addObsoleted(installed_pkg, obsoleting_pkg) tx_return.append(txmbr) for (new, old) in updates: if self.tsInfo.isObsoleted(pkgtup=old): self.verbose_logger.log(logginglevels.DEBUG_2, 'Not Updating Package that is already obsoleted: %s.%s %s:%s-%s', old) else: updating_pkg = self.getPackageObject(new) updated_pkg = self.rpmdb.searchPkgTuple(old)[0] txmbr = self.tsInfo.addUpdate(updating_pkg, updated_pkg) tx_return.append(txmbr) return tx_return else: instpkgs = [] availpkgs = [] if po: # just a po if po.repoid == 'installed': instpkgs.append(po) else: availpkgs.append(po) else: # we have kwargs, sort them out. nevra_dict = self._nevra_kwarg_parse(kwargs) availpkgs = self.pkgSack.searchNevra(name=nevra_dict['name'], epoch=nevra_dict['epoch'], arch=nevra_dict['arch'], ver=nevra_dict['version'], rel=nevra_dict['release']) instpkgs = self.rpmdb.searchNevra(name=nevra_dict['name'], epoch=nevra_dict['epoch'], arch=nevra_dict['arch'], ver=nevra_dict['version'], rel=nevra_dict['release']) # for any thing specified # get the list of available pkgs matching it (or take the po) # get the list of installed pkgs matching it (or take the po) # go through each list and look for: # things obsoleting it if it is an installed pkg # things it updates if it is an available pkg # things updating it if it is an installed pkg # in that order # all along checking to make sure we: # don't update something that's already been obsoleted # TODO: we should search the updates and obsoletes list and # mark the package being updated or obsoleted away appropriately # and the package relationship in the tsInfo for installed_pkg in instpkgs: if self.up.obsoleted_dict.has_key(installed_pkg.pkgtup) and self.conf.obsoletes: obsoleting = self.up.obsoleted_dict[installed_pkg.pkgtup][0] obsoleting_pkg = self.getPackageObject(obsoleting) # FIXME check for what might be in there here txmbr = self.tsInfo.addObsoleting(obsoleting_pkg, installed_pkg) self.tsInfo.addObsoleted(installed_pkg, obsoleting_pkg) tx_return.append(txmbr) for available_pkg in availpkgs: if self.up.updating_dict.has_key(available_pkg.pkgtup): updated = self.up.updating_dict[available_pkg.pkgtup][0] if self.tsInfo.isObsoleted(updated): self.verbose_logger.log(logginglevels.DEBUG_2, 'Not Updating Package that is already obsoleted: %s.%s %s:%s-%s', updated) else: updated_pkg = self.rpmdb.searchPkgTuple(updated)[0] txmbr = self.tsInfo.addUpdate(available_pkg, updated_pkg) tx_return.append(txmbr) for installed_pkg in instpkgs: if self.up.updatesdict.has_key(installed_pkg.pkgtup): updating = self.up.updatesdict[installed_pkg.pkgtup][0] updating_pkg = self.getPackageObject(updating) if self.tsInfo.isObsoleted(installed_pkg.pkgtup): self.verbose_logger.log(logginglevels.DEBUG_2, 'Not Updating Package that is already obsoleted: %s.%s %s:%s-%s', installed_pkg.pkgtup) else: txmbr = self.tsInfo.addUpdate(updating_pkg, installed_pkg) tx_return.append(txmbr) return tx_return def remove(self, po=None, **kwargs): """try to find and mark for remove the specified package(s) - if po is specified then that package object (if it is installed) will be marked for removal. if no po then look at kwargs, if neither then raise an exception""" if not po and not kwargs.keys(): raise Errors.RemoveError, 'Nothing specified to remove' self.doRpmDBSetup() tx_return = [] pkgs = [] if po: pkgs = [po] else: nevra_dict = self._nevra_kwarg_parse(kwargs) pkgs = self.rpmdb.searchNevra(name=nevra_dict['name'], epoch=nevra_dict['epoch'], arch=nevra_dict['arch'], ver=nevra_dict['version'], rel=nevra_dict['release']) if len(pkgs) == 0: # should this even be happening? self.logger.warning("No package matched to remove") for po in pkgs: txmbr = self.tsInfo.addErase(po) tx_return.append(txmbr) return tx_return def _nevra_kwarg_parse(self, kwargs): returndict = {} try: returndict['name'] = kwargs['name'] except KeyError: returndict['name'] = None try: returndict['epoch'] = kwargs['epoch'] except KeyError: returndict['epoch'] = None try: returndict['arch'] = kwargs['arch'] except KeyError: returndict['arch'] = None # get them as ver, version and rel, release - if someone # specifies one of each then that's kinda silly. try: returndict['version'] = kwargs['version'] except KeyError: returndict['version'] = None if returndict['version'] is None: try: returndict['version'] = kwargs['ver'] except KeyError: returndict['version'] = None try: returndict['release'] = kwargs['release'] except KeyError: returndict['release'] = None if returndict['release'] is None: try: release = kwargs['rel'] except KeyError: returndict['release'] = None return returndict def _isPackageInstalled(self, pkgname): # FIXME: Taken from anaconda/pirut # clean up and make public installed = False if self.rpmdb.installed(name = pkgname): installed = True lst = self.tsInfo.matchNaevr(name = pkgname) for txmbr in lst: if txmbr.output_state in TS_INSTALL_STATES: return True if installed and len(lst) > 0: # if we get here, then it was installed, but it's in the tsInfo # for an erase or obsoleted --> not going to be installed at end return False return installed def getKeyForPackage(self, po, askcb = None, fullaskcb = None): """Retrieve a key for a package. If needed, prompt for if the key should be imported using askcb. @po: Package object to retrieve the key of. @askcb: Callback function to use for asking for verification. Takes arguments of the po, the userid for the key, and the keyid. @fullaskcb: Callback function to use for asking for verification of a key. Differs from askcb in that it gets passed a dictionary so that we can expand the values passed. """ repo = self.repos.getRepo(po.repoid) keyurls = repo.gpgkey key_installed = False ts = rpmUtils.transaction.TransactionWrapper(self.conf.installroot) for keyurl in keyurls: self.logger.info('Retrieving GPG key from %s' % keyurl) # Go get the GPG key from the given URL try: rawkey = urlgrabber.urlread(keyurl, limit=9999) except urlgrabber.grabber.URLGrabError, e: raise Errors.YumBaseError('GPG key retrieval failed: ' + str(e)) # Parse the key try: keyinfo = misc.getgpgkeyinfo(rawkey) keyid = keyinfo['keyid'] hexkeyid = misc.keyIdToRPMVer(keyid).upper() timestamp = keyinfo['timestamp'] userid = keyinfo['userid'] except ValueError, e: raise Errors.YumBaseError, \ 'GPG key parsing failed: ' + str(e) # Check if key is already installed if misc.keyInstalled(ts, keyid, timestamp) >= 0: self.logger.info('GPG key at %s (0x%s) is already installed' % ( keyurl, hexkeyid)) continue # Try installing/updating GPG key self.logger.critical('Importing GPG key 0x%s "%s" from %s' % (hexkeyid, userid, keyurl.replace("file://",""))) rc = False if self.conf.assumeyes: rc = True elif fullaskcb: rc = fullaskcb({"po": po, "userid": userid, "hexkeyid": hexkeyid, "keyurl": keyurl}) elif askcb: rc = askcb(po, userid, hexkeyid) if not rc: raise Errors.YumBaseError, "Not installing key" # Import the key result = ts.pgpImportPubkey(misc.procgpgkey(rawkey)) if result != 0: raise Errors.YumBaseError, \ 'Key import failed (code %d)' % result self.logger.info('Key imported successfully') key_installed = True if not key_installed: raise Errors.YumBaseError, \ 'The GPG keys listed for the "%s" repository are ' \ 'already installed but they are not correct for this ' \ 'package.\n' \ 'Check that the correct key URLs are configured for ' \ 'this repository.' % (repo.name) # Check if the newly installed keys helped result, errmsg = self.sigCheckPkg(po) if result != 0: self.logger.info("Import of key(s) didn't help, wrong key(s)?") raise Errors.YumBaseError, errmsg