# Copyright (C) 2003 - 2005 Michael 'Mickey' Lauer
# Copyright (C) 2005 Holger Hans Peter Freyther
# Copyright (C) 2005 ROAD GmbH
+# Copyright (C) 2006 Richard Purdie
#
# This program is free software; you can redistribute it and/or modify it under
# the terms of the GNU General Public License as published by the Free Software
import sys, os, getopt, glob, copy, os.path, re, time
sys.path.insert(0,os.path.join(os.path.dirname(os.path.dirname(sys.argv[0])), 'lib'))
import bb
-from bb import utils, data, parse, event, cache, providers
+from bb import utils, data, parse, event, cache, providers, taskdata, runqueue
from sets import Set
import itertools, optparse
parsespin = itertools.cycle( r'|/-\\' )
-__version__ = "1.7.0"
-
-#============================================================================#
-# BBParsingStatus
-#============================================================================#
-class BBParsingStatus:
- """
- The initial idea for this status class is to use the data when it is
- already loaded instead of loading it from various place over and over
- again.
- """
-
- def __init__(self):
- self.providers = {}
- self.rproviders = {}
- self.packages = {}
- self.packages_dynamic = {}
- self.bbfile_priority = {}
- self.bbfile_config_priorities = []
- self.ignored_dependencies = None
- self.possible_world = []
- self.world_target = Set()
- self.pkg_pn = {}
- self.pkg_fn = {}
- self.pkg_pvpr = {}
- self.pkg_dp = {}
- self.pn_provides = {}
- self.all_depends = Set()
- self.build_all = {}
- self.deps = {}
- self.rundeps = {}
- self.runrecs = {}
- self.task_queues = {}
- self.task_deps = {}
- self.stamp = {}
- self.preferred = {}
-
- def handle_bb_data(self, file_name, bb_cache, cached):
- """
- We will fill the dictionaries with the stuff we
- need for building the tree more fast
- """
-
- pn = bb_cache.getVar('PN', file_name, True)
- pv = bb_cache.getVar('PV', file_name, True)
- pr = bb_cache.getVar('PR', file_name, True)
- dp = int(bb_cache.getVar('DEFAULT_PREFERENCE', file_name, True) or "0")
- provides = Set([pn] + (bb_cache.getVar("PROVIDES", file_name, True) or "").split())
- depends = (bb_cache.getVar("DEPENDS", file_name, True) or "").split()
- packages = (bb_cache.getVar('PACKAGES', file_name, True) or "").split()
- packages_dynamic = (bb_cache.getVar('PACKAGES_DYNAMIC', file_name, True) or "").split()
- rprovides = (bb_cache.getVar("RPROVIDES", file_name, True) or "").split()
-
- self.task_queues[file_name] = bb_cache.getVar("_task_graph", file_name, True)
- self.task_deps[file_name] = bb_cache.getVar("_task_deps", file_name, True)
-
- # build PackageName to FileName lookup table
- if pn not in self.pkg_pn:
- self.pkg_pn[pn] = []
- self.pkg_pn[pn].append(file_name)
-
- self.build_all[file_name] = int(bb_cache.getVar('BUILD_ALL_DEPS', file_name, True) or "0")
- self.stamp[file_name] = bb_cache.getVar('STAMP', file_name, True)
-
- # build FileName to PackageName lookup table
- self.pkg_fn[file_name] = pn
- self.pkg_pvpr[file_name] = (pv,pr)
- self.pkg_dp[file_name] = dp
-
- # Build forward and reverse provider hashes
- # Forward: virtual -> [filenames]
- # Reverse: PN -> [virtuals]
- if pn not in self.pn_provides:
- self.pn_provides[pn] = Set()
- self.pn_provides[pn] |= provides
-
- for provide in provides:
- if provide not in self.providers:
- self.providers[provide] = []
- self.providers[provide].append(file_name)
-
- self.deps[file_name] = Set()
- for dep in depends:
- self.all_depends.add(dep)
- self.deps[file_name].add(dep)
-
- # Build reverse hash for PACKAGES, so runtime dependencies
- # can be be resolved (RDEPENDS, RRECOMMENDS etc.)
- for package in packages:
- if not package in self.packages:
- self.packages[package] = []
- self.packages[package].append(file_name)
- rprovides += (bb_cache.getVar("RPROVIDES_%s" % package, file_name, 1) or "").split()
-
- for package in packages_dynamic:
- if not package in self.packages_dynamic:
- self.packages_dynamic[package] = []
- self.packages_dynamic[package].append(file_name)
-
- for rprovide in rprovides:
- if not rprovide in self.rproviders:
- self.rproviders[rprovide] = []
- self.rproviders[rprovide].append(file_name)
-
- # Build hash of runtime depends and rececommends
-
- def add_dep(deplist, deps):
- for dep in deps:
- if not dep in deplist:
- deplist[dep] = ""
-
- if not file_name in self.rundeps:
- self.rundeps[file_name] = {}
- if not file_name in self.runrecs:
- self.runrecs[file_name] = {}
-
- for package in packages + [pn]:
- if not package in self.rundeps[file_name]:
- self.rundeps[file_name][package] = {}
- if not package in self.runrecs[file_name]:
- self.runrecs[file_name][package] = {}
-
- add_dep(self.rundeps[file_name][package], bb.utils.explode_deps(bb_cache.getVar('RDEPENDS', file_name, True) or ""))
- add_dep(self.runrecs[file_name][package], bb.utils.explode_deps(bb_cache.getVar('RRECOMMENDS', file_name, True) or ""))
- add_dep(self.rundeps[file_name][package], bb.utils.explode_deps(bb_cache.getVar("RDEPENDS_%s" % package, file_name, True) or ""))
- add_dep(self.runrecs[file_name][package], bb.utils.explode_deps(bb_cache.getVar("RRECOMMENDS_%s" % package, file_name, True) or ""))
-
- # Collect files we may need for possible world-dep
- # calculations
- if not bb_cache.getVar('BROKEN', file_name, True) and not bb_cache.getVar('EXCLUDE_FROM_WORLD', file_name, True):
- self.possible_world.append(file_name)
-
+__version__ = "1.7.4"
#============================================================================#
# BBStatistics
Manages one bitbake build run
"""
- ParsingStatus = BBParsingStatus # make it visible from the shell
Statistics = BBStatistics # make it visible from the shell
def __init__( self ):
self.build_cache_fail = []
self.build_cache = []
- self.rbuild_cache = []
- self.building_list = []
- self.build_path = []
- self.consider_msgs_cache = []
self.stats = BBStatistics()
self.status = None
self.cache = None
self.bb_cache = None
- def tryBuildPackage( self, fn, item, the_data ):
- """Build one package"""
+ def tryBuildPackage(self, fn, item, task, the_data, build_depends):
+ """
+ Build one task of a package, optionally build following task depends
+ """
bb.event.fire(bb.event.PkgStarted(item, the_data))
try:
self.stats.attempt += 1
if self.configuration.force:
- bb.data.setVarFlag('do_%s' % self.configuration.cmd, 'force', 1, the_data)
+ bb.data.setVarFlag('do_%s' % task, 'force', 1, the_data)
+ if not build_depends:
+ bb.data.setVarFlag('do_%s' % task, 'dontrundeps', 1, the_data)
if not self.configuration.dry_run:
- bb.build.exec_task('do_%s' % self.configuration.cmd, the_data)
+ bb.build.exec_task('do_%s' % task, the_data)
bb.event.fire(bb.event.PkgSucceeded(item, the_data))
self.build_cache.append(fn)
return True
self.build_cache_fail.append(fn)
raise
- def tryBuild( self, fn, virtual , buildAllDeps , build_depends = []):
+ def tryBuild( self, fn, build_depends):
"""
Build a provider and its dependencies.
build_depends is a list of previous build dependencies (not runtime)
If build_depends is empty, we're dealing with a runtime depends
"""
- the_data = self.bb_cache.loadDataFull(fn, self)
-
- # Only follow all (runtime) dependencies if doing a build
- if not buildAllDeps and self.configuration.cmd is "build":
- buildAllDeps = self.status.build_all[fn]
-
- # Error on build time dependency loops
- if build_depends and build_depends.count(fn) > 1:
- bb.msg.error(bb.msg.domain.Depends, "%s depends on itself (eventually)" % fn)
- bb.msg.error(bb.msg.domain.Depends, "upwards chain is: %s" % (" -> ".join(self.build_path)))
- return False
-
- # See if this is a runtime dependency we've already built
- # Or a build dependency being handled in a different build chain
- if fn in self.building_list:
- return self.addRunDeps(fn, virtual , buildAllDeps)
+ the_data = self.bb_cache.loadDataFull(fn, self.configuration.data)
item = self.status.pkg_fn[fn]
- self.building_list.append(fn)
-
- pathstr = "%s (%s)" % (item, virtual)
- self.build_path.append(pathstr)
-
- depends_list = (bb.data.getVar('DEPENDS', the_data, True) or "").split()
-
- bb.msg.note(2, bb.msg.domain.Depends, "current path: %s" % (" -> ".join(self.build_path)))
- bb.msg.note(2, bb.msg.domain.Depends, "dependencies for %s are: %s" % (item, " ".join(depends_list)))
-
- try:
- failed = False
-
- depcmd = self.configuration.cmd
- bbdepcmd = bb.data.getVarFlag('do_%s' % self.configuration.cmd, 'bbdepcmd', the_data)
- if bbdepcmd is not None:
- if bbdepcmd == "":
- depcmd = None
- else:
- depcmd = bbdepcmd
-
- if depcmd:
- oldcmd = self.configuration.cmd
- self.configuration.cmd = depcmd
-
- for dependency in depends_list:
- if dependency in self.status.ignored_dependencies:
- continue
- if not depcmd:
- continue
- if self.buildProvider( dependency , buildAllDeps , build_depends ) == 0:
- bb.msg.error(bb.msg.domain.Depends, "dependency %s (for %s) not satisfied" % (dependency,item))
- failed = True
- if self.configuration.abort:
- break
-
- if depcmd:
- self.configuration.cmd = oldcmd
-
- if failed:
- self.stats.deps += 1
- return False
-
- if not self.addRunDeps(fn, virtual , buildAllDeps):
- return False
-
- if bb.build.stamp_is_current('do_%s' % self.configuration.cmd, the_data):
- self.build_cache.append(fn)
- return True
-
- return self.tryBuildPackage( fn, item, the_data )
-
- finally:
- self.building_list.remove(fn)
- self.build_path.remove(pathstr)
+ if bb.build.stamp_is_current('do_%s' % self.configuration.cmd, the_data) and not self.configuration.force:
+ self.build_cache.append(fn)
+ return True
+ return self.tryBuildPackage(fn, item, self.configuration.cmd, the_data, build_depends)
def showVersions( self ):
pkg_pn = self.status.pkg_pn
self.cb = None
self.bb_cache = bb.cache.init(self)
try:
- self.configuration.data = self.bb_cache.loadDataFull(self.configuration.buildfile, self)
+ self.configuration.data = self.bb_cache.loadDataFull(self.configuration.buildfile, self.configuration.data)
except IOError, e:
bb.msg.fatal(bb.msg.domain.Parsing, "Unable to read %s: %s" % ( self.configuration.buildfile, e ))
except Exception, e:
except Exception, e:
bb.msg.fatal(bb.msg.domain.Parsing, "%s" % e)
# emit the metadata which isnt valid shell
+ data.expandKeys( self.configuration.data )
for e in self.configuration.data.keys():
if data.getVarFlag( e, 'python', self.configuration.data ):
sys.__stdout__.write("\npython %s () {\n%s}\n" % (e, data.getVar(e, self.configuration.data, 1)))
bb.msg.debug(1, bb.msg.domain.Provider, "providers for %s are: %s" % (item, pkg_pn.keys()))
for pn in pkg_pn.keys():
- preferred_versions[pn] = bb.providers.findBestProvider(pn, pkg_pn)[2:4]
+ preferred_versions[pn] = bb.providers.findBestProvider(pn, self.configuration.data, self.status, pkg_pn)[2:4]
eligible.append(preferred_versions[pn][1])
for p in eligible:
return eligible
-
-
# try to avoid adding the same rdepends over an over again
seen_depends = []
seen_rdepends = []
We have not seen this name -> error in
dependency handling
"""
- bb.note( "ERROR with provider: %(package)s" % vars() )
+ bb.msg.note(1, bb.msg.domain.Depends, "ERROR with provider: %(package)s" % vars() )
print >> depends_file, '"%(package)s" -> ERROR' % vars()
continue
# for the package.
if package in the_depends:
if not package in self.status.providers:
- bb.note( "ERROR with provider: %(package)s" % vars() )
+ bb.msg.note(1, bb.msg.domain.Depends, "ERROR with provider: %(package)s" % vars() )
print >> alldepends_file, '"%(package)s" -> ERROR' % vars()
continue
providers = self.status.providers[package]
elif package in the_rdepends:
if len(bb.providers.getRuntimeProviders(self.status, package)) == 0:
- bb.note( "ERROR with rprovider: %(package)s" % vars() )
+ bb.msg.note(1, bb.msg.domain.Depends, "ERROR with rprovider: %(package)s" % vars() )
print >> alldepends_file, '"%(package)s" -> ERROR [style="dashed"]' % vars()
continue
add_all_depends( pkgs_to_build, [] )
print >> alldepends_file, "}"
- def buildProvider( self, item , buildAllDeps , build_depends = [] ):
- """
- Build something to provide a named build requirement
- (takes item names from DEPENDS namespace)
- """
-
- fn = None
- discriminated = False
-
- if not item in self.status.providers:
- bb.msg.error(bb.msg.domain.Depends, "Nothing provides dependency %s" % item)
- bb.event.fire(bb.event.NoProvider(item,self.configuration.data))
- return 0
-
- all_p = self.status.providers[item]
-
- for p in all_p:
- if p in self.build_cache:
- bb.msg.debug(1, bb.msg.domain.Provider, "already built %s in this run" % p)
- return 1
-
- eligible = bb.providers.filterProviders(all_p, item, self.configuration.data, self.status, self.build_cache_fail)
-
- if not eligible:
- return 0
-
- prefervar = bb.data.getVar('PREFERRED_PROVIDER_%s' % item, self.configuration.data, 1)
- if prefervar:
- self.status.preferred[item] = prefervar
-
- if item in self.status.preferred:
- for p in eligible:
- pn = self.status.pkg_fn[p]
- if self.status.preferred[item] == pn:
- bb.msg.note(2, bb.msg.domain.Provider, "selecting %s to satisfy %s due to PREFERRED_PROVIDERS" % (pn, item))
- eligible.remove(p)
- eligible = [p] + eligible
- discriminated = True
- break
-
- if len(eligible) > 1 and discriminated == False:
- if item not in self.consider_msgs_cache:
- providers_list = []
- for fn in eligible:
- providers_list.append(self.status.pkg_fn[fn])
- bb.msg.note(1, bb.msg.domain.Provider, "multiple providers are available (%s);" % ", ".join(providers_list))
- bb.msg.note(1, bb.msg.domain.Provider, "consider defining PREFERRED_PROVIDER_%s" % item)
- bb.event.fire(bb.event.MultipleProviders(item,providers_list,self.configuration.data))
- self.consider_msgs_cache.append(item)
-
-
- # run through the list until we find one that we can build
- for fn in eligible:
- bb.msg.debug(2, bb.msg.domain.Provider, "selecting %s to satisfy %s" % (fn, item))
- if self.tryBuild(fn, item, buildAllDeps, build_depends + [fn]):
- return 1
-
- bb.msg.note(1, bb.msg.domain.Provider, "no buildable providers for %s" % item)
- bb.event.fire(bb.event.NoProvider(item,self.configuration.data))
- return 0
-
- def buildRProvider( self, item , buildAllDeps ):
- """
- Build something to provide a named runtime requirement
- (takes item names from RDEPENDS/PACKAGES namespace)
- """
-
- fn = None
- all_p = []
- discriminated = False
-
- if not buildAllDeps:
- return True
-
- all_p = bb.providers.getRuntimeProviders(self.status, item)
-
- if not all_p:
- bb.msg.error(bb.msg.domain.Provider, "Nothing provides runtime dependency %s" % (item))
- bb.event.fire(bb.event.NoProvider(item,self.configuration.data,runtime=True))
- return False
-
- for p in all_p:
- if p in self.rbuild_cache:
- bb.msg.debug(2, bb.msg.domain.Provider, "Already built %s providing runtime %s" % (p,item))
- return True
- if p in self.build_cache:
- bb.msg.debug(2, bb.msg.domain.Provider, "Already built %s but adding any further RDEPENDS for %s" % (p, item))
- return self.addRunDeps(p, item , buildAllDeps)
-
- eligible = bb.providers.filterProviders(all_p, item, self.configuration.data, self.status, self.build_cache_fail)
- if not eligible:
- return 0
-
- preferred = []
- for p in eligible:
- pn = self.status.pkg_fn[p]
- provides = self.status.pn_provides[pn]
- for provide in provides:
- prefervar = bb.data.getVar('PREFERRED_PROVIDER_%s' % provide, self.configuration.data, 1)
- if prefervar == pn:
- bb.msg.note(2, bb.msg.domain.Provider, "selecting %s to satisfy runtime %s due to PREFERRED_PROVIDERS" % (pn, item))
- eligible.remove(p)
- eligible = [p] + eligible
- preferred.append(p)
-
- if len(eligible) > 1 and len(preferred) == 0:
- if item not in self.consider_msgs_cache:
- providers_list = []
- for fn in eligible:
- providers_list.append(self.status.pkg_fn[fn])
- bb.msg.note(1, bb.msg.domain.Provider, "multiple providers are available (%s);" % ", ".join(providers_list))
- bb.msg.note(1, bb.msg.domain.Provider, "consider defining a PREFERRED_PROVIDER to match runtime %s" % item)
- bb.event.fire(bb.event.MultipleProviders(item,providers_list,self.configuration.data,runtime=True))
- self.consider_msgs_cache.append(item)
-
- if len(preferred) > 1:
- if item not in self.consider_msgs_cache:
- providers_list = []
- for fn in preferred:
- providers_list.append(self.status.pkg_fn[fn])
- bb.msg.note(1, bb.msg.domain.Provider, "multiple preferred providers are available (%s);" % ", ".join(providers_list))
- bb.msg.note(1, bb.msg.domain.Provider, "consider defining only one PREFERRED_PROVIDER to match runtime %s" % item)
- bb.event.fire(bb.event.MultipleProviders(item,providers_list,self.configuration.data,runtime=True))
- self.consider_msgs_cache.append(item)
-
- # run through the list until we find one that we can build
- for fn in eligible:
- bb.msg.debug(2, bb.msg.domain.Provider, "selecting %s to satisfy runtime %s" % (fn, item))
- if self.tryBuild(fn, item, buildAllDeps):
- return True
-
- bb.msg.error(bb.msg.domain.Provider, "No buildable providers for runtime %s" % item)
- bb.event.fire(bb.event.NoProvider(item,self.configuration.data))
- return False
-
- def addRunDeps(self , fn, item , buildAllDeps):
- """
- Add any runtime dependencies of runtime item provided by fn
- as long as item has't previously been processed by this function.
- """
-
- if item in self.rbuild_cache:
- return True
-
- if not buildAllDeps:
- return True
-
- rdepends = []
- self.rbuild_cache.append(item)
-
- if fn in self.status.rundeps and item in self.status.rundeps[fn]:
- rdepends += self.status.rundeps[fn][item].keys()
- if fn in self.status.runrecs and item in self.status.runrecs[fn]:
- rdepends += self.status.runrecs[fn][item].keys()
-
- bb.msg.debug(2, bb.msg.domain.Provider, "Additional runtime dependencies for %s are: %s" % (item, " ".join(rdepends)))
-
- for rdepend in rdepends:
- if rdepend in self.status.ignored_dependencies:
- continue
- if not self.buildRProvider(rdepend, buildAllDeps):
- return False
- return True
-
def buildDepgraph( self ):
all_depends = self.status.all_depends
pn_provides = self.status.pn_provides
localdata = data.createCopy(self.configuration.data)
bb.data.update_data(localdata)
+ bb.data.expandKeys(localdata)
def calc_bbfile_priority(filename):
for (regex, pri) in self.status.bbfile_config_priorities:
self.status.possible_world = None
self.status.all_depends = None
- def myProgressCallback( self, x, y, f, bb_cache, from_cache ):
- # feed the status with new input
-
- self.status.handle_bb_data(f, bb_cache, from_cache)
-
+ def myProgressCallback( self, x, y, f, from_cache ):
+ """Update any tty with the progress change"""
if os.isatty(sys.stdout.fileno()):
sys.stdout.write("\rNOTE: Handling BitBake files: %s (%04d/%04d) [%2d %%]" % ( parsespin.next(), x, y, x*100/y ) )
sys.stdout.flush()
bb.msg.fatal(bb.msg.domain.Parsing, "Sorry, shell not available (%s)" % details )
else:
bb.data.update_data( self.configuration.data )
+ bb.data.expandKeys(localdata)
shell.start( self )
sys.exit( 0 )
if self.configuration.verbose:
bb.msg.set_verbose(True)
- if not self.configuration.cmd:
- self.configuration.cmd = "build"
-
if self.configuration.debug:
bb.msg.set_debug_level(self.configuration.debug)
+ else:
+ bb.msg.set_debug_level(0)
+
+ if self.configuration.debug_domains:
+ bb.msg.set_debug_domains(self.configuration.debug_domains)
self.configuration.data = bb.data.init()
self.parseConfigurationFile( os.path.join( "conf", "bitbake.conf" ) )
+ if not self.configuration.cmd:
+ self.configuration.cmd = bb.data.getVar("BB_DEFAULT_TASK", self.configuration.data)
+
+ # For backwards compatibility - REMOVE ME
+ if not self.configuration.cmd:
+ self.configuration.cmd = "build"
#
# Special updated configuration we use for firing events
if self.configuration.buildfile is not None:
bf = os.path.abspath( self.configuration.buildfile )
try:
- bbfile_data = bb.parse.handle(bf, self.configuration.data)
- except IOError:
- bb.msg.fatal(bb.msg.domain.Parsing, "Unable to open %s" % bf)
+ os.stat(bf)
+ except OSError:
+ (filelist, masked) = self.collect_bbfiles()
+ regexp = re.compile(self.configuration.buildfile)
+ matches = []
+ for f in filelist:
+ if regexp.search(f) and os.path.isfile(f):
+ bf = f
+ matches.append(f)
+ if len(matches) != 1:
+ bb.msg.error(bb.msg.domain.Parsing, "Unable to match %s (%s matches found):" % (self.configuration.buildfile, len(matches)))
+ for f in matches:
+ bb.msg.error(bb.msg.domain.Parsing, " %s" % f)
+ sys.exit(1)
+ bf = matches[0]
+
+ bbfile_data = bb.parse.handle(bf, self.configuration.data)
item = bb.data.getVar('PN', bbfile_data, 1)
try:
- self.tryBuildPackage( bf, item, bbfile_data )
+ self.tryBuildPackage(bf, item, self.configuration.cmd, bbfile_data, True)
except bb.build.EventException:
bb.msg.error(bb.msg.domain.Build, "Build of '%s' failed" % item )
sys.exit( self.stats.show() )
# initialise the parsing status now we know we will need deps
- self.status = BBParsingStatus()
+ self.status = bb.cache.CacheData()
ignore = bb.data.getVar("ASSUME_PROVIDED", self.configuration.data, 1) or ""
self.status.ignored_dependencies = Set( ignore.split() )
except ImportError:
bb.msg.note(1, bb.msg.domain.Collection, "Psyco JIT Compiler (http://psyco.sf.net) not available. Install it to increase performance.")
else:
- psyco.bind( self.collect_bbfiles )
+ psyco.bind( self.parse_bbfiles )
else:
bb.msg.note(1, bb.msg.domain.Collection, "You have disabled Psyco. This decreases performance.")
try:
bb.msg.debug(1, bb.msg.domain.Collection, "collecting .bb files")
- self.collect_bbfiles( self.myProgressCallback )
+ (filelist, masked) = self.collect_bbfiles()
+ self.parse_bbfiles(filelist, masked, self.myProgressCallback)
bb.msg.debug(1, bb.msg.domain.Collection, "parsing complete")
print
if self.configuration.parse_only:
self.generateDotGraph( pkgs_to_build, self.configuration.ignored_dot_deps )
sys.exit( 0 )
-
bb.event.fire(bb.event.BuildStarted(buildname, pkgs_to_build, self.configuration.event_data))
- failures = 0
- for k in pkgs_to_build:
- failed = False
- try:
- if self.buildProvider( k , False ) == 0:
- # already diagnosed
- failed = True
- except bb.build.EventException:
- bb.msg.error(bb.msg.domain.Build, "Build of " + k + " failed")
- failed = True
-
- if failed:
- failures += failures
- if self.configuration.abort:
- sys.exit(1)
+ localdata = data.createCopy(self.configuration.data)
+ bb.data.update_data(localdata)
+ bb.data.expandKeys(localdata)
+ taskdata = bb.taskdata.TaskData(self.configuration.abort)
+
+ runlist = []
+ try:
+ for k in pkgs_to_build:
+ taskdata.add_provider(localdata, self.status, k)
+ runlist.append([k, "do_%s" % self.configuration.cmd])
+ taskdata.add_unresolved(localdata, self.status)
+ except bb.providers.NoProvider:
+ sys.exit(1)
+
+ rq = bb.runqueue.RunQueue()
+ rq.prepare_runqueue(self.configuration.data, self.status, taskdata, runlist)
+ try:
+ failures = rq.execute_runqueue(self, self.configuration.data, self.status, taskdata, runlist)
+ except runqueue.TaskFailure, fnids:
+ for fnid in fnids:
+ bb.msg.error(bb.msg.domain.Build, "'%s' failed" % taskdata.fn_index[fnid])
+ sys.exit(1)
bb.event.fire(bb.event.BuildCompleted(buildname, pkgs_to_build, self.configuration.event_data, failures))
sys.exit( self.stats.show() )
return []
return finddata.readlines()
- def collect_bbfiles( self, progressCallback ):
+ def collect_bbfiles( self ):
"""Collect all available .bb build files"""
- self.cb = progressCallback
parsed, cached, skipped, masked = 0, 0, 0, 0
self.bb_cache = bb.cache.init(self)
continue
newfiles += glob.glob(f) or [ f ]
- bbmask = bb.data.getVar('BBMASK', self.configuration.data, 1) or ""
+ bbmask = bb.data.getVar('BBMASK', self.configuration.data, 1)
+
+ if not bbmask:
+ return (newfiles, 0)
+
try:
bbmask_compiled = re.compile(bbmask)
except sre_constants.error:
bb.msg.fatal(bb.msg.domain.Collection, "BBMASK is not a valid regular expression.")
+ finalfiles = []
for i in xrange( len( newfiles ) ):
f = newfiles[i]
if bbmask and bbmask_compiled.search(f):
- bb.msg.debug(1, bb.msg.domain.Collection, "bbmake: skipping %s" % f, f)
+ bb.msg.debug(1, bb.msg.domain.Collection, "skipping masked file %s" % f)
masked += 1
continue
- bb.msg.debug(1, bb.msg.domain.Collection, "bbmake: parsing %s" % f, f)
+ finalfiles.append(f)
+
+ return (finalfiles, masked)
+
+ def parse_bbfiles(self, filelist, masked, progressCallback = None):
+ parsed, cached, skipped = 0, 0, 0
+ for i in xrange( len( filelist ) ):
+ f = filelist[i]
+
+ bb.msg.debug(1, bb.msg.domain.Collection, "parsing %s" % f)
# read a file's metadata
try:
- fromCache, skip = self.bb_cache.loadData(f, self)
+ fromCache, skip = self.bb_cache.loadData(f, self.configuration.data)
if skip:
skipped += 1
- bb.msg.debug(2, bb.msg.domain.Collection, "Skipping %s" % f, f)
+ bb.msg.debug(2, bb.msg.domain.Collection, "skipping %s" % f)
self.bb_cache.skip(f)
continue
elif fromCache: cached += 1
else: parsed += 1
deps = None
+ # Disabled by RP as was no longer functional
# allow metadata files to add items to BBFILES
#data.update_data(self.pkgdata[f])
- addbbfiles = self.bb_cache.getVar('BBFILES', f, False) or None
- if addbbfiles:
- for aof in addbbfiles.split():
- if not files.count(aof):
- if not os.path.isabs(aof):
- aof = os.path.join(os.path.dirname(f),aof)
- files.append(aof)
+ #addbbfiles = self.bb_cache.getVar('BBFILES', f, False) or None
+ #if addbbfiles:
+ # for aof in addbbfiles.split():
+ # if not files.count(aof):
+ # if not os.path.isabs(aof):
+ # aof = os.path.join(os.path.dirname(f),aof)
+ # files.append(aof)
+
+ self.bb_cache.handle_data(f, self.status)
# now inform the caller
- if self.cb is not None:
- self.cb( i + 1, len( newfiles ), f, self.bb_cache, fromCache )
+ if progressCallback is not None:
+ progressCallback( i + 1, len( filelist ), f, fromCache )
except IOError, e:
self.bb_cache.remove(f)
- bb.msg.error(bb.msg.domain.Collection, "opening %s: %s" % (f, e), f)
+ bb.msg.error(bb.msg.domain.Collection, "opening %s: %s" % (f, e))
pass
except KeyboardInterrupt:
self.bb_cache.sync()
raise
except Exception, e:
self.bb_cache.remove(f)
- bb.msg.error(bb.msg.domain.Collection, "%s while parsing %s" % (e, f), f)
+ bb.msg.error(bb.msg.domain.Collection, "%s while parsing %s" % (e, f))
except:
self.bb_cache.remove(f)
raise
- if self.cb is not None:
+ if progressCallback is not None:
print "\r" # need newline after Handling Bitbake files message
bb.msg.note(1, bb.msg.domain.Collection, "Parsing finished. %d cached, %d parsed, %d skipped, %d masked." % ( cached, parsed, skipped, masked ))
action = "store_true", dest = "interactive", default = False )
parser.add_option( "-c", "--cmd", help = "Specify task to execute. Note that this only executes the specified task for the providee and the packages it depends on, i.e. 'compile' does not implicitly call stage for the dependencies (IOW: use only if you know what you are doing). Depending on the base.bbclass a listtaks tasks is defined and will show available tasks",
- action = "store", dest = "cmd", default = "build" )
+ action = "store", dest = "cmd" )
parser.add_option( "-r", "--read", help = "read the specified file before bitbake.conf",
action = "append", dest = "file", default = [] )
parser.add_option( "-g", "--graphviz", help = "emit the dependency trees of the specified packages in the dot syntax",
action = "store_true", dest = "dot_graph", default = False )
+
parser.add_option( "-I", "--ignore-deps", help = """Stop processing at the given list of dependencies when generating dependency graphs. This can help to make the graph more appealing""",
action = "append", dest = "ignored_dot_deps", default = [] )
+ parser.add_option( "-l", "--log-domains", help = """Show debug logging for the specified logging domains""",
+ action = "append", dest = "debug_domains", default = [] )
+
options, args = parser.parse_args( sys.argv )
if __name__ == "__main__":
print """WARNING, WARNING, WARNING
This is a Bitbake from the Unstable/Development Branch.
-You might want to use the bitbake-1.4 stable branch (if you are not a BitBake developer or tester). I'm going to sleep 5 seconds now to make sure you see that."""
+You might want to use the bitbake-1.6 stable branch (if you are not a BitBake developer or tester). I'm going to sleep 5 seconds now to make sure you see that."""
import time
time.sleep(5)
main()