builder.py: further refactoring to run more stuff in the parent process

instead of inside the fork.

Still doesn't seem to affect runtime.  Good.

One nice side effect is jwack.py no longer needs to know anything about our
locks.
This commit is contained in:
Avery Pennarun 2010-11-22 00:03:43 -08:00
commit dcc2edba0c
2 changed files with 72 additions and 70 deletions

View file

@ -41,78 +41,86 @@ def _nice(t):
return os.path.normpath(os.path.join(vars.PWD, t)) return os.path.normpath(os.path.join(vars.PWD, t))
def _build(t):
if (os.path.exists(t) and not state.is_generated(t)
and not os.path.exists('%s.do' % t)):
# an existing source file that is not marked as a generated file.
# This step is mentioned by djb in his notes. It turns out to be
# important to prevent infinite recursion. For example, a rule
# called default.c.do could be used to try to produce hello.c,
# which is undesirable since hello.c existed already.
state.stamp(t)
return # success
state.start(t)
(dofile, basename, ext) = _find_do_file(t)
if not dofile:
err('no rule to make %r\n' % t)
return 1
state.stamp(dofile)
tmpname = '%s.redo.tmp' % t
unlink(tmpname)
f = open(tmpname, 'w+')
# this will run in the dofile's directory, so use only basenames here
argv = ['sh', '-e',
os.path.basename(dofile),
os.path.basename(basename), # target name (extension removed)
ext, # extension (if any), including leading dot
os.path.basename(tmpname) # randomized output file name
]
if vars.VERBOSE: argv[1] += 'v'
if vars.XTRACE: argv[1] += 'x'
if vars.VERBOSE or vars.XTRACE: log_('\n')
log('%s\n' % _nice(t))
rv = subprocess.call(argv, preexec_fn=lambda: _preexec(t),
stdout=f.fileno())
if rv==0:
if os.path.exists(tmpname) and os.stat(tmpname).st_size:
# there's a race condition here, but if the tmpfile disappears
# at *this* point you deserve to get an error, because you're
# doing something totally scary.
os.rename(tmpname, t)
else:
unlink(tmpname)
state.stamp(t)
else:
unlink(tmpname)
state.unstamp(t)
f.close()
if rv != 0:
err('%s: exit code %d\n' % (_nice(t),rv))
return 1
if vars.VERBOSE or vars.XTRACE:
log('%s (done)\n\n' % _nice(t))
class BuildJob: class BuildJob:
def __init__(self, t, lock, shouldbuildfunc, donefunc): def __init__(self, t, lock, shouldbuildfunc, donefunc):
self.t = t self.t = t
self.tmpname = '%s.redo.tmp' % t
self.lock = lock self.lock = lock
self.shouldbuildfunc = shouldbuildfunc self.shouldbuildfunc = shouldbuildfunc
self.donefunc = donefunc self.donefunc = donefunc
def start(self): def start(self):
if not self.shouldbuildfunc(self.t): assert(self.lock.owned)
t = self.t
tmpname = self.tmpname
if not self.shouldbuildfunc(t):
# target doesn't need to be built; skip the whole task # target doesn't need to be built; skip the whole task
self.done(self.t, 0) return self._after2(0)
return if (os.path.exists(t) and not state.is_generated(t)
jwack.start_job(self.t, self.lock, self.build, self.done) and not os.path.exists('%s.do' % t)):
# an existing source file that is not marked as a generated file.
# This step is mentioned by djb in his notes. It turns out to be
# important to prevent infinite recursion. For example, a rule
# called default.c.do could be used to try to produce hello.c,
# which is undesirable since hello.c existed already.
state.stamp(t)
return self._after2(0)
state.start(t)
(dofile, basename, ext) = _find_do_file(t)
if not dofile:
err('no rule to make %r\n' % t)
return self._after2(1)
state.stamp(dofile)
unlink(tmpname)
self.f = open(tmpname, 'w+')
# this will run in the dofile's directory, so use only basenames here
argv = ['sh', '-e',
os.path.basename(dofile),
os.path.basename(basename), # target name (extension removed)
ext, # extension (if any), including leading dot
os.path.basename(tmpname) # randomized output file name
]
if vars.VERBOSE: argv[1] += 'v'
if vars.XTRACE: argv[1] += 'x'
if vars.VERBOSE or vars.XTRACE: log_('\n')
log('%s\n' % _nice(t))
self.argv = argv
jwack.start_job(t, self._do_subproc, self._after)
def build(self): def _do_subproc(self):
return _build(self.t) t = self.t
argv = self.argv
f = self.f
return subprocess.call(argv, preexec_fn=lambda: _preexec(t),
stdout=f.fileno())
def done(self, name, rv): def _after(self, t, rv):
return self.donefunc(self.t, rv) f = self.f
tmpname = self.tmpname
if rv==0:
if os.path.exists(tmpname) and os.stat(tmpname).st_size:
# there's a race condition here, but if the tmpfile disappears
# at *this* point you deserve to get an error, because you're
# doing something totally scary.
os.rename(tmpname, t)
else:
unlink(tmpname)
state.stamp(t)
else:
unlink(tmpname)
state.unstamp(t)
f.close()
if rv != 0:
err('%s: exit code %d\n' % (_nice(t),rv))
else:
if vars.VERBOSE or vars.XTRACE:
log('%s (done)\n\n' % _nice(t))
return self._after2(rv)
def _after2(self, rv):
self.donefunc(self.t, rv)
assert(self.lock.owned)
self.lock.unlock()
def main(targets, shouldbuildfunc): def main(targets, shouldbuildfunc):
@ -145,8 +153,7 @@ def main(targets, shouldbuildfunc):
log('%s (locked...)\n' % _nice(t)) log('%s (locked...)\n' % _nice(t))
locked.append(t) locked.append(t)
else: else:
j = BuildJob(t, lock, shouldbuildfunc, done) BuildJob(t, lock, shouldbuildfunc, done).start()
j.start()
# Now we've built all the "easy" ones. Go back and just wait on the # Now we've built all the "easy" ones. Go back and just wait on the
# remaining ones one by one. This is technically non-optimal; we could # remaining ones one by one. This is technically non-optimal; we could
@ -170,6 +177,5 @@ def main(targets, shouldbuildfunc):
retcode[0] = 2 retcode[0] = 2
lock.unlock() lock.unlock()
else: else:
j = BuildJob(t, lock, shouldbuildfunc, done) BuildJob(t, lock, shouldbuildfunc, done).start()
j.start()
return retcode[0] return retcode[0]

View file

@ -182,9 +182,8 @@ class Job:
return 'Job(%s,%d)' % (self.name, self.pid) return 'Job(%s,%d)' % (self.name, self.pid)
def start_job(reason, lock, jobfunc, donefunc): def start_job(reason, jobfunc, donefunc):
global _mytokens global _mytokens
assert(lock.owned)
assert(_mytokens <= 1) assert(_mytokens <= 1)
get_token(reason) get_token(reason)
assert(_mytokens >= 1) assert(_mytokens >= 1)
@ -203,12 +202,9 @@ def start_job(reason, lock, jobfunc, donefunc):
except Exception: except Exception:
import traceback import traceback
traceback.print_exc() traceback.print_exc()
lock.unlock()
finally: finally:
_debug('exit: %d\n' % rv) _debug('exit: %d\n' % rv)
os._exit(rv) os._exit(rv)
# else we're the parent process
lock.owned = False # child owns it now
os.close(w) os.close(w)
pd = Job(reason, pid, donefunc) pd = Job(reason, pid, donefunc)
_waitfds[r] = pd _waitfds[r] = pd