basics.py   [plain text]


#!/usr/bin/python2.4
#
# Copyright 2007 Google Inc.
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; either version 2
# of the License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301,
# USA.
#
#
"""Common and low-level stuff for include server."""

__author__ = 'Nils Klarlund'

import glob
import os.path
import resource
import signal
import shutil
import sys
import tempfile


# MANAGEMENT OF TEMPORARY LOCATIONS FOR GENERATIONS OF COMPRESSED FILES


class ClientRootKeeper(object):
  """Determine the tmp directory to use for compressed files.

  Use the RAM disk-like /dev/shm as default place to store compressed files if
  available.  The protocol between the include server and distcc client
  stipulates that the top three directories constitute the prefix prepended to
  absolute file paths.

  Instance vars:
    client_tmp: a path, the place for creation of temporary directories.
    client_root: a path, the current such temporary directory
    _client_root_before_padding: a path kept for testing purposes

  A typical client root looks like:

  -  /tmp/tmpBDoZQV.include_server-6642-13/padding, or
  -  /dev/shm/tmpBDoZQV.include_server-6642-19

  Note that each path has exactly three directory components to it.  This is an
  invariant.  Some client roots are padded with '/padding' to satisfy the
  invariant.
  """
  
  # This constant is embedded in names of client root directories.
  INCLUDE_SERVER_NAME = 'include_server'

  def __init__(self):
    """Constructor."""
    if 'DISTCC_CLIENT_TMP' in os.environ:
      self.client_tmp = os.environ['DISTCC_CLIENT_TMP']
    elif os.path.isdir('/dev/shm') and os.access('/dev/shm',
                                                 os.X_OK + os.W_OK + os.R_OK):
      self.client_tmp = '/dev/shm'
    else:
      self.client_tmp = '/tmp'
    if not self.client_tmp or self.client_tmp[0] != '/':
      sys.exit("""DISTCC_CLIENT_TMP must start with '/'.""")
    self.client_tmp = self.client_tmp.rstrip('/')
    # To have room to make a temp directory, we'll need to have less than two
    # levels at this point.  Note: '/a/b'.split('/') == ['', 'a', 'b'].
    if len(self.client_tmp.split('/')) > 3:
      sys.exit('DISTCC_CLIENT_TMP must have at most two directory levels.')
    self.number_missing_levels = 3 - len(self.client_tmp.split('/'))
    self.client_root = None

  def Glob(self, pid_expr):
    """Glob unpadded client roots whose pid is matched by pid expression."""
    return glob.glob('%s/*.%s-%s-*'
                     % (self.client_tmp, self.INCLUDE_SERVER_NAME,
                        pid_expr))
  
  def ClientRootMakedir(self, generation):
    """Make a new client directory for a generation of compressed files.

    Arguments:
      generation: a natural number, usually 1 or slightly bigger; this number,
        minus 1, indicates how many times a reset of the caches has taken place.
    """
    try:
      # Create a unique identifier that will never repeat. Use pid as suffix for
      # cleanout mechanism that wipes files not associated with a running pid.
      self._client_root_before_padding = tempfile.mkdtemp(
          '.%s-%s-%d' %
          (self.INCLUDE_SERVER_NAME,
           os.getpid(), generation),
          dir=self.client_tmp)
      self.client_root = (self._client_root_before_padding
                          + '/padding' * self.number_missing_levels)
      if not os.path.isdir(self.client_root):
        os.makedirs(self.client_root)
    except (IOError, OSError), why:
      sys.exit('Could not create client root directory %s: %s' %
               (self.client_root, why))

  def CleanOutClientRoots(self, pid=None):
    """Delete client root directories pertaining to this process.
    Args:
      pid: None (which means 'pid of current process') or an integer
    """
    if not pid:
      pid = os.getpid()
    for client_root in self.Glob(str(pid)):
      shutil.rmtree(client_root, ignore_errors=True)

  def CleanOutOthers(self):
    """Search for left-overs from include servers that have passed away."""
    # Find all client root subdirectories whether abandoned or not.
    distcc_directories = self.Glob('*')
    for directory in distcc_directories:
      # Fish out pid from end of directory name.
      hyphen_ultimate_position = directory.rfind('-')
      assert hyphen_ultimate_position != -1
      hyphen_penultimate_position = directory.rfind('-', 0,
                                                    hyphen_ultimate_position)
      assert hyphen_penultimate_position != -1
      pid_str = directory[hyphen_penultimate_position + 1:
                          hyphen_ultimate_position]
      try:
        pid = int(pid_str)
      except ValueError:
        continue  # Happens only if a spoofer is around.
      try:
        # Got a pid; does it still exist?
        os.getpgid(pid)
        continue
      except OSError:
        # Process pid does not exist. Nuke its associated files. This will
        # of course only succeed if the files belong the current uid of
        # this process.
        if not os.access(directory, os.W_OK):
          continue  # no access, not ours
        Debug(DEBUG_TRACE,
              "Cleaning out '%s' after defunct include server." % directory)
        self.CleanOutClientRoots(pid)

# COMPILATION SERVER

# An upper bound on the number of directory components in the temporary
# directory on the distccd server that corresponds to the root directory
# on the client machine.  Normally the server root is /tmp/distccd_XXXXXX.
# But it could be something different if TMPDIR is set when distccd is
# started.  See dcc_get_tmp_top() in ../src/tempfile.c.
# It turns out that our own test infrastructure (test/testdistcc.py)
# sets TMPDIR before invoking distccd, so this needs to be reasonably
# high, otherwise 'make distcheck' will fail.
MAX_COMPONENTS_IN_SERVER_ROOT = 20

# EMAILS

# For automated emails, see also src/emaillog.h.
DCC_EMAILLOG_WHOM_TO_BLAME = os.getenv('DISTCC_EMAILLOG_WHOM_TO_BLAME',
                                       'distcc-pump-errors')
EMAIL_SUBJECT = 'distcc-pump include server email'
CANT_SEND_MESSAGE = """Please notify %s that the distcc-pump include server
tried to send them email but failed.""" % DCC_EMAILLOG_WHOM_TO_BLAME
MAX_EMAILS_TO_SEND = 3

# TIME QUOTAS (SOLVING THE HALTING PROBLEM)

# The maximum user time the include server is allowed handling one request. This
# is a critical parameter because all caches are reset if this time is
# exceeded. And if all caches are reset, then the next request may take much
# longer time, possibly again exceeding the quota.  The parameter is also of
# importance to builds that involve compilations that distcc-pump does not grok:
# an amount of time roughly equal to this quota is wasted before CPP is invoked
# instead.
USER_TIME_QUOTA = 6.0  # seconds

# How often the following question is answered: has too much user time been
# spent in the include handler servicing the current request?
#
# FIXME(klarlund): SIGALRM should not be raised in code that has I/O. Fix
# include server so that this is guaranteed not to happen. Until then, we are
# careful to wait a full 4 s before issuing SIGALRM.
USER_TIME_QUOTA_CHECK_INTERVAL_TIME = 4  # seconds, an integer

# ALGORITHMS

SIMPLE = 0     # not implemented
MEMOIZING = 1  # only one currently implemented
ALGORITHMS = [SIMPLE, MEMOIZING]

# PYTHON TUNING

# The default for the first parameter of gc.set_threshold is 700; see
# http://www.isi.edu/~chiang/python.html for a discussion of why this parameter
# can be bumped up considerably for speed-up.  The new default of 10000 was
# tested on a very large application, where include server time CPU time drops
# from 151s to 118s (best times out of 10 runs). There was no seeming changes to
# memory usage.  Trying with 100,000 did not speed up the application further.
GC_THRESHOLD = 10000  

# INCLUDE DIR COMMAND LINE HANDLING

INCLUDE_DIR_NORMAL     = 1  # -I
INCLUDE_DIR_FRAMEWORKS = 2  # -F

# FLAGS FOR COMMAND LINE OPTIONS

opt_algorithm = MEMOIZING  # currently, only choice
opt_debug_pattern = 1  # see DEBUG below
opt_email_bound = MAX_EMAILS_TO_SEND
opt_exact_analysis = False         # use CPP instead of include analyzer
opt_print_times = False
opt_path_observation_re = None
opt_send_email = False
opt_simple_algorithm = False
opt_stat_reset_triggers = {}
opt_statistics = False
opt_unsafe_absolute_includes = False
opt_verify = False     # whether to compare calculated include closure to that
                       # produced by compiler
opt_write_include_closure = False  # write include closures to file

# HELPER FUNCTION FOR STAT_RESET_TRIGGERS


def Stamp(path):
  """Return a stamp characterizing a file and its modification time."""
  try:
    st_inf = os.stat(path)
    # The inode and device identify a file uniquely.
    return (st_inf.st_mtime, st_inf.st_ino, st_inf.st_dev)
  except OSError:
    return None


# LANGUAGES AND FILE EXTENSIONS

# The languages that we recognize.
LANGUAGES = set(['c', 'c++', 'objective-c', 'objective-c++'])

# The suffixes, following last period, used for source files and
# preprocessed files, each with their corresponding source language.
TRANSLATION_UNIT_MAP = {
    # C
    'c': 'c', 'i': 'c',
    # C++
    'cc': 'c++', 'cpp': 'c++', 'cxx': 'c++', 'C': 'c++', 'CXX': 'c++',
    'ii': 'c++',
    # Objective C
    'm': 'objective-c', 'mi': 'objective-c',
    # Objective C++
    'mm': 'objective-c++', 'M': 'objective-c++', 'mii': 'objective-c++',
    }

# All languages are described by suffixes.
assert set(TRANSLATION_UNIT_MAP.values()) == LANGUAGES


# DEBUG

# Debugging is controlled by the 5 least significant bits of
# opt_debug_pattern.
DEBUG_WARNING = 1   # For warnings
DEBUG_TRACE = 2     # For tracing functions (upper level)
DEBUG_TRACE1 = 4    # For tracing functions (medium level)
DEBUG_TRACE2 = 8    # For tracing functions (lower level)
DEBUG_DATA = 16     # For printing data
DEBUG_NUM_BITS = 5  # The cardinality of {1,2,4,8,16}


def Debug(trigger_pattern, message, *params):
  """Print message to stderr depending on trigger pattern.

  Args:
    trigger_pattern: a bit vector (as an integer)
    message: a format string
    params: arguments to message
  """
  # TODO(klarlund): use Python's logging module.
  triggered = opt_debug_pattern & trigger_pattern
  if triggered:
    i = 1
    for unused_j in range(DEBUG_NUM_BITS):
      if i & DEBUG_WARNING & triggered:
        print >> sys.stderr, 'WARNING include server:', message % params
      if i & DEBUG_TRACE & triggered:
        print >> sys.stderr, 'TRACE:', message % params
      elif i & DEBUG_TRACE1 & triggered:
        print >> sys.stderr, 'TRACE1:', message % params
      elif i & DEBUG_TRACE2 & triggered:
        print >> sys.stderr, 'TRACE2:', message % params
      elif i & DEBUG_DATA & triggered:
        print >> sys.stderr, 'DATA:', message % params
      i *= 2
    sys.stderr.flush()


# EXCEPTIONS


class Error(Exception):
  """For include server errors."""
  pass


class NotCoveredError(Error):
  """Exception for included file not covered by include processing."""

  def __init__(self, message,
               source_file=None,
               line_number=None,
               send_email=True):
    """Constructor.

    Arguments:
      message: text of error message
      source_file: name of source_file if known
      line_number: an integer, if known
      send_email: a Boolean, if False then never send email

    These arguments are all stored in the exception. However, the source_file
    and line_number are appended, in a syntax defined here, to the message
    before it is stored as self.args[0] through invocation of the Error
    constructor.
    """
    assert not line_number or source_file
    self.source_file = None
    self.line_number = None
    self.send_email = send_email
    if source_file:
      # Mark this exception as mentioning the source_file.
      self.source_file = source_file
      # Line numbers are not currently used.
      if line_number:
        self.line_number = line_number
        message = ("""File: '%s', line: %s: %s"""
                   % (source_file, line_number, message))
      else:
        message = """File: '%s': %s""" % (source_file, message)
    # Message, a string, becomes self.args[0]    
    Error.__init__(self, message)


class NotCoveredTimeOutError(NotCoveredError):
  """Raised when spending too much time analyzing dependencies."""
  pass


class IncludeAnalyzerTimer(object):
  """Start a timer limiting CPU time for servicing a single request.

  We use user time so that a network hiccup will not entail a cache reset if,
  say, we are using NFS.

  An object of this class must be instantiated so that, no matter what, the
  Cancel method is eventually called. This reinstates the original timer (if
  present).
  """

  def __init__(self):
    self.start_utime = resource.getrusage(resource.RUSAGE_SELF).ru_utime
    self.old = signal.signal(signal.SIGALRM, self._TimeIsUp)
    signal.alarm(USER_TIME_QUOTA_CHECK_INTERVAL_TIME)

  def _TimeIsUp(self, unused_sig_number, unused_frame):
    """Check CPU time spent and raise exception or reschedule."""
    if (resource.getrusage(resource.RUSAGE_SELF).ru_utime
        > self.start_utime + USER_TIME_QUOTA):
      raise NotCoveredTimeOutError(('Bailing out because include server '
                                    + 'spent more than %3.1fs user time '
                                    + 'handling request') %
                                   USER_TIME_QUOTA)
    else:
      # Reschedule ourselves.
      signal.alarm(USER_TIME_QUOTA_CHECK_INTERVAL_TIME)

  def Stop(self):
    signal.alarm(0)

  def Start(self):
    signal.alarm(USER_TIME_QUOTA_CHECK_INTERVAL_TIME)

  def Cancel(self):
    """Must be called eventually. See class documentation."""
    sys.stdout.flush()
    signal.alarm(0)
    signal.signal(signal.SIGALRM, self.old)
    
  
class SignalSIGTERM(Error):
  pass


def RaiseSignalSIGTERM(*unused_args):
  """Raise SignalSIGTERM.

  Use signal.signal for binding this function to SIGTERM.
  """
  raise SignalSIGTERM


# COMMON FUNCTIONS


def SafeNormPath(path):
  """Safe, but limited, version of os.path.normpath.

  Args:
    path: a string

  Returns:
    a string
    
  Python's os.path.normpath is an unsafe operation; the result may not point to
  the same file as the argument. Instead, this function just removes
  initial './'s and a final '/'s if present.
  """
  if path == '.':
    return ''
  else:
    while path.startswith('./'):
      path = path[2:]
    return path.rstrip('/')

def PathFromDirMapEntryAndInclude(dirmap_entry, leaf):
  """Takes a DirectoryMap entry and leaf path and builds a full path for it.
  
  DirectoryMap entries already have a trailing slash, so the strings
  can just be appended together.  *BUT* when we put framework search paths
  into the DirectoryMap, we list them twice prefixed with "*H" and "*P" to
  tell the rest of the code that the leaf must be processed according to
  framework rules to build the path.
  
  If the leaf is invalid for a framework search, and dirmap_entry is
  a framework search directory, None is returned.
  """
  if not dirmap_entry or dirmap_entry[0] != '*':
    return dirmap_entry + leaf

  if not '/' in leaf:
    # Frameworks must be #included with at least one slash separating
    # the framework name from the header name.
    return None

  (i_fwk, i_hdr) = leaf.split('/', 1)
  i_fwk = i_fwk + '.framework/'
  if dirmap_entry[1] == 'H':
    return dirmap_entry[2:] + i_fwk + 'Headers/' + i_hdr

  assert dirmap_entry[1] == 'P'
  return dirmap_entry[2:] + i_fwk + 'PrivateHeaders/' + i_hdr