# Please keep this code python 2.4 compatible and stand alone.
import logging, os, shutil, sys, tempfile, time, urllib2
import subprocess, re
from distutils.version import LooseVersion
from autotest_lib.client.common_lib import autotemp, revision_control, utils
_READ_SIZE = 64*1024
_MAX_PACKAGE_SIZE = 100*1024*1024
_CHROMEOS_MIRROR = ('http://commondatastorage.googleapis.com/'
'chromeos-mirror/gentoo/distfiles/')
class Error(Exception):
"""Local exception to be raised by code in this file."""
class FetchError(Error):
"""Failed to fetch a package from any of its listed URLs."""
def _checksum_file(full_path):
"""@returns The hex checksum of a file given its pathname."""
inputfile = open(full_path, 'rb')
try:
hex_sum = utils.hash('sha1', inputfile.read()).hexdigest()
finally:
inputfile.close()
return hex_sum
def system(commandline):
"""Same as os.system(commandline) but logs the command first.
@param commandline: commandline to be called.
"""
logging.info(commandline)
return os.system(commandline)
def find_top_of_autotest_tree():
"""@returns The full path to the top of the autotest directory tree."""
dirname = os.path.dirname(__file__)
autotest_dir = os.path.abspath(os.path.join(dirname, '..'))
return autotest_dir
class ExternalPackage(object):
"""
Defines an external package with URLs to fetch its sources from and
a build_and_install() method to unpack it, build it and install it
beneath our own autotest/site-packages directory.
Base Class. Subclass this to define packages.
Note: Unless your subclass has a specific reason to, it should not
re-install the package every time build_externals is invoked, as this
happens periodically through the scheduler. To avoid doing so the is_needed
method needs to return an appropriate value.
Attributes:
@attribute urls - A tuple of URLs to try fetching the package from.
@attribute local_filename - A local filename to use when saving the
fetched package.
@attribute dist_name - The name of the Python distribution. For example,
the package MySQLdb is included in the distribution named
MySQL-python. This is generally the PyPI name. Defaults to the
name part of the local_filename.
@attribute hex_sum - The hex digest (currently SHA1) of this package
to be used to verify its contents.
@attribute module_name - The installed python module name to be used for
for a version check. Defaults to the lower case class name with
the word Package stripped off.
@attribute extracted_package_path - The path to package directory after
extracting.
@attribute version - The desired minimum package version.
@attribute os_requirements - A dictionary mapping pathname tuples on the
the OS distribution to a likely name of a package the user
needs to install on their system in order to get this file.
One of the files in the tuple must exist.
@attribute name - Read only, the printable name of the package.
@attribute subclasses - This class attribute holds a list of all defined
subclasses. It is constructed dynamically using the metaclass.
"""
# Modules that are meant to be installed in system directory, rather than
# autotest/site-packages. These modules should be skipped if the module
# is already installed in system directory. This prevents an older version
# of the module from being installed in system directory.
SYSTEM_MODULES = ['setuptools']
subclasses = []
urls = ()
local_filename = None
dist_name = None
hex_sum = None
module_name = None
version = None
os_requirements = None
class __metaclass__(type):
"""Any time a subclass is defined, add it to our list."""
def __init__(mcs, name, bases, dict):
if name != 'ExternalPackage' and not name.startswith('_'):
mcs.subclasses.append(mcs)
def __init__(self):
self.verified_package = ''
if not self.module_name:
self.module_name = self.name.lower()
if not self.dist_name and self.local_filename:
self.dist_name = self.local_filename[:self.local_filename.rindex('-')]
self.installed_version = ''
@property
def extracted_package_path(self):
"""Return the package path after extracting.
If the package has assigned its own extracted_package_path, use it.
Or use part of its local_filename as the extracting path.
"""
return self.local_filename[:-len(self._get_extension(
self.local_filename))]
@property
def name(self):
"""Return the class name with any trailing 'Package' stripped off."""
class_name = self.__class__.__name__
if class_name.endswith('Package'):
return class_name[:-len('Package')]
return class_name
def is_needed(self, install_dir):
"""
Check to see if we need to reinstall a package. This is contingent on:
1. Module name: If the name of the module is different from the package,
the class that installs it needs to specify a module_name string,
so we can try importing the module.
2. Installed version: If the module doesn't contain a __version__ the
class that installs it needs to override the
_get_installed_version_from_module method to return an appropriate
version string.
3. Version/Minimum version: The class that installs the package should
contain a version string, and an optional minimum version string.
4. install_dir: If the module exists in a different directory, e.g.,
/usr/lib/python2.7/dist-packages/, the module will be forced to be
installed in install_dir.
@param install_dir: install directory.
@returns True if self.module_name needs to be built and installed.
"""
if not self.module_name or not self.version:
logging.warning('version and module_name required for '
'is_needed() check to work.')
return True
try:
module = __import__(self.module_name)
except ImportError, e:
logging.info("%s isn't present. Will install.", self.module_name)
return True
if (not module.__file__.startswith(install_dir) and
not self.module_name in self.SYSTEM_MODULES):
logging.info('Module %s is installed in %s, rather than %s. The '
'module will be forced to be installed in %s.',
self.module_name, module.__file__, install_dir,
install_dir)
return True
self.installed_version = self._get_installed_version_from_module(module)
if not self.installed_version:
return True
logging.info('imported %s version %s.', self.module_name,
self.installed_version)
if hasattr(self, 'minimum_version'):
return LooseVersion(self.minimum_version) > LooseVersion(
self.installed_version)
else:
return LooseVersion(self.version) > LooseVersion(
self.installed_version)
def _get_installed_version_from_module(self, module):
"""Ask our module its version string and return it or '' if unknown."""
try:
return module.__version__
except AttributeError:
logging.error('could not get version from %s', module)
return ''
def _build_and_install(self, install_dir):
"""Subclasses MUST provide their own implementation."""
raise NotImplementedError
def _build_and_install_current_dir(self, install_dir):
"""
Subclasses that use _build_and_install_from_package() MUST provide
their own implementation of this method.
"""
raise NotImplementedError
def build_and_install(self, install_dir):
"""
Builds and installs the package. It must have been fetched already.
@param install_dir - The package installation directory. If it does
not exist it will be created.
"""
if not self.verified_package:
raise Error('Must call fetch() first. - %s' % self.name)
self._check_os_requirements()
return self._build_and_install(install_dir)
def _check_os_requirements(self):
if not self.os_requirements:
return
failed = False
for file_names, package_name in self.os_requirements.iteritems():
if not any(os.path.exists(file_name) for file_name in file_names):
failed = True
logging.error('Can\'t find %s, %s probably needs it.',
' or '.join(file_names), self.name)
logging.error('Perhaps you need to install something similar '
'to the %s package for OS first.', package_name)
if failed:
raise Error('Missing OS requirements for %s. (see above)' %
self.name)
def _build_and_install_current_dir_setup_py(self, install_dir):
"""For use as a _build_and_install_current_dir implementation."""
egg_path = self._build_egg_using_setup_py(setup_py='setup.py')
if not egg_path:
return False
return self._install_from_egg(install_dir, egg_path)
def _build_and_install_current_dir_setupegg_py(self, install_dir):
"""For use as a _build_and_install_current_dir implementation."""
egg_path = self._build_egg_using_setup_py(setup_py='setupegg.py')
if not egg_path:
return False
return self._install_from_egg(install_dir, egg_path)
def _build_and_install_current_dir_noegg(self, install_dir):
if not self._build_using_setup_py():
return False
return self._install_using_setup_py_and_rsync(install_dir)
def _get_extension(self, package):
"""Get extension of package."""
valid_package_extensions = ['.tar.gz', '.tar.bz2', '.zip']
extension = None
for ext in valid_package_extensions:
if package.endswith(ext):
extension = ext
break
if not extension:
raise Error('Unexpected package file extension on %s' % package)
return extension
def _build_and_install_from_package(self, install_dir):
"""
This method may be used as a _build_and_install() implementation
for subclasses if they implement _build_and_install_current_dir().
Extracts the .tar.gz file, chdirs into the extracted directory
(which is assumed to match the tar filename) and calls
_build_and_isntall_current_dir from there.
Afterwards the build (regardless of failure) extracted .tar.gz
directory is cleaned up.
@returns True on success, False otherwise.
@raises OSError If the expected extraction directory does not exist.
"""
self._extract_compressed_package()
extension = self._get_extension(self.verified_package)
os.chdir(os.path.dirname(self.verified_package))
os.chdir(self.extracted_package_path)
extracted_dir = os.getcwd()
try:
return self._build_and_install_current_dir(install_dir)
finally:
os.chdir(os.path.join(extracted_dir, '..'))
shutil.rmtree(extracted_dir)
def _extract_compressed_package(self):
"""Extract the fetched compressed .tar or .zip within its directory."""
if not self.verified_package:
raise Error('Package must have been fetched first.')
os.chdir(os.path.dirname(self.verified_package))
if self.verified_package.endswith('gz'):
status = system("tar -xzf '%s'" % self.verified_package)
elif self.verified_package.endswith('bz2'):
status = system("tar -xjf '%s'" % self.verified_package)
elif self.verified_package.endswith('zip'):
status = system("unzip '%s'" % self.verified_package)
else:
raise Error('Unknown compression suffix on %s.' %
self.verified_package)
if status:
raise Error('tar failed with %s' % (status,))
def _build_using_setup_py(self, setup_py='setup.py'):
"""
Assuming the cwd is the extracted python package, execute a simple
python setup.py build.
@param setup_py - The name of the setup.py file to execute.
@returns True on success, False otherwise.
"""
if not os.path.exists(setup_py):
raise Error('%s does not exist in %s' % (setup_py, os.getcwd()))
status = system("'%s' %s build" % (sys.executable, setup_py))
if status:
logging.error('%s build failed.', self.name)
return False
return True
def _build_egg_using_setup_py(self, setup_py='setup.py'):
"""
Assuming the cwd is the extracted python package, execute a simple
python setup.py bdist_egg.
@param setup_py - The name of the setup.py file to execute.
@returns The relative path to the resulting egg file or '' on failure.
"""
if not os.path.exists(setup_py):
raise Error('%s does not exist in %s' % (setup_py, os.getcwd()))
egg_subdir = 'dist'
if os.path.isdir(egg_subdir):
shutil.rmtree(egg_subdir)
status = system("'%s' %s bdist_egg" % (sys.executable, setup_py))
if status:
logging.error('bdist_egg of setuptools failed.')
return ''
# I've never seen a bdist_egg lay multiple .egg files.
for filename in os.listdir(egg_subdir):
if filename.endswith('.egg'):
return os.path.join(egg_subdir, filename)
def _install_from_egg(self, install_dir, egg_path):
"""
Install a module from an egg file by unzipping the necessary parts
into install_dir.
@param install_dir - The installation directory.
@param egg_path - The pathname of the egg file.
"""
status = system("unzip -q -o -d '%s' '%s'" % (install_dir, egg_path))
if status:
logging.error('unzip of %s failed', egg_path)
return False
egg_info_dir = os.path.join(install_dir, 'EGG-INFO')
if os.path.isdir(egg_info_dir):
egg_info_new_path = self._get_egg_info_path(install_dir)
if egg_info_new_path:
if os.path.exists(egg_info_new_path):
shutil.rmtree(egg_info_new_path)
os.rename(egg_info_dir, egg_info_new_path)
else:
shutil.rmtree(egg_info_dir)
return True
def _get_egg_info_path(self, install_dir):
"""Get egg-info path for this package.
Example path: install_dir/MySQL_python-1.2.3.egg-info
"""
if self.dist_name:
egg_info_name_part = self.dist_name.replace('-', '_')
if self.version:
egg_info_filename = '%s-%s.egg-info' % (egg_info_name_part,
self.version)
else:
egg_info_filename = '%s.egg-info' % (egg_info_name_part,)
return os.path.join(install_dir, egg_info_filename)
else:
return None
def _get_temp_dir(self):
return tempfile.mkdtemp(dir='/var/tmp')
def _site_packages_path(self, temp_dir):
# This makes assumptions about what python setup.py install
# does when given a prefix. Is this always correct?
python_xy = 'python%s' % sys.version[:3]
return os.path.join(temp_dir, 'lib', python_xy, 'site-packages')
def _rsync (self, temp_site_dir, install_dir):
"""Rsync contents. """
status = system("rsync -r '%s/' '%s/'" %
(os.path.normpath(temp_site_dir),
os.path.normpath(install_dir)))
if status:
logging.error('%s rsync to install_dir failed.', self.name)
return False
return True
def _install_using_setup_py_and_rsync(self, install_dir,
setup_py='setup.py',
temp_dir=None):
"""
Assuming the cwd is the extracted python package, execute a simple:
python setup.py install --prefix=BLA
BLA will be a temporary directory that everything installed will
be picked out of and rsynced to the appropriate place under
install_dir afterwards.
Afterwards, it deconstructs the extra lib/pythonX.Y/site-packages/
directory tree that setuptools created and moves all installed
site-packages directly up into install_dir itself.
@param install_dir the directory for the install to happen under.
@param setup_py - The name of the setup.py file to execute.
@returns True on success, False otherwise.
"""
if not os.path.exists(setup_py):
raise Error('%s does not exist in %s' % (setup_py, os.getcwd()))
if temp_dir is None:
temp_dir = self._get_temp_dir()
try:
status = system("'%s' %s install --no-compile --prefix='%s'"
% (sys.executable, setup_py, temp_dir))
if status:
logging.error('%s install failed.', self.name)
return False
if os.path.isdir(os.path.join(temp_dir, 'lib')):
# NOTE: This ignores anything outside of the lib/ dir that
# was installed.
temp_site_dir = self._site_packages_path(temp_dir)
else:
temp_site_dir = temp_dir
return self._rsync(temp_site_dir, install_dir)
finally:
shutil.rmtree(temp_dir)
def _build_using_make(self, install_dir):
"""Build the current package using configure/make.
@returns True on success, False otherwise.
"""
install_prefix = os.path.join(install_dir, 'usr', 'local')
status = system('./configure --prefix=%s' % install_prefix)
if status:
logging.error('./configure failed for %s', self.name)
return False
status = system('make')
if status:
logging.error('make failed for %s', self.name)
return False
status = system('make check')
if status:
logging.error('make check failed for %s', self.name)
return False
return True
def _install_using_make(self):
"""Install the current package using make install.
Assumes the install path was set up while running ./configure (in
_build_using_make()).
@returns True on success, False otherwise.
"""
status = system('make install')
return status == 0
def fetch(self, dest_dir):
"""
Fetch the package from one its URLs and save it in dest_dir.
If the the package already exists in dest_dir and the checksum
matches this code will not fetch it again.
Sets the 'verified_package' attribute with the destination pathname.
@param dest_dir - The destination directory to save the local file.
If it does not exist it will be created.
@returns A boolean indicating if we the package is now in dest_dir.
@raises FetchError - When something unexpected happens.
"""
if not os.path.exists(dest_dir):
os.makedirs(dest_dir)
local_path = os.path.join(dest_dir, self.local_filename)
# If the package exists, verify its checksum and be happy if it is good.
if os.path.exists(local_path):
actual_hex_sum = _checksum_file(local_path)
if self.hex_sum == actual_hex_sum:
logging.info('Good checksum for existing %s package.',
self.name)
self.verified_package = local_path
return True
logging.warning('Bad checksum for existing %s package. '
'Re-downloading', self.name)
os.rename(local_path, local_path + '.wrong-checksum')
# Download the package from one of its urls, rejecting any if the
# checksum does not match.
for url in self.urls:
logging.info('Fetching %s', url)
try:
url_file = urllib2.urlopen(url)
except (urllib2.URLError, EnvironmentError):
logging.warning('Could not fetch %s package from %s.',
self.name, url)
continue
data_length = int(url_file.info().get('Content-Length',
_MAX_PACKAGE_SIZE))
if data_length <= 0 or data_length > _MAX_PACKAGE_SIZE:
raise FetchError('%s from %s fails Content-Length %d '
'sanity check.' % (self.name, url,
data_length))
checksum = utils.hash('sha1')
total_read = 0
output = open(local_path, 'wb')
try:
while total_read < data_length:
data = url_file.read(_READ_SIZE)
if not data:
break
output.write(data)
checksum.update(data)
total_read += len(data)
finally:
output.close()
if self.hex_sum != checksum.hexdigest():
logging.warning('Bad checksum for %s fetched from %s.',
self.name, url)
logging.warning('Got %s', checksum.hexdigest())
logging.warning('Expected %s', self.hex_sum)
os.unlink(local_path)
continue
logging.info('Good checksum.')
self.verified_package = local_path
return True
else:
return False
# NOTE: This class definition must come -before- all other ExternalPackage
# classes that need to use this version of setuptools so that is is inserted
# into the ExternalPackage.subclasses list before them.
class SetuptoolsPackage(ExternalPackage):
"""setuptools package"""
# For all known setuptools releases a string compare works for the
# version string. Hopefully they never release a 0.10. (Their own
# version comparison code would break if they did.)
# Any system with setuptools > 18.0.1 is fine. If none installed, then
# try to install the latest found on the upstream.
minimum_version = '18.0.1'
version = '18.0.1'
urls = (_CHROMEOS_MIRROR + 'setuptools-%s.tar.gz' % (version,),)
local_filename = 'setuptools-%s.tar.gz' % version
hex_sum = 'ebc4fe81b7f6d61d923d9519f589903824044f52'
SUDO_SLEEP_DELAY = 15
def _build_and_install(self, install_dir):
"""Install setuptools on the system."""
logging.info('NOTE: setuptools install does not use install_dir.')
return self._build_and_install_from_package(install_dir)
def _build_and_install_current_dir(self, install_dir):
egg_path = self._build_egg_using_setup_py()
if not egg_path:
return False
print '!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\n'
print 'About to run sudo to install setuptools', self.version
print 'on your system for use by', sys.executable, '\n'
print '!! ^C within', self.SUDO_SLEEP_DELAY, 'seconds to abort.\n'
time.sleep(self.SUDO_SLEEP_DELAY)
# Copy the egg to the local filesystem /var/tmp so that root can
# access it properly (avoid NFS squashroot issues).
temp_dir = self._get_temp_dir()
try:
shutil.copy(egg_path, temp_dir)
egg_name = os.path.split(egg_path)[1]
temp_egg = os.path.join(temp_dir, egg_name)
p = subprocess.Popen(['sudo', '/bin/sh', temp_egg],
stdout=subprocess.PIPE)
regex = re.compile('Copying (.*?) to (.*?)\n')
match = regex.search(p.communicate()[0])
status = p.wait()
if match:
compiled = os.path.join(match.group(2), match.group(1))
os.system("sudo chmod a+r '%s'" % compiled)
finally:
shutil.rmtree(temp_dir)
if status:
logging.error('install of setuptools from egg failed.')
return False
return True
class MySQLdbPackage(ExternalPackage):
"""mysql package, used in scheduler."""
module_name = 'MySQLdb'
version = '1.2.3'
local_filename = 'MySQL-python-%s.tar.gz' % version
urls = ('http://commondatastorage.googleapis.com/chromeos-mirror/gentoo/'
'distfiles/%s' % local_filename,)
hex_sum = '3511bb8c57c6016eeafa531d5c3ea4b548915e3c'
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
def _build_and_install(self, install_dir):
if not os.path.exists('/usr/bin/mysql_config'):
error_msg = '''\
You need to install /usr/bin/mysql_config.
On recent Debian based distros, run: \
sudo apt-get install libmariadbclient-dev-compat
On older Debian based distros, run: sudo apt-get install libmysqlclient15-dev
'''
logging.error(error_msg)
return False, error_msg
return self._build_and_install_from_package(install_dir)
class DjangoPackage(ExternalPackage):
"""django package."""
version = '1.5.1'
local_filename = 'Django-%s.tar.gz' % version
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '0ab97b90c4c79636e56337f426f1e875faccbba1'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_noegg)
def _get_installed_version_from_module(self, module):
try:
return module.get_version().split()[0]
except AttributeError:
return '0.9.6'
class NumpyPackage(ExternalPackage):
"""numpy package, required by matploglib."""
version = '1.7.0'
local_filename = 'numpy-%s.tar.gz' % version
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'ba328985f20390b0f969a5be2a6e1141d5752cf9'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setupegg_py)
class JsonRPCLib(ExternalPackage):
"""jsonrpclib package"""
version = '0.1.3'
module_name = 'jsonrpclib'
local_filename = '%s-%s.tar.gz' % (module_name, version)
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '431714ed19ab677f641ce5d678a6a95016f5c452'
def _get_installed_version_from_module(self, module):
# jsonrpclib doesn't contain a proper version
return self.version
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_noegg)
class GwtPackage(ExternalPackage):
"""Fetch and extract a local copy of GWT used to build the frontend."""
version = '2.3.0'
local_filename = 'gwt-%s.zip' % version
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'd51fce9166e6b31349659ffca89baf93e39bc84b'
name = 'gwt'
about_filename = 'about.txt'
module_name = None # Not a Python module.
def is_needed(self, install_dir):
gwt_dir = os.path.join(install_dir, self.name)
about_file = os.path.join(install_dir, self.name, self.about_filename)
if not os.path.exists(gwt_dir) or not os.path.exists(about_file):
logging.info('gwt not installed for autotest')
return True
f = open(about_file, 'r')
version_line = f.readline()
f.close()
match = re.match(r'Google Web Toolkit (.*)', version_line)
if not match:
logging.info('did not find gwt version')
return True
logging.info('found gwt version %s', match.group(1))
return match.group(1) != self.version
def _build_and_install(self, install_dir):
os.chdir(install_dir)
self._extract_compressed_package()
extracted_dir = self.local_filename[:-len('.zip')]
target_dir = os.path.join(install_dir, self.name)
if os.path.exists(target_dir):
shutil.rmtree(target_dir)
os.rename(extracted_dir, target_dir)
return True
class PyudevPackage(ExternalPackage):
"""
pyudev module
Used in unittests.
"""
version = '0.16.1'
url_filename = 'pyudev-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'b36bc5c553ce9b56d32a5e45063a2c88156771c0'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class PyMoxPackage(ExternalPackage):
"""
mox module
Used in unittests.
"""
module_name = 'mox'
version = '0.5.3'
# Note: url_filename does not match local_filename, because of
# an uncontrolled fork at some point in time of mox versions.
url_filename = 'mox-%s-autotest.tar.gz' % version
local_filename = 'mox-%s.tar.gz' % version
urls = (_CHROMEOS_MIRROR + url_filename,)
hex_sum = '1c502d2c0a8aefbba2c7f385a83d33e7d822452a'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_noegg)
def _get_installed_version_from_module(self, module):
# mox doesn't contain a proper version
return self.version
class PySeleniumPackage(ExternalPackage):
"""
selenium module
Used in wifi_interop suite.
"""
module_name = 'selenium'
version = '2.37.2'
url_filename = 'selenium-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '66946d5349e36d946daaad625c83c30c11609e36'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class FaultHandlerPackage(ExternalPackage):
"""
faulthandler module
"""
module_name = 'faulthandler'
version = '2.3'
url_filename = '%s-%s.tar.gz' % (module_name, version)
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'efb30c068414fba9df892e48fcf86170cbf53589'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_noegg)
class PsutilPackage(ExternalPackage):
"""
psutil module
"""
module_name = 'psutil'
version = '2.1.1'
url_filename = '%s-%s.tar.gz' % (module_name, version)
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '0c20a20ed316e69f2b0881530439213988229916'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class ElasticSearchPackage(ExternalPackage):
"""elasticsearch-py package."""
version = '1.6.0'
url_filename = 'elasticsearch-%s.tar.gz' % version
local_filename = url_filename
urls = ('https://pypi.python.org/packages/source/e/elasticsearch/%s' %
(url_filename),)
hex_sum = '3e676c96f47935b1f52df82df3969564bd356b1c'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
def _get_installed_version_from_module(self, module):
# Elastic's version format is like tuple (1, 6, 0), which needs to be
# transferred to 1.6.0.
try:
return '.'.join(str(i) for i in module.__version__)
except:
return self.version
class Urllib3Package(ExternalPackage):
"""elasticsearch-py package."""
version = '1.9'
url_filename = 'urllib3-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '9522197efb2a2b49ce804de3a515f06d97b6602f'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class ImagingLibraryPackage(ExternalPackage):
"""Python Imaging Library (PIL)."""
version = '1.1.7'
url_filename = 'Imaging-%s.tar.gz' % version
local_filename = url_filename
urls = ('http://commondatastorage.googleapis.com/chromeos-mirror/gentoo/'
'distfiles/%s' % url_filename,)
hex_sum = '76c37504251171fda8da8e63ecb8bc42a69a5c81'
def _build_and_install(self, install_dir):
#The path of zlib library might be different from what PIL setup.py is
#expected. Following change does the best attempt to link the library
#to a path PIL setup.py will try.
libz_possible_path = '/usr/lib/x86_64-linux-gnu/libz.so'
libz_expected_path = '/usr/lib/libz.so'
if (os.path.exists(libz_possible_path) and
not os.path.exists(libz_expected_path)):
utils.run('sudo ln -s %s %s' %
(libz_possible_path, libz_expected_path))
return self._build_and_install_from_package(install_dir)
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_noegg)
class AstroidPackage(ExternalPackage):
"""astroid package."""
version = '1.5.3'
url_filename = 'astroid-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'e654225ab5bd2788e5e246b156910990bf33cde6'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class LazyObjectProxyPackage(ExternalPackage):
"""lazy-object-proxy package (dependency for astroid)."""
version = '1.3.1'
url_filename = 'lazy-object-proxy-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '984828d8f672986ca926373986214d7057b772fb'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class SingleDispatchPackage(ExternalPackage):
"""singledispatch package (dependency for astroid)."""
version = '3.4.0.3'
url_filename = 'singledispatch-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'f93241b06754a612af8bb7aa208c4d1805637022'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class Enum34Package(ExternalPackage):
"""enum34 package (dependency for astroid)."""
version = '1.1.6'
url_filename = 'enum34-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '014ef5878333ff91099893d615192c8cd0b1525a'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class WraptPackage(ExternalPackage):
"""wrapt package (dependency for astroid)."""
version = '1.10.10'
url_filename = 'wrapt-%s.tar.gz' % version
local_filename = url_filename
#md5=97365e906afa8b431f266866ec4e2e18
urls = ('https://pypi.python.org/packages/a3/bb/'
'525e9de0a220060394f4aa34fdf6200853581803d92714ae41fc3556e7d7/%s' %
(url_filename),)
hex_sum = '6be4f1bb50db879863f4247692360eb830a3eb33'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_noegg)
class SixPackage(ExternalPackage):
"""six package (dependency for astroid)."""
version = '1.10.0'
url_filename = 'six-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '30d480d2e352e8e4c2aae042cf1bf33368ff0920'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class LruCachePackage(ExternalPackage):
"""backports.functools_lru_cache package (dependency for astroid)."""
version = '1.4'
url_filename = 'backports.functools_lru_cache-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '8a546e7887e961c2873c9b053f4e2cd2a96bd71d'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class LogilabCommonPackage(ExternalPackage):
"""logilab-common package."""
version = '1.2.2'
module_name = 'logilab'
url_filename = 'logilab-common-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'ecad2d10c31dcf183c8bed87b6ec35e7ed397d27'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class PyLintPackage(ExternalPackage):
"""pylint package."""
version = '1.7.2'
url_filename = 'pylint-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '42d8b9394e5a485377ae128b01350f25d8b131e0'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class ConfigParserPackage(ExternalPackage):
"""configparser package (dependency for pylint)."""
version = '3.5.0'
url_filename = 'configparser-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '8ee6b29c6a11977c0e094da1d4f5f71e7e7ac78b'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class IsortPackage(ExternalPackage):
"""isort package (dependency for pylint)."""
version = '4.2.15'
url_filename = 'isort-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'acacc36e476b70e13e6fda812c193f4c3c187781'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class Pytz(ExternalPackage):
"""Pytz package."""
version = '2016.10'
url_filename = 'pytz-%s.tar.gz' % version
local_filename = url_filename
#md5=cc9f16ba436efabdcef3c4d32ae4919c
urls = ('https://pypi.python.org/packages/42/00/'
'5c89fc6c9b305df84def61863528e899e9dccb196f8438f6cbe960758fc5/%s' %
(url_filename),)
hex_sum = '8d63f1e9b1ee862841b990a7d8ad1d4508d9f0be'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class Tzlocal(ExternalPackage):
"""Tzlocal package."""
version = '1.3'
url_filename = 'tzlocal-%s.tar.gz' % version
local_filename = url_filename
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = '730e9d7112335865a1dcfabec69c8c3086be424f'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_setup_py)
class PyYAMLPackage(ExternalPackage):
"""pyyaml package."""
version = '3.12'
local_filename = 'PyYAML-%s.tar.gz' % version
urls = (_CHROMEOS_MIRROR + local_filename,)
hex_sum = 'cb7fd3e58c129494ee86e41baedfec69eb7dafbe'
_build_and_install = ExternalPackage._build_and_install_from_package
_build_and_install_current_dir = (
ExternalPackage._build_and_install_current_dir_noegg)
class _ExternalGitRepo(ExternalPackage):
"""
Parent class for any package which needs to pull a git repo.
This class inherits from ExternalPackage only so we can sync git
repos through the build_externals script. We do not reuse any of
ExternalPackage's other methods. Any package that needs a git repo
should subclass this and override build_and_install or fetch as
they see appropriate.
"""
os_requirements = {('/usr/bin/git') : 'git-core'}
# All the chromiumos projects used on the lab servers should have a 'prod'
# branch used to track the software version deployed in prod.
PROD_BRANCH = 'prod'
MASTER_BRANCH = 'master'
def is_needed(self, unused_install_dir):
"""Tell build_externals that we need to fetch."""
# TODO(beeps): check if we're already upto date.
return True
def build_and_install(self, unused_install_dir):
"""
Fall through method to install a package.
Overwritten in base classes to pull a git repo.
"""
raise NotImplementedError
def fetch(self, unused_dest_dir):
"""Fallthrough method to fetch a package."""
return True
class HdctoolsRepo(_ExternalGitRepo):
"""Clones or updates the hdctools repo."""
module_name = 'servo'
temp_hdctools_dir = tempfile.mktemp(suffix='hdctools')
_GIT_URL = ('https://chromium.googlesource.com/'
'chromiumos/third_party/hdctools')
def fetch(self, unused_dest_dir):
"""
Fetch repo to a temporary location.
We use an intermediate temp directory to stage our
installation because we only care about the servo package.
If we can't get at the top commit hash after fetching
something is wrong. This can happen when we've cloned/pulled
an empty repo. Not something we expect to do.
@parma unused_dest_dir: passed in because we inherit from
ExternalPackage.
@return: True if repo sync was successful.
"""
git_repo = revision_control.GitRepo(
self.temp_hdctools_dir,
self._GIT_URL,
None,
abs_work_tree=self.temp_hdctools_dir)
git_repo.reinit_repo_at(self.PROD_BRANCH)
if git_repo.get_latest_commit_hash():
return True
return False
def build_and_install(self, install_dir):
"""Reach into the hdctools repo and rsync only the servo directory."""
servo_dir = os.path.join(self.temp_hdctools_dir, 'servo')
if not os.path.exists(servo_dir):
return False
rv = self._rsync(servo_dir, os.path.join(install_dir, 'servo'))
shutil.rmtree(self.temp_hdctools_dir)
return rv
class ChromiteRepo(_ExternalGitRepo):
"""Clones or updates the chromite repo."""
_GIT_URL = ('https://chromium.googlesource.com/chromiumos/chromite')
def build_and_install(self, install_dir, master_branch=False):
"""
Clone if the repo isn't initialized, pull clean bits if it is.
Unlike it's hdctools counterpart the chromite repo clones master
directly into site-packages. It doesn't use an intermediate temp
directory because it doesn't need installation.
@param install_dir: destination directory for chromite installation.
@param master_branch: if True, install master branch. Otherwise,
install prod branch.
"""
init_branch = (self.MASTER_BRANCH if master_branch
else self.PROD_BRANCH)
local_chromite_dir = os.path.join(install_dir, 'chromite')
git_repo = revision_control.GitRepo(
local_chromite_dir,
self._GIT_URL,
abs_work_tree=local_chromite_dir)
git_repo.reinit_repo_at(init_branch)
if git_repo.get_latest_commit_hash():
return True
return False
class BtsocketRepo(_ExternalGitRepo):
"""Clones or updates the btsocket repo."""
_GIT_URL = ('https://chromium.googlesource.com/'
'chromiumos/platform/btsocket')
def fetch(self, unused_dest_dir):
"""
Fetch repo to a temporary location.
We use an intermediate temp directory because we have to build an
egg for installation. If we can't get at the top commit hash after
fetching something is wrong. This can happen when we've cloned/pulled
an empty repo. Not something we expect to do.
@parma unused_dest_dir: passed in because we inherit from
ExternalPackage.
@return: True if repo sync was successful.
"""
self.temp_btsocket_dir = autotemp.tempdir(unique_id='btsocket')
try:
git_repo = revision_control.GitRepo(
self.temp_btsocket_dir.name,
self._GIT_URL,
None,
abs_work_tree=self.temp_btsocket_dir.name)
git_repo.reinit_repo_at(self.PROD_BRANCH)
if git_repo.get_latest_commit_hash():
return True
except:
self.temp_btsocket_dir.clean()
raise
self.temp_btsocket_dir.clean()
return False
def build_and_install(self, install_dir):
"""
Install the btsocket module using setup.py
@param install_dir: Target installation directory.
@return: A boolean indicating success of failure.
"""
work_dir = os.getcwd()
try:
os.chdir(self.temp_btsocket_dir.name)
rv = self._build_and_install_current_dir_setup_py(install_dir)
finally:
os.chdir(work_dir)
self.temp_btsocket_dir.clean()
return rv