diff options
Diffstat (limited to 'bower_components/bootstrap/test-infra/s3_cache.py')
| -rwxr-xr-x | bower_components/bootstrap/test-infra/s3_cache.py | 107 |
1 files changed, 0 insertions, 107 deletions
diff --git a/bower_components/bootstrap/test-infra/s3_cache.py b/bower_components/bootstrap/test-infra/s3_cache.py deleted file mode 100755 index 472963a..0000000 --- a/bower_components/bootstrap/test-infra/s3_cache.py +++ /dev/null @@ -1,107 +0,0 @@ -#!/usr/bin/env python2.7 -from __future__ import absolute_import, unicode_literals, print_function, division - -from sys import argv -from os import environ, stat, remove as _delete_file -from os.path import isfile, dirname, basename, abspath -from hashlib import sha256 -from subprocess import check_call as run - -from boto.s3.connection import S3Connection -from boto.s3.key import Key -from boto.exception import S3ResponseError - - -NEED_TO_UPLOAD_MARKER = '.need-to-upload' -BYTES_PER_MB = 1024 * 1024 -try: - BUCKET_NAME = environ['TWBS_S3_BUCKET'] -except KeyError: - raise SystemExit("TWBS_S3_BUCKET environment variable not set!") - - -def _sha256_of_file(filename): - hasher = sha256() - with open(filename, 'rb') as input_file: - hasher.update(input_file.read()) - file_hash = hasher.hexdigest() - print('sha256({}) = {}'.format(filename, file_hash)) - return file_hash - - -def _delete_file_quietly(filename): - try: - _delete_file(filename) - except (OSError, IOError): - pass - - -def _tarball_size(directory): - kib = stat(_tarball_filename_for(directory)).st_size // BYTES_PER_MB - return "{} MiB".format(kib) - - -def _tarball_filename_for(directory): - return abspath('./{}.tar.gz'.format(basename(directory))) - - -def _create_tarball(directory): - print("Creating tarball of {}...".format(directory)) - run(['tar', '-czf', _tarball_filename_for(directory), '-C', dirname(directory), basename(directory)]) - - -def _extract_tarball(directory): - print("Extracting tarball of {}...".format(directory)) - run(['tar', '-xzf', _tarball_filename_for(directory), '-C', dirname(directory)]) - - -def download(directory): - _delete_file_quietly(NEED_TO_UPLOAD_MARKER) - try: - print("Downloading {} tarball from S3...".format(friendly_name)) - key.get_contents_to_filename(_tarball_filename_for(directory)) - except S3ResponseError as err: - open(NEED_TO_UPLOAD_MARKER, 'a').close() - print(err) - raise SystemExit("Cached {} download failed!".format(friendly_name)) - print("Downloaded {}.".format(_tarball_size(directory))) - _extract_tarball(directory) - print("{} successfully installed from cache.".format(friendly_name)) - - -def upload(directory): - _create_tarball(directory) - print("Uploading {} tarball to S3... ({})".format(friendly_name, _tarball_size(directory))) - key.set_contents_from_filename(_tarball_filename_for(directory)) - print("{} cache successfully updated.".format(friendly_name)) - _delete_file_quietly(NEED_TO_UPLOAD_MARKER) - - -if __name__ == '__main__': - # Uses environment variables: - # AWS_ACCESS_KEY_ID -- AWS Access Key ID - # AWS_SECRET_ACCESS_KEY -- AWS Secret Access Key - argv.pop(0) - if len(argv) != 4: - raise SystemExit("USAGE: s3_cache.py <download | upload> <friendly name> <dependencies file> <directory>") - mode, friendly_name, dependencies_file, directory = argv - - conn = S3Connection() - bucket = conn.lookup(BUCKET_NAME, validate=False) - if bucket is None: - raise SystemExit("Could not access bucket!") - - dependencies_file_hash = _sha256_of_file(dependencies_file) - - key = Key(bucket, dependencies_file_hash) - key.storage_class = 'REDUCED_REDUNDANCY' - - if mode == 'download': - download(directory) - elif mode == 'upload': - if isfile(NEED_TO_UPLOAD_MARKER): # FIXME - upload(directory) - else: - print("No need to upload anything.") - else: - raise SystemExit("Unrecognized mode {!r}".format(mode)) |
