from sys import argv
from os import environ, stat, remove as _delete_file
-from os.path import isfile
+from os.path import isfile, dirname, basename, abspath
from hashlib import sha256
from subprocess import check_call as run
from boto.exception import S3ResponseError
-NODE_MODULES_TARBALL = 'node_modules.tar.gz'
NEED_TO_UPLOAD_MARKER = '.need-to-upload'
BYTES_PER_MB = 1024 * 1024
try:
hasher = sha256()
with open(filename, 'rb') as input_file:
hasher.update(input_file.read())
- return hasher.hexdigest()
+ file_hash = hasher.hexdigest()
+ print('sha256({}) = {}'.format(filename, file_hash))
+ return file_hash
def _delete_file_quietly(filename):
pass
-def _tarball_size():
- kib = stat(NODE_MODULES_TARBALL).st_size // BYTES_PER_MB
+def _tarball_size(directory):
+ kib = stat(_tarball_filename_for(directory)).st_size // BYTES_PER_MB
return "{} MiB".format(kib)
+def _tarball_filename_for(directory):
+ return abspath('./{}.tar.gz'.format(basename(directory)))
+
+
+def _create_tarball(directory):
+ print("Creating tarball of {}...".format(directory))
+ run(['tar', '-czf', _tarball_filename_for(directory), '-C', dirname(directory), basename(directory)])
+
+
+def _extract_tarball(directory):
+ print("Extracting tarball of {}...".format(directory))
+ run(['tar', '-xzf', _tarball_filename_for(directory), '-C', dirname(directory)])
+
+
+def download(directory):
+ _delete_file_quietly(NEED_TO_UPLOAD_MARKER)
+ try:
+ print("Downloading {} tarball from S3...".format(basename(directory)))
+ key.get_contents_to_filename(_tarball_filename_for(directory))
+ except S3ResponseError as err:
+ open(NEED_TO_UPLOAD_MARKER, 'a').close()
+ print(err)
+ raise SystemExit("Cached {} download failed!".format(basename(directory)))
+ print("Downloaded {}.".format(_tarball_size(directory)))
+ _extract_tarball(directory)
+ print("{} successfully installed from cache.".format(directory))
+
+
+def upload(directory):
+ _create_tarball(directory)
+ print("Uploading {} tarball to S3... ({})".format(basename(directory), _tarball_size(directory)))
+ key.set_contents_from_filename(_tarball_filename_for(directory))
+ print("{} cache successfully updated.".format(directory))
+ _delete_file_quietly(NEED_TO_UPLOAD_MARKER)
+
+
if __name__ == '__main__':
# Uses environment variables:
# AWS_ACCESS_KEY_ID - AWS Access Key ID
# AWS_SECRET_ACCESS_KEY - AWS Secret Access Key
argv.pop(0)
- if len(argv) != 1:
- raise SystemExit("USAGE: node_modules_cache.py <download | upload>")
- mode = argv.pop()
+ if len(argv) != 3:
+ raise SystemExit("USAGE: node_modules_cache.py <download | upload> <dependencies file> <directory>")
+ mode, dependencies_file, directory = argv
conn = S3Connection()
bucket = conn.lookup(BUCKET_NAME)
if bucket is None:
raise SystemExit("Could not access bucket!")
- package_json_hash = _sha256_of_file('package.json')
- print('sha256(package.json) = ' + package_json_hash)
+ dependencies_file_hash = _sha256_of_file(dependencies_file)
- key = Key(bucket, package_json_hash)
+ key = Key(bucket, dependencies_file_hash)
key.storage_class = 'REDUCED_REDUNDANCY'
if mode == 'download':
- _delete_file_quietly(NEED_TO_UPLOAD_MARKER)
- try:
- print("Downloading tarball from S3...")
- key.get_contents_to_filename(NODE_MODULES_TARBALL)
- except S3ResponseError as err:
- open(NEED_TO_UPLOAD_MARKER, 'a').close()
- print(err)
- raise SystemExit("Cached node_modules download failed!")
- print("Downloaded {}.".format(_tarball_size()))
- print("Extracting tarball...")
- run(['tar', 'xzf', NODE_MODULES_TARBALL])
- print("node_modules successfully installed from cache.")
+ download(directory)
elif mode == 'upload':
- if isfile(NEED_TO_UPLOAD_MARKER):
- print("Creating tarball...")
- run(['tar', 'czf', NODE_MODULES_TARBALL, 'node_modules'])
- print("Uploading tarball to S3... ({})".format(_tarball_size()))
- key.set_contents_from_filename(NODE_MODULES_TARBALL)
- print("node_modules cache successfully updated.")
- _delete_file_quietly(NEED_TO_UPLOAD_MARKER)
+ if isfile(NEED_TO_UPLOAD_MARKER): # FIXME
+ upload(directory)
else:
print("No need to upload anything.")
else: