| // Copyright (c) 2018, Compiler Explorer Authors |
| // |
| // All rights reserved. |
| // |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are met: |
| // |
| // * Redistributions of source code must retain the above copyright notice, |
| // this list of conditions and the following disclaimer. |
| // * Redistributions in binary form must reproduce the above copyright |
| // notice, this list of conditions and the following disclaimer in the |
| // documentation and/or other materials provided with the distribution. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| // AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| // ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE |
| // LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| // CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| // SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| // INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| // CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| // ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| // POSSIBILITY OF SUCH DAMAGE. |
| |
| const StorageBase = require('./storage').StorageBase, |
| logger = require('../logger').logger, |
| AWS = require('aws-sdk'), |
| _ = require('underscore'), |
| S3Bucket = require('../s3-handler'), |
| anonymizeIp = require('../utils').anonymizeIp; |
| |
| |
| const MIN_STORED_ID_LENGTH = 6; |
| |
| class StorageS3 extends StorageBase { |
| constructor(httpRootDir, compilerProps, awsProps) { |
| super(httpRootDir, compilerProps); |
| const region = awsProps('region'); |
| const bucket = awsProps('storageBucket'); |
| this.prefix = awsProps('storagePrefix'); |
| this.table = awsProps('storageDynamoTable'); |
| logger.info(`Using s3 storage solution on ${region}, bucket ${bucket}, ` + |
| `prefix ${this.prefix}, dynamo table ${this.table}`); |
| AWS.config.update({region: region}); |
| this.s3 = new S3Bucket(bucket, region); |
| this.dynamoDb = new AWS.DynamoDB(); |
| } |
| |
| async storeItem(item, req) { |
| logger.info(`Storing item ${item.prefix}`); |
| const now = new Date(); |
| let ip = req.get('X-Forwarded-For') || anonymizeIp(req.ip); |
| const commaIndex = ip.indexOf(','); |
| if (commaIndex > 0) { |
| // Anonymize only client IP |
| ip = `${anonymizeIp(ip.substring(0, commaIndex))}${ip.substring(commaIndex, ip.length)}`; |
| } |
| now.setSeconds(0, 0); |
| try { |
| await Promise.all([ |
| this.dynamoDb.putItem({ |
| TableName: this.table, |
| Item: { |
| prefix: {S: item.prefix}, |
| unique_subhash: {S: item.uniqueSubHash}, |
| full_hash: {S: item.fullHash}, |
| stats: { |
| M: {clicks: {N: '0'}}, |
| }, |
| creation_ip: {S: ip}, |
| creation_date: {S: now.toISOString()}, |
| }, |
| }).promise(), |
| this.s3.put(item.fullHash, item.config, this.prefix, {}), |
| ]); |
| return item; |
| } catch (err) { |
| logger.error('Unable to store item', {item, err}); |
| throw err; |
| } |
| } |
| |
| async findUniqueSubhash(hash) { |
| const prefix = hash.substring(0, MIN_STORED_ID_LENGTH); |
| const data = await this.dynamoDb.query({ |
| TableName: this.table, |
| ProjectionExpression: 'unique_subhash, full_hash', |
| KeyConditionExpression: 'prefix = :prefix', |
| ExpressionAttributeValues: {':prefix': {S: prefix}}, |
| }).promise(); |
| const subHashes = _.chain(data.Items) |
| .pluck('unique_subhash') |
| .pluck('S') |
| .value(); |
| const fullHashes = _.chain(data.Items) |
| .pluck('full_hash') |
| .pluck('S') |
| .value(); |
| for (let i = MIN_STORED_ID_LENGTH; i < hash.length - 1; i++) { |
| let subHash = hash.substring(0, i); |
| // Check if the current base is present in the subHashes array |
| const index = _.indexOf(subHashes, subHash, true); |
| if (index === -1) { |
| // Current base is not present, we have a new config in our hands |
| return { |
| prefix: prefix, |
| uniqueSubHash: subHash, |
| alreadyPresent: false, |
| }; |
| } else { |
| const itemHash = fullHashes[index]; |
| /* If the hashes coincide, it means this config has already been stored. |
| * Else, keep looking |
| */ |
| if (itemHash === hash) { |
| return { |
| prefix: prefix, |
| uniqueSubHash: subHash, |
| alreadyPresent: true, |
| }; |
| } |
| } |
| } |
| throw new Error(`Could not find unique subhash for hash "${hash}"`); |
| } |
| |
| getKeyStruct(id) { |
| return { |
| prefix: {S: id.substring(0, MIN_STORED_ID_LENGTH)}, |
| unique_subhash: {S: id}, |
| }; |
| } |
| |
| async expandId(id) { |
| // By just getting the item and not trying to update it, we save an update when the link does not exist |
| // for which we have less resources allocated, but get one extra read (But we do have more reserved for it) |
| const item = await this.dynamoDb.getItem({ |
| TableName: this.table, |
| Key: this.getKeyStruct(id), |
| }).promise(); |
| |
| const attributes = item.Item; |
| if (!attributes) |
| throw new Error(`Missing attributes on ${id}`); |
| const result = await this.s3.get(attributes.full_hash.S, this.prefix); |
| // If we're here, we are pretty confident there is a match. But never hurts to double check |
| if (!result.hit) |
| throw new Error(`ID ${id} not present in storage`); |
| const metadata = attributes.named_metadata ? attributes.named_metadata.M : null; |
| return { |
| config: result.data.toString(), |
| specialMetadata: metadata, |
| }; |
| } |
| |
| async incrementViewCount(id) { |
| try { |
| await this.dynamoDb.updateItem({ |
| TableName: this.table, |
| Key: this.getKeyStruct(id), |
| UpdateExpression: 'SET stats.clicks = stats.clicks + :inc', |
| ExpressionAttributeValues: { |
| ':inc': {N: '1'}, |
| }, |
| ReturnValues: 'NONE', |
| }).promise(); |
| } catch (err) { |
| // Swallow up errors |
| logger.error(`Error when incrementing view count for ${id}`, err); |
| } |
| } |
| } |
| |
| module.exports = StorageS3; |