blob: ffb612a89c285c0b38c866179d84a14dae6e3d3a [file] [log] [blame] [raw]
// Copyright (c) 2018, Compiler Explorer Authors
// All rights reserved.
//
// Redistribution and use in source and binary forms, with or without
// modification, are permitted provided that the following conditions are met:
//
// * Redistributions of source code must retain the above copyright notice,
// this list of conditions and the following disclaimer.
// * Redistributions in binary form must reproduce the above copyright
// notice, this list of conditions and the following disclaimer in the
// documentation and/or other materials provided with the distribution.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
// AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
// ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
// LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
// CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
// SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
// INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
// CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
// ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
// POSSIBILITY OF SUCH DAMAGE.
import AWS from 'aws-sdk-mock';
import {unwrap} from '../../lib/assert';
import * as properties from '../../lib/properties';
import {StorageS3} from '../../lib/storage';
type Handler = (q: any) => any;
// NB!!! Anything using mocked AWS calls needs to be initialised in the `it(...)` block! If you initialise it in the
// `describe()` top level code then it won't be mocked in time. We only mock and de-mock before/after else we end up
// fighting over the global AWS mocking stuff. I hate mocha...there's probably a better way...
function mockerise(service, method) {
const handlers: Handler[] = [];
beforeEach(() => {
AWS.mock(service, method, (q, callback) => {
const qh = unwrap(handlers.shift());
try {
callback(undefined, qh(q));
} catch (e) {
callback(e as any, undefined);
}
});
});
afterEach(() => {
AWS.restore(service, method);
});
return handlers;
}
describe('Find unique subhash tests', () => {
const dynamoDbQueryHandlers = mockerise('DynamoDB', 'query');
const compilerProps = properties.fakeProps({});
const httpRootDir = '/';
const awsProps = properties.fakeProps({
region: 'not-a-region',
storageBucket: 'bucket',
storagePrefix: 'prefix',
storageDynamoTable: 'table',
});
it('works when empty', () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
dynamoDbQueryHandlers.push(q => {
q.TableName.should.equal('table');
return {};
});
return storage.findUniqueSubhash('ABCDEFGHIJKLMNOPQRSTUV').should.eventually.deep.equal({
alreadyPresent: false,
prefix: 'ABCDEF',
uniqueSubHash: 'ABCDEFGHI',
});
});
it('works when not empty', () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
dynamoDbQueryHandlers.push(() => {
return {
Items: [
{
full_hash: {S: 'ZZVZT'},
unique_subhash: {S: 'ZZVZT'},
},
],
};
});
return storage.findUniqueSubhash('ABCDEFGHIJKLMNOPQRSTUV').should.eventually.deep.equal({
alreadyPresent: false,
prefix: 'ABCDEF',
uniqueSubHash: 'ABCDEFGHI',
});
});
it("works when there' a collision", () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
dynamoDbQueryHandlers.push(() => {
return {
Items: [
{
full_hash: {S: 'ABCDEFGHIZZ'},
unique_subhash: {S: 'ABCDEFGHI'},
},
],
};
});
return storage.findUniqueSubhash('ABCDEFGHIJKLMNOPQRSTUV').should.eventually.deep.equal({
alreadyPresent: false,
prefix: 'ABCDEF',
uniqueSubHash: 'ABCDEFGHIJ',
});
});
it('finds an existing match', () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
dynamoDbQueryHandlers.push(() => {
return {
Items: [
{
full_hash: {S: 'ABCDEFGHIJKLMNOPQRSTUV'},
unique_subhash: {S: 'ABCDEFGHI'},
},
],
};
});
return storage.findUniqueSubhash('ABCDEFGHIJKLMNOPQRSTUV').should.eventually.deep.equal({
alreadyPresent: true,
prefix: 'ABCDEF',
uniqueSubHash: 'ABCDEFGHI',
});
});
});
describe('Stores to s3', () => {
const dynamoDbPutItemHandlers = mockerise('DynamoDB', 'putItem');
const s3PutObjectHandlers = mockerise('S3', 'putObject');
const httpRootDir = '/';
const compilerProps = properties.fakeProps({});
const awsProps = properties.fakeProps({
region: 'not-a-region',
storageBucket: 'bucket',
storagePrefix: 'prefix',
storageDynamoTable: 'table',
});
it('and works ok', () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
const object = {
prefix: 'ABCDEF',
uniqueSubHash: 'ABCDEFG',
fullHash: 'ABCDEFGHIJKLMNOP',
config: 'yo',
};
const ran = {s3: false, dynamo: false};
s3PutObjectHandlers.push(q => {
q.Bucket.should.equal('bucket');
q.Key.should.equal('prefix/ABCDEFGHIJKLMNOP');
q.Body.should.equal('yo');
ran.s3 = true;
return {};
});
dynamoDbPutItemHandlers.push(q => {
q.TableName.should.equals('table');
q.Item.should.deep.equals({
prefix: {S: 'ABCDEF'},
unique_subhash: {S: 'ABCDEFG'},
full_hash: {S: 'ABCDEFGHIJKLMNOP'},
stats: {M: {clicks: {N: '0'}}},
creation_ip: {S: 'localhost'},
// Cheat the date
creation_date: {S: q.Item.creation_date.S},
});
ran.dynamo = true;
return {};
});
return storage.storeItem(object, {get: () => 'localhost'}).then(() => {
ran.should.deep.equal({s3: true, dynamo: true});
});
});
});
describe('Retrieves from s3', () => {
const dynamoDbGetItemHandlers = mockerise('DynamoDB', 'getItem');
const s3GetObjectHandlers = mockerise('S3', 'getObject');
const httpRootDir = '/';
const compilerProps = properties.fakeProps({});
const awsProps = properties.fakeProps({
region: 'not-a-region',
storageBucket: 'bucket',
storagePrefix: 'prefix',
storageDynamoTable: 'table',
});
it('fetches in the happy path', async () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
const ran = {s3: false, dynamo: false};
dynamoDbGetItemHandlers.push(q => {
q.TableName.should.equals('table');
q.Key.should.deep.equals({
prefix: {S: 'ABCDEF'},
unique_subhash: {S: 'ABCDEF'},
});
ran.dynamo = true;
return {Item: {full_hash: {S: 'ABCDEFGHIJKLMNOP'}}};
});
s3GetObjectHandlers.push(q => {
q.Bucket.should.equal('bucket');
q.Key.should.equal('prefix/ABCDEFGHIJKLMNOP');
ran.s3 = true;
return {
Body: 'I am a monkey',
};
});
const result = await storage.expandId('ABCDEF');
ran.should.deep.equal({s3: true, dynamo: true});
result.should.deep.equal({config: 'I am a monkey', specialMetadata: null});
});
it('should handle failures', async () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
for (let i = 0; i < 10; ++i) dynamoDbGetItemHandlers.push(() => ({}));
return storage.expandId('ABCDEF').should.be.rejectedWith(Error, 'ID ABCDEF not present in links table');
});
});
describe('Updates counts in s3', async () => {
const dynamoDbUpdateItemHandlers = mockerise('DynamoDB', 'updateItem');
const httpRootDir = '/';
const compilerProps = properties.fakeProps({});
const awsProps = properties.fakeProps({
region: 'not-a-region',
storageBucket: 'bucket',
storagePrefix: 'prefix',
storageDynamoTable: 'table',
});
it('should increment for simple cases', async () => {
const storage = new StorageS3(httpRootDir, compilerProps, awsProps);
let called = false;
dynamoDbUpdateItemHandlers.push(q => {
q.should.deep.equals({
ExpressionAttributeValues: {':inc': {N: '1'}},
Key: {
prefix: {S: 'ABCDEF'},
unique_subhash: {S: 'ABCDEF'},
},
ReturnValues: 'NONE',
TableName: 'table',
UpdateExpression: 'SET stats.clicks = stats.clicks + :inc',
});
called = true;
});
await storage.incrementViewCount('ABCDEF');
called.should.be.true;
});
});