From c2b557c34cd00864fb3d5f26ff090d78dc51ea77 Mon Sep 17 00:00:00 2001 From: Aleksandr Statciuk Date: Mon, 7 Feb 2022 06:51:59 +0300 Subject: [PATCH] Update save-results.js --- scripts/commands/save-results.js | 64 +++++++++++-------- .../cluster_1.log | 0 tests/commands/save-results.test.js | 14 ++-- 3 files changed, 45 insertions(+), 33 deletions(-) rename tests/__data__/input/logs/{load-streams => load-cluster}/cluster_1.log (100%) diff --git a/scripts/commands/save-results.js b/scripts/commands/save-results.js index d77b6db69c..93b2be85c9 100644 --- a/scripts/commands/save-results.js +++ b/scripts/commands/save-results.js @@ -2,48 +2,53 @@ const _ = require('lodash') const statuses = require('../data/statuses') const { db, store, parser, file, logger } = require('../core') -let streams = [] -let results = {} -const origins = {} const items = [] -const LOGS_PATH = process.env.LOGS_PATH || 'scripts/logs/load-streams' +const LOGS_DIR = process.env.LOGS_DIR || 'scripts/logs/load-cluster' async function main() { - await loadDatabase() - await loadResults() - await findStreamOrigins() - await updateStreams() - await updateDatabase() + let streams = await loadStreams() + const results = await loadResults() + const origins = await findOrigins(results) + streams = await updateStreams(streams, results, origins) + + await updateDatabase(streams) } main() -async function loadDatabase() { - logger.info('loading database...') +async function loadStreams() { + logger.info('loading streams...') - streams = await db.find({}) + await db.streams.load() + const streams = await db.streams.find({}) logger.info(`found ${streams.length} streams`) + + return streams } async function loadResults() { - logger.info('loading results from logs/load-streams...') + logger.info('loading results from logs...') - const files = await file.list(`${LOGS_PATH}/cluster_*.log`) + const results = {} + const files = await file.list(`${LOGS_DIR}/cluster_*.log`) for (const filepath of files) { const parsed = await parser.parseLogs(filepath) - for (const result of parsed) { - results[result._id] = result + for (const item of parsed) { + results[item._id] = item } } logger.info(`found ${Object.values(results).length} results`) + + return results } -async function findStreamOrigins() { +async function findOrigins(results = {}) { logger.info('searching for stream origins...') + const origins = {} for (const { error, requests } of Object.values(results)) { if (error || !Array.isArray(requests) || !requests.length) continue @@ -59,20 +64,23 @@ async function findStreamOrigins() { } logger.info(`found ${_.uniq(Object.values(origins)).length} origins`) + + return origins } -async function updateStreams() { +async function updateStreams(items = [], results = {}, origins = {}) { logger.info('updating streams...') let updated = 0 - for (const item of streams) { + const output = [] + for (const item of items) { const stream = store.create(item) const result = results[item._id] if (result) { const { error, streams, requests } = result const resolution = parseResolution(streams) - const origin = findOrigin(requests) + const origin = findOrigin(requests, origins) let status = parseStatus(error) if (status) { @@ -105,26 +113,28 @@ async function updateStreams() { if (stream.changed) { stream.set('updated', true) - items.push(stream.data()) + output.push(stream.data()) updated++ } } - logger.info(`updated ${updated} items`) + logger.info(`updated ${updated} streams`) + + return output } -async function updateDatabase() { +async function updateDatabase(streams = []) { logger.info('updating database...') - for (const item of items) { - await db.update({ _id: item._id }, item) + for (const stream of streams) { + await db.streams.update({ _id: stream._id }, stream) } - db.compact() + db.streams.compact() logger.info('done') } -function findOrigin(requests) { +function findOrigin(requests = [], origins = {}) { if (origins && Array.isArray(requests)) { requests = requests.map(r => r.url.replace(/(^\w+:|^)/, '')) for (const url of requests) { diff --git a/tests/__data__/input/logs/load-streams/cluster_1.log b/tests/__data__/input/logs/load-cluster/cluster_1.log similarity index 100% rename from tests/__data__/input/logs/load-streams/cluster_1.log rename to tests/__data__/input/logs/load-cluster/cluster_1.log diff --git a/tests/commands/save-results.test.js b/tests/commands/save-results.test.js index 73af2cd220..7375b9d4a1 100644 --- a/tests/commands/save-results.test.js +++ b/tests/commands/save-results.test.js @@ -4,19 +4,21 @@ const { execSync } = require('child_process') beforeEach(() => { fs.emptyDirSync('tests/__data__/temp') - fs.copyFileSync('tests/__data__/input/save-results.streams.db', 'tests/__data__/temp/streams.db') + fs.copyFileSync( + 'tests/__data__/input/database/save-results.streams.db', + 'tests/__data__/temp/streams.db' + ) const stdout = execSync( - 'DB_FILEPATH=tests/__data__/temp/streams.db LOGS_PATH=tests/__data__/input/logs/load-streams node scripts/commands/save-results.js', + 'DB_DIR=tests/__data__/temp LOGS_DIR=tests/__data__/input/logs/load-cluster node scripts/commands/save-results.js', { encoding: 'utf8' } ) }) it('can save results', () => { - const output = content('tests/__data__/temp/streams.db') - const expected = content('tests/__data__/expected/save-results.streams.db') - - expect(output).toEqual(expected) + expect(content('tests/__data__/temp/streams.db')).toEqual( + content('tests/__data__/expected/save-results.streams.db') + ) }) function content(filepath) {