172 lines
5.3 KiB
TypeScript
172 lines
5.3 KiB
TypeScript
import { Queue, Worker } from 'bullmq';
|
|
import IORedis from 'ioredis';
|
|
import { deleteFile, readFile, readFiles, saveFile } from "../storage";
|
|
import { createRemoteBrowserForRun, destroyRemoteBrowser, getActiveBrowserId } from '../../browser-management/controller';
|
|
import { RemoteBrowser } from '../../browser-management/classes/RemoteBrowser';
|
|
import logger from '../../logger';
|
|
import { browserPool, io } from "../../server";
|
|
import fs from "fs";
|
|
import { uuid } from "uuidv4";
|
|
import { chromium } from "playwright";
|
|
|
|
const connection = new IORedis({
|
|
host: 'localhost',
|
|
port: 6379,
|
|
maxRetriesPerRequest: null,
|
|
});
|
|
|
|
connection.on('connect', () => {
|
|
console.log('Connected to Redis!');
|
|
});
|
|
|
|
connection.on('error', (err) => {
|
|
console.error('Redis connection error:', err);
|
|
});
|
|
|
|
const workflowQueue = new Queue('workflow', { connection });
|
|
|
|
export const worker = new Worker('workflow', async job => {
|
|
const { fileName, runId } = job.data;
|
|
try {
|
|
const result = await scheduleWorkflow(fileName, runId);
|
|
return result;
|
|
} catch (error) {
|
|
console.error('Error running workflow:', error);
|
|
throw error;
|
|
}
|
|
}, { connection });
|
|
|
|
// Listen for job completion and close worker/queue
|
|
worker.on('completed', async (job: any) => {
|
|
console.log(`Job ${job.id} completed for ${job.data.fileName}_${job.data.runId}`);
|
|
|
|
// Gracefully close the worker and queue
|
|
await worker.close();
|
|
await workflowQueue.close();
|
|
console.log('Worker and queue have been closed.');
|
|
});
|
|
|
|
// Listen for job failure and close worker/queue
|
|
worker.on('failed', async (job: any, err) => {
|
|
console.error(`Job ${job.id} failed for ${job.data.fileName}_${job.data.runId}:`, err);
|
|
|
|
// Gracefully close the worker and queue
|
|
await worker.close();
|
|
await workflowQueue.close();
|
|
console.log('Worker and queue have been closed after failure.');
|
|
});
|
|
|
|
async function scheduleWorkflow(fileName: string, runId: string) {
|
|
if (!runId) {
|
|
runId = uuid();
|
|
}
|
|
|
|
try {
|
|
const browserId = createRemoteBrowserForRun({
|
|
browser: chromium,
|
|
launchOptions: { headless: true }
|
|
});
|
|
const run_meta = {
|
|
status: 'SCHEDULED',
|
|
name: fileName,
|
|
startedAt: new Date().toLocaleString(),
|
|
finishedAt: '',
|
|
duration: '',
|
|
task: '', // Optionally set based on workflow
|
|
browserId: browserId,
|
|
interpreterSettings: { maxConcurrency: 1, maxRepeats: 1, debug: true },
|
|
log: '',
|
|
runId: runId,
|
|
};
|
|
|
|
fs.mkdirSync('../storage/runs', { recursive: true });
|
|
await saveFile(
|
|
`../storage/runs/${fileName}_${runId}.json`,
|
|
JSON.stringify(run_meta, null, 2)
|
|
);
|
|
|
|
logger.log('debug', `Scheduled run with name: ${fileName}_${runId}.json`);
|
|
|
|
} catch (e) {
|
|
const { message } = e as Error;
|
|
logger.log('info', `Error while scheduling a run with name: ${fileName}_${runId}.json`);
|
|
console.log(message);
|
|
return {
|
|
success: false,
|
|
error: message,
|
|
};
|
|
}
|
|
}
|
|
|
|
async function executeRun(fileName: string, runId: string) {
|
|
try {
|
|
// Read the recording from storage
|
|
const recording = await readFile(`./../storage/recordings/${fileName}.waw.json`);
|
|
const parsedRecording = JSON.parse(recording);
|
|
|
|
// Read the run from storage
|
|
const run = await readFile(`./../storage/runs/${fileName}_${runId}.json`);
|
|
const parsedRun = JSON.parse(run);
|
|
|
|
// Update status to RUNNING
|
|
parsedRun.status = 'RUNNING';
|
|
await saveFile(
|
|
`../storage/runs/${fileName}_${runId}.json`,
|
|
JSON.stringify(parsedRun, null, 2)
|
|
);
|
|
|
|
// Interpret the run in active browser
|
|
|
|
const browser = browserPool.getRemoteBrowser(parsedRun.browserId);
|
|
if (!browser) {
|
|
throw new Error('Could not access browser');
|
|
}
|
|
|
|
const currentPage = await browser.getCurrentPage();
|
|
if (!currentPage) {
|
|
throw new Error('Could not create a new page');
|
|
}
|
|
|
|
const interpretationInfo = await browser.interpreter.InterpretRecording(
|
|
parsedRecording.recording, currentPage, parsedRun.interpreterSettings);
|
|
|
|
const duration = Math.round((new Date().getTime() - new Date(parsedRun.startedAt).getTime()) / 1000);
|
|
const durString = duration < 60 ? `${duration} s` : `${Math.floor(duration / 60)} m ${duration % 60} s`;
|
|
|
|
await destroyRemoteBrowser(parsedRun.browserId);
|
|
|
|
const updated_run_meta = {
|
|
...parsedRun,
|
|
status: interpretationInfo.result,
|
|
finishedAt: new Date().toLocaleString(),
|
|
duration: durString,
|
|
browserId: null,
|
|
log: interpretationInfo.log.join('\n'),
|
|
serializableOutput: interpretationInfo.serializableOutput,
|
|
binaryOutput: interpretationInfo.binaryOutput,
|
|
};
|
|
|
|
await saveFile(
|
|
`../storage/runs/${fileName}_${runId}.json`,
|
|
JSON.stringify(updated_run_meta, null, 2)
|
|
);
|
|
|
|
return true;
|
|
} catch (error: any) {
|
|
logger.log('info', `Error while running a recording with name: ${fileName}_${runId}.json`);
|
|
console.log(error.message);
|
|
|
|
// Update run status to ERROR
|
|
const errorRun = await readFile(`./../storage/runs/${fileName}_${runId}.json`);
|
|
const parsedErrorRun = JSON.parse(errorRun);
|
|
parsedErrorRun.status = 'ERROR';
|
|
parsedErrorRun.log += `\nError: ${error.message}`;
|
|
await saveFile(
|
|
`../storage/runs/${fileName}_${runId}.json`,
|
|
JSON.stringify(parsedErrorRun, null, 2)
|
|
);
|
|
|
|
return false;
|
|
}
|
|
}
|
|
export { workflowQueue, scheduleWorkflow }; |