feat: recorder revamp server changes

This commit is contained in:
Rohit Rajan
2025-10-21 00:43:08 +05:30
parent eafe11aef4
commit 5be2b3175b
9 changed files with 879 additions and 369 deletions

View File

@@ -1,16 +1,14 @@
import { readFile, readFiles } from "../workflow-management/storage";
import { Router, Request, Response } from 'express'; import { Router, Request, Response } from 'express';
import { chromium } from "playwright-extra"; import { chromium } from "playwright-extra";
import stealthPlugin from 'puppeteer-extra-plugin-stealth'; import stealthPlugin from 'puppeteer-extra-plugin-stealth';
import { requireAPIKey } from "../middlewares/api"; import { requireAPIKey } from "../middlewares/api";
import Robot from "../models/Robot"; import Robot from "../models/Robot";
import Run from "../models/Run"; import Run from "../models/Run";
const router = Router();
import { getDecryptedProxyConfig } from "../routes/proxy"; import { getDecryptedProxyConfig } from "../routes/proxy";
import { v4 as uuid } from "uuid"; import { v4 as uuid } from "uuid";
import { createRemoteBrowserForRun, destroyRemoteBrowser } from "../browser-management/controller"; import { createRemoteBrowserForRun, destroyRemoteBrowser } from "../browser-management/controller";
import logger from "../logger"; import logger from "../logger";
import { browserPool } from "../server"; import { browserPool, io as serverIo } from "../server";
import { io, Socket } from "socket.io-client"; import { io, Socket } from "socket.io-client";
import { BinaryOutputService } from "../storage/mino"; import { BinaryOutputService } from "../storage/mino";
import { AuthenticatedRequest } from "../routes/record" import { AuthenticatedRequest } from "../routes/record"
@@ -20,8 +18,11 @@ import { WorkflowFile } from "maxun-core";
import { googleSheetUpdateTasks, processGoogleSheetUpdates } from "../workflow-management/integrations/gsheet"; import { googleSheetUpdateTasks, processGoogleSheetUpdates } from "../workflow-management/integrations/gsheet";
import { airtableUpdateTasks, processAirtableUpdates } from "../workflow-management/integrations/airtable"; import { airtableUpdateTasks, processAirtableUpdates } from "../workflow-management/integrations/airtable";
import { sendWebhook } from "../routes/webhook"; import { sendWebhook } from "../routes/webhook";
chromium.use(stealthPlugin()); chromium.use(stealthPlugin());
const router = Router();
const formatRecording = (recordingData: any) => { const formatRecording = (recordingData: any) => {
const recordingMeta = recordingData.recording_meta; const recordingMeta = recordingData.recording_meta;
const workflow = recordingData.recording.workflow || []; const workflow = recordingData.recording.workflow || [];
@@ -334,7 +335,7 @@ function formatRunResponse(run: any) {
id: run.id, id: run.id,
status: run.status, status: run.status,
name: run.name, name: run.name,
robotId: run.robotMetaId, // Renaming robotMetaId to robotId robotId: run.robotMetaId,
startedAt: run.startedAt, startedAt: run.startedAt,
finishedAt: run.finishedAt, finishedAt: run.finishedAt,
runId: run.runId, runId: run.runId,
@@ -342,20 +343,20 @@ function formatRunResponse(run: any) {
runByScheduleId: run.runByScheduleId, runByScheduleId: run.runByScheduleId,
runByAPI: run.runByAPI, runByAPI: run.runByAPI,
data: { data: {
textData: [], textData: {},
listData: [] listData: {}
}, },
screenshots: [] as any[], screenshots: [] as any[],
}; };
if (run.serializableOutput) { const output = run.serializableOutput || {};
if (run.serializableOutput.scrapeSchema && run.serializableOutput.scrapeSchema.length > 0) {
formattedRun.data.textData = run.serializableOutput.scrapeSchema;
}
if (run.serializableOutput.scrapeList && run.serializableOutput.scrapeList.length > 0) { if (output.scrapeSchema && typeof output.scrapeSchema === 'object') {
formattedRun.data.listData = run.serializableOutput.scrapeList; formattedRun.data.textData = output.scrapeSchema;
} }
if (output.scrapeList && typeof output.scrapeList === 'object') {
formattedRun.data.listData = output.scrapeList;
} }
if (run.binaryOutput) { if (run.binaryOutput) {
@@ -505,10 +506,30 @@ async function createWorkflowAndStoreMetadata(id: string, userId: string) {
runByAPI: true, runByAPI: true,
serializableOutput: {}, serializableOutput: {},
binaryOutput: {}, binaryOutput: {},
retryCount: 0
}); });
const plainRun = run.toJSON(); const plainRun = run.toJSON();
try {
const runStartedData = {
runId: plainRun.runId,
robotMetaId: plainRun.robotMetaId,
robotName: plainRun.name,
status: 'running',
startedAt: plainRun.startedAt,
runByUserId: plainRun.runByUserId,
runByScheduleId: plainRun.runByScheduleId,
runByAPI: plainRun.runByAPI || false,
browserId: plainRun.browserId
};
serverIo.of('/queued-run').to(`user-${userId}`).emit('run-started', runStartedData);
logger.log('info', `API run started notification sent for run: ${plainRun.runId} to user-${userId}`);
} catch (socketError: any) {
logger.log('warn', `Failed to send run-started notification for API run ${plainRun.runId}: ${socketError.message}`);
}
return { return {
browserId, browserId,
runId: plainRun.runId, runId: plainRun.runId,
@@ -525,6 +546,29 @@ async function createWorkflowAndStoreMetadata(id: string, userId: string) {
} }
} }
async function triggerIntegrationUpdates(runId: string, robotMetaId: string): Promise<void> {
try {
googleSheetUpdateTasks[runId] = {
robotId: robotMetaId,
runId: runId,
status: 'pending',
retries: 5,
};
airtableUpdateTasks[runId] = {
robotId: robotMetaId,
runId: runId,
status: 'pending',
retries: 5,
};
processAirtableUpdates().catch(err => logger.log('error', `Airtable update error: ${err.message}`));
processGoogleSheetUpdates().catch(err => logger.log('error', `Google Sheets update error: ${err.message}`));
} catch (err: any) {
logger.log('error', `Failed to update integrations for run: ${runId}: ${err.message}`);
}
}
async function readyForRunHandler(browserId: string, id: string, userId: string){ async function readyForRunHandler(browserId: string, id: string, userId: string){
try { try {
const result = await executeRun(id, userId); const result = await executeRun(id, userId);
@@ -565,6 +609,8 @@ function AddGeneratedFlags(workflow: WorkflowFile) {
}; };
async function executeRun(id: string, userId: string) { async function executeRun(id: string, userId: string) {
let browser: any = null;
try { try {
const run = await Run.findOne({ where: { runId: id } }); const run = await Run.findOne({ where: { runId: id } });
if (!run) { if (!run) {
@@ -576,6 +622,27 @@ async function executeRun(id: string, userId: string) {
const plainRun = run.toJSON(); const plainRun = run.toJSON();
if (run.status === 'aborted' || run.status === 'aborting') {
logger.log('info', `API Run ${id} has status ${run.status}, skipping execution`);
return { success: true };
}
if (run.status === 'queued') {
logger.log('info', `API Run ${id} has status 'queued', skipping stale execution - will be handled by recovery`);
return { success: true };
}
const retryCount = plainRun.retryCount || 0;
if (retryCount >= 3) {
logger.log('warn', `API Run ${id} has exceeded max retries (${retryCount}/3), marking as failed`);
await run.update({
status: 'failed',
finishedAt: new Date().toLocaleString(),
log: `Max retries exceeded (${retryCount}/3) - Run permanently failed`
});
return { success: false, error: 'Max retries exceeded' };
}
const recording = await Robot.findOne({ where: { 'recording_meta.id': plainRun.robotMetaId }, raw: true }); const recording = await Robot.findOne({ where: { 'recording_meta.id': plainRun.robotMetaId }, raw: true });
if (!recording) { if (!recording) {
return { return {
@@ -586,7 +653,7 @@ async function executeRun(id: string, userId: string) {
plainRun.status = 'running'; plainRun.status = 'running';
const browser = browserPool.getRemoteBrowser(plainRun.browserId); browser = browserPool.getRemoteBrowser(plainRun.browserId);
if (!browser) { if (!browser) {
throw new Error('Could not access browser'); throw new Error('Could not access browser');
} }
@@ -597,41 +664,33 @@ async function executeRun(id: string, userId: string) {
} }
const workflow = AddGeneratedFlags(recording.recording); const workflow = AddGeneratedFlags(recording.recording);
browser.interpreter.setRunId(id); browser.interpreter.setRunId(plainRun.runId);
const interpretationInfo = await browser.interpreter.InterpretRecording( const interpretationInfo = await browser.interpreter.InterpretRecording(
workflow, currentPage, (newPage: Page) => currentPage = newPage, plainRun.interpreterSettings workflow, currentPage, (newPage: Page) => currentPage = newPage, plainRun.interpreterSettings
); );
const binaryOutputService = new BinaryOutputService('maxun-run-screenshots');
const uploadedBinaryOutput = await binaryOutputService.uploadAndStoreBinaryOutput(run, interpretationInfo.binaryOutput);
await destroyRemoteBrowser(plainRun.browserId, userId); await destroyRemoteBrowser(plainRun.browserId, userId);
const updatedRun = await run.update({ const updatedRun = await run.update({
status: 'success', status: 'success',
finishedAt: new Date().toLocaleString(), finishedAt: new Date().toLocaleString(),
log: interpretationInfo.log.join('\n'), log: interpretationInfo.log.join('\n'),
binaryOutput: uploadedBinaryOutput,
}); });
// Upload binary output to MinIO and update run with MinIO URLs
const finalRun = await Run.findOne({ where: { runId: id } });
if (finalRun && finalRun.binaryOutput && Object.keys(finalRun.binaryOutput).length > 0) {
try {
const binaryService = new BinaryOutputService('maxun-run-screenshots');
await binaryService.uploadAndStoreBinaryOutput(finalRun, finalRun.binaryOutput);
logger.log('info', `Uploaded binary output to MinIO for API run ${id}`);
} catch (minioError: any) {
logger.log('error', `Failed to upload binary output to MinIO for API run ${id}: ${minioError.message}`);
}
}
let totalSchemaItemsExtracted = 0; let totalSchemaItemsExtracted = 0;
let totalListItemsExtracted = 0; let totalListItemsExtracted = 0;
let extractedScreenshotsCount = 0; let extractedScreenshotsCount = 0;
if (finalRun) { if (updatedRun) {
if (finalRun.serializableOutput) { if (updatedRun.dataValues.serializableOutput) {
if (finalRun.serializableOutput.scrapeSchema) { if (updatedRun.dataValues.serializableOutput.scrapeSchema) {
Object.values(finalRun.serializableOutput.scrapeSchema).forEach((schemaResult: any) => { Object.values(updatedRun.dataValues.serializableOutput.scrapeSchema).forEach((schemaResult: any) => {
if (Array.isArray(schemaResult)) { if (Array.isArray(schemaResult)) {
totalSchemaItemsExtracted += schemaResult.length; totalSchemaItemsExtracted += schemaResult.length;
} else if (schemaResult && typeof schemaResult === 'object') { } else if (schemaResult && typeof schemaResult === 'object') {
@@ -640,8 +699,8 @@ async function executeRun(id: string, userId: string) {
}); });
} }
if (finalRun.serializableOutput.scrapeList) { if (updatedRun.dataValues.serializableOutput.scrapeList) {
Object.values(finalRun.serializableOutput.scrapeList).forEach((listResult: any) => { Object.values(updatedRun.dataValues.serializableOutput.scrapeList).forEach((listResult: any) => {
if (Array.isArray(listResult)) { if (Array.isArray(listResult)) {
totalListItemsExtracted += listResult.length; totalListItemsExtracted += listResult.length;
} }
@@ -649,8 +708,8 @@ async function executeRun(id: string, userId: string) {
} }
} }
if (finalRun.binaryOutput) { if (updatedRun.dataValues.binaryOutput) {
extractedScreenshotsCount = Object.keys(finalRun.binaryOutput).length; extractedScreenshotsCount = Object.keys(updatedRun.dataValues.binaryOutput).length;
} }
} }
@@ -667,17 +726,31 @@ async function executeRun(id: string, userId: string) {
} }
) )
const parsedOutput =
typeof updatedRun.dataValues.serializableOutput === "string"
? JSON.parse(updatedRun.dataValues.serializableOutput)
: updatedRun.dataValues.serializableOutput || {};
const parsedList =
typeof parsedOutput.scrapeList === "string"
? JSON.parse(parsedOutput.scrapeList)
: parsedOutput.scrapeList || {};
const parsedSchema =
typeof parsedOutput.scrapeSchema === "string"
? JSON.parse(parsedOutput.scrapeSchema)
: parsedOutput.scrapeSchema || {};
const webhookPayload = { const webhookPayload = {
robot_id: plainRun.robotMetaId, robot_id: plainRun.robotMetaId,
run_id: plainRun.runId, run_id: plainRun.runId,
robot_name: recording.recording_meta.name, robot_name: recording.recording_meta.name,
status: 'success', status: "success",
started_at: plainRun.startedAt, started_at: plainRun.startedAt,
finished_at: new Date().toLocaleString(), finished_at: new Date().toLocaleString(),
extracted_data: { extracted_data: {
captured_texts: finalRun?.serializableOutput?.scrapeSchema ? Object.values(finalRun.serializableOutput.scrapeSchema).flat() : [], captured_texts: parsedSchema || {},
captured_lists: finalRun?.serializableOutput?.scrapeList || {}, captured_lists: parsedList || {},
total_rows: totalRowsExtracted,
captured_texts_count: totalSchemaItemsExtracted, captured_texts_count: totalSchemaItemsExtracted,
captured_lists_count: totalListItemsExtracted, captured_lists_count: totalListItemsExtracted,
screenshots_count: extractedScreenshotsCount screenshots_count: extractedScreenshotsCount
@@ -685,7 +758,7 @@ async function executeRun(id: string, userId: string) {
metadata: { metadata: {
browser_id: plainRun.browserId, browser_id: plainRun.browserId,
user_id: userId, user_id: userId,
} },
}; };
try { try {
@@ -695,26 +768,7 @@ async function executeRun(id: string, userId: string) {
logger.log('error', `Failed to send webhooks for run ${plainRun.runId}: ${webhookError.message}`); logger.log('error', `Failed to send webhooks for run ${plainRun.runId}: ${webhookError.message}`);
} }
try { await triggerIntegrationUpdates(plainRun.runId, plainRun.robotMetaId);
googleSheetUpdateTasks[id] = {
robotId: plainRun.robotMetaId,
runId: id,
status: 'pending',
retries: 5,
};
airtableUpdateTasks[id] = {
robotId: plainRun.robotMetaId,
runId: id,
status: 'pending',
retries: 5,
};
processAirtableUpdates().catch(err => logger.log('error', `Airtable update error: ${err.message}`));
processGoogleSheetUpdates().catch(err => logger.log('error', `Google Sheets update error: ${err.message}`));
} catch (err: any) {
logger.log('error', `Failed to update Google Sheet for run: ${plainRun.runId}: ${err.message}`);
}
return { return {
success: true, success: true,
@@ -728,8 +782,29 @@ async function executeRun(id: string, userId: string) {
await run.update({ await run.update({
status: 'failed', status: 'failed',
finishedAt: new Date().toLocaleString(), finishedAt: new Date().toLocaleString(),
log: (run.log ? run.log + '\n' : '') + `Error: ${error.message}\n` + (error.stack ? error.stack : ''),
}); });
try {
const recording = await Robot.findOne({ where: { 'recording_meta.id': run.robotMetaId }, raw: true });
const failureData = {
runId: run.runId,
robotMetaId: run.robotMetaId,
robotName: recording ? recording.recording_meta.name : 'Unknown Robot',
status: 'failed',
finishedAt: new Date().toLocaleString(),
runByUserId: run.runByUserId,
runByScheduleId: run.runByScheduleId,
runByAPI: run.runByAPI || false,
browserId: run.browserId
};
serverIo.of('/queued-run').to(`user-${userId}`).emit('run-completed', failureData);
logger.log('info', `API run permanently failed notification sent for run: ${run.runId} to user-${userId}`);
} catch (socketError: any) {
logger.log('warn', `Failed to send run-completed notification for permanently failed API run ${run.runId}: ${socketError.message}`);
}
const recording = await Robot.findOne({ where: { 'recording_meta.id': run.robotMetaId }, raw: true }); const recording = await Robot.findOne({ where: { 'recording_meta.id': run.robotMetaId }, raw: true });
// Trigger webhooks for run failure // Trigger webhooks for run failure
@@ -814,7 +889,7 @@ async function waitForRunCompletion(runId: string, interval: number = 2000) {
if (!run) throw new Error('Run not found'); if (!run) throw new Error('Run not found');
if (run.status === 'success') { if (run.status === 'success') {
return run; return run.toJSON();
} else if (run.status === 'failed') { } else if (run.status === 'failed') {
throw new Error('Run failed'); throw new Error('Run failed');
} }

View File

@@ -65,6 +65,7 @@ const handleWrapper = async (
interface CustomActionEventData { interface CustomActionEventData {
action: CustomActions; action: CustomActions;
settings: any; settings: any;
actionId?: string;
} }
/** /**
@@ -84,23 +85,24 @@ const onGenerateAction = async (customActionEventData: CustomActionEventData, us
* @param page The active page * @param page The active page
* @param action The custom action * @param action The custom action
* @param settings The custom action settings * @param settings The custom action settings
* @param actionId Optional action ID for tracking and updating specific actions
* @category BrowserManagement * @category BrowserManagement
*/ */
const handleGenerateAction = const handleGenerateAction =
async (activeBrowser: RemoteBrowser, page: Page, { action, settings }: CustomActionEventData) => { async (activeBrowser: RemoteBrowser, page: Page, { action, settings, actionId }: CustomActionEventData) => {
try { try {
if (page.isClosed()) { if (page.isClosed()) {
logger.log("debug", `Ignoring generate action event: page is closed`); logger.log("debug", `Ignoring generate action event: page is closed`);
return; return;
} }
const generator = activeBrowser.generator; const generator = activeBrowser.generator;
await generator.customAction(action, settings, page); await generator.customAction(action, actionId || '', settings, page);
} catch (e) { } catch (e) {
const { message } = e as Error; const { message } = e as Error;
logger.log("warn", `Error handling generate action event: ${message}`); logger.log("warn", `Error handling generate action event: ${message}`);
}
} }
}
/** /**
* A wrapper function for handling mousedown event. * A wrapper function for handling mousedown event.
@@ -819,6 +821,49 @@ const onDOMWorkflowPair = async (
await handleWrapper(handleWorkflowPair, userId, data); await handleWrapper(handleWorkflowPair, userId, data);
}; };
/**
* Handles the remove action event.
* This is called when a user discards a capture action (list or text) that was already emitted to the backend.
* @param activeBrowser - the active remote browser instance
* @param page - the active page of the remote browser
* @param data - the data containing the actionId to remove
* @category BrowserManagement
*/
const handleRemoveAction = async (
activeBrowser: RemoteBrowser,
page: Page,
data: { actionId: string }
) => {
try {
const { actionId } = data;
const generator = activeBrowser.generator;
const removed = generator.removeAction(actionId);
if (removed) {
logger.log("info", `Action ${actionId} successfully removed from workflow`);
} else {
logger.log("debug", `Action ${actionId} not found in workflow`);
}
} catch (e) {
const { message } = e as Error;
logger.log("warn", `Error handling remove action event: ${message}`);
}
};
/**
* A wrapper function for handling the remove action event.
* @param data - the data containing the actionId to remove
* @param userId - the user ID
* @category HelperFunctions
*/
const onRemoveAction = async (
data: { actionId: string },
userId: string
) => {
logger.log("debug", "Handling remove action event emitted from client");
await handleWrapper(handleRemoveAction, userId, data);
};
/** /**
* Helper function for registering the handlers onto established websocket connection. * Helper function for registering the handlers onto established websocket connection.
* Registers various input handlers. * Registers various input handlers.
@@ -831,7 +876,7 @@ const onDOMWorkflowPair = async (
* @returns void * @returns void
* @category BrowserManagement * @category BrowserManagement
*/ */
const registerInputHandlers = (socket: Socket, userId: string) => { const registerInputHandlers = (socket: Socket, userId: string) => {
// Register handlers with the socket // Register handlers with the socket
socket.on("input:mousedown", (data) => onMousedown(data, userId)); socket.on("input:mousedown", (data) => onMousedown(data, userId));
socket.on("input:wheel", (data) => onWheel(data, userId)); socket.on("input:wheel", (data) => onWheel(data, userId));
@@ -847,6 +892,7 @@ const registerInputHandlers = (socket: Socket, userId: string) => {
socket.on("input:time", (data) => onTimeSelection(data, userId)); socket.on("input:time", (data) => onTimeSelection(data, userId));
socket.on("input:datetime-local", (data) => onDateTimeLocalSelection(data, userId)); socket.on("input:datetime-local", (data) => onDateTimeLocalSelection(data, userId));
socket.on("action", (data) => onGenerateAction(data, userId)); socket.on("action", (data) => onGenerateAction(data, userId));
socket.on("removeAction", (data) => onRemoveAction(data, userId));
socket.on("dom:click", (data) => onDOMClickAction(data, userId)); socket.on("dom:click", (data) => onDOMClickAction(data, userId));
socket.on("dom:keypress", (data) => onDOMKeyboardAction(data, userId)); socket.on("dom:keypress", (data) => onDOMKeyboardAction(data, userId));

View File

@@ -149,14 +149,20 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
let browser = browserPool.getRemoteBrowser(browserId); let browser = browserPool.getRemoteBrowser(browserId);
const browserWaitStart = Date.now(); const browserWaitStart = Date.now();
let lastLogTime = 0; let lastLogTime = 0;
let pollAttempts = 0;
const MAX_POLL_ATTEMPTS = 15;
while (!browser && (Date.now() - browserWaitStart) < BROWSER_INIT_TIMEOUT) { while (!browser && (Date.now() - browserWaitStart) < BROWSER_INIT_TIMEOUT && pollAttempts < MAX_POLL_ATTEMPTS) {
const currentTime = Date.now(); const currentTime = Date.now();
pollAttempts++;
const browserStatus = browserPool.getBrowserStatus(browserId); const browserStatus = browserPool.getBrowserStatus(browserId);
if (browserStatus === null) { if (browserStatus === null) {
throw new Error(`Browser slot ${browserId} does not exist in pool`); throw new Error(`Browser slot ${browserId} does not exist in pool`);
} }
if (browserStatus === "failed") {
throw new Error(`Browser ${browserId} initialization failed`);
}
if (currentTime - lastLogTime > 10000) { if (currentTime - lastLogTime > 10000) {
logger.log('info', `Browser ${browserId} not ready yet (status: ${browserStatus}), waiting... (${Math.round((currentTime - browserWaitStart) / 1000)}s elapsed)`); logger.log('info', `Browser ${browserId} not ready yet (status: ${browserStatus}), waiting... (${Math.round((currentTime - browserWaitStart) / 1000)}s elapsed)`);
@@ -183,17 +189,25 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
} }
const isRunAborted = async (): Promise<boolean> => { const isRunAborted = async (): Promise<boolean> => {
const currentRun = await Run.findOne({ where: { runId: data.runId } }); try {
return currentRun ? (currentRun.status === 'aborted' || currentRun.status === 'aborting') : false; const currentRun = await Run.findOne({ where: { runId: data.runId } });
return currentRun ? (currentRun.status === 'aborted' || currentRun.status === 'aborting') : false;
} catch (error: any) {
logger.log('error', `Error checking if run ${data.runId} is aborted: ${error.message}`);
return false;
}
}; };
let currentPage = browser.getCurrentPage(); let currentPage = browser.getCurrentPage();
const pageWaitStart = Date.now(); const pageWaitStart = Date.now();
let lastPageLogTime = 0; let lastPageLogTime = 0;
let pageAttempts = 0;
const MAX_PAGE_ATTEMPTS = 15;
while (!currentPage && (Date.now() - pageWaitStart) < BROWSER_PAGE_TIMEOUT) { while (!currentPage && (Date.now() - pageWaitStart) < BROWSER_PAGE_TIMEOUT && pageAttempts < MAX_PAGE_ATTEMPTS) {
const currentTime = Date.now(); const currentTime = Date.now();
pageAttempts++;
if (currentTime - lastPageLogTime > 5000) { if (currentTime - lastPageLogTime > 5000) {
logger.log('info', `Page not ready for browser ${browserId}, waiting... (${Math.round((currentTime - pageWaitStart) / 1000)}s elapsed)`); logger.log('info', `Page not ready for browser ${browserId}, waiting... (${Math.round((currentTime - pageWaitStart) / 1000)}s elapsed)`);
@@ -209,6 +223,26 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
} }
logger.log('info', `Starting workflow execution for run ${data.runId}`); logger.log('info', `Starting workflow execution for run ${data.runId}`);
await run.update({
status: 'running',
log: 'Workflow execution started'
});
try {
const startedData = {
runId: data.runId,
robotMetaId: plainRun.robotMetaId,
robotName: recording.recording_meta.name,
status: 'running',
startedAt: new Date().toLocaleString()
};
serverIo.of(browserId).emit('run-started', startedData);
serverIo.of('/queued-run').to(`user-${data.userId}`).emit('run-started', startedData);
} catch (socketError: any) {
logger.log('warn', `Failed to send run-started notification for API run ${plainRun.runId}: ${socketError.message}`);
}
// Execute the workflow // Execute the workflow
const workflow = AddGeneratedFlags(recording.recording); const workflow = AddGeneratedFlags(recording.recording);
@@ -231,6 +265,19 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
} }
logger.log('info', `Workflow execution completed for run ${data.runId}`); logger.log('info', `Workflow execution completed for run ${data.runId}`);
const binaryOutputService = new BinaryOutputService('maxuncloud-run-screenshots');
const uploadedBinaryOutput = await binaryOutputService.uploadAndStoreBinaryOutput(
run,
interpretationInfo.binaryOutput
);
// Get the already persisted and credit-validated data from the run record
const finalRun = await Run.findByPk(run.id);
const categorizedOutput = {
scrapeSchema: finalRun?.serializableOutput?.scrapeSchema || {},
scrapeList: finalRun?.serializableOutput?.scrapeList || {}
};
if (await isRunAborted()) { if (await isRunAborted()) {
logger.log('info', `Run ${data.runId} was aborted while processing results, not updating status`); logger.log('info', `Run ${data.runId} was aborted while processing results, not updating status`);
@@ -240,48 +287,39 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
await run.update({ await run.update({
status: 'success', status: 'success',
finishedAt: new Date().toLocaleString(), finishedAt: new Date().toLocaleString(),
log: interpretationInfo.log.join('\n') log: interpretationInfo.log.join('\n'),
serializableOutput: JSON.parse(JSON.stringify({
scrapeSchema: categorizedOutput.scrapeSchema || {},
scrapeList: categorizedOutput.scrapeList || {},
})),
binaryOutput: uploadedBinaryOutput,
}); });
// Upload binary output to MinIO and update run with MinIO URLs
const updatedRun = await Run.findOne({ where: { runId: data.runId } });
if (updatedRun && updatedRun.binaryOutput && Object.keys(updatedRun.binaryOutput).length > 0) {
try {
const binaryService = new BinaryOutputService('maxun-run-screenshots');
await binaryService.uploadAndStoreBinaryOutput(updatedRun, updatedRun.binaryOutput);
logger.log('info', `Uploaded binary output to MinIO for run ${data.runId}`);
} catch (minioError: any) {
logger.log('error', `Failed to upload binary output to MinIO for run ${data.runId}: ${minioError.message}`);
}
}
let totalSchemaItemsExtracted = 0; let totalSchemaItemsExtracted = 0;
let totalListItemsExtracted = 0; let totalListItemsExtracted = 0;
let extractedScreenshotsCount = 0; let extractedScreenshotsCount = 0;
if (updatedRun) { if (categorizedOutput) {
if (updatedRun.serializableOutput) { if (categorizedOutput.scrapeSchema) {
if (updatedRun.serializableOutput.scrapeSchema) { Object.values(categorizedOutput.scrapeSchema).forEach((schemaResult: any) => {
Object.values(updatedRun.serializableOutput.scrapeSchema).forEach((schemaResult: any) => { if (Array.isArray(schemaResult)) {
if (Array.isArray(schemaResult)) { totalSchemaItemsExtracted += schemaResult.length;
totalSchemaItemsExtracted += schemaResult.length; } else if (schemaResult && typeof schemaResult === 'object') {
} else if (schemaResult && typeof schemaResult === 'object') { totalSchemaItemsExtracted += 1;
totalSchemaItemsExtracted += 1; }
} });
});
}
if (updatedRun.serializableOutput.scrapeList) {
Object.values(updatedRun.serializableOutput.scrapeList).forEach((listResult: any) => {
if (Array.isArray(listResult)) {
totalListItemsExtracted += listResult.length;
}
});
}
} }
if (updatedRun.binaryOutput) { if (categorizedOutput.scrapeList) {
extractedScreenshotsCount = Object.keys(updatedRun.binaryOutput).length; Object.values(categorizedOutput.scrapeList).forEach((listResult: any) => {
if (Array.isArray(listResult)) {
totalListItemsExtracted += listResult.length;
}
});
}
if (run.binaryOutput) {
extractedScreenshotsCount = Object.keys(run.binaryOutput).length;
} }
} }
@@ -302,6 +340,21 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
} }
); );
try {
const completionData = {
runId: data.runId,
robotMetaId: plainRun.robotMetaId,
robotName: recording.recording_meta.name,
status: 'success',
finishedAt: new Date().toLocaleString()
};
serverIo.of(browserId).emit('run-completed', completionData);
serverIo.of('/queued-run').to(`user-${data.userId}`).emit('run-completed', completionData);
} catch (socketError: any) {
logger.log('warn', `Failed to send run-completed notification for API run ${plainRun.runId}: ${socketError.message}`);
}
const webhookPayload = { const webhookPayload = {
robot_id: plainRun.robotMetaId, robot_id: plainRun.robotMetaId,
run_id: data.runId, run_id: data.runId,
@@ -310,12 +363,16 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
started_at: plainRun.startedAt, started_at: plainRun.startedAt,
finished_at: new Date().toLocaleString(), finished_at: new Date().toLocaleString(),
extracted_data: { extracted_data: {
captured_texts: updatedRun?.serializableOutput?.scrapeSchema ? Object.values(updatedRun.serializableOutput.scrapeSchema).flat() : [], captured_texts: Object.keys(categorizedOutput.scrapeSchema || {}).length > 0
captured_lists: updatedRun?.serializableOutput?.scrapeList || {}, ? Object.entries(categorizedOutput.scrapeSchema).reduce((acc, [name, value]) => {
total_rows: totalRowsExtracted, acc[name] = Array.isArray(value) ? value : [value];
return acc;
}, {} as Record<string, any[]>)
: {},
captured_lists: categorizedOutput.scrapeList,
captured_texts_count: totalSchemaItemsExtracted, captured_texts_count: totalSchemaItemsExtracted,
captured_lists_count: totalListItemsExtracted, captured_lists_count: totalListItemsExtracted,
screenshots_count: extractedScreenshotsCount, screenshots_count: extractedScreenshotsCount
}, },
metadata: { metadata: {
browser_id: plainRun.browserId, browser_id: plainRun.browserId,
@@ -330,26 +387,8 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
logger.log('error', `Failed to send webhooks for run ${data.runId}: ${webhookError.message}`); logger.log('error', `Failed to send webhooks for run ${data.runId}: ${webhookError.message}`);
} }
// Schedule updates for Google Sheets and Airtable
await triggerIntegrationUpdates(plainRun.runId, plainRun.robotMetaId); await triggerIntegrationUpdates(plainRun.runId, plainRun.robotMetaId);
// Flush any remaining persistence buffer before emitting socket event
if (browser && browser.interpreter) {
await browser.interpreter.flushPersistenceBuffer();
logger.log('debug', `Flushed persistence buffer before emitting run-completed for run ${data.runId}`);
}
const completionData = {
runId: data.runId,
robotMetaId: plainRun.robotMetaId,
robotName: recording.recording_meta.name,
status: 'success',
finishedAt: new Date().toLocaleString()
};
serverIo.of(browserId).emit('run-completed', completionData);
serverIo.of('/queued-run').to(`user-${data.userId}`).emit('run-completed', completionData);
await destroyRemoteBrowser(browserId, data.userId); await destroyRemoteBrowser(browserId, data.userId);
logger.log('info', `Browser ${browserId} destroyed after successful run ${data.runId}`); logger.log('info', `Browser ${browserId} destroyed after successful run ${data.runId}`);
@@ -416,9 +455,13 @@ async function processRunExecution(job: Job<ExecuteRunData>) {
}, },
partial_data_extracted: partialDataExtracted, partial_data_extracted: partialDataExtracted,
extracted_data: partialDataExtracted ? { extracted_data: partialDataExtracted ? {
captured_texts: Object.values(partialUpdateData.serializableOutput?.scrapeSchema || []).flat() || [], captured_texts: Object.keys(partialUpdateData.serializableOutput?.scrapeSchema || {}).length > 0
? Object.entries(partialUpdateData.serializableOutput.scrapeSchema).reduce((acc, [name, value]) => {
acc[name] = Array.isArray(value) ? value : [value];
return acc;
}, {} as Record<string, any[]>)
: {},
captured_lists: partialUpdateData.serializableOutput?.scrapeList || {}, captured_lists: partialUpdateData.serializableOutput?.scrapeList || {},
total_data_points_extracted: partialData?.totalDataPointsExtracted || 0,
captured_texts_count: partialData?.totalSchemaItemsExtracted || 0, captured_texts_count: partialData?.totalSchemaItemsExtracted || 0,
captured_lists_count: partialData?.totalListItemsExtracted || 0, captured_lists_count: partialData?.totalListItemsExtracted || 0,
screenshots_count: partialData?.extractedScreenshotsCount || 0 screenshots_count: partialData?.extractedScreenshotsCount || 0

View File

@@ -726,34 +726,108 @@ export class WorkflowGenerator {
/** /**
* Generates a pair for the custom action event. * Generates a pair for the custom action event.
*
* @param action The type of the custom action. * @param action The type of the custom action.
* @param settings The settings of the custom action. * @param actionId The unique identifier for this action (for updates)
* @param settings The settings of the custom action (may include name and actionId).
* @param page The page to use for obtaining the needed data. * @param page The page to use for obtaining the needed data.
*/ */
public customAction = async (action: CustomActions, settings: any, page: Page) => { public customAction = async (action: CustomActions, actionId: string, settings: any, page: Page) => {
const pair: WhereWhatPair = { try {
where: { url: this.getBestUrl(page.url()) }, let actionSettings = settings;
what: [{ let actionName: string | undefined;
action,
args: settings ? Array.isArray(settings) ? settings : [settings] : [],
}],
}
await this.addPairToWorkflowAndNotifyClient(pair, page); if (settings && !Array.isArray(settings)) {
actionName = settings.name;
actionSettings = JSON.parse(JSON.stringify(settings));
delete actionSettings.name;
}
if (this.generatedData.lastUsedSelector) { const pair: WhereWhatPair = {
const elementInfo = await this.getLastUsedSelectorInfo(page, this.generatedData.lastUsedSelector); where: { url: this.getBestUrl(page.url()) },
what: [{
action,
args: actionSettings
? Array.isArray(actionSettings)
? actionSettings
: [actionSettings]
: [],
...(actionName ? { name: actionName } : {}),
...(actionId ? { actionId } : {}),
}],
};
this.socket.emit('decision', { if (actionId) {
pair, actionType: 'customAction', const existingIndex = this.workflowRecord.workflow.findIndex(
lastData: { (workflowPair) =>
selector: this.generatedData.lastUsedSelector, Array.isArray(workflowPair.what) &&
action: this.generatedData.lastAction, workflowPair.what.some((whatItem: any) => whatItem.actionId === actionId)
tagName: elementInfo.tagName, );
innerText: elementInfo.innerText,
if (existingIndex !== -1) {
const existingPair = this.workflowRecord.workflow[existingIndex];
const existingAction = existingPair.what.find((whatItem: any) => whatItem.actionId === actionId);
const updatedAction = {
...existingAction,
action,
args: Array.isArray(actionSettings)
? actionSettings
: [actionSettings],
name: actionName || existingAction?.name || '',
actionId,
};
this.workflowRecord.workflow[existingIndex] = {
where: JSON.parse(JSON.stringify(existingPair.where)),
what: existingPair.what.map((whatItem: any) =>
whatItem.actionId === actionId ? updatedAction : whatItem
),
};
if (action === 'scrapeSchema' && actionName) {
this.workflowRecord.workflow.forEach((pair, index) => {
pair.what.forEach((whatItem: any, whatIndex: number) => {
if (whatItem.action === 'scrapeSchema' && whatItem.actionId !== actionId) {
this.workflowRecord.workflow[index].what[whatIndex] = {
...whatItem,
name: actionName
};
}
});
});
}
} else {
await this.addPairToWorkflowAndNotifyClient(pair, page);
logger.log("debug", `Added new workflow action: ${action} with actionId: ${actionId}`);
} }
}); } else {
} await this.addPairToWorkflowAndNotifyClient(pair, page);
logger.log("debug", `Added new workflow action: ${action} without actionId`);
}
if (this.generatedData.lastUsedSelector) {
const elementInfo = await this.getLastUsedSelectorInfo(
page,
this.generatedData.lastUsedSelector
);
this.socket.emit('decision', {
pair,
actionType: 'customAction',
lastData: {
selector: this.generatedData.lastUsedSelector,
action: this.generatedData.lastAction,
tagName: elementInfo.tagName,
innerText: elementInfo.innerText,
},
});
}
} catch (e) {
const { message } = e as Error;
logger.log("warn", `Error handling customAction: ${message}`);
}
}; };
/** /**
@@ -810,6 +884,48 @@ export class WorkflowGenerator {
} }
}; };
/**
* Removes an action with the given actionId from the workflow.
* Only removes the specific action from the what array, not the entire pair.
* If the what array becomes empty after removal, then the entire pair is removed.
* @param actionId The actionId of the action to remove
* @returns boolean indicating whether an action was removed
*/
public removeAction = (actionId: string): boolean => {
let actionWasRemoved = false;
this.workflowRecord.workflow = this.workflowRecord.workflow
.map((pair) => {
const filteredWhat = pair.what.filter(
(whatItem: any) => whatItem.actionId !== actionId
);
if (filteredWhat.length < pair.what.length) {
actionWasRemoved = true;
if (filteredWhat.length > 0) {
return {
...pair,
what: filteredWhat
};
}
return null;
}
return pair;
})
.filter((pair) => pair !== null) as WhereWhatPair[]; // Remove null entries
if (actionWasRemoved) {
logger.log("info", `Action with actionId ${actionId} removed from workflow`);
} else {
logger.log("debug", `No action found with actionId ${actionId}`);
}
return actionWasRemoved;
};
/** /**
* Updates the socket used for communication with the client. * Updates the socket used for communication with the client.
* @param socket The socket to be used for communication. * @param socket The socket to be used for communication.

View File

@@ -91,13 +91,16 @@ export class WorkflowInterpreter {
* Storage for different types of serializable data * Storage for different types of serializable data
*/ */
public serializableDataByType: { public serializableDataByType: {
scrapeSchema: any[], scrapeSchema: Record<string, any>;
scrapeList: any[], scrapeList: Record<string, any>;
[key: string]: any;
} = { } = {
scrapeSchema: [], scrapeSchema: {},
scrapeList: [], scrapeList: {},
}; };
private currentActionName: string | null = null;
/** /**
* Track the current action type being processed * Track the current action type being processed
*/ */
@@ -106,7 +109,7 @@ export class WorkflowInterpreter {
/** /**
* An array of all the binary data extracted from the run. * An array of all the binary data extracted from the run.
*/ */
public binaryData: { mimetype: string, data: string }[] = []; public binaryData: { name: string; mimeType: string; data: string }[] = [];
/** /**
* Track current scrapeList index * Track current scrapeList index
@@ -259,14 +262,19 @@ export class WorkflowInterpreter {
} }
}, },
binaryCallback: async (data: string, mimetype: string) => { binaryCallback: async (data: string, mimetype: string) => {
const binaryItem = { mimetype, data: JSON.stringify(data) }; // For editor mode, we don't have the name yet, so use a timestamp-based name
const binaryItem = {
name: `Screenshot ${Date.now()}`,
mimeType: mimetype,
data: JSON.stringify(data)
};
this.binaryData.push(binaryItem); this.binaryData.push(binaryItem);
// Persist binary data to database // Persist binary data to database
await this.persistBinaryDataToDatabase(binaryItem); await this.persistBinaryDataToDatabase(binaryItem);
this.socket.emit('binaryCallback', { this.socket.emit('binaryCallback', {
data, data,
mimetype, mimetype,
type: 'captureScreenshot' type: 'captureScreenshot'
}); });
@@ -364,9 +372,10 @@ export class WorkflowInterpreter {
this.breakpoints = []; this.breakpoints = [];
this.interpretationResume = null; this.interpretationResume = null;
this.currentActionType = null; this.currentActionType = null;
this.currentActionName = null;
this.serializableDataByType = { this.serializableDataByType = {
scrapeSchema: [], scrapeSchema: {},
scrapeList: [], scrapeList: {},
}; };
this.binaryData = []; this.binaryData = [];
this.currentScrapeListIndex = 0; this.currentScrapeListIndex = 0;
@@ -409,7 +418,7 @@ export class WorkflowInterpreter {
* Persists binary data to database in real-time * Persists binary data to database in real-time
* @private * @private
*/ */
private persistBinaryDataToDatabase = async (binaryItem: { mimetype: string, data: string }): Promise<void> => { private persistBinaryDataToDatabase = async (binaryItem: { name: string; mimeType: string; data: string }): Promise<void> => {
if (!this.currentRunId) { if (!this.currentRunId) {
logger.log('debug', 'No run ID available for binary data persistence'); logger.log('debug', 'No run ID available for binary data persistence');
return; return;
@@ -422,22 +431,29 @@ export class WorkflowInterpreter {
return; return;
} }
const currentBinaryOutput = run.binaryOutput ? const currentBinaryOutput =
JSON.parse(JSON.stringify(run.binaryOutput)) : run.binaryOutput && typeof run.binaryOutput === 'object'
{}; ? JSON.parse(JSON.stringify(run.binaryOutput))
: {};
const uniqueKey = `item-${Date.now()}-${Object.keys(currentBinaryOutput).length}`;
const baseName = binaryItem.name?.trim() || `Screenshot ${Object.keys(currentBinaryOutput).length + 1}`;
let uniqueName = baseName;
let counter = 1;
while (currentBinaryOutput[uniqueName]) {
uniqueName = `${baseName} (${counter++})`;
}
const updatedBinaryOutput = { const updatedBinaryOutput = {
...currentBinaryOutput, ...currentBinaryOutput,
[uniqueKey]: binaryItem [uniqueName]: binaryItem,
}; };
await run.update({ await run.update({
binaryOutput: updatedBinaryOutput binaryOutput: updatedBinaryOutput
}); });
logger.log('debug', `Persisted binary data for run ${this.currentRunId}: ${binaryItem.mimetype}`); logger.log('debug', `Persisted binary data for run ${this.currentRunId}: ${binaryItem.name} (${binaryItem.mimeType})`);
} catch (error: any) { } catch (error: any) {
logger.log('error', `Failed to persist binary data in real-time for run ${this.currentRunId}: ${error.message}`); logger.log('error', `Failed to persist binary data in real-time for run ${this.currentRunId}: ${error.message}`);
} }
@@ -478,41 +494,101 @@ export class WorkflowInterpreter {
}, },
incrementScrapeListIndex: () => { incrementScrapeListIndex: () => {
this.currentScrapeListIndex++; this.currentScrapeListIndex++;
} },
setActionName: (name: string) => {
this.currentActionName = name;
},
}, },
serializableCallback: async (data: any) => { serializableCallback: async (data: any) => {
if (this.currentActionType === 'scrapeSchema') { try {
if (Array.isArray(data) && data.length > 0) { if (!data || typeof data !== "object") return;
mergedScrapeSchema = { ...mergedScrapeSchema, ...data[0] };
this.serializableDataByType.scrapeSchema.push(data); if (!this.currentActionType && Array.isArray(data) && data.length > 0) {
} else { const first = data[0];
mergedScrapeSchema = { ...mergedScrapeSchema, ...data }; if (first && Object.keys(first).some(k => k.toLowerCase().includes("label") || k.toLowerCase().includes("text"))) {
this.serializableDataByType.scrapeSchema.push([data]); this.currentActionType = "scrapeSchema";
}
} }
// Persist the cumulative scrapeSchema data let typeKey = this.currentActionType || "unknown";
const cumulativeScrapeSchemaData = Object.keys(mergedScrapeSchema).length > 0 ? [mergedScrapeSchema] : [];
if (cumulativeScrapeSchemaData.length > 0) { if (this.currentActionType === "scrapeList") {
await this.persistDataToDatabase('scrapeSchema', cumulativeScrapeSchemaData); typeKey = "scrapeList";
} else if (this.currentActionType === "scrapeSchema") {
typeKey = "scrapeSchema";
} }
} else if (this.currentActionType === 'scrapeList') {
if (data && Array.isArray(data) && data.length > 0) { if (this.currentActionType === "scrapeList" && data.scrapeList) {
// Use the current index for persistence data = data.scrapeList;
await this.persistDataToDatabase('scrapeList', data, this.currentScrapeListIndex); } else if (this.currentActionType === "scrapeSchema" && data.scrapeSchema) {
data = data.scrapeSchema;
} }
this.serializableDataByType.scrapeList[this.currentScrapeListIndex] = data;
} let actionName = this.currentActionName || "";
this.socket.emit('serializableCallback', data); if (!actionName) {
if (!Array.isArray(data) && Object.keys(data).length === 1) {
const soleKey = Object.keys(data)[0];
const soleValue = data[soleKey];
if (Array.isArray(soleValue) || typeof soleValue === "object") {
actionName = soleKey;
data = soleValue;
}
}
}
if (!actionName) {
actionName = "Unnamed Action";
}
const flattened = Array.isArray(data)
? data
: (data?.List ?? (data && typeof data === 'object' ? Object.values(data).flat?.() ?? data : []));
if (!this.serializableDataByType[typeKey]) {
this.serializableDataByType[typeKey] = {};
}
this.serializableDataByType[typeKey][actionName] = flattened;
await this.persistDataToDatabase(typeKey, { [actionName]: flattened });
this.socket.emit("serializableCallback", {
type: typeKey,
name: actionName,
data: flattened,
});
this.currentActionType = null;
this.currentActionName = null;
} catch (err: any) {
logger.log('error', `serializableCallback handler failed: ${err.message}`);
}
}, },
binaryCallback: async (data: string, mimetype: string) => { binaryCallback: async (payload: { name: string; data: Buffer; mimeType: string }) => {
const binaryItem = { mimetype, data: JSON.stringify(data) }; try {
this.binaryData.push(binaryItem); const { name, data, mimeType } = payload;
// Persist binary data to database const base64Data = data.toString("base64");
await this.persistBinaryDataToDatabase(binaryItem);
const binaryItem = {
this.socket.emit('binaryCallback', { data, mimetype }); name,
mimeType,
data: base64Data
};
this.binaryData.push(binaryItem);
await this.persistBinaryDataToDatabase(binaryItem);
this.socket.emit("binaryCallback", {
name,
data: base64Data,
mimeType
});
} catch (err: any) {
logger.log("error", `binaryCallback handler failed: ${err.message}`);
}
} }
} }
@@ -542,20 +618,13 @@ export class WorkflowInterpreter {
const result = { const result = {
log: this.debugMessages, log: this.debugMessages,
result: status, result: status,
scrapeSchemaOutput: Object.keys(mergedScrapeSchema).length > 0 scrapeSchemaOutput: this.serializableDataByType.scrapeSchema,
? { "schema_merged": [mergedScrapeSchema] } scrapeListOutput: this.serializableDataByType.scrapeList,
: this.serializableDataByType.scrapeSchema.reduce((reducedObject, item, index) => { binaryOutput: this.binaryData.reduce<Record<string, { data: string; mimeType: string }>>((acc, item) => {
reducedObject[`schema_${index}`] = item; const key = item.name || `Screenshot ${Object.keys(acc).length + 1}`;
return reducedObject; acc[key] = { data: item.data, mimeType: item.mimeType };
}, {} as Record<string, any>), return acc;
scrapeListOutput: this.serializableDataByType.scrapeList.reduce((reducedObject, item, index) => { }, {})
reducedObject[`list_${index}`] = item;
return reducedObject;
}, {} as Record<string, any>),
binaryOutput: this.binaryData.reduce((reducedObject, item, index) => {
reducedObject[`item_${index}`] = item;
return reducedObject;
}, {} as Record<string, any>)
} }
logger.log('debug', `Interpretation finished`); logger.log('debug', `Interpretation finished`);
@@ -642,19 +711,37 @@ export class WorkflowInterpreter {
const currentSerializableOutput = run.serializableOutput ? const currentSerializableOutput = run.serializableOutput ?
JSON.parse(JSON.stringify(run.serializableOutput)) : JSON.parse(JSON.stringify(run.serializableOutput)) :
{ scrapeSchema: [], scrapeList: [] }; { scrapeSchema: [], scrapeList: [] };
if (Array.isArray(currentSerializableOutput.scrapeList)) {
currentSerializableOutput.scrapeList = {};
}
if (Array.isArray(currentSerializableOutput.scrapeSchema)) {
currentSerializableOutput.scrapeSchema = {};
}
let hasUpdates = false; let hasUpdates = false;
const mergeLists = (target: Record<string, any>, updates: Record<string, any>) => {
for (const [key, val] of Object.entries(updates)) {
const flattened = Array.isArray(val)
? val
: (val?.List ?? (val && typeof val === 'object' ? Object.values(val).flat?.() ?? val : []));
target[key] = flattened;
}
};
for (const item of batchToProcess) { for (const item of batchToProcess) {
if (item.actionType === 'scrapeSchema') { if (item.actionType === 'scrapeSchema') {
const newSchemaData = Array.isArray(item.data) ? item.data : [item.data]; if (!currentSerializableOutput.scrapeSchema || typeof currentSerializableOutput.scrapeSchema !== 'object') {
currentSerializableOutput.scrapeSchema = newSchemaData; currentSerializableOutput.scrapeSchema = {};
hasUpdates = true;
} else if (item.actionType === 'scrapeList' && typeof item.listIndex === 'number') {
if (!Array.isArray(currentSerializableOutput.scrapeList)) {
currentSerializableOutput.scrapeList = [];
} }
currentSerializableOutput.scrapeList[item.listIndex] = item.data; mergeLists(currentSerializableOutput.scrapeSchema, item.data);
hasUpdates = true;
} else if (item.actionType === 'scrapeList') {
if (!currentSerializableOutput.scrapeList || typeof currentSerializableOutput.scrapeList !== 'object') {
currentSerializableOutput.scrapeList = {};
}
mergeLists(currentSerializableOutput.scrapeList, item.data);
hasUpdates = true; hasUpdates = true;
} }
} }

View File

@@ -12,8 +12,8 @@ interface AirtableUpdateTask {
} }
interface SerializableOutput { interface SerializableOutput {
scrapeSchema?: any[]; scrapeSchema?: Record<string, any[]>;
scrapeList?: any[]; scrapeList?: Record<string, any[]>;
} }
const MAX_RETRIES = 3; const MAX_RETRIES = 3;
@@ -48,47 +48,94 @@ async function refreshAirtableToken(refreshToken: string) {
function mergeRelatedData(serializableOutput: SerializableOutput, binaryOutput: Record<string, string>) { function mergeRelatedData(serializableOutput: SerializableOutput, binaryOutput: Record<string, string>) {
const allRecords: Record<string, any>[] = []; const allRecords: Record<string, any>[] = [];
const schemaData: Array<{key: string, value: any}> = []; const schemaData: Array<{ Group: string; Field: string; Value: any }> = [];
const listData: any[] = []; const listData: any[] = [];
const screenshotData: Array<{key: string, url: string}> = []; const screenshotData: Array<{key: string, url: string}> = [];
// Collect schema data // Collect schema data
if (serializableOutput.scrapeSchema) { if (serializableOutput.scrapeSchema) {
for (const schemaArray of serializableOutput.scrapeSchema) { if (Array.isArray(serializableOutput.scrapeSchema)) {
if (!Array.isArray(schemaArray)) continue; for (const schemaArray of serializableOutput.scrapeSchema) {
for (const schemaItem of schemaArray) { if (!Array.isArray(schemaArray)) continue;
Object.entries(schemaItem).forEach(([key, value]) => { for (const schemaItem of schemaArray) {
if (key && key.trim() !== '' && value !== null && value !== undefined && value !== '') { Object.entries(schemaItem || {}).forEach(([key, value]) => {
schemaData.push({key, value}); if (key && key.trim() !== "" && value !== null && value !== undefined && value !== "") {
} schemaData.push({ Group: "Default", Field: key, Value: value });
}); }
});
}
}
} else if (typeof serializableOutput.scrapeSchema === "object") {
for (const [groupName, schemaArray] of Object.entries(serializableOutput.scrapeSchema)) {
if (!Array.isArray(schemaArray)) continue;
for (const schemaItem of schemaArray) {
Object.entries(schemaItem || {}).forEach(([fieldName, value]) => {
if (fieldName && fieldName.trim() !== "" && value !== null && value !== undefined && value !== "") {
schemaData.push({
Group: groupName,
Field: fieldName,
Value: value,
});
}
});
}
} }
} }
} }
// Collect list data // Collect list data
if (serializableOutput.scrapeList) { if (serializableOutput.scrapeList) {
for (const listArray of serializableOutput.scrapeList) { if (Array.isArray(serializableOutput.scrapeList)) {
if (!Array.isArray(listArray)) continue; for (const listArray of serializableOutput.scrapeList) {
listArray.forEach(listItem => { if (!Array.isArray(listArray)) continue;
const hasContent = Object.values(listItem).some(value => listArray.forEach((listItem) => {
value !== null && value !== undefined && value !== '' const hasContent = Object.values(listItem || {}).some(
); (value) => value !== null && value !== undefined && value !== ""
if (hasContent) { );
listData.push(listItem); if (hasContent) listData.push(listItem);
} });
}); }
} else if (typeof serializableOutput.scrapeList === "object") {
for (const [listName, listArray] of Object.entries(serializableOutput.scrapeList)) {
if (!Array.isArray(listArray)) continue;
listArray.forEach((listItem) => {
const hasContent = Object.values(listItem || {}).some(
(value) => value !== null && value !== undefined && value !== ""
);
if (hasContent) listData.push({ List: listName, ...listItem });
});
}
} }
} }
// Collect screenshot data // Collect screenshot data
if (binaryOutput && Object.keys(binaryOutput).length > 0) { // if (binaryOutput && Object.keys(binaryOutput).length > 0) {
Object.entries(binaryOutput).forEach(([key, url]) => { // Object.entries(binaryOutput).forEach(([key, rawValue]: [string, any]) => {
if (key && key.trim() !== '' && url && url.trim() !== '') { // if (!key || key.trim() === "") return;
screenshotData.push({key, url});
} // let urlString = "";
});
} // // Case 1: old format (string URL)
// if (typeof rawValue === "string") {
// urlString = rawValue;
// }
// // Case 2: new format (object with { url?, data?, mimeType? })
// else if (rawValue && typeof rawValue === "object") {
// const valueObj = rawValue as { url?: string; data?: string; mimeType?: string };
// if (typeof valueObj.url === "string") {
// urlString = valueObj.url;
// } else if (typeof valueObj.data === "string") {
// const mime = valueObj.mimeType || "image/png";
// urlString = `data:${mime};base64,${valueObj.data}`;
// }
// }
// if (typeof urlString === "string" && urlString.trim() !== "") {
// screenshotData.push({ key, url: urlString });
// }
// });
// }
// Mix all data types together to create consecutive records // Mix all data types together to create consecutive records
const maxLength = Math.max(schemaData.length, listData.length, screenshotData.length); const maxLength = Math.max(schemaData.length, listData.length, screenshotData.length);
@@ -97,8 +144,9 @@ function mergeRelatedData(serializableOutput: SerializableOutput, binaryOutput:
const record: Record<string, any> = {}; const record: Record<string, any> = {};
if (i < schemaData.length) { if (i < schemaData.length) {
record.Label = schemaData[i].key; record.Group = schemaData[i].Group;
record.Value = schemaData[i].value; record.Label = schemaData[i].Field;
record.Value = schemaData[i].Value;
} }
if (i < listData.length) { if (i < listData.length) {
@@ -120,20 +168,15 @@ function mergeRelatedData(serializableOutput: SerializableOutput, binaryOutput:
} }
for (let i = maxLength; i < schemaData.length; i++) { for (let i = maxLength; i < schemaData.length; i++) {
allRecords.push({ allRecords.push({ Label: schemaData[i].Field, Value: schemaData[i].Value });
Label: schemaData[i].key,
Value: schemaData[i].value
});
} }
for (let i = maxLength; i < listData.length; i++) { for (let i = maxLength; i < listData.length; i++) {
allRecords.push(listData[i]); allRecords.push(listData[i]);
} }
for (let i = maxLength; i < screenshotData.length; i++) { for (let i = maxLength; i < screenshotData.length; i++) {
allRecords.push({ allRecords.push({
Key: screenshotData[i].key, Key: screenshotData[i].key,
Screenshot: screenshotData[i].url Screenshot: screenshotData[i].url,
}); });
} }

View File

@@ -49,25 +49,34 @@ export async function updateGoogleSheet(robotId: string, runId: string) {
const serializableOutput = plainRun.serializableOutput as SerializableOutput; const serializableOutput = plainRun.serializableOutput as SerializableOutput;
if (serializableOutput) { if (serializableOutput) {
if (serializableOutput.scrapeSchema && serializableOutput.scrapeSchema.length > 0) { if (serializableOutput.scrapeSchema && typeof serializableOutput.scrapeSchema === "object") {
await processOutputType( for (const [groupName, schemaArray] of Object.entries(serializableOutput.scrapeSchema)) {
robotId, if (!Array.isArray(schemaArray) || schemaArray.length === 0) continue;
spreadsheetId,
'Text', await processOutputType(
serializableOutput.scrapeSchema, robotId,
plainRobot spreadsheetId,
); `Schema - ${groupName}`,
schemaArray,
plainRobot
);
}
} }
if (serializableOutput.scrapeList && serializableOutput.scrapeList.length > 0) { if (serializableOutput.scrapeList && typeof serializableOutput.scrapeList === "object") {
await processOutputType( for (const [listName, listArray] of Object.entries(serializableOutput.scrapeList)) {
robotId, if (!Array.isArray(listArray) || listArray.length === 0) continue;
spreadsheetId,
'List', await processOutputType(
serializableOutput.scrapeList, robotId,
plainRobot spreadsheetId,
); `List - ${listName}`,
listArray,
plainRobot
);
}
} }
} }
if (plainRun.binaryOutput && Object.keys(plainRun.binaryOutput).length > 0) { if (plainRun.binaryOutput && Object.keys(plainRun.binaryOutput).length > 0) {
@@ -102,30 +111,27 @@ async function processOutputType(
outputData: any[], outputData: any[],
robotConfig: any robotConfig: any
) { ) {
for (let i = 0; i < outputData.length; i++) { const data = outputData;
const data = outputData[i]; const sheetName = outputType;
if (!data || data.length === 0) { if (!Array.isArray(data) || data.length === 0) {
console.log(`No data to write for ${outputType}-${i}. Skipping.`); console.log(`No data to write for ${sheetName}. Skipping.`);
continue; return;
}
const sheetName = `${outputType}-${i}`;
await ensureSheetExists(spreadsheetId, sheetName, robotConfig);
let formattedData = data;
if (outputType === 'Text' && data.length > 0) {
const schemaItem = data[0];
formattedData = Object.entries(schemaItem).map(([key, value]) => ({
Label: key,
Value: value
}));
}
await writeDataToSheet(robotId, spreadsheetId, formattedData, sheetName, robotConfig);
console.log(`Data written to ${sheetName} sheet for ${outputType} data`);
} }
await ensureSheetExists(spreadsheetId, sheetName, robotConfig);
const formattedData = data.map(item => {
const flatRow: Record<string, any> = {};
for (const [key, value] of Object.entries(item || {})) {
flatRow[key] =
typeof value === "object" && value !== null ? JSON.stringify(value) : value;
}
return flatRow;
});
await writeDataToSheet(robotId, spreadsheetId, formattedData, sheetName, robotConfig);
console.log(`Data written to ${sheetName} sheet for ${outputType} data`);
} }
async function ensureSheetExists(spreadsheetId: string, sheetName: string, robotConfig: any) { async function ensureSheetExists(spreadsheetId: string, sheetName: string, robotConfig: any) {

View File

@@ -4,7 +4,7 @@ import stealthPlugin from 'puppeteer-extra-plugin-stealth';
import { io, Socket } from "socket.io-client"; import { io, Socket } from "socket.io-client";
import { createRemoteBrowserForRun, destroyRemoteBrowser } from '../../browser-management/controller'; import { createRemoteBrowserForRun, destroyRemoteBrowser } from '../../browser-management/controller';
import logger from '../../logger'; import logger from '../../logger';
import { browserPool } from "../../server"; import { browserPool, io as serverIo } from "../../server";
import { googleSheetUpdateTasks, processGoogleSheetUpdates } from "../integrations/gsheet"; import { googleSheetUpdateTasks, processGoogleSheetUpdates } from "../integrations/gsheet";
import Robot from "../../models/Robot"; import Robot from "../../models/Robot";
import Run from "../../models/Run"; import Run from "../../models/Run";
@@ -46,7 +46,7 @@ async function createWorkflowAndStoreMetadata(id: string, userId: string) {
}; };
} }
const browserId = createRemoteBrowserForRun( userId); const browserId = createRemoteBrowserForRun(userId);
const runId = uuid(); const runId = uuid();
const run = await Run.create({ const run = await Run.create({
@@ -63,10 +63,30 @@ async function createWorkflowAndStoreMetadata(id: string, userId: string) {
runByScheduleId: uuid(), runByScheduleId: uuid(),
serializableOutput: {}, serializableOutput: {},
binaryOutput: {}, binaryOutput: {},
retryCount: 0
}); });
const plainRun = run.toJSON(); const plainRun = run.toJSON();
try {
const runScheduledData = {
runId: plainRun.runId,
robotMetaId: plainRun.robotMetaId,
robotName: plainRun.name,
status: 'scheduled',
startedAt: plainRun.startedAt,
runByUserId: plainRun.runByUserId,
runByScheduleId: plainRun.runByScheduleId,
runByAPI: plainRun.runByAPI || false,
browserId: plainRun.browserId
};
serverIo.of('/queued-run').to(`user-${userId}`).emit('run-scheduled', runScheduledData);
logger.log('info', `Scheduled run notification sent for run: ${plainRun.runId} to user-${userId}`);
} catch (socketError: any) {
logger.log('warn', `Failed to send run-scheduled notification for run ${plainRun.runId}: ${socketError.message}`);
}
return { return {
browserId, browserId,
runId: plainRun.runId, runId: plainRun.runId,
@@ -83,6 +103,29 @@ async function createWorkflowAndStoreMetadata(id: string, userId: string) {
} }
} }
async function triggerIntegrationUpdates(runId: string, robotMetaId: string): Promise<void> {
try {
googleSheetUpdateTasks[runId] = {
robotId: robotMetaId,
runId: runId,
status: 'pending',
retries: 5,
};
airtableUpdateTasks[runId] = {
robotId: robotMetaId,
runId: runId,
status: 'pending',
retries: 5,
};
processAirtableUpdates().catch(err => logger.log('error', `Airtable update error: ${err.message}`));
processGoogleSheetUpdates().catch(err => logger.log('error', `Google Sheets update error: ${err.message}`));
} catch (err: any) {
logger.log('error', `Failed to update integrations for run: ${runId}: ${err.message}`);
}
}
function AddGeneratedFlags(workflow: WorkflowFile) { function AddGeneratedFlags(workflow: WorkflowFile) {
const copy = JSON.parse(JSON.stringify(workflow)); const copy = JSON.parse(JSON.stringify(workflow));
for (let i = 0; i < workflow.workflow.length; i++) { for (let i = 0; i < workflow.workflow.length; i++) {
@@ -95,6 +138,8 @@ function AddGeneratedFlags(workflow: WorkflowFile) {
}; };
async function executeRun(id: string, userId: string) { async function executeRun(id: string, userId: string) {
let browser: any = null;
try { try {
const run = await Run.findOne({ where: { runId: id } }); const run = await Run.findOne({ where: { runId: id } });
if (!run) { if (!run) {
@@ -133,6 +178,21 @@ async function executeRun(id: string, userId: string) {
log: plainRun.log ? `${plainRun.log}\nMax retries exceeded (3/3) - Run failed after multiple attempts.` : `Max retries exceeded (3/3) - Run failed after multiple attempts.` log: plainRun.log ? `${plainRun.log}\nMax retries exceeded (3/3) - Run failed after multiple attempts.` : `Max retries exceeded (3/3) - Run failed after multiple attempts.`
}); });
try {
const failureSocketData = {
runId: plainRun.runId,
robotMetaId: plainRun.robotMetaId,
robotName: recording ? recording.recording_meta.name : 'Unknown Robot',
status: 'failed',
finishedAt: new Date().toLocaleString()
};
serverIo.of(run.browserId).emit('run-completed', failureSocketData);
serverIo.of('/queued-run').to(`user-${userId}`).emit('run-completed', failureSocketData);
} catch (socketError: any) {
logger.log('warn', `Failed to emit failure event in main catch: ${socketError.message}`);
}
return { return {
success: false, success: false,
error: 'Max retries exceeded' error: 'Max retries exceeded'
@@ -149,7 +209,22 @@ async function executeRun(id: string, userId: string) {
plainRun.status = 'running'; plainRun.status = 'running';
const browser = browserPool.getRemoteBrowser(plainRun.browserId); try {
const runStartedData = {
runId: plainRun.runId,
robotMetaId: plainRun.robotMetaId,
robotName: recording ? recording.recording_meta.name : 'Unknown Robot',
status: 'running',
startedAt: plainRun.startedAt
};
serverIo.of('/queued-run').to(`user-${userId}`).emit('run-started', runStartedData);
logger.log('info', `Run started notification sent for run: ${plainRun.runId} to user-${userId}`);
} catch (socketError: any) {
logger.log('warn', `Failed to send run-started notification for run ${plainRun.runId}: ${socketError.message}`);
}
browser = browserPool.getRemoteBrowser(plainRun.browserId);
if (!browser) { if (!browser) {
throw new Error('Could not access browser'); throw new Error('Could not access browser');
} }
@@ -168,56 +243,52 @@ async function executeRun(id: string, userId: string) {
workflow, currentPage, (newPage: Page) => currentPage = newPage, plainRun.interpreterSettings workflow, currentPage, (newPage: Page) => currentPage = newPage, plainRun.interpreterSettings
); );
const binaryOutputService = new BinaryOutputService('maxun-run-screenshots');
const uploadedBinaryOutput = await binaryOutputService.uploadAndStoreBinaryOutput(run, interpretationInfo.binaryOutput);
const finalRun = await Run.findByPk(run.id);
const categorizedOutput = {
scrapeSchema: finalRun?.serializableOutput?.scrapeSchema || {},
scrapeList: finalRun?.serializableOutput?.scrapeList || {},
};
await destroyRemoteBrowser(plainRun.browserId, userId); await destroyRemoteBrowser(plainRun.browserId, userId);
await run.update({ await run.update({
status: 'success', status: 'success',
finishedAt: new Date().toLocaleString(), finishedAt: new Date().toLocaleString(),
log: interpretationInfo.log.join('\n'), log: interpretationInfo.log.join('\n'),
binaryOutput: uploadedBinaryOutput
}); });
// Upload binary output to MinIO and update run with MinIO URLs
const updatedRun = await Run.findOne({ where: { runId: id } });
if (updatedRun && updatedRun.binaryOutput && Object.keys(updatedRun.binaryOutput).length > 0) {
try {
const binaryService = new BinaryOutputService('maxun-run-screenshots');
await binaryService.uploadAndStoreBinaryOutput(updatedRun, updatedRun.binaryOutput);
logger.log('info', `Uploaded binary output to MinIO for scheduled run ${id}`);
} catch (minioError: any) {
logger.log('error', `Failed to upload binary output to MinIO for scheduled run ${id}: ${minioError.message}`);
}
}
// Get metrics from persisted data for analytics and webhooks // Get metrics from persisted data for analytics and webhooks
let totalSchemaItemsExtracted = 0; let totalSchemaItemsExtracted = 0;
let totalListItemsExtracted = 0; let totalListItemsExtracted = 0;
let extractedScreenshotsCount = 0; let extractedScreenshotsCount = 0;
if (updatedRun) { if (categorizedOutput) {
if (updatedRun.serializableOutput) { if (categorizedOutput.scrapeSchema) {
if (updatedRun.serializableOutput.scrapeSchema) { Object.values(categorizedOutput.scrapeSchema).forEach((schemaResult: any) => {
Object.values(updatedRun.serializableOutput.scrapeSchema).forEach((schemaResult: any) => { if (Array.isArray(schemaResult)) {
if (Array.isArray(schemaResult)) { totalSchemaItemsExtracted += schemaResult.length;
totalSchemaItemsExtracted += schemaResult.length; } else if (schemaResult && typeof schemaResult === 'object') {
} else if (schemaResult && typeof schemaResult === 'object') { totalSchemaItemsExtracted += 1;
totalSchemaItemsExtracted += 1; }
} });
});
}
if (updatedRun.serializableOutput.scrapeList) {
Object.values(updatedRun.serializableOutput.scrapeList).forEach((listResult: any) => {
if (Array.isArray(listResult)) {
totalListItemsExtracted += listResult.length;
}
});
}
} }
if (updatedRun.binaryOutput) { if (categorizedOutput.scrapeList) {
extractedScreenshotsCount = Object.keys(updatedRun.binaryOutput).length; Object.values(categorizedOutput.scrapeList).forEach((listResult: any) => {
if (Array.isArray(listResult)) {
totalListItemsExtracted += listResult.length;
}
});
} }
} }
if (run.binaryOutput) {
extractedScreenshotsCount = Object.keys(run.binaryOutput).length;
}
const totalRowsExtracted = totalSchemaItemsExtracted + totalListItemsExtracted; const totalRowsExtracted = totalSchemaItemsExtracted + totalListItemsExtracted;
@@ -234,6 +305,21 @@ async function executeRun(id: string, userId: string) {
} }
); );
try {
const completionData = {
runId: plainRun.runId,
robotMetaId: plainRun.robotMetaId,
robotName: recording.recording_meta.name,
status: 'success',
finishedAt: new Date().toLocaleString()
};
serverIo.of(plainRun.browserId).emit('run-completed', completionData);
serverIo.of('/queued-run').to(`user-${userId}`).emit('run-completed', completionData);
} catch (emitError: any) {
logger.log('warn', `Failed to emit success event: ${emitError.message}`);
}
const webhookPayload = { const webhookPayload = {
robot_id: plainRun.robotMetaId, robot_id: plainRun.robotMetaId,
run_id: plainRun.runId, run_id: plainRun.runId,
@@ -242,16 +328,20 @@ async function executeRun(id: string, userId: string) {
started_at: plainRun.startedAt, started_at: plainRun.startedAt,
finished_at: new Date().toLocaleString(), finished_at: new Date().toLocaleString(),
extracted_data: { extracted_data: {
captured_texts: updatedRun?.serializableOutput?.scrapeSchema ? Object.values(updatedRun.serializableOutput.scrapeSchema).flat() : [], captured_texts: Object.keys(categorizedOutput.scrapeSchema || {}).length > 0
captured_lists: updatedRun?.serializableOutput?.scrapeList || {}, ? Object.entries(categorizedOutput.scrapeSchema).reduce((acc, [name, value]) => {
total_rows: totalRowsExtracted, acc[name] = Array.isArray(value) ? value : [value];
return acc;
}, {} as Record<string, any[]>)
: {},
captured_lists: categorizedOutput.scrapeList,
captured_texts_count: totalSchemaItemsExtracted, captured_texts_count: totalSchemaItemsExtracted,
captured_lists_count: totalListItemsExtracted, captured_lists_count: totalListItemsExtracted,
screenshots_count: extractedScreenshotsCount screenshots_count: extractedScreenshotsCount
}, },
metadata: { metadata: {
browser_id: plainRun.browserId, browser_id: plainRun.browserId,
user_id: userId user_id: userId,
} }
}; };
@@ -262,26 +352,7 @@ async function executeRun(id: string, userId: string) {
logger.log('error', `Failed to send webhooks for run ${plainRun.runId}: ${webhookError.message}`); logger.log('error', `Failed to send webhooks for run ${plainRun.runId}: ${webhookError.message}`);
} }
try { await triggerIntegrationUpdates(plainRun.runId, plainRun.robotMetaId);
googleSheetUpdateTasks[plainRun.runId] = {
robotId: plainRun.robotMetaId,
runId: plainRun.runId,
status: 'pending',
retries: 5,
};
airtableUpdateTasks[plainRun.runId] = {
robotId: plainRun.robotMetaId,
runId: plainRun.runId,
status: 'pending',
retries: 5,
};
processAirtableUpdates().catch(err => logger.log('error', `Airtable update error: ${err.message}`));
processGoogleSheetUpdates().catch(err => logger.log('error', `Google Sheets update error: ${err.message}`));
} catch (err: any) {
logger.log('error', `Failed to update Google Sheet for run: ${plainRun.runId}: ${err.message}`);
}
return true; return true;
} catch (error: any) { } catch (error: any) {
logger.log('info', `Error while running a robot with id: ${id} - ${error.message}`); logger.log('info', `Error while running a robot with id: ${id} - ${error.message}`);
@@ -320,6 +391,21 @@ async function executeRun(id: string, userId: string) {
} catch (webhookError: any) { } catch (webhookError: any) {
logger.log('error', `Failed to send failure webhooks for run ${run.runId}: ${webhookError.message}`); logger.log('error', `Failed to send failure webhooks for run ${run.runId}: ${webhookError.message}`);
} }
try {
const failureSocketData = {
runId: run.runId,
robotMetaId: run.robotMetaId,
robotName: recording ? recording.recording_meta.name : 'Unknown Robot',
status: 'failed',
finishedAt: new Date().toLocaleString()
};
serverIo.of(run.browserId).emit('run-completed', failureSocketData);
serverIo.of('/queued-run').to(`user-${userId}`).emit('run-completed', failureSocketData);
} catch (socketError: any) {
logger.log('warn', `Failed to emit failure event in main catch: ${socketError.message}`);
}
} }
capture( capture(
'maxun-oss-run-created-scheduled', 'maxun-oss-run-created-scheduled',

View File

@@ -1,6 +1,7 @@
import React, { createContext, useContext, useEffect, useRef, useState } from 'react'; import React, { createContext, useContext, useEffect, useRef, useState } from 'react';
import { useSocketStore } from "./socket"; import { useSocketStore } from "./socket";
import { useGlobalInfoStore } from "./globalInfo"; import { useGlobalInfoStore } from "./globalInfo";
import { useActionContext } from './browserActions';
export interface TextStep { export interface TextStep {
id: number; id: number;
@@ -96,6 +97,7 @@ export const BrowserStepsProvider: React.FC<{ children: React.ReactNode }> = ({
const { currentTextGroupName } = useGlobalInfoStore(); const { currentTextGroupName } = useGlobalInfoStore();
const [browserSteps, setBrowserSteps] = useState<BrowserStep[]>([]); const [browserSteps, setBrowserSteps] = useState<BrowserStep[]>([]);
const [discardedFields, setDiscardedFields] = useState<Set<string>>(new Set()); const [discardedFields, setDiscardedFields] = useState<Set<string>>(new Set());
const { paginationType, limitType, customLimit } = useActionContext();
const browserStepsRef = useRef<BrowserStep[]>(browserSteps); const browserStepsRef = useRef<BrowserStep[]>(browserSteps);
useEffect(() => { useEffect(() => {
@@ -127,15 +129,21 @@ export const BrowserStepsProvider: React.FC<{ children: React.ReactNode }> = ({
} }
}); });
const livePaginationType = paginationType || listStep.pagination?.type || "";
const liveLimit =
limitType === "custom"
? parseInt(customLimit || "0", 10)
: parseInt(limitType || "0", 10);
return { return {
listSelector: listStep.listSelector, listSelector: listStep.listSelector,
fields: fields, fields: fields,
pagination: { pagination: {
type: listStep.pagination?.type || "", type: livePaginationType,
selector: listStep.pagination?.selector, selector: listStep.pagination?.selector,
isShadow: listStep.isShadow isShadow: listStep.isShadow
}, },
limit: listStep.limit, limit: liveLimit > 0 ? liveLimit : listStep.limit,
isShadow: listStep.isShadow isShadow: listStep.isShadow
}; };
}; };