Compare commits
12 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
4e9aa0efc3 | ||
| e5e4b1316c | |||
| e8d511b4de | |||
|
|
c4bbf5c251 | ||
| 32a9e6732b | |||
| e7c076e2ed | |||
|
|
dbe8e72efe | ||
| 38bd193042 | |||
|
|
57215e2778 | ||
| 2c1de24e9a | |||
| c8baff7aac | |||
| de3f21a7ec |
@@ -88,7 +88,7 @@ module.exports = {
|
||||
// --- General Worker ---
|
||||
name: 'flyer-crawler-worker',
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'src/services/queueService.server.ts', // tsx will execute this file
|
||||
args: 'src/worker.ts', // tsx will execute this file
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production',
|
||||
@@ -164,7 +164,7 @@ module.exports = {
|
||||
// --- Analytics Worker ---
|
||||
name: 'flyer-crawler-analytics-worker',
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'src/services/queueService.server.ts', // tsx will execute this file
|
||||
args: 'src/worker.ts', // tsx will execute this file
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production',
|
||||
|
||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.1.3",
|
||||
"version": "0.1.7",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.1.3",
|
||||
"version": "0.1.7",
|
||||
"dependencies": {
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"private": true,
|
||||
"version": "0.1.3",
|
||||
"version": "0.1.7",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||
|
||||
@@ -44,7 +44,7 @@ export const FlyerCorrectionTool: React.FC<FlyerCorrectionToolProps> = ({
|
||||
})
|
||||
.catch((err) => {
|
||||
console.error('[DEBUG] FlyerCorrectionTool: Failed to fetch image.', { err });
|
||||
logger.error('Failed to fetch image for correction tool', { error: err });
|
||||
logger.error({ error: err }, 'Failed to fetch image for correction tool');
|
||||
notifyError('Could not load the image for correction.');
|
||||
});
|
||||
}
|
||||
@@ -164,7 +164,7 @@ export const FlyerCorrectionTool: React.FC<FlyerCorrectionToolProps> = ({
|
||||
const msg = err instanceof Error ? err.message : 'An unknown error occurred.';
|
||||
console.error('[DEBUG] handleRescan: Caught an error.', { error: err });
|
||||
notifyError(msg);
|
||||
logger.error('Error during rescan:', { error: err });
|
||||
logger.error({ error: err }, 'Error during rescan:');
|
||||
} finally {
|
||||
console.debug('[DEBUG] handleRescan: Finished. Setting isProcessing=false.');
|
||||
setIsProcessing(false);
|
||||
|
||||
@@ -112,7 +112,7 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
break;
|
||||
}
|
||||
} catch (error) {
|
||||
logger.error('Error during polling:', { error });
|
||||
logger.error({ error }, 'Error during polling:');
|
||||
setErrorMessage(
|
||||
error instanceof Error ? error.message : 'An unexpected error occurred during polling.',
|
||||
);
|
||||
@@ -157,7 +157,7 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
setProcessingState('polling');
|
||||
} catch (error: any) {
|
||||
// Handle the structured error thrown by the API client.
|
||||
logger.error('An error occurred during file upload:', { error });
|
||||
logger.error({ error }, 'An error occurred during file upload:');
|
||||
// Handle 409 Conflict for duplicate flyers
|
||||
if (error?.status === 409 && error.body?.flyerId) {
|
||||
setErrorMessage(`This flyer has already been processed. You can view it here:`);
|
||||
|
||||
@@ -1,94 +1,68 @@
|
||||
// src/middleware/errorHandler.ts
|
||||
import { Request, Response, NextFunction } from 'express';
|
||||
import { ZodError } from 'zod';
|
||||
import {
|
||||
DatabaseError,
|
||||
UniqueConstraintError,
|
||||
ForeignKeyConstraintError,
|
||||
NotFoundError,
|
||||
UniqueConstraintError,
|
||||
ValidationError,
|
||||
ValidationIssue,
|
||||
} from '../services/db/errors.db';
|
||||
import crypto from 'crypto';
|
||||
import { logger } from '../services/logger.server';
|
||||
|
||||
interface HttpError extends Error {
|
||||
status?: number;
|
||||
}
|
||||
|
||||
export const errorHandler = (err: HttpError, req: Request, res: Response, next: NextFunction) => {
|
||||
// If the response headers have already been sent, we must delegate to the default Express error handler.
|
||||
/**
|
||||
* A centralized error handling middleware for the Express application.
|
||||
* This middleware should be the LAST `app.use()` call to catch all errors from previous routes and middleware.
|
||||
*
|
||||
* It standardizes error responses and ensures consistent logging.
|
||||
*/
|
||||
export const errorHandler = (err: Error, req: Request, res: Response, next: NextFunction) => {
|
||||
// If headers have already been sent, delegate to the default Express error handler.
|
||||
if (res.headersSent) {
|
||||
return next(err);
|
||||
}
|
||||
|
||||
// The pino-http middleware guarantees that `req.log` will be available.
|
||||
const log = req.log;
|
||||
// Use the request-scoped logger if available, otherwise fall back to the global logger.
|
||||
const log = req.log || logger;
|
||||
|
||||
// --- 1. Determine Final Status Code and Message ---
|
||||
let statusCode = err.status ?? 500;
|
||||
const message = err.message;
|
||||
let validationIssues: ValidationIssue[] | undefined;
|
||||
let errorId: string | undefined;
|
||||
// --- Handle Zod Validation Errors ---
|
||||
if (err instanceof ZodError) {
|
||||
log.warn({ err: err.flatten() }, 'Request validation failed');
|
||||
return res.status(400).json({
|
||||
message: 'The request data is invalid.',
|
||||
errors: err.issues.map((e) => ({ path: e.path, message: e.message })),
|
||||
});
|
||||
}
|
||||
|
||||
// --- Handle Custom Operational Errors ---
|
||||
if (err instanceof NotFoundError) {
|
||||
log.info({ err }, 'Resource not found');
|
||||
return res.status(404).json({ message: err.message });
|
||||
}
|
||||
|
||||
if (err instanceof ValidationError) {
|
||||
log.warn({ err }, 'Validation error occurred');
|
||||
return res.status(400).json({ message: err.message, errors: err.validationErrors });
|
||||
}
|
||||
|
||||
// Refine the status code for known error types. Check for most specific types first.
|
||||
if (err instanceof UniqueConstraintError) {
|
||||
statusCode = 409; // Conflict
|
||||
} else if (err instanceof NotFoundError) {
|
||||
statusCode = 404;
|
||||
} else if (err instanceof ForeignKeyConstraintError) {
|
||||
statusCode = 400;
|
||||
} else if (err instanceof ValidationError) {
|
||||
statusCode = 400;
|
||||
validationIssues = err.validationErrors;
|
||||
} else if (err instanceof DatabaseError) {
|
||||
// This is a generic fallback for other database errors that are not the specific subclasses above.
|
||||
statusCode = err.status;
|
||||
} else if (err.name === 'UnauthorizedError') {
|
||||
statusCode = err.status || 401;
|
||||
log.warn({ err }, 'Constraint error occurred');
|
||||
return res.status(409).json({ message: err.message }); // Use 409 Conflict for unique constraints
|
||||
}
|
||||
|
||||
// --- 2. Log Based on Final Status Code ---
|
||||
// Log the full error details for debugging, especially for server errors.
|
||||
if (statusCode >= 500) {
|
||||
errorId = crypto.randomBytes(4).toString('hex');
|
||||
// The request-scoped logger already contains user, IP, and request_id.
|
||||
// We add the full error and the request object itself.
|
||||
// Pino's `redact` config will automatically sanitize sensitive fields in `req`.
|
||||
log.error(
|
||||
{
|
||||
err,
|
||||
errorId,
|
||||
req: { method: req.method, url: req.originalUrl, headers: req.headers, body: req.body },
|
||||
},
|
||||
`Unhandled API Error (ID: ${errorId})`,
|
||||
);
|
||||
} else {
|
||||
// For 4xx errors, log at a lower level (e.g., 'warn') to avoid flooding error trackers.
|
||||
// We include the validation errors in the log context if they exist.
|
||||
log.warn(
|
||||
{
|
||||
err,
|
||||
validationErrors: validationIssues, // Add validation issues to the log object
|
||||
statusCode,
|
||||
},
|
||||
`Client Error on ${req.method} ${req.path}: ${message}`,
|
||||
);
|
||||
if (err instanceof ForeignKeyConstraintError) {
|
||||
log.warn({ err }, 'Foreign key constraint violation');
|
||||
return res.status(400).json({ message: err.message });
|
||||
}
|
||||
|
||||
// --- TEST ENVIRONMENT DEBUGGING ---
|
||||
if (process.env.NODE_ENV === 'test') {
|
||||
console.error('--- [TEST] UNHANDLED ERROR ---', err);
|
||||
// --- Handle Generic Errors ---
|
||||
// Log the full error object for debugging. The pino logger will handle redaction.
|
||||
log.error({ err }, 'An unhandled error occurred in an Express route');
|
||||
|
||||
// In production, send a generic message to avoid leaking implementation details.
|
||||
if (process.env.NODE_ENV === 'production') {
|
||||
return res.status(500).json({ message: 'An internal server error occurred.' });
|
||||
}
|
||||
|
||||
// --- 3. Send Response ---
|
||||
// In production, send a generic message for 5xx errors.
|
||||
// In dev/test, send the actual error message for easier debugging.
|
||||
const responseMessage =
|
||||
statusCode >= 500 && process.env.NODE_ENV === 'production'
|
||||
? `An unexpected server error occurred. Please reference error ID: ${errorId}`
|
||||
: message;
|
||||
|
||||
res.status(statusCode).json({
|
||||
message: responseMessage,
|
||||
...(validationIssues && { errors: validationIssues }), // Conditionally add the 'errors' array if it exists
|
||||
});
|
||||
};
|
||||
// In development, send more details for easier debugging.
|
||||
return res.status(500).json({ message: err.message, stack: err.stack });
|
||||
};
|
||||
@@ -5,7 +5,16 @@ import type { Request, Response, NextFunction } from 'express';
|
||||
import { createMockUserProfile, createMockActivityLogItem } from '../tests/utils/mockFactories';
|
||||
import type { UserProfile } from '../types';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
import { mockLogger } from '../tests/utils/mockLogger';
|
||||
|
||||
const { mockLogger } = vi.hoisted(() => ({
|
||||
mockLogger: {
|
||||
info: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
error: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
child: vi.fn().mockReturnThis(),
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('../lib/queue', () => ({
|
||||
serverAdapter: {
|
||||
@@ -27,19 +36,22 @@ vi.mock('../services/db/index.db', () => ({
|
||||
notificationRepo: {},
|
||||
}));
|
||||
|
||||
// Mock the queue service to control worker statuses
|
||||
// Mock the queue service for queue status checks
|
||||
vi.mock('../services/queueService.server', () => ({
|
||||
flyerQueue: { name: 'flyer-processing', getJobCounts: vi.fn() },
|
||||
emailQueue: { name: 'email-sending', getJobCounts: vi.fn() },
|
||||
analyticsQueue: { name: 'analytics-reporting', getJobCounts: vi.fn() },
|
||||
cleanupQueue: { name: 'file-cleanup', getJobCounts: vi.fn() },
|
||||
weeklyAnalyticsQueue: { name: 'weekly-analytics-reporting', getJobCounts: vi.fn() },
|
||||
}));
|
||||
|
||||
// Mock the worker service for worker status checks
|
||||
vi.mock('../services/workers.server', () => ({
|
||||
flyerWorker: { name: 'flyer-processing', isRunning: vi.fn() },
|
||||
emailWorker: { name: 'email-sending', isRunning: vi.fn() },
|
||||
analyticsWorker: { name: 'analytics-reporting', isRunning: vi.fn() },
|
||||
cleanupWorker: { name: 'file-cleanup', isRunning: vi.fn() },
|
||||
weeklyAnalyticsWorker: { name: 'weekly-analytics-reporting', isRunning: vi.fn() },
|
||||
flyerQueue: { name: 'flyer-processing', getJobCounts: vi.fn() },
|
||||
emailQueue: { name: 'email-sending', getJobCounts: vi.fn() },
|
||||
analyticsQueue: { name: 'analytics-reporting', getJobCounts: vi.fn() },
|
||||
cleanupQueue: { name: 'file-cleanup', getJobCounts: vi.fn() },
|
||||
// FIX: Add the missing weeklyAnalyticsQueue to prevent import errors in admin.routes.ts
|
||||
weeklyAnalyticsQueue: { name: 'weekly-analytics-reporting', getJobCounts: vi.fn() },
|
||||
}));
|
||||
|
||||
// Mock other dependencies that are part of the adminRouter setup but not directly tested here
|
||||
@@ -67,8 +79,10 @@ import adminRouter from './admin.routes';
|
||||
|
||||
// Import the mocked modules to control them
|
||||
import * as queueService from '../services/queueService.server';
|
||||
import * as workerService from '../services/workers.server';
|
||||
import { adminRepo } from '../services/db/index.db';
|
||||
const mockedQueueService = queueService as Mocked<typeof queueService>;
|
||||
const mockedWorkerService = workerService as Mocked<typeof workerService>;
|
||||
|
||||
// Mock the logger
|
||||
vi.mock('../services/logger.server', () => ({
|
||||
@@ -137,11 +151,11 @@ describe('Admin Monitoring Routes (/api/admin)', () => {
|
||||
describe('GET /workers/status', () => {
|
||||
it('should return the status of all registered workers', async () => {
|
||||
// Arrange: Set the mock status for each worker
|
||||
vi.mocked(mockedQueueService.flyerWorker.isRunning).mockReturnValue(true);
|
||||
vi.mocked(mockedQueueService.emailWorker.isRunning).mockReturnValue(true);
|
||||
vi.mocked(mockedQueueService.analyticsWorker.isRunning).mockReturnValue(false); // Simulate one worker being stopped
|
||||
vi.mocked(mockedQueueService.cleanupWorker.isRunning).mockReturnValue(true);
|
||||
vi.mocked(mockedQueueService.weeklyAnalyticsWorker.isRunning).mockReturnValue(true);
|
||||
vi.mocked(mockedWorkerService.flyerWorker.isRunning).mockReturnValue(true);
|
||||
vi.mocked(mockedWorkerService.emailWorker.isRunning).mockReturnValue(true);
|
||||
vi.mocked(mockedWorkerService.analyticsWorker.isRunning).mockReturnValue(false); // Simulate one worker being stopped
|
||||
vi.mocked(mockedWorkerService.cleanupWorker.isRunning).mockReturnValue(true);
|
||||
vi.mocked(mockedWorkerService.weeklyAnalyticsWorker.isRunning).mockReturnValue(true);
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).get('/api/admin/workers/status');
|
||||
|
||||
@@ -25,12 +25,14 @@ import {
|
||||
analyticsQueue,
|
||||
cleanupQueue,
|
||||
weeklyAnalyticsQueue,
|
||||
flyerWorker,
|
||||
emailWorker,
|
||||
} from '../services/queueService.server'; // Import your queues
|
||||
import {
|
||||
analyticsWorker,
|
||||
cleanupWorker,
|
||||
emailWorker,
|
||||
flyerWorker,
|
||||
weeklyAnalyticsWorker,
|
||||
} from '../services/queueService.server'; // Import your queues
|
||||
} from '../services/workers.server';
|
||||
import { getSimpleWeekAndYear } from '../utils/dateUtils';
|
||||
import {
|
||||
requiredString,
|
||||
@@ -135,6 +137,7 @@ router.get('/corrections', async (req, res, next: NextFunction) => {
|
||||
const corrections = await db.adminRepo.getSuggestedCorrections(req.log);
|
||||
res.json(corrections);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching suggested corrections');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -144,6 +147,7 @@ router.get('/brands', async (req, res, next: NextFunction) => {
|
||||
const brands = await db.flyerRepo.getAllBrands(req.log);
|
||||
res.json(brands);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching brands');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -153,6 +157,7 @@ router.get('/stats', async (req, res, next: NextFunction) => {
|
||||
const stats = await db.adminRepo.getApplicationStats(req.log);
|
||||
res.json(stats);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching application stats');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -162,6 +167,7 @@ router.get('/stats/daily', async (req, res, next: NextFunction) => {
|
||||
const dailyStats = await db.adminRepo.getDailyStatsForLast30Days(req.log);
|
||||
res.json(dailyStats);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching daily stats');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -176,6 +182,7 @@ router.post(
|
||||
await db.adminRepo.approveCorrection(params.id, req.log); // params.id is now safely typed as number
|
||||
res.status(200).json({ message: 'Correction approved successfully.' });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error approving correction');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -191,6 +198,7 @@ router.post(
|
||||
await db.adminRepo.rejectCorrection(params.id, req.log); // params.id is now safely typed as number
|
||||
res.status(200).json({ message: 'Correction rejected successfully.' });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error rejecting correction');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -210,6 +218,7 @@ router.put(
|
||||
);
|
||||
res.status(200).json(updatedCorrection);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating suggested correction');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -225,6 +234,7 @@ router.put(
|
||||
const updatedRecipe = await db.adminRepo.updateRecipeStatus(params.id, body.status, req.log); // This is still a standalone function in admin.db.ts
|
||||
res.status(200).json(updatedRecipe);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating recipe status');
|
||||
next(error); // Pass all errors to the central error handler
|
||||
}
|
||||
},
|
||||
@@ -250,6 +260,7 @@ router.post(
|
||||
logger.info({ brandId: params.id, logoUrl }, `Brand logo updated for brand ID: ${params.id}`);
|
||||
res.status(200).json({ message: 'Brand logo updated successfully.', logoUrl });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating brand logo');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -260,6 +271,7 @@ router.get('/unmatched-items', async (req, res, next: NextFunction) => {
|
||||
const items = await db.adminRepo.getUnmatchedFlyerItems(req.log);
|
||||
res.json(items);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching unmatched items');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -279,6 +291,7 @@ router.delete(
|
||||
await db.recipeRepo.deleteRecipe(params.recipeId, userProfile.user.user_id, true, req.log);
|
||||
res.status(204).send();
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error deleting recipe');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -297,6 +310,7 @@ router.delete(
|
||||
await db.flyerRepo.deleteFlyer(params.flyerId, req.log);
|
||||
res.status(204).send();
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error deleting flyer');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -316,6 +330,7 @@ router.put(
|
||||
); // This is still a standalone function in admin.db.ts
|
||||
res.status(200).json(updatedComment);
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error updating comment status');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -326,6 +341,7 @@ router.get('/users', async (req, res, next: NextFunction) => {
|
||||
const users = await db.adminRepo.getAllUsers(req.log);
|
||||
res.json(users);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching users');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -345,6 +361,7 @@ router.get(
|
||||
const logs = await db.adminRepo.getActivityLog(limit, offset, req.log);
|
||||
res.json(logs);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching activity log');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -360,6 +377,7 @@ router.get(
|
||||
const user = await db.userRepo.findUserProfileById(params.id, req.log);
|
||||
res.json(user);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching user profile');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -395,6 +413,7 @@ router.delete(
|
||||
await db.userRepo.deleteUserById(params.id, req.log);
|
||||
res.status(204).send();
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error deleting user');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -478,6 +497,7 @@ router.post(
|
||||
.status(202)
|
||||
.json({ message: `File cleanup job for flyer ID ${params.flyerId} has been enqueued.` });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error enqueuing cleanup job');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -500,6 +520,7 @@ router.post('/trigger/failing-job', async (req: Request, res: Response, next: Ne
|
||||
.status(202)
|
||||
.json({ message: `Failing test job has been enqueued successfully. Job ID: ${job.id}` });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error enqueuing failing job');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -572,6 +593,7 @@ router.get('/queues/status', async (req: Request, res: Response, next: NextFunct
|
||||
);
|
||||
res.json(queueStatuses);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching queue statuses');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -620,6 +642,7 @@ router.post(
|
||||
);
|
||||
res.status(200).json({ message: `Job ${jobId} has been successfully marked for retry.` });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error retrying job');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -651,6 +674,7 @@ router.post(
|
||||
.status(202)
|
||||
.json({ message: 'Successfully enqueued weekly analytics job.', jobId: job.id });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error enqueuing weekly analytics job');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -134,8 +134,8 @@ router.post(
|
||||
// If the email is a duplicate, return a 409 Conflict status.
|
||||
return res.status(409).json({ message: error.message });
|
||||
}
|
||||
// The createUser method now handles its own transaction logging, so we just log the route failure.
|
||||
logger.error({ error }, `User registration route failed for email: ${email}.`);
|
||||
// Pass the error to the centralized handler
|
||||
return next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -108,6 +108,7 @@ router.post(
|
||||
const items = await db.flyerRepo.getFlyerItemsForFlyers(body.flyerIds, req.log);
|
||||
res.json(items);
|
||||
} catch (error) {
|
||||
req.log.error({ error }, 'Error fetching batch flyer items');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -127,6 +128,7 @@ router.post(
|
||||
const count = await db.flyerRepo.countFlyerItemsForFlyers(body.flyerIds ?? [], req.log);
|
||||
res.json({ count });
|
||||
} catch (error) {
|
||||
req.log.error({ error }, 'Error counting batch flyer items');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -39,10 +39,7 @@ router.get('/db-schema', validateRequest(emptySchema), async (req, res, next: Ne
|
||||
}
|
||||
return res.status(200).json({ success: true, message: 'All required database tables exist.' });
|
||||
} catch (error: unknown) {
|
||||
logger.error(
|
||||
{ error: error instanceof Error ? error.message : error },
|
||||
'Error during DB schema check:',
|
||||
);
|
||||
logger.error({ error }, 'Error during DB schema check:');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
@@ -133,6 +130,7 @@ router.get(
|
||||
}
|
||||
throw new Error(`Unexpected Redis ping response: ${reply}`); // This will be caught below
|
||||
} catch (error: unknown) {
|
||||
logger.error({ error }, 'Error checking Redis health');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -46,7 +46,6 @@ router.get(
|
||||
}
|
||||
|
||||
// Check if there was output to stderr, even if the exit code was 0 (success).
|
||||
// This handles warnings or non-fatal errors that should arguably be treated as failures in this context.
|
||||
if (stderr && stderr.trim().length > 0) {
|
||||
logger.error({ stderr }, '[API /pm2-status] PM2 executed but produced stderr:');
|
||||
return next(new Error(`PM2 command produced an error: ${stderr}`));
|
||||
@@ -86,6 +85,7 @@ router.post(
|
||||
|
||||
res.json(coordinates);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error geocoding address');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -77,7 +77,7 @@ router.use(passport.authenticate('jwt', { session: false }));
|
||||
// Ensure the directory for avatar uploads exists.
|
||||
const avatarUploadDir = path.join(process.cwd(), 'public', 'uploads', 'avatars');
|
||||
fs.mkdir(avatarUploadDir, { recursive: true }).catch((err) => {
|
||||
logger.error('Failed to create avatar upload directory:', err);
|
||||
logger.error({ err }, 'Failed to create avatar upload directory');
|
||||
});
|
||||
|
||||
// Define multer storage configuration. The `req.user` object will be available
|
||||
@@ -122,6 +122,7 @@ router.post(
|
||||
);
|
||||
res.json(updatedProfile);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error uploading avatar');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -151,6 +152,7 @@ router.get(
|
||||
);
|
||||
res.json(notifications);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching notifications');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -168,6 +170,7 @@ router.post(
|
||||
await db.notificationRepo.markAllNotificationsAsRead(userProfile.user.user_id, req.log);
|
||||
res.status(204).send(); // No Content
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error marking all notifications as read');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -193,6 +196,7 @@ router.post(
|
||||
);
|
||||
res.status(204).send(); // Success, no content to return
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error marking notification as read');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -345,11 +349,7 @@ router.post(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to add watched item');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -453,11 +453,7 @@ router.post(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to create shopping list');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -516,12 +512,7 @@ router.post(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
params: req.params,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, params: req.params, body: req.body }, 'Failed to add shopping list item');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -661,11 +652,7 @@ router.put(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to set user dietary restrictions');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -709,11 +696,7 @@ router.put(
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
return res.status(400).json({ message: error.message });
|
||||
}
|
||||
const errorMessage = error instanceof Error ? error.message : 'An unknown error occurred';
|
||||
logger.error({
|
||||
errorMessage,
|
||||
body: req.body,
|
||||
});
|
||||
logger.error({ error, body: req.body }, 'Failed to set user appliances');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -743,6 +726,7 @@ router.get(
|
||||
const address = await db.addressRepo.getAddressById(addressId, req.log); // This will throw NotFoundError if not found
|
||||
res.json(address);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching user address');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -781,6 +765,7 @@ router.put(
|
||||
const addressId = await userService.upsertUserAddress(userProfile, addressData, req.log); // This was a duplicate, fixed.
|
||||
res.status(200).json({ message: 'Address updated successfully', address_id: addressId });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error updating user address');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -51,9 +51,7 @@ export class AiAnalysisService {
|
||||
// Normalize sources to a consistent format.
|
||||
const mappedSources = (response.sources || []).map(
|
||||
(s: RawSource) =>
|
||||
(s.web
|
||||
? { uri: s.web.uri || '', title: 'Untitled' }
|
||||
: { uri: '', title: 'Untitled' }) as Source,
|
||||
(s.web ? { uri: s.web.uri || '', title: s.web.title || 'Untitled' } : { uri: '', title: 'Untitled' }) as Source,
|
||||
);
|
||||
return { ...response, sources: mappedSources };
|
||||
}
|
||||
@@ -84,9 +82,7 @@ export class AiAnalysisService {
|
||||
// Normalize sources to a consistent format.
|
||||
const mappedSources = (response.sources || []).map(
|
||||
(s: RawSource) =>
|
||||
(s.web
|
||||
? { uri: s.web.uri || '', title: 'Untitled' }
|
||||
: { uri: '', title: 'Untitled' }) as Source,
|
||||
(s.web ? { uri: s.web.uri || '', title: s.web.title || 'Untitled' } : { uri: '', title: 'Untitled' }) as Source,
|
||||
);
|
||||
return { ...response, sources: mappedSources };
|
||||
}
|
||||
|
||||
@@ -239,8 +239,8 @@ describe('AI Service (Server)', () => {
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(1);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
`[AIService Adapter] Model 'gemini-2.5-flash' failed with a non-retriable error.`,
|
||||
{ error: nonRetriableError },
|
||||
`[AIService Adapter] Model 'gemini-2.5-flash' failed with a non-retriable error.`,
|
||||
);
|
||||
});
|
||||
|
||||
@@ -281,8 +281,8 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
'[AIService Adapter] All AI models failed. Throwing last known error.',
|
||||
{ lastError: quotaError3 },
|
||||
'[AIService Adapter] All AI models failed. Throwing last known error.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -186,13 +186,13 @@ export class AIService {
|
||||
return result;
|
||||
} catch (error: unknown) {
|
||||
lastError = error instanceof Error ? error : new Error(String(error));
|
||||
const errorMessage = lastError.message || '';
|
||||
const errorMessage = (lastError.message || '').toLowerCase(); // Make case-insensitive
|
||||
|
||||
// Check for specific error messages indicating quota issues or model unavailability.
|
||||
if (
|
||||
errorMessage.includes('quota') ||
|
||||
errorMessage.includes('429') || // HTTP 429 Too Many Requests
|
||||
errorMessage.includes('RESOURCE_EXHAUSTED') ||
|
||||
errorMessage.includes('resource_exhausted') || // Make case-insensitive
|
||||
errorMessage.includes('model is overloaded')
|
||||
) {
|
||||
this.logger.warn(
|
||||
|
||||
@@ -176,15 +176,13 @@ describe('API Client', () => {
|
||||
// We expect the promise to still resolve with the bad response, but log an error.
|
||||
await apiClient.apiFetch('/some/failing/endpoint');
|
||||
|
||||
// FIX: Use stringContaining to be resilient to port numbers (e.g., localhost:3001)
|
||||
// This checks for the essential parts of the log message without being brittle.
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.stringContaining('apiFetch: Request to http://'),
|
||||
'Internal Server Error',
|
||||
);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.stringContaining('/api/some/failing/endpoint failed with status 500'),
|
||||
'Internal Server Error',
|
||||
expect.objectContaining({
|
||||
status: 500,
|
||||
body: 'Internal Server Error',
|
||||
url: expect.stringContaining('/some/failing/endpoint'),
|
||||
}),
|
||||
'apiFetch: Request failed',
|
||||
);
|
||||
});
|
||||
|
||||
@@ -242,10 +240,6 @@ describe('API Client', () => {
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
});
|
||||
|
||||
it('logSearchQuery should log a warning on failure', async () => {
|
||||
@@ -259,8 +253,6 @@ describe('API Client', () => {
|
||||
was_successful: false,
|
||||
});
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
// src/services/apiClient.ts
|
||||
import { Profile, ShoppingListItem, SearchQuery, Budget, Address } from '../types';
|
||||
import { logger } from './logger.client';
|
||||
import { eventBus } from './eventBus';
|
||||
|
||||
// This constant should point to your backend API.
|
||||
// It's often a good practice to store this in an environment variable.
|
||||
@@ -62,12 +63,12 @@ const refreshToken = async (): Promise<string> => {
|
||||
logger.info('Successfully refreshed access token.');
|
||||
return data.token;
|
||||
} catch (error) {
|
||||
logger.error('Failed to refresh token. User will be logged out.', { error });
|
||||
logger.error({ error }, 'Failed to refresh token. User session has expired.');
|
||||
// Only perform browser-specific actions if in the browser environment.
|
||||
if (typeof window !== 'undefined') {
|
||||
localStorage.removeItem('authToken');
|
||||
// A hard redirect is a simple way to reset the app state to logged-out.
|
||||
// window.location.href = '/'; // Removed to allow the caller to handle session expiry.
|
||||
// Dispatch a global event that the UI layer can listen for to handle session expiry.
|
||||
eventBus.dispatch('sessionExpired');
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
@@ -144,9 +145,8 @@ export const apiFetch = async (
|
||||
// --- DEBUG LOGGING for failed requests ---
|
||||
if (!response.ok) {
|
||||
const responseText = await response.clone().text();
|
||||
logger.error(
|
||||
`apiFetch: Request to ${fullUrl} failed with status ${response.status}. Response body:`,
|
||||
responseText,
|
||||
logger.error({ url: fullUrl, status: response.status, body: responseText },
|
||||
'apiFetch: Request failed',
|
||||
);
|
||||
}
|
||||
// --- END DEBUG LOGGING ---
|
||||
|
||||
31
src/services/eventBus.ts
Normal file
31
src/services/eventBus.ts
Normal file
@@ -0,0 +1,31 @@
|
||||
// src/services/eventBus.ts
|
||||
|
||||
/**
|
||||
* A simple, generic event bus for cross-component communication without direct coupling.
|
||||
* This is particularly useful for broadcasting application-wide events, such as session expiry.
|
||||
*/
|
||||
|
||||
type EventCallback = (data?: any) => void;
|
||||
|
||||
class EventBus {
|
||||
private listeners: { [key: string]: EventCallback[] } = {};
|
||||
|
||||
on(event: string, callback: EventCallback): void {
|
||||
if (!this.listeners[event]) {
|
||||
this.listeners[event] = [];
|
||||
}
|
||||
this.listeners[event].push(callback);
|
||||
}
|
||||
|
||||
off(event: string, callback: EventCallback): void {
|
||||
if (!this.listeners[event]) return;
|
||||
this.listeners[event] = this.listeners[event].filter((l) => l !== callback);
|
||||
}
|
||||
|
||||
dispatch(event: string, data?: any): void {
|
||||
if (!this.listeners[event]) return;
|
||||
this.listeners[event].forEach((callback) => callback(data));
|
||||
}
|
||||
}
|
||||
|
||||
export const eventBus = new EventBus();
|
||||
@@ -87,7 +87,7 @@ describe('Geocoding Service', () => {
|
||||
// Assert
|
||||
expect(result).toEqual(coordinates);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: 'Redis down', cacheKey: expect.any(String) },
|
||||
{ err: expect.any(Error), cacheKey: expect.any(String) },
|
||||
'Redis GET or JSON.parse command failed. Proceeding without cache.',
|
||||
);
|
||||
expect(mockGoogleService.geocode).toHaveBeenCalled(); // Should still proceed to fetch
|
||||
@@ -107,7 +107,7 @@ describe('Geocoding Service', () => {
|
||||
expect(mocks.mockRedis.get).toHaveBeenCalledWith(cacheKey);
|
||||
// The service should log the JSON parsing error and continue
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: expect.any(String), cacheKey: expect.any(String) },
|
||||
{ err: expect.any(SyntaxError), cacheKey: expect.any(String) },
|
||||
'Redis GET or JSON.parse command failed. Proceeding without cache.',
|
||||
);
|
||||
expect(mockGoogleService.geocode).toHaveBeenCalledTimes(1);
|
||||
@@ -185,7 +185,7 @@ describe('Geocoding Service', () => {
|
||||
// Assert
|
||||
expect(result).toEqual(coordinates);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: 'Network Error' },
|
||||
{ err: expect.any(Error) },
|
||||
expect.stringContaining('An error occurred while calling the Google Maps Geocoding API'),
|
||||
);
|
||||
expect(mockNominatimService.geocode).toHaveBeenCalledWith(address, logger);
|
||||
@@ -223,7 +223,7 @@ describe('Geocoding Service', () => {
|
||||
expect(mockGoogleService.geocode).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.mockRedis.set).toHaveBeenCalledTimes(1);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: 'Redis SET failed', cacheKey: expect.any(String) },
|
||||
{ err: expect.any(Error), cacheKey: expect.any(String) },
|
||||
'Redis SET command failed. Result will not be cached.',
|
||||
);
|
||||
});
|
||||
@@ -271,7 +271,7 @@ describe('Geocoding Service', () => {
|
||||
// Act & Assert
|
||||
await expect(geocodingService.clearGeocodeCache(logger)).rejects.toThrow(redisError);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: redisError.message },
|
||||
{ err: expect.any(Error) },
|
||||
'Failed to clear geocode cache from Redis.',
|
||||
);
|
||||
expect(mocks.mockRedis.del).not.toHaveBeenCalled();
|
||||
|
||||
@@ -25,10 +25,7 @@ export class GeocodingService {
|
||||
return JSON.parse(cached);
|
||||
}
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error, cacheKey },
|
||||
'Redis GET or JSON.parse command failed. Proceeding without cache.',
|
||||
);
|
||||
logger.error({ err: error, cacheKey }, 'Redis GET or JSON.parse command failed. Proceeding without cache.');
|
||||
}
|
||||
|
||||
if (process.env.GOOGLE_MAPS_API_KEY) {
|
||||
@@ -44,8 +41,8 @@ export class GeocodingService {
|
||||
);
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error },
|
||||
'An error occurred while calling the Google Maps Geocoding API. Falling back to Nominatim.',
|
||||
{ err: error },
|
||||
'An error occurred while calling the Google Maps Geocoding API. Falling back to Nominatim.'
|
||||
);
|
||||
}
|
||||
} else {
|
||||
@@ -72,10 +69,7 @@ export class GeocodingService {
|
||||
try {
|
||||
await redis.set(cacheKey, JSON.stringify(result), 'EX', 60 * 60 * 24 * 30); // Cache for 30 days
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error, cacheKey },
|
||||
'Redis SET command failed. Result will not be cached.',
|
||||
);
|
||||
logger.error({ err: error, cacheKey }, 'Redis SET command failed. Result will not be cached.');
|
||||
}
|
||||
}
|
||||
|
||||
@@ -98,10 +92,7 @@ export class GeocodingService {
|
||||
logger.info(`Successfully deleted ${totalDeleted} geocode cache entries.`);
|
||||
return totalDeleted;
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error instanceof Error ? error.message : error },
|
||||
'Failed to clear geocode cache from Redis.',
|
||||
);
|
||||
logger.error({ err: error }, 'Failed to clear geocode cache from Redis.');
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,437 +1,32 @@
|
||||
// src/services/queueService.server.ts
|
||||
import { Queue, Worker, Job, UnrecoverableError } from 'bullmq';
|
||||
import IORedis from 'ioredis'; // Correctly imported
|
||||
import fsPromises from 'node:fs/promises';
|
||||
import { exec } from 'child_process';
|
||||
import { promisify } from 'util';
|
||||
|
||||
import { logger } from './logger.server';
|
||||
import { aiService } from './aiService.server';
|
||||
import * as emailService from './emailService.server';
|
||||
import * as db from './db/index.db';
|
||||
import { connection } from './redis.server';
|
||||
import {
|
||||
FlyerProcessingService,
|
||||
type FlyerJobData,
|
||||
type IFileSystem,
|
||||
} from './flyerProcessingService.server';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
flyerQueue,
|
||||
emailQueue,
|
||||
analyticsQueue,
|
||||
weeklyAnalyticsQueue,
|
||||
cleanupQueue,
|
||||
tokenCleanupQueue,
|
||||
} from './queues.server';
|
||||
|
||||
export const connection = new IORedis(process.env.REDIS_URL!, {
|
||||
maxRetriesPerRequest: null, // Important for BullMQ
|
||||
password: process.env.REDIS_PASSWORD, // Add the password from environment variables
|
||||
});
|
||||
// Re-export everything for backward compatibility where possible
|
||||
export { connection } from './redis.server';
|
||||
export * from './queues.server';
|
||||
|
||||
// --- Redis Connection Event Listeners ---
|
||||
connection.on('connect', () => {
|
||||
logger.info('[Redis] Connection established successfully.');
|
||||
});
|
||||
|
||||
connection.on('error', (err) => {
|
||||
// This is crucial for diagnosing Redis connection issues. // The patch requested this specific error handling.
|
||||
logger.error({ err }, '[Redis] Connection error.');
|
||||
});
|
||||
|
||||
const execAsync = promisify(exec);
|
||||
// --- Queues ---
|
||||
export const flyerQueue = new Queue<FlyerJobData>('flyer-processing', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 3, // Attempt a job 3 times before marking it as failed.
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 5000, // Start with a 5-second delay for the first retry
|
||||
},
|
||||
},
|
||||
});
|
||||
|
||||
export const emailQueue = new Queue<EmailJobData>('email-sending', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 5, // Emails can be retried more aggressively
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 10000, // Start with a 10-second delay
|
||||
},
|
||||
},
|
||||
});
|
||||
|
||||
export const analyticsQueue = new Queue<AnalyticsJobData>('analytics-reporting', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 2, // Analytics can be intensive, so fewer retries might be desired.
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 60000, // Wait a minute before retrying.
|
||||
},
|
||||
// Remove job from queue on completion to save space, as results are in the DB.
|
||||
removeOnComplete: true,
|
||||
removeOnFail: 50, // Keep the last 50 failed jobs for inspection.
|
||||
},
|
||||
});
|
||||
|
||||
export const weeklyAnalyticsQueue = new Queue<WeeklyAnalyticsJobData>(
|
||||
'weekly-analytics-reporting',
|
||||
{
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 2,
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 3600000, // 1 hour delay for retries
|
||||
},
|
||||
removeOnComplete: true,
|
||||
removeOnFail: 50,
|
||||
},
|
||||
},
|
||||
);
|
||||
|
||||
export const cleanupQueue = new Queue<CleanupJobData>('file-cleanup', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 3,
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 30000, // Retry cleanup after 30 seconds
|
||||
},
|
||||
removeOnComplete: true, // No need to keep successful cleanup jobs
|
||||
},
|
||||
});
|
||||
|
||||
export const tokenCleanupQueue = new Queue<TokenCleanupJobData>('token-cleanup', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 2,
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 3600000, // 1 hour delay
|
||||
},
|
||||
removeOnComplete: true,
|
||||
removeOnFail: 10,
|
||||
},
|
||||
});
|
||||
// --- Job Data Interfaces ---
|
||||
|
||||
interface EmailJobData {
|
||||
to: string;
|
||||
subject: string;
|
||||
text: string;
|
||||
html: string;
|
||||
}
|
||||
// We do NOT export workers here anymore to prevent side effects.
|
||||
// Consumers needing workers must import from './workers.server'.
|
||||
|
||||
/**
|
||||
* Defines the data for an analytics job.
|
||||
*/
|
||||
interface AnalyticsJobData {
|
||||
reportDate: string; // e.g., '2024-10-26'
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data for a weekly analytics job.
|
||||
*/
|
||||
interface WeeklyAnalyticsJobData {
|
||||
reportYear: number;
|
||||
reportWeek: number; // ISO week number (1-53)
|
||||
}
|
||||
|
||||
interface CleanupJobData {
|
||||
flyerId: number;
|
||||
// An array of absolute file paths to be deleted. Made optional for manual cleanup triggers.
|
||||
paths?: string[];
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data for a token cleanup job.
|
||||
*/
|
||||
interface TokenCleanupJobData {
|
||||
timestamp: string; // ISO string to ensure the job is unique per run
|
||||
}
|
||||
|
||||
// --- Worker Instantiation ---
|
||||
|
||||
// Create an adapter for fsPromises to match the IFileSystem interface.
|
||||
const fsAdapter: IFileSystem = {
|
||||
readdir: (path: string, options: { withFileTypes: true }) => fsPromises.readdir(path, options),
|
||||
unlink: (path: string) => fsPromises.unlink(path),
|
||||
};
|
||||
|
||||
// Instantiate the service with its real dependencies
|
||||
const flyerProcessingService = new FlyerProcessingService(
|
||||
aiService,
|
||||
db,
|
||||
fsAdapter,
|
||||
execAsync,
|
||||
cleanupQueue, // Inject the cleanup queue to break the circular dependency
|
||||
new FlyerDataTransformer(), // Inject the new transformer
|
||||
);
|
||||
|
||||
/**
|
||||
* A generic function to attach logging event listeners to any worker.
|
||||
* This centralizes logging for job completion and final failure.
|
||||
* @param worker The BullMQ worker instance.
|
||||
*/
|
||||
const attachWorkerEventListeners = (worker: Worker) => {
|
||||
worker.on('completed', (job: Job, returnValue: unknown) => {
|
||||
logger.info({ returnValue }, `[${worker.name}] Job ${job.id} completed successfully.`);
|
||||
});
|
||||
|
||||
worker.on('failed', (job: Job | undefined, error: Error) => {
|
||||
// This event fires after all retries have failed.
|
||||
logger.error(
|
||||
{ err: error, jobData: job?.data },
|
||||
`[${worker.name}] Job ${job?.id} has ultimately failed after all attempts.`,
|
||||
);
|
||||
});
|
||||
};
|
||||
|
||||
export const flyerWorker = new Worker<FlyerJobData>(
|
||||
'flyer-processing', // Must match the queue name
|
||||
async (job) => {
|
||||
try {
|
||||
// The processJob method creates its own job-specific logger internally.
|
||||
return await flyerProcessingService.processJob(job);
|
||||
} catch (error: any) {
|
||||
// Check for quota errors or other unrecoverable errors from the AI service
|
||||
const errorMessage = error?.message || '';
|
||||
if (
|
||||
errorMessage.includes('quota') ||
|
||||
errorMessage.includes('429') ||
|
||||
errorMessage.includes('RESOURCE_EXHAUSTED')
|
||||
) {
|
||||
logger.error(
|
||||
{ err: error, jobId: job.id },
|
||||
'[FlyerWorker] Unrecoverable quota error detected. Failing job immediately.',
|
||||
);
|
||||
throw new UnrecoverableError(errorMessage);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.WORKER_CONCURRENCY || '1', 10),
|
||||
},
|
||||
);
|
||||
/**
|
||||
* A dedicated worker process for sending emails.
|
||||
*/
|
||||
export const emailWorker = new Worker<EmailJobData>(
|
||||
'email-sending',
|
||||
async (job: Job<EmailJobData>) => {
|
||||
const { to, subject } = job.data;
|
||||
// Create a job-specific logger instance
|
||||
const jobLogger = logger.child({ jobId: job.id, jobName: job.name });
|
||||
jobLogger.info({ to, subject }, `[EmailWorker] Sending email for job ${job.id}`);
|
||||
try {
|
||||
await emailService.sendEmail(job.data, jobLogger);
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging to capture the full error object, including the stack trace.
|
||||
// This provides more context for debugging than just logging the message.
|
||||
logger.error(
|
||||
{
|
||||
// Log the full error object for better diagnostics. // The patch requested this specific error handling.
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
// Also include the job data for context.
|
||||
jobData: job.data,
|
||||
},
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
// Re-throw to let BullMQ handle the failure and retry.
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.EMAIL_WORKER_CONCURRENCY || '10', 10),
|
||||
},
|
||||
);
|
||||
|
||||
/**
|
||||
* A dedicated worker for generating daily analytics reports.
|
||||
* This is a placeholder for the actual report generation logic.
|
||||
*/
|
||||
export const analyticsWorker = new Worker<AnalyticsJobData>(
|
||||
'analytics-reporting',
|
||||
async (job: Job<AnalyticsJobData>) => {
|
||||
const { reportDate } = job.data;
|
||||
logger.info({ reportDate }, `[AnalyticsWorker] Starting report generation for job ${job.id}`);
|
||||
try {
|
||||
// Special case for testing the retry mechanism
|
||||
if (reportDate === 'FAIL') {
|
||||
throw new Error('This is a test failure for the analytics job.');
|
||||
}
|
||||
|
||||
// In a real implementation, you would call a database function here.
|
||||
// For example: await db.generateDailyAnalyticsReport(reportDate);
|
||||
await new Promise((resolve) => setTimeout(resolve, 10000)); // Simulate a 10-second task
|
||||
logger.info(`[AnalyticsWorker] Successfully generated report for ${reportDate}.`);
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging.
|
||||
logger.error(
|
||||
{
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
jobData: job.data,
|
||||
},
|
||||
`[AnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw error; // Re-throw to let BullMQ handle the failure and retry.
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.ANALYTICS_WORKER_CONCURRENCY || '1', 10),
|
||||
},
|
||||
);
|
||||
|
||||
/**
|
||||
* A dedicated worker for cleaning up flyer-related files from the filesystem.
|
||||
* This is triggered manually by an admin after a flyer has been reviewed.
|
||||
*/
|
||||
export const cleanupWorker = new Worker<CleanupJobData>(
|
||||
// This worker now handles two types of cleanup jobs.
|
||||
'file-cleanup', // The queue name
|
||||
async (job: Job<CleanupJobData>) => {
|
||||
// Destructure the data from the job payload.
|
||||
const { flyerId, paths } = job.data;
|
||||
logger.info(
|
||||
{ paths },
|
||||
`[CleanupWorker] Starting file cleanup for job ${job.id} (Flyer ID: ${flyerId})`,
|
||||
);
|
||||
|
||||
try {
|
||||
if (!paths || paths.length === 0) {
|
||||
logger.warn(
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${flyerId} received no paths to clean. Skipping.`,
|
||||
);
|
||||
return;
|
||||
}
|
||||
|
||||
// Iterate over the file paths provided in the job data and delete each one.
|
||||
for (const filePath of paths) {
|
||||
try {
|
||||
await fsAdapter.unlink(filePath);
|
||||
logger.info(`[CleanupWorker] Deleted temporary file: ${filePath}`);
|
||||
} catch (unlinkError: unknown) {
|
||||
// If the file doesn't exist, it's a success from our perspective.
|
||||
// We can log it as a warning and continue without failing the job.
|
||||
if (
|
||||
unlinkError instanceof Error &&
|
||||
'code' in unlinkError &&
|
||||
unlinkError.code === 'ENOENT'
|
||||
) {
|
||||
logger.warn(
|
||||
`[CleanupWorker] File not found during cleanup (already deleted?): ${filePath}`,
|
||||
);
|
||||
} else {
|
||||
throw unlinkError; // For any other error (e.g., permissions), re-throw to fail the job.
|
||||
}
|
||||
}
|
||||
}
|
||||
logger.info(
|
||||
`[CleanupWorker] Successfully cleaned up ${paths.length} file(s) for flyer ${flyerId}.`,
|
||||
);
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging.
|
||||
logger.error(
|
||||
{
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
},
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${flyerId} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw error; // Re-throw to let BullMQ handle the failure and retry.
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.CLEANUP_WORKER_CONCURRENCY || '10', 10),
|
||||
},
|
||||
);
|
||||
|
||||
/**
|
||||
* A dedicated worker for generating weekly analytics reports.
|
||||
* This is a placeholder for the actual report generation logic.
|
||||
*/
|
||||
export const weeklyAnalyticsWorker = new Worker<WeeklyAnalyticsJobData>(
|
||||
'weekly-analytics-reporting',
|
||||
async (job: Job<WeeklyAnalyticsJobData>) => {
|
||||
const { reportYear, reportWeek } = job.data;
|
||||
logger.info(
|
||||
{ reportYear, reportWeek },
|
||||
`[WeeklyAnalyticsWorker] Starting weekly report generation for job ${job.id}`,
|
||||
);
|
||||
try {
|
||||
// Simulate a longer-running task for weekly reports
|
||||
await new Promise((resolve) => setTimeout(resolve, 30000)); // Simulate 30-second task
|
||||
logger.info(
|
||||
`[WeeklyAnalyticsWorker] Successfully generated weekly report for week ${reportWeek}, ${reportYear}.`,
|
||||
);
|
||||
} catch (error: unknown) {
|
||||
// Standardize error logging.
|
||||
logger.error(
|
||||
{
|
||||
err: error instanceof Error ? error : new Error(String(error)),
|
||||
jobData: job.data,
|
||||
},
|
||||
`[WeeklyAnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw error; // Re-throw to let BullMQ handle the failure and retry.
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.WEEKLY_ANALYTICS_WORKER_CONCURRENCY || '1', 10),
|
||||
},
|
||||
);
|
||||
|
||||
/**
|
||||
* A dedicated worker for cleaning up expired password reset tokens.
|
||||
*/
|
||||
export const tokenCleanupWorker = new Worker<TokenCleanupJobData>(
|
||||
'token-cleanup',
|
||||
async (job: Job<TokenCleanupJobData>) => {
|
||||
const jobLogger = logger.child({ jobId: job.id, jobName: job.name });
|
||||
jobLogger.info('[TokenCleanupWorker] Starting cleanup of expired password reset tokens.');
|
||||
try {
|
||||
const deletedCount = await db.userRepo.deleteExpiredResetTokens(jobLogger);
|
||||
jobLogger.info(`[TokenCleanupWorker] Successfully deleted ${deletedCount} expired tokens.`);
|
||||
return { deletedCount };
|
||||
} catch (error: unknown) {
|
||||
jobLogger.error({ err: error }, `[TokenCleanupWorker] Job ${job.id} failed.`);
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: 1, // This is a low-priority, non-intensive task.
|
||||
},
|
||||
);
|
||||
|
||||
// --- Attach Event Listeners to All Workers ---
|
||||
attachWorkerEventListeners(flyerWorker);
|
||||
attachWorkerEventListeners(emailWorker);
|
||||
attachWorkerEventListeners(analyticsWorker);
|
||||
attachWorkerEventListeners(cleanupWorker);
|
||||
attachWorkerEventListeners(weeklyAnalyticsWorker);
|
||||
attachWorkerEventListeners(tokenCleanupWorker);
|
||||
|
||||
logger.info('All workers started and listening for jobs.');
|
||||
|
||||
/**
|
||||
* A function to gracefully shut down all queue workers and connections.
|
||||
* This is essential for preventing jobs from getting stuck in an 'active' state
|
||||
* when the application process is terminated.
|
||||
* @param signal The signal that triggered the shutdown (e.g., 'SIGINT').
|
||||
* A function to gracefully shut down all queues and connections.
|
||||
* This is for the API process which only uses queues.
|
||||
* For worker processes, use the gracefulShutdown from workers.server.ts
|
||||
*/
|
||||
export const gracefulShutdown = async (signal: string) => {
|
||||
logger.info(`[Shutdown] Received ${signal}. Closing all workers and queues...`);
|
||||
logger.info(`[Shutdown] Received ${signal}. Closing all queues...`);
|
||||
let exitCode = 0; // Default to success
|
||||
|
||||
const resources = [
|
||||
{ name: 'flyerWorker', close: () => flyerWorker.close() },
|
||||
{ name: 'emailWorker', close: () => emailWorker.close() },
|
||||
{ name: 'analyticsWorker', close: () => analyticsWorker.close() },
|
||||
{ name: 'cleanupWorker', close: () => cleanupWorker.close() },
|
||||
{ name: 'weeklyAnalyticsWorker', close: () => weeklyAnalyticsWorker.close() },
|
||||
{ name: 'tokenCleanupWorker', close: () => tokenCleanupWorker.close() },
|
||||
{ name: 'flyerQueue', close: () => flyerQueue.close() },
|
||||
{ name: 'emailQueue', close: () => emailQueue.close() },
|
||||
{ name: 'analyticsQueue', close: () => analyticsQueue.close() },
|
||||
@@ -454,7 +49,7 @@ export const gracefulShutdown = async (signal: string) => {
|
||||
});
|
||||
|
||||
if (exitCode === 0) {
|
||||
logger.info('[Shutdown] All workers, queues, and connections closed successfully.');
|
||||
logger.info('[Shutdown] All queues and connections closed successfully.');
|
||||
} else {
|
||||
logger.warn('[Shutdown] Graceful shutdown completed with errors.');
|
||||
}
|
||||
|
||||
@@ -175,7 +175,7 @@ describe('Queue Workers', () => {
|
||||
const emailError = 'SMTP server is down'; // Reject with a string
|
||||
mocks.sendEmail.mockRejectedValue(emailError);
|
||||
|
||||
await expect(emailProcessor(job)).rejects.toBe(emailError);
|
||||
await expect(emailProcessor(job)).rejects.toThrow(emailError);
|
||||
|
||||
// The worker should wrap the string in an Error object for logging
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
|
||||
96
src/services/queues.server.ts
Normal file
96
src/services/queues.server.ts
Normal file
@@ -0,0 +1,96 @@
|
||||
import { Queue } from 'bullmq';
|
||||
import { connection } from './redis.server';
|
||||
import type { FlyerJobData } from './flyerProcessingService.server';
|
||||
|
||||
// --- Job Data Interfaces ---
|
||||
|
||||
export interface EmailJobData {
|
||||
to: string;
|
||||
subject: string;
|
||||
text: string;
|
||||
html: string;
|
||||
}
|
||||
|
||||
export interface AnalyticsJobData {
|
||||
reportDate: string; // e.g., '2024-10-26'
|
||||
}
|
||||
|
||||
export interface WeeklyAnalyticsJobData {
|
||||
reportYear: number;
|
||||
reportWeek: number; // ISO week number (1-53)
|
||||
}
|
||||
|
||||
export interface CleanupJobData {
|
||||
flyerId: number;
|
||||
paths?: string[];
|
||||
}
|
||||
|
||||
export interface TokenCleanupJobData {
|
||||
timestamp: string;
|
||||
}
|
||||
|
||||
// --- Queues ---
|
||||
|
||||
export const flyerQueue = new Queue<FlyerJobData>('flyer-processing', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 3,
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 5000,
|
||||
},
|
||||
},
|
||||
});
|
||||
|
||||
export const emailQueue = new Queue<EmailJobData>('email-sending', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 5,
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 10000,
|
||||
},
|
||||
},
|
||||
});
|
||||
|
||||
export const analyticsQueue = new Queue<AnalyticsJobData>('analytics-reporting', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 2,
|
||||
backoff: {
|
||||
type: 'exponential',
|
||||
delay: 60000,
|
||||
},
|
||||
removeOnComplete: true,
|
||||
removeOnFail: 50,
|
||||
},
|
||||
});
|
||||
|
||||
export const weeklyAnalyticsQueue = new Queue<WeeklyAnalyticsJobData>('weekly-analytics-reporting', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 2,
|
||||
backoff: { type: 'exponential', delay: 3600000 },
|
||||
removeOnComplete: true,
|
||||
removeOnFail: 50,
|
||||
},
|
||||
});
|
||||
|
||||
export const cleanupQueue = new Queue<CleanupJobData>('file-cleanup', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 3,
|
||||
backoff: { type: 'exponential', delay: 30000 },
|
||||
removeOnComplete: true,
|
||||
},
|
||||
});
|
||||
|
||||
export const tokenCleanupQueue = new Queue<TokenCleanupJobData>('token-cleanup', {
|
||||
connection,
|
||||
defaultJobOptions: {
|
||||
attempts: 2,
|
||||
backoff: { type: 'exponential', delay: 3600000 },
|
||||
removeOnComplete: true,
|
||||
removeOnFail: 10,
|
||||
},
|
||||
});
|
||||
16
src/services/redis.server.ts
Normal file
16
src/services/redis.server.ts
Normal file
@@ -0,0 +1,16 @@
|
||||
import IORedis from 'ioredis';
|
||||
import { logger } from './logger.server';
|
||||
|
||||
export const connection = new IORedis(process.env.REDIS_URL!, {
|
||||
maxRetriesPerRequest: null, // Important for BullMQ
|
||||
password: process.env.REDIS_PASSWORD,
|
||||
});
|
||||
|
||||
// --- Redis Connection Event Listeners ---
|
||||
connection.on('connect', () => {
|
||||
logger.info('[Redis] Connection established successfully.');
|
||||
});
|
||||
|
||||
connection.on('error', (err) => {
|
||||
logger.error({ err }, '[Redis] Connection error.');
|
||||
});
|
||||
30
src/services/worker.ts
Normal file
30
src/services/worker.ts
Normal file
@@ -0,0 +1,30 @@
|
||||
import { gracefulShutdown } from './workers.server';
|
||||
import { logger } from './logger.server';
|
||||
|
||||
logger.info('[Worker] Initializing worker process...');
|
||||
|
||||
// The workers are instantiated as side effects of importing workers.server.ts.
|
||||
// This pattern ensures they start immediately upon import.
|
||||
|
||||
// Handle graceful shutdown
|
||||
const handleShutdown = (signal: string) => {
|
||||
logger.info(`[Worker] Received ${signal}. Initiating graceful shutdown...`);
|
||||
gracefulShutdown(signal).catch((error: unknown) => {
|
||||
logger.error({ err: error }, '[Worker] Error during shutdown.');
|
||||
process.exit(1);
|
||||
});
|
||||
};
|
||||
|
||||
process.on('SIGINT', () => handleShutdown('SIGINT'));
|
||||
process.on('SIGTERM', () => handleShutdown('SIGTERM'));
|
||||
|
||||
// Catch unhandled errors to log them before crashing
|
||||
process.on('uncaughtException', (err) => {
|
||||
logger.error({ err }, '[Worker] Uncaught exception');
|
||||
});
|
||||
|
||||
process.on('unhandledRejection', (reason, promise) => {
|
||||
logger.error({ reason, promise }, '[Worker] Unhandled Rejection');
|
||||
});
|
||||
|
||||
logger.info('[Worker] Worker process is running and listening for jobs.');
|
||||
346
src/services/workers.server.test.ts
Normal file
346
src/services/workers.server.test.ts
Normal file
@@ -0,0 +1,346 @@
|
||||
// src/services/workers.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import type { Job } from 'bullmq';
|
||||
|
||||
// --- Hoisted Mocks ---
|
||||
const mocks = vi.hoisted(() => {
|
||||
// This object will store the processor functions captured from the worker constructors.
|
||||
const capturedProcessors: Record<string, (job: Job) => Promise<unknown>> = {};
|
||||
|
||||
return {
|
||||
sendEmail: vi.fn(),
|
||||
unlink: vi.fn(),
|
||||
processFlyerJob: vi.fn(),
|
||||
capturedProcessors,
|
||||
deleteExpiredResetTokens: vi.fn(),
|
||||
// Mock the Worker constructor to capture the processor function. It must be a
|
||||
// `function` and not an arrow function so it can be called with `new`.
|
||||
MockWorker: vi.fn(function (name: string, processor: (job: Job) => Promise<unknown>) {
|
||||
if (processor) {
|
||||
capturedProcessors[name] = processor;
|
||||
}
|
||||
// Return a mock worker instance, though it's not used in this test file.
|
||||
return { on: vi.fn(), close: vi.fn() };
|
||||
}),
|
||||
};
|
||||
});
|
||||
|
||||
// --- Mock Modules ---
|
||||
vi.mock('./emailService.server', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('./emailService.server')>();
|
||||
return {
|
||||
...actual,
|
||||
// We only need to mock the specific function being called by the worker.
|
||||
// The rest of the module can retain its original implementation if needed elsewhere.
|
||||
sendEmail: mocks.sendEmail,
|
||||
};
|
||||
});
|
||||
|
||||
// The workers use an `fsAdapter`. We can mock the underlying `fsPromises`
|
||||
// that the adapter is built from in queueService.server.ts.
|
||||
vi.mock('node:fs/promises', () => ({
|
||||
default: {
|
||||
unlink: mocks.unlink,
|
||||
// Add other fs functions if needed by other tests
|
||||
readdir: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./logger.server', () => ({
|
||||
logger: {
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
child: vi.fn().mockReturnThis(),
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./db/index.db', () => ({
|
||||
userRepo: {
|
||||
deleteExpiredResetTokens: mocks.deleteExpiredResetTokens,
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock bullmq to capture the processor functions passed to the Worker constructor
|
||||
import { logger as mockLogger } from './logger.server';
|
||||
vi.mock('bullmq', () => ({
|
||||
Worker: mocks.MockWorker,
|
||||
// FIX: Use a standard function for the mock constructor to allow `new Queue(...)` to work.
|
||||
Queue: vi.fn(function () {
|
||||
return { add: vi.fn() };
|
||||
}),
|
||||
}));
|
||||
|
||||
// Mock flyerProcessingService.server as flyerWorker depends on it
|
||||
vi.mock('./flyerProcessingService.server', () => ({
|
||||
FlyerProcessingService: class {
|
||||
processJob = mocks.processFlyerJob;
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock flyerDataTransformer as it's a dependency of FlyerProcessingService
|
||||
vi.mock('./flyerDataTransformer', () => ({
|
||||
FlyerDataTransformer: class {
|
||||
transform = vi.fn(); // Mock transform method
|
||||
},
|
||||
}));
|
||||
|
||||
// Helper to create a mock BullMQ Job object
|
||||
const createMockJob = <T>(data: T): Job<T> => {
|
||||
return {
|
||||
id: 'job-1',
|
||||
data,
|
||||
updateProgress: vi.fn().mockResolvedValue(undefined),
|
||||
log: vi.fn().mockResolvedValue(undefined),
|
||||
opts: { attempts: 3 },
|
||||
attemptsMade: 1,
|
||||
trace: vi.fn().mockResolvedValue(undefined),
|
||||
moveToCompleted: vi.fn().mockResolvedValue(undefined),
|
||||
moveToFailed: vi.fn().mockResolvedValue(undefined),
|
||||
} as unknown as Job<T>;
|
||||
};
|
||||
|
||||
describe('Queue Workers', () => {
|
||||
// These will hold the captured processor functions for each test.
|
||||
let flyerProcessor: (job: Job) => Promise<unknown>;
|
||||
let emailProcessor: (job: Job) => Promise<unknown>;
|
||||
let analyticsProcessor: (job: Job) => Promise<unknown>;
|
||||
let cleanupProcessor: (job: Job) => Promise<unknown>;
|
||||
let weeklyAnalyticsProcessor: (job: Job) => Promise<unknown>;
|
||||
let tokenCleanupProcessor: (job: Job) => Promise<unknown>;
|
||||
|
||||
beforeEach(async () => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
// Reset default mock implementations for hoisted mocks
|
||||
mocks.sendEmail.mockResolvedValue(undefined);
|
||||
mocks.unlink.mockResolvedValue(undefined);
|
||||
mocks.processFlyerJob.mockResolvedValue({ flyerId: 123 }); // Default success for flyer processing
|
||||
mocks.deleteExpiredResetTokens.mockResolvedValue(5);
|
||||
|
||||
// Reset modules to re-evaluate the workers.server.ts file with fresh mocks.
|
||||
// This ensures that new worker instances are created and their processors are captured for each test.
|
||||
vi.resetModules();
|
||||
|
||||
// Dynamically import the module under test AFTER mocks are reset.
|
||||
// This will trigger the instantiation of the workers, and our mocked Worker constructor will capture the processors.
|
||||
await import('./workers.server');
|
||||
|
||||
// Re-capture the processors for each test to ensure isolation.
|
||||
flyerProcessor = mocks.capturedProcessors['flyer-processing'];
|
||||
emailProcessor = mocks.capturedProcessors['email-sending'];
|
||||
analyticsProcessor = mocks.capturedProcessors['analytics-reporting'];
|
||||
cleanupProcessor = mocks.capturedProcessors['file-cleanup'];
|
||||
weeklyAnalyticsProcessor = mocks.capturedProcessors['weekly-analytics-reporting'];
|
||||
tokenCleanupProcessor = mocks.capturedProcessors['token-cleanup'];
|
||||
});
|
||||
|
||||
describe('flyerWorker', () => {
|
||||
it('should call flyerProcessingService.processJob with the job data', async () => {
|
||||
const jobData = {
|
||||
filePath: '/tmp/flyer.pdf',
|
||||
originalFileName: 'flyer.pdf',
|
||||
checksum: 'abc',
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
|
||||
await flyerProcessor(job);
|
||||
|
||||
expect(mocks.processFlyerJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processFlyerJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should re-throw an error if flyerProcessingService.processJob fails', async () => {
|
||||
const job = createMockJob({
|
||||
filePath: '/tmp/fail.pdf',
|
||||
originalFileName: 'fail.pdf',
|
||||
checksum: 'def',
|
||||
});
|
||||
const processingError = new Error('Flyer processing failed');
|
||||
mocks.processFlyerJob.mockRejectedValue(processingError);
|
||||
|
||||
await expect(flyerProcessor(job)).rejects.toThrow('Flyer processing failed');
|
||||
});
|
||||
});
|
||||
|
||||
describe('emailWorker', () => {
|
||||
it('should call emailService.sendEmail with the job data', async () => {
|
||||
const jobData = {
|
||||
to: 'test@example.com',
|
||||
subject: 'Test Email',
|
||||
html: '<p>Hello</p>',
|
||||
text: 'Hello',
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
|
||||
await emailProcessor(job);
|
||||
|
||||
expect(mocks.sendEmail).toHaveBeenCalledTimes(1);
|
||||
// The implementation passes the logger as the second argument
|
||||
expect(mocks.sendEmail).toHaveBeenCalledWith(jobData, expect.anything());
|
||||
});
|
||||
|
||||
it('should log and re-throw an error if sendEmail fails with a non-Error object', async () => {
|
||||
const job = createMockJob({ to: 'fail@example.com', subject: 'fail', html: '', text: '' });
|
||||
const emailError = 'SMTP server is down'; // Reject with a string
|
||||
mocks.sendEmail.mockRejectedValue(emailError);
|
||||
|
||||
await expect(emailProcessor(job)).rejects.toThrow(emailError);
|
||||
|
||||
// The worker should wrap the string in an Error object for logging
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: new Error(emailError), jobData: job.data },
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
});
|
||||
|
||||
it('should re-throw an error if sendEmail fails', async () => {
|
||||
const job = createMockJob({ to: 'fail@example.com', subject: 'fail', html: '', text: '' });
|
||||
const emailError = new Error('SMTP server is down');
|
||||
mocks.sendEmail.mockRejectedValue(emailError);
|
||||
|
||||
await expect(emailProcessor(job)).rejects.toThrow('SMTP server is down');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: emailError, jobData: job.data },
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('analyticsWorker', () => {
|
||||
it('should complete successfully for a valid report date', async () => {
|
||||
vi.useFakeTimers();
|
||||
const job = createMockJob({ reportDate: '2024-01-01' });
|
||||
|
||||
const promise = analyticsProcessor(job);
|
||||
// Advance timers to simulate the 10-second task completing
|
||||
await vi.advanceTimersByTimeAsync(10000);
|
||||
await promise; // Wait for the promise to resolve
|
||||
|
||||
// No error should be thrown
|
||||
expect(true).toBe(true);
|
||||
vi.useRealTimers();
|
||||
});
|
||||
|
||||
it('should throw an error if reportDate is "FAIL"', async () => {
|
||||
const job = createMockJob({ reportDate: 'FAIL' });
|
||||
|
||||
await expect(analyticsProcessor(job)).rejects.toThrow(
|
||||
'This is a test failure for the analytics job.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('cleanupWorker', () => {
|
||||
it('should call unlink for each path provided in the job data', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/file1.jpg', '/tmp/file2.pdf'],
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
mocks.unlink.mockResolvedValue(undefined);
|
||||
|
||||
await cleanupProcessor(job);
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file1.jpg');
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file2.pdf');
|
||||
});
|
||||
|
||||
it('should not throw an error if a file is already deleted (ENOENT)', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/existing.jpg', '/tmp/already-deleted.jpg'],
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
// Use the built-in NodeJS.ErrnoException type for mock system errors.
|
||||
const enoentError: NodeJS.ErrnoException = new Error('File not found');
|
||||
enoentError.code = 'ENOENT';
|
||||
|
||||
// First call succeeds, second call fails with ENOENT
|
||||
mocks.unlink.mockResolvedValueOnce(undefined).mockRejectedValueOnce(enoentError);
|
||||
|
||||
// The processor should complete without throwing
|
||||
await expect(cleanupProcessor(job)).resolves.toBeUndefined();
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
});
|
||||
|
||||
it('should re-throw an error for issues other than ENOENT (e.g., permissions)', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/protected-file.jpg'],
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
// Use the built-in NodeJS.ErrnoException type for mock system errors.
|
||||
const permissionError: NodeJS.ErrnoException = new Error('Permission denied');
|
||||
permissionError.code = 'EACCES';
|
||||
|
||||
mocks.unlink.mockRejectedValue(permissionError);
|
||||
|
||||
await expect(cleanupProcessor(job)).rejects.toThrow('Permission denied');
|
||||
|
||||
// Verify the error was logged by the worker's catch block
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: permissionError },
|
||||
expect.stringContaining(
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${job.data.flyerId} failed.`,
|
||||
),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('weeklyAnalyticsWorker', () => {
|
||||
it('should complete successfully for a valid report date', async () => {
|
||||
vi.useFakeTimers();
|
||||
const job = createMockJob({ reportYear: 2024, reportWeek: 1 });
|
||||
|
||||
const promise = weeklyAnalyticsProcessor(job);
|
||||
// Advance timers to simulate the 30-second task completing
|
||||
await vi.advanceTimersByTimeAsync(30000);
|
||||
await promise; // Wait for the promise to resolve
|
||||
|
||||
// No error should be thrown
|
||||
expect(true).toBe(true);
|
||||
vi.useRealTimers();
|
||||
});
|
||||
|
||||
it('should re-throw an error if the job fails', async () => {
|
||||
vi.useFakeTimers();
|
||||
const job = createMockJob({ reportYear: 2024, reportWeek: 1 });
|
||||
// Mock the internal logic to throw an error
|
||||
const originalSetTimeout = setTimeout;
|
||||
vi.spyOn(global, 'setTimeout').mockImplementation((callback, ms) => {
|
||||
if (ms === 30000) {
|
||||
// Target the simulated delay
|
||||
throw new Error('Weekly analytics job failed');
|
||||
}
|
||||
return originalSetTimeout(callback, ms);
|
||||
});
|
||||
|
||||
await expect(weeklyAnalyticsProcessor(job)).rejects.toThrow('Weekly analytics job failed');
|
||||
vi.useRealTimers();
|
||||
vi.restoreAllMocks(); // Restore setTimeout mock
|
||||
});
|
||||
});
|
||||
|
||||
describe('tokenCleanupWorker', () => {
|
||||
it('should call userRepo.deleteExpiredResetTokens and return the count', async () => {
|
||||
const job = createMockJob({ timestamp: new Date().toISOString() });
|
||||
mocks.deleteExpiredResetTokens.mockResolvedValue(10);
|
||||
|
||||
const result = await tokenCleanupProcessor(job);
|
||||
|
||||
expect(mocks.deleteExpiredResetTokens).toHaveBeenCalledTimes(1);
|
||||
expect(result).toEqual({ deletedCount: 10 });
|
||||
});
|
||||
|
||||
it('should re-throw an error if the database call fails', async () => {
|
||||
const job = createMockJob({ timestamp: new Date().toISOString() });
|
||||
const dbError = new Error('DB cleanup failed');
|
||||
mocks.deleteExpiredResetTokens.mockRejectedValue(dbError);
|
||||
await expect(tokenCleanupProcessor(job)).rejects.toThrow(dbError);
|
||||
});
|
||||
});
|
||||
});
|
||||
344
src/services/workers.server.ts
Normal file
344
src/services/workers.server.ts
Normal file
@@ -0,0 +1,344 @@
|
||||
import { Worker, Job, UnrecoverableError } from 'bullmq';
|
||||
import fsPromises from 'node:fs/promises';
|
||||
import { exec } from 'child_process';
|
||||
import { promisify } from 'util';
|
||||
|
||||
import { logger } from './logger.server';
|
||||
import { connection } from './redis.server';
|
||||
import { aiService } from './aiService.server';
|
||||
import * as emailService from './emailService.server';
|
||||
import * as db from './db/index.db';
|
||||
import {
|
||||
FlyerProcessingService,
|
||||
type FlyerJobData,
|
||||
type IFileSystem,
|
||||
} from './flyerProcessingService.server';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
import {
|
||||
flyerQueue,
|
||||
emailQueue,
|
||||
analyticsQueue,
|
||||
weeklyAnalyticsQueue,
|
||||
cleanupQueue,
|
||||
tokenCleanupQueue,
|
||||
type EmailJobData,
|
||||
type AnalyticsJobData,
|
||||
type CleanupJobData,
|
||||
type WeeklyAnalyticsJobData,
|
||||
type TokenCleanupJobData,
|
||||
} from './queues.server';
|
||||
|
||||
const execAsync = promisify(exec);
|
||||
|
||||
// --- Worker Instantiation ---
|
||||
|
||||
const fsAdapter: IFileSystem = {
|
||||
readdir: (path: string, options: { withFileTypes: true }) => fsPromises.readdir(path, options),
|
||||
unlink: (path: string) => fsPromises.unlink(path),
|
||||
};
|
||||
|
||||
const flyerProcessingService = new FlyerProcessingService(
|
||||
aiService,
|
||||
db,
|
||||
fsAdapter,
|
||||
execAsync,
|
||||
cleanupQueue,
|
||||
new FlyerDataTransformer(),
|
||||
);
|
||||
|
||||
const normalizeError = (error: unknown): Error => {
|
||||
return error instanceof Error ? error : new Error(String(error));
|
||||
};
|
||||
|
||||
const attachWorkerEventListeners = (worker: Worker) => {
|
||||
worker.on('completed', (job: Job, returnValue: unknown) => {
|
||||
logger.info({ returnValue }, `[${worker.name}] Job ${job.id} completed successfully.`);
|
||||
});
|
||||
|
||||
worker.on('failed', (job: Job | undefined, error: Error) => {
|
||||
logger.error(
|
||||
{ err: error, jobData: job?.data },
|
||||
`[${worker.name}] Job ${job?.id} has ultimately failed after all attempts.`,
|
||||
);
|
||||
});
|
||||
};
|
||||
|
||||
export const flyerWorker = new Worker<FlyerJobData>(
|
||||
'flyer-processing',
|
||||
async (job) => {
|
||||
try {
|
||||
return await flyerProcessingService.processJob(job);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
const errorMessage = wrappedError.message || '';
|
||||
if (
|
||||
errorMessage.includes('quota') ||
|
||||
errorMessage.includes('429') ||
|
||||
errorMessage.includes('RESOURCE_EXHAUSTED')
|
||||
) {
|
||||
logger.error(
|
||||
{ err: wrappedError, jobId: job.id },
|
||||
'[FlyerWorker] Unrecoverable quota error detected. Failing job immediately.',
|
||||
);
|
||||
throw new UnrecoverableError(errorMessage);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.WORKER_CONCURRENCY || '1', 10),
|
||||
},
|
||||
);
|
||||
|
||||
export const emailWorker = new Worker<EmailJobData>(
|
||||
'email-sending',
|
||||
async (job: Job<EmailJobData>) => {
|
||||
const { to, subject } = job.data;
|
||||
const jobLogger = logger.child({ jobId: job.id, jobName: job.name });
|
||||
jobLogger.info({ to, subject }, `[EmailWorker] Sending email for job ${job.id}`);
|
||||
try {
|
||||
await emailService.sendEmail(job.data, jobLogger);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error(
|
||||
{
|
||||
err: wrappedError,
|
||||
jobData: job.data,
|
||||
},
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.EMAIL_WORKER_CONCURRENCY || '10', 10),
|
||||
},
|
||||
);
|
||||
|
||||
export const analyticsWorker = new Worker<AnalyticsJobData>(
|
||||
'analytics-reporting',
|
||||
async (job: Job<AnalyticsJobData>) => {
|
||||
const { reportDate } = job.data;
|
||||
logger.info({ reportDate }, `[AnalyticsWorker] Starting report generation for job ${job.id}`);
|
||||
try {
|
||||
if (reportDate === 'FAIL') {
|
||||
throw new Error('This is a test failure for the analytics job.');
|
||||
}
|
||||
await new Promise((resolve) => setTimeout(resolve, 10000));
|
||||
logger.info(`[AnalyticsWorker] Successfully generated report for ${reportDate}.`);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error({ err: wrappedError, jobData: job.data },
|
||||
`[AnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.ANALYTICS_WORKER_CONCURRENCY || '1', 10),
|
||||
},
|
||||
);
|
||||
|
||||
export const cleanupWorker = new Worker<CleanupJobData>(
|
||||
'file-cleanup',
|
||||
async (job: Job<CleanupJobData>) => {
|
||||
const { flyerId, paths } = job.data;
|
||||
logger.info(
|
||||
{ paths },
|
||||
`[CleanupWorker] Starting file cleanup for job ${job.id} (Flyer ID: ${flyerId})`,
|
||||
);
|
||||
|
||||
try {
|
||||
if (!paths || paths.length === 0) {
|
||||
logger.warn(
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${flyerId} received no paths to clean. Skipping.`,
|
||||
);
|
||||
return;
|
||||
}
|
||||
|
||||
for (const filePath of paths) {
|
||||
try {
|
||||
await fsAdapter.unlink(filePath);
|
||||
logger.info(`[CleanupWorker] Deleted temporary file: ${filePath}`);
|
||||
} catch (unlinkError: unknown) {
|
||||
if (
|
||||
unlinkError instanceof Error &&
|
||||
'code' in unlinkError &&
|
||||
(unlinkError as any).code === 'ENOENT'
|
||||
) {
|
||||
logger.warn(
|
||||
`[CleanupWorker] File not found during cleanup (already deleted?): ${filePath}`,
|
||||
);
|
||||
} else {
|
||||
throw unlinkError;
|
||||
}
|
||||
}
|
||||
}
|
||||
logger.info(
|
||||
`[CleanupWorker] Successfully cleaned up ${paths.length} file(s) for flyer ${flyerId}.`,
|
||||
);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error(
|
||||
{ err: wrappedError },
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${flyerId} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.CLEANUP_WORKER_CONCURRENCY || '10', 10),
|
||||
},
|
||||
);
|
||||
|
||||
export const weeklyAnalyticsWorker = new Worker<WeeklyAnalyticsJobData>(
|
||||
'weekly-analytics-reporting',
|
||||
async (job: Job<WeeklyAnalyticsJobData>) => {
|
||||
const { reportYear, reportWeek } = job.data;
|
||||
logger.info(
|
||||
{ reportYear, reportWeek },
|
||||
`[WeeklyAnalyticsWorker] Starting weekly report generation for job ${job.id}`,
|
||||
);
|
||||
try {
|
||||
await new Promise((resolve) => setTimeout(resolve, 30000));
|
||||
logger.info(
|
||||
`[WeeklyAnalyticsWorker] Successfully generated weekly report for week ${reportWeek}, ${reportYear}.`,
|
||||
);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error(
|
||||
{ err: wrappedError, jobData: job.data },
|
||||
`[WeeklyAnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.WEEKLY_ANALYTICS_WORKER_CONCURRENCY || '1', 10),
|
||||
},
|
||||
);
|
||||
|
||||
export const tokenCleanupWorker = new Worker<TokenCleanupJobData>(
|
||||
'token-cleanup',
|
||||
async (job: Job<TokenCleanupJobData>) => {
|
||||
const jobLogger = logger.child({ jobId: job.id, jobName: job.name });
|
||||
jobLogger.info('[TokenCleanupWorker] Starting cleanup of expired password reset tokens.');
|
||||
try {
|
||||
const deletedCount = await db.userRepo.deleteExpiredResetTokens(jobLogger);
|
||||
jobLogger.info(`[TokenCleanupWorker] Successfully deleted ${deletedCount} expired tokens.`);
|
||||
return { deletedCount };
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
jobLogger.error({ err: wrappedError }, `[TokenCleanupWorker] Job ${job.id} failed.`);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
{
|
||||
connection,
|
||||
concurrency: 1,
|
||||
},
|
||||
);
|
||||
|
||||
attachWorkerEventListeners(flyerWorker);
|
||||
attachWorkerEventListeners(emailWorker);
|
||||
attachWorkerEventListeners(analyticsWorker);
|
||||
attachWorkerEventListeners(cleanupWorker);
|
||||
attachWorkerEventListeners(weeklyAnalyticsWorker);
|
||||
attachWorkerEventListeners(tokenCleanupWorker);
|
||||
|
||||
logger.info('All workers started and listening for jobs.');
|
||||
|
||||
const SHUTDOWN_TIMEOUT = 30000; // 30 seconds
|
||||
|
||||
export const gracefulShutdown = async (signal: string) => {
|
||||
logger.info(
|
||||
`[Shutdown] Received ${signal}. Initiating graceful shutdown (timeout: ${SHUTDOWN_TIMEOUT / 1000}s)...`,
|
||||
);
|
||||
|
||||
const shutdownPromise = (async () => {
|
||||
let hasErrors = false;
|
||||
|
||||
// Helper function to close a group of resources and log results
|
||||
const closeResources = async (resources: { name: string; close: () => Promise<any> }[], type: string) => {
|
||||
logger.info(`[Shutdown] Closing all ${type}...`);
|
||||
const results = await Promise.allSettled(resources.map((r) => r.close()));
|
||||
let groupHasErrors = false;
|
||||
|
||||
results.forEach((result, index) => {
|
||||
if (result.status === 'rejected') {
|
||||
groupHasErrors = true;
|
||||
logger.error(
|
||||
{ err: result.reason, resource: resources[index].name },
|
||||
`[Shutdown] Error closing ${resources[index].name}.`,
|
||||
);
|
||||
}
|
||||
});
|
||||
|
||||
if (!groupHasErrors) logger.info(`[Shutdown] All ${type} closed successfully.`);
|
||||
return groupHasErrors;
|
||||
};
|
||||
|
||||
// Define resource groups for sequential shutdown
|
||||
const workerResources = [
|
||||
{ name: 'flyerWorker', close: () => flyerWorker.close() },
|
||||
{ name: 'emailWorker', close: () => emailWorker.close() },
|
||||
{ name: 'analyticsWorker', close: () => analyticsWorker.close() },
|
||||
{ name: 'cleanupWorker', close: () => cleanupWorker.close() },
|
||||
{ name: 'weeklyAnalyticsWorker', close: () => weeklyAnalyticsWorker.close() },
|
||||
{ name: 'tokenCleanupWorker', close: () => tokenCleanupWorker.close() },
|
||||
];
|
||||
|
||||
const queueResources = [
|
||||
{ name: 'flyerQueue', close: () => flyerQueue.close() },
|
||||
{ name: 'emailQueue', close: () => emailQueue.close() },
|
||||
{ name: 'analyticsQueue', close: () => analyticsQueue.close() },
|
||||
{ name: 'cleanupQueue', close: () => cleanupQueue.close() },
|
||||
{ name: 'weeklyAnalyticsQueue', close: () => weeklyAnalyticsQueue.close() },
|
||||
{ name: 'tokenCleanupQueue', close: () => tokenCleanupQueue.close() },
|
||||
];
|
||||
|
||||
// 1. Close workers first
|
||||
if (await closeResources(workerResources, 'workers')) hasErrors = true;
|
||||
|
||||
// 2. Then close queues
|
||||
if (await closeResources(queueResources, 'queues')) hasErrors = true;
|
||||
|
||||
// 3. Finally, close the Redis connection
|
||||
logger.info('[Shutdown] Closing Redis connection...');
|
||||
try {
|
||||
await connection.quit();
|
||||
logger.info('[Shutdown] Redis connection closed successfully.');
|
||||
} catch (err) {
|
||||
hasErrors = true;
|
||||
logger.error({ err, resource: 'redisConnection' }, `[Shutdown] Error closing Redis connection.`);
|
||||
}
|
||||
|
||||
return hasErrors;
|
||||
})();
|
||||
|
||||
const timeoutPromise = new Promise<string>((resolve) =>
|
||||
setTimeout(() => resolve('timeout'), SHUTDOWN_TIMEOUT),
|
||||
);
|
||||
|
||||
const result = await Promise.race([shutdownPromise, timeoutPromise]);
|
||||
|
||||
if (result === 'timeout') {
|
||||
logger.error(
|
||||
`[Shutdown] Graceful shutdown timed out after ${SHUTDOWN_TIMEOUT / 1000} seconds. Forcing exit.`,
|
||||
);
|
||||
process.exit(1);
|
||||
} else {
|
||||
const hasErrors = result as boolean;
|
||||
if (!hasErrors) {
|
||||
logger.info('[Shutdown] All resources closed successfully.');
|
||||
} else {
|
||||
logger.warn('[Shutdown] Graceful shutdown completed with errors.');
|
||||
}
|
||||
process.exit(hasErrors ? 1 : 0);
|
||||
}
|
||||
};
|
||||
Reference in New Issue
Block a user