Compare commits
8 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
c182ef6d30 | ||
| fdb3b76cbd | |||
|
|
01e7c843cb | ||
| a0dbefbfa0 | |||
|
|
ab3fc318a0 | ||
| e658b35e43 | |||
|
|
67e106162a | ||
| b7f3182fd6 |
@@ -127,7 +127,7 @@ jobs:
|
|||||||
|
|
||||||
# --- Increase Node.js memory limit to prevent heap out of memory errors ---
|
# --- Increase Node.js memory limit to prevent heap out of memory errors ---
|
||||||
# This is crucial for memory-intensive tasks like running tests and coverage.
|
# This is crucial for memory-intensive tasks like running tests and coverage.
|
||||||
NODE_OPTIONS: '--max-old-space-size=8192'
|
NODE_OPTIONS: '--max-old-space-size=8192 --trace-warnings --unhandled-rejections=strict'
|
||||||
|
|
||||||
run: |
|
run: |
|
||||||
# Fail-fast check to ensure secrets are configured in Gitea for testing.
|
# Fail-fast check to ensure secrets are configured in Gitea for testing.
|
||||||
@@ -376,7 +376,7 @@ jobs:
|
|||||||
|
|
||||||
# Application Secrets
|
# Application Secrets
|
||||||
FRONTEND_URL: 'https://flyer-crawler-test.projectium.com'
|
FRONTEND_URL: 'https://flyer-crawler-test.projectium.com'
|
||||||
JWT_SECRET: ${{ secrets.JWT_SECRET_TEST }}
|
JWT_SECRET: ${{ secrets.JWT_SECRET }}
|
||||||
GEMINI_API_KEY: ${{ secrets.VITE_GOOGLE_GENAI_API_KEY_TEST }}
|
GEMINI_API_KEY: ${{ secrets.VITE_GOOGLE_GENAI_API_KEY_TEST }}
|
||||||
GOOGLE_MAPS_API_KEY: ${{ secrets.GOOGLE_MAPS_API_KEY }}
|
GOOGLE_MAPS_API_KEY: ${{ secrets.GOOGLE_MAPS_API_KEY }}
|
||||||
|
|
||||||
|
|||||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
|||||||
{
|
{
|
||||||
"name": "flyer-crawler",
|
"name": "flyer-crawler",
|
||||||
"version": "0.2.22",
|
"version": "0.2.26",
|
||||||
"lockfileVersion": 3,
|
"lockfileVersion": 3,
|
||||||
"requires": true,
|
"requires": true,
|
||||||
"packages": {
|
"packages": {
|
||||||
"": {
|
"": {
|
||||||
"name": "flyer-crawler",
|
"name": "flyer-crawler",
|
||||||
"version": "0.2.22",
|
"version": "0.2.26",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@bull-board/api": "^6.14.2",
|
"@bull-board/api": "^6.14.2",
|
||||||
"@bull-board/express": "^6.14.2",
|
"@bull-board/express": "^6.14.2",
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
{
|
{
|
||||||
"name": "flyer-crawler",
|
"name": "flyer-crawler",
|
||||||
"private": true,
|
"private": true,
|
||||||
"version": "0.2.22",
|
"version": "0.2.26",
|
||||||
"type": "module",
|
"type": "module",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||||
|
|||||||
@@ -11,6 +11,8 @@ import { createTestApp } from '../tests/utils/createTestApp';
|
|||||||
vi.mock('../services/backgroundJobService', () => ({
|
vi.mock('../services/backgroundJobService', () => ({
|
||||||
backgroundJobService: {
|
backgroundJobService: {
|
||||||
runDailyDealCheck: vi.fn(),
|
runDailyDealCheck: vi.fn(),
|
||||||
|
triggerAnalyticsReport: vi.fn(),
|
||||||
|
triggerWeeklyAnalyticsReport: vi.fn(),
|
||||||
},
|
},
|
||||||
}));
|
}));
|
||||||
|
|
||||||
@@ -142,22 +144,17 @@ describe('Admin Job Trigger Routes (/api/admin/trigger)', () => {
|
|||||||
|
|
||||||
describe('POST /trigger/analytics-report', () => {
|
describe('POST /trigger/analytics-report', () => {
|
||||||
it('should trigger the analytics report job and return 202 Accepted', async () => {
|
it('should trigger the analytics report job and return 202 Accepted', async () => {
|
||||||
const mockJob = { id: 'manual-report-job-123' } as Job;
|
vi.mocked(backgroundJobService.triggerAnalyticsReport).mockResolvedValue('manual-report-job-123');
|
||||||
vi.mocked(analyticsQueue.add).mockResolvedValue(mockJob);
|
|
||||||
|
|
||||||
const response = await supertest(app).post('/api/admin/trigger/analytics-report');
|
const response = await supertest(app).post('/api/admin/trigger/analytics-report');
|
||||||
|
|
||||||
expect(response.status).toBe(202);
|
expect(response.status).toBe(202);
|
||||||
expect(response.body.message).toContain('Analytics report generation job has been enqueued');
|
expect(response.body.message).toContain('Analytics report generation job has been enqueued');
|
||||||
expect(analyticsQueue.add).toHaveBeenCalledWith(
|
expect(backgroundJobService.triggerAnalyticsReport).toHaveBeenCalledTimes(1);
|
||||||
'generate-daily-report',
|
|
||||||
expect.objectContaining({ reportDate: expect.any(String) }),
|
|
||||||
expect.any(Object),
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 if enqueuing the analytics job fails', async () => {
|
it('should return 500 if enqueuing the analytics job fails', async () => {
|
||||||
vi.mocked(analyticsQueue.add).mockRejectedValue(new Error('Queue error'));
|
vi.mocked(backgroundJobService.triggerAnalyticsReport).mockRejectedValue(new Error('Queue error'));
|
||||||
const response = await supertest(app).post('/api/admin/trigger/analytics-report');
|
const response = await supertest(app).post('/api/admin/trigger/analytics-report');
|
||||||
expect(response.status).toBe(500);
|
expect(response.status).toBe(500);
|
||||||
});
|
});
|
||||||
@@ -165,22 +162,17 @@ describe('Admin Job Trigger Routes (/api/admin/trigger)', () => {
|
|||||||
|
|
||||||
describe('POST /trigger/weekly-analytics', () => {
|
describe('POST /trigger/weekly-analytics', () => {
|
||||||
it('should trigger the weekly analytics job and return 202 Accepted', async () => {
|
it('should trigger the weekly analytics job and return 202 Accepted', async () => {
|
||||||
const mockJob = { id: 'manual-weekly-report-job-123' } as Job;
|
vi.mocked(backgroundJobService.triggerWeeklyAnalyticsReport).mockResolvedValue('manual-weekly-report-job-123');
|
||||||
vi.mocked(weeklyAnalyticsQueue.add).mockResolvedValue(mockJob);
|
|
||||||
|
|
||||||
const response = await supertest(app).post('/api/admin/trigger/weekly-analytics');
|
const response = await supertest(app).post('/api/admin/trigger/weekly-analytics');
|
||||||
|
|
||||||
expect(response.status).toBe(202);
|
expect(response.status).toBe(202);
|
||||||
expect(response.body.message).toContain('Successfully enqueued weekly analytics job');
|
expect(response.body.message).toContain('Successfully enqueued weekly analytics job');
|
||||||
expect(weeklyAnalyticsQueue.add).toHaveBeenCalledWith(
|
expect(backgroundJobService.triggerWeeklyAnalyticsReport).toHaveBeenCalledTimes(1);
|
||||||
'generate-weekly-report',
|
|
||||||
expect.objectContaining({ reportYear: expect.any(Number), reportWeek: expect.any(Number) }),
|
|
||||||
expect.any(Object),
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 if enqueuing the weekly analytics job fails', async () => {
|
it('should return 500 if enqueuing the weekly analytics job fails', async () => {
|
||||||
vi.mocked(weeklyAnalyticsQueue.add).mockRejectedValue(new Error('Queue error'));
|
vi.mocked(backgroundJobService.triggerWeeklyAnalyticsReport).mockRejectedValue(new Error('Queue error'));
|
||||||
const response = await supertest(app).post('/api/admin/trigger/weekly-analytics');
|
const response = await supertest(app).post('/api/admin/trigger/weekly-analytics');
|
||||||
expect(response.status).toBe(500);
|
expect(response.status).toBe(500);
|
||||||
});
|
});
|
||||||
@@ -242,15 +234,17 @@ describe('Admin Job Trigger Routes (/api/admin/trigger)', () => {
|
|||||||
expect(response.status).toBe(400);
|
expect(response.status).toBe(400);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 404 if the queue name is valid but not in the retry map', async () => {
|
it('should return 404 if the job ID is not found in the weekly-analytics-reporting queue', async () => {
|
||||||
const queueName = 'weekly-analytics-reporting'; // This is in the Zod enum but not the queueMap
|
const queueName = 'weekly-analytics-reporting';
|
||||||
const jobId = 'some-job-id';
|
const jobId = 'some-job-id';
|
||||||
|
|
||||||
|
// Ensure getJob returns undefined (not found)
|
||||||
|
vi.mocked(weeklyAnalyticsQueue.getJob).mockResolvedValue(undefined);
|
||||||
|
|
||||||
const response = await supertest(app).post(`/api/admin/jobs/${queueName}/${jobId}/retry`);
|
const response = await supertest(app).post(`/api/admin/jobs/${queueName}/${jobId}/retry`);
|
||||||
|
|
||||||
// The route throws a NotFoundError, which the error handler should convert to a 404.
|
|
||||||
expect(response.status).toBe(404);
|
expect(response.status).toBe(404);
|
||||||
expect(response.body.message).toBe(`Queue 'weekly-analytics-reporting' not found.`);
|
expect(response.body.message).toBe(`Job with ID '${jobId}' not found in queue '${queueName}'.`);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 404 if the job ID is not found in the queue', async () => {
|
it('should return 404 if the job ID is not found in the queue', async () => {
|
||||||
|
|||||||
@@ -20,49 +20,25 @@ import { validateRequest } from '../middleware/validation.middleware';
|
|||||||
import { createBullBoard } from '@bull-board/api';
|
import { createBullBoard } from '@bull-board/api';
|
||||||
import { BullMQAdapter } from '@bull-board/api/bullMQAdapter';
|
import { BullMQAdapter } from '@bull-board/api/bullMQAdapter';
|
||||||
import { ExpressAdapter } from '@bull-board/express';
|
import { ExpressAdapter } from '@bull-board/express';
|
||||||
|
|
||||||
import type { Queue } from 'bullmq';
|
|
||||||
import { backgroundJobService } from '../services/backgroundJobService';
|
import { backgroundJobService } from '../services/backgroundJobService';
|
||||||
import {
|
import { flyerQueue, emailQueue, analyticsQueue, cleanupQueue, weeklyAnalyticsQueue } from '../services/queueService.server';
|
||||||
flyerQueue,
|
|
||||||
emailQueue,
|
|
||||||
analyticsQueue,
|
|
||||||
cleanupQueue,
|
|
||||||
weeklyAnalyticsQueue,
|
|
||||||
} from '../services/queueService.server'; // Import your queues
|
|
||||||
import {
|
|
||||||
analyticsWorker,
|
|
||||||
cleanupWorker,
|
|
||||||
emailWorker,
|
|
||||||
flyerWorker,
|
|
||||||
weeklyAnalyticsWorker,
|
|
||||||
} from '../services/workers.server';
|
|
||||||
import { getSimpleWeekAndYear } from '../utils/dateUtils';
|
import { getSimpleWeekAndYear } from '../utils/dateUtils';
|
||||||
import {
|
import {
|
||||||
requiredString,
|
requiredString,
|
||||||
numericIdParam,
|
numericIdParam,
|
||||||
uuidParamSchema,
|
uuidParamSchema,
|
||||||
optionalNumeric,
|
optionalNumeric,
|
||||||
|
optionalString,
|
||||||
} from '../utils/zodUtils';
|
} from '../utils/zodUtils';
|
||||||
import { logger } from '../services/logger.server';
|
import { logger } from '../services/logger.server'; // This was a duplicate, fixed.
|
||||||
import fs from 'node:fs/promises';
|
import { monitoringService } from '../services/monitoringService.server';
|
||||||
|
import { userService } from '../services/userService';
|
||||||
/**
|
import { cleanupUploadedFile } from '../utils/fileUtils';
|
||||||
* Safely deletes a file from the filesystem, ignoring errors if the file doesn't exist.
|
import { brandService } from '../services/brandService';
|
||||||
* @param file The multer file object to delete.
|
|
||||||
*/
|
|
||||||
const cleanupUploadedFile = async (file?: Express.Multer.File) => {
|
|
||||||
if (!file) return;
|
|
||||||
try {
|
|
||||||
await fs.unlink(file.path);
|
|
||||||
} catch (err) {
|
|
||||||
logger.warn({ err, filePath: file.path }, 'Failed to clean up uploaded logo file.');
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const updateCorrectionSchema = numericIdParam('id').extend({
|
const updateCorrectionSchema = numericIdParam('id').extend({
|
||||||
body: z.object({
|
body: z.object({
|
||||||
suggested_value: requiredString('A new suggested_value is required.'),
|
suggested_value: z.string().trim().min(1, 'A new suggested_value is required.'),
|
||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -100,10 +76,12 @@ const jobRetrySchema = z.object({
|
|||||||
'file-cleanup',
|
'file-cleanup',
|
||||||
'weekly-analytics-reporting',
|
'weekly-analytics-reporting',
|
||||||
]),
|
]),
|
||||||
jobId: requiredString('A valid Job ID is required.'),
|
jobId: z.string().trim().min(1, 'A valid Job ID is required.'),
|
||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
|
const emptySchema = z.object({});
|
||||||
|
|
||||||
const router = Router();
|
const router = Router();
|
||||||
|
|
||||||
const upload = createUploadMiddleware({ storageType: 'flyer' });
|
const upload = createUploadMiddleware({ storageType: 'flyer' });
|
||||||
@@ -138,7 +116,7 @@ router.use(passport.authenticate('jwt', { session: false }), isAdmin);
|
|||||||
|
|
||||||
// --- Admin Routes ---
|
// --- Admin Routes ---
|
||||||
|
|
||||||
router.get('/corrections', async (req, res, next: NextFunction) => {
|
router.get('/corrections', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const corrections = await db.adminRepo.getSuggestedCorrections(req.log);
|
const corrections = await db.adminRepo.getSuggestedCorrections(req.log);
|
||||||
res.json(corrections);
|
res.json(corrections);
|
||||||
@@ -148,7 +126,7 @@ router.get('/corrections', async (req, res, next: NextFunction) => {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
router.get('/review/flyers', async (req, res, next: NextFunction) => {
|
router.get('/review/flyers', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
req.log.debug('Fetching flyers for review via adminRepo');
|
req.log.debug('Fetching flyers for review via adminRepo');
|
||||||
const flyers = await db.adminRepo.getFlyersForReview(req.log);
|
const flyers = await db.adminRepo.getFlyersForReview(req.log);
|
||||||
@@ -160,7 +138,7 @@ router.get('/review/flyers', async (req, res, next: NextFunction) => {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
router.get('/brands', async (req, res, next: NextFunction) => {
|
router.get('/brands', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const brands = await db.flyerRepo.getAllBrands(req.log);
|
const brands = await db.flyerRepo.getAllBrands(req.log);
|
||||||
res.json(brands);
|
res.json(brands);
|
||||||
@@ -170,7 +148,7 @@ router.get('/brands', async (req, res, next: NextFunction) => {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
router.get('/stats', async (req, res, next: NextFunction) => {
|
router.get('/stats', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const stats = await db.adminRepo.getApplicationStats(req.log);
|
const stats = await db.adminRepo.getApplicationStats(req.log);
|
||||||
res.json(stats);
|
res.json(stats);
|
||||||
@@ -180,7 +158,7 @@ router.get('/stats', async (req, res, next: NextFunction) => {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
router.get('/stats/daily', async (req, res, next: NextFunction) => {
|
router.get('/stats/daily', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const dailyStats = await db.adminRepo.getDailyStatsForLast30Days(req.log);
|
const dailyStats = await db.adminRepo.getDailyStatsForLast30Days(req.log);
|
||||||
res.json(dailyStats);
|
res.json(dailyStats);
|
||||||
@@ -264,7 +242,6 @@ router.post(
|
|||||||
upload.single('logoImage'),
|
upload.single('logoImage'),
|
||||||
requireFileUpload('logoImage'),
|
requireFileUpload('logoImage'),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
// Apply ADR-003 pattern for type safety
|
|
||||||
const { params } = req as unknown as z.infer<ReturnType<typeof numericIdParam>>;
|
const { params } = req as unknown as z.infer<ReturnType<typeof numericIdParam>>;
|
||||||
try {
|
try {
|
||||||
// Although requireFileUpload middleware should ensure the file exists,
|
// Although requireFileUpload middleware should ensure the file exists,
|
||||||
@@ -272,9 +249,8 @@ router.post(
|
|||||||
if (!req.file) {
|
if (!req.file) {
|
||||||
throw new ValidationError([], 'Logo image file is missing.');
|
throw new ValidationError([], 'Logo image file is missing.');
|
||||||
}
|
}
|
||||||
// The storage path is 'flyer-images', so the URL should reflect that for consistency.
|
|
||||||
const logoUrl = `/flyer-images/${req.file.filename}`;
|
const logoUrl = await brandService.updateBrandLogo(params.id, req.file, req.log);
|
||||||
await db.adminRepo.updateBrandLogo(params.id, logoUrl, req.log);
|
|
||||||
|
|
||||||
logger.info({ brandId: params.id, logoUrl }, `Brand logo updated for brand ID: ${params.id}`);
|
logger.info({ brandId: params.id, logoUrl }, `Brand logo updated for brand ID: ${params.id}`);
|
||||||
res.status(200).json({ message: 'Brand logo updated successfully.', logoUrl });
|
res.status(200).json({ message: 'Brand logo updated successfully.', logoUrl });
|
||||||
@@ -288,7 +264,7 @@ router.post(
|
|||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|
||||||
router.get('/unmatched-items', async (req, res, next: NextFunction) => {
|
router.get('/unmatched-items', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const items = await db.adminRepo.getUnmatchedFlyerItems(req.log);
|
const items = await db.adminRepo.getUnmatchedFlyerItems(req.log);
|
||||||
res.json(items);
|
res.json(items);
|
||||||
@@ -358,7 +334,7 @@ router.put(
|
|||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|
||||||
router.get('/users', async (req, res, next: NextFunction) => {
|
router.get('/users', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const users = await db.adminRepo.getAllUsers(req.log);
|
const users = await db.adminRepo.getAllUsers(req.log);
|
||||||
res.json(users);
|
res.json(users);
|
||||||
@@ -373,14 +349,11 @@ router.get(
|
|||||||
validateRequest(activityLogSchema),
|
validateRequest(activityLogSchema),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
// Apply ADR-003 pattern for type safety.
|
// Apply ADR-003 pattern for type safety.
|
||||||
// We explicitly coerce query params here because the validation middleware might not
|
// We parse the query here to apply Zod's coercions (string to number) and defaults.
|
||||||
// replace req.query with the coerced values in all environments.
|
const { limit, offset } = activityLogSchema.shape.query.parse(req.query);
|
||||||
const query = req.query as unknown as { limit?: string; offset?: string };
|
|
||||||
const limit = query.limit ? Number(query.limit) : 50;
|
|
||||||
const offset = query.offset ? Number(query.offset) : 0;
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const logs = await db.adminRepo.getActivityLog(limit, offset, req.log);
|
const logs = await db.adminRepo.getActivityLog(limit!, offset!, req.log);
|
||||||
res.json(logs);
|
res.json(logs);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, 'Error fetching activity log');
|
logger.error({ error }, 'Error fetching activity log');
|
||||||
@@ -429,10 +402,7 @@ router.delete(
|
|||||||
// Apply ADR-003 pattern for type safety
|
// Apply ADR-003 pattern for type safety
|
||||||
const { params } = req as unknown as z.infer<ReturnType<typeof uuidParamSchema>>;
|
const { params } = req as unknown as z.infer<ReturnType<typeof uuidParamSchema>>;
|
||||||
try {
|
try {
|
||||||
if (userProfile.user.user_id === params.id) {
|
await userService.deleteUserAsAdmin(userProfile.user.user_id, params.id, req.log);
|
||||||
throw new ValidationError([], 'Admins cannot delete their own account.');
|
|
||||||
}
|
|
||||||
await db.userRepo.deleteUserById(params.id, req.log);
|
|
||||||
res.status(204).send();
|
res.status(204).send();
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, 'Error deleting user');
|
logger.error({ error }, 'Error deleting user');
|
||||||
@@ -447,6 +417,7 @@ router.delete(
|
|||||||
*/
|
*/
|
||||||
router.post(
|
router.post(
|
||||||
'/trigger/daily-deal-check',
|
'/trigger/daily-deal-check',
|
||||||
|
validateRequest(emptySchema),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
const userProfile = req.user as UserProfile;
|
const userProfile = req.user as UserProfile;
|
||||||
logger.info(
|
logger.info(
|
||||||
@@ -474,6 +445,7 @@ router.post(
|
|||||||
*/
|
*/
|
||||||
router.post(
|
router.post(
|
||||||
'/trigger/analytics-report',
|
'/trigger/analytics-report',
|
||||||
|
validateRequest(emptySchema),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
const userProfile = req.user as UserProfile;
|
const userProfile = req.user as UserProfile;
|
||||||
logger.info(
|
logger.info(
|
||||||
@@ -481,14 +453,9 @@ router.post(
|
|||||||
);
|
);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const reportDate = new Date().toISOString().split('T')[0]; // YYYY-MM-DD
|
const jobId = await backgroundJobService.triggerAnalyticsReport();
|
||||||
// Use a unique job ID for manual triggers to distinguish them from scheduled jobs.
|
|
||||||
const jobId = `manual-report-${reportDate}-${Date.now()}`;
|
|
||||||
|
|
||||||
const job = await analyticsQueue.add('generate-daily-report', { reportDate }, { jobId });
|
|
||||||
|
|
||||||
res.status(202).json({
|
res.status(202).json({
|
||||||
message: `Analytics report generation job has been enqueued successfully. Job ID: ${job.id}`,
|
message: `Analytics report generation job has been enqueued successfully. Job ID: ${jobId}`,
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, '[Admin] Failed to enqueue analytics report job.');
|
logger.error({ error }, '[Admin] Failed to enqueue analytics report job.');
|
||||||
@@ -529,7 +496,10 @@ router.post(
|
|||||||
* POST /api/admin/trigger/failing-job - Enqueue a test job designed to fail.
|
* POST /api/admin/trigger/failing-job - Enqueue a test job designed to fail.
|
||||||
* This is for testing the retry mechanism and Bull Board UI.
|
* This is for testing the retry mechanism and Bull Board UI.
|
||||||
*/
|
*/
|
||||||
router.post('/trigger/failing-job', async (req: Request, res: Response, next: NextFunction) => {
|
router.post(
|
||||||
|
'/trigger/failing-job',
|
||||||
|
validateRequest(emptySchema),
|
||||||
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
const userProfile = req.user as UserProfile;
|
const userProfile = req.user as UserProfile;
|
||||||
logger.info(
|
logger.info(
|
||||||
`[Admin] Manual trigger for a failing job received from user: ${userProfile.user.user_id}`,
|
`[Admin] Manual trigger for a failing job received from user: ${userProfile.user.user_id}`,
|
||||||
@@ -545,7 +515,8 @@ router.post('/trigger/failing-job', async (req: Request, res: Response, next: Ne
|
|||||||
logger.error({ error }, 'Error enqueuing failing job');
|
logger.error({ error }, 'Error enqueuing failing job');
|
||||||
next(error);
|
next(error);
|
||||||
}
|
}
|
||||||
});
|
}
|
||||||
|
);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* POST /api/admin/system/clear-geocode-cache - Clears the Redis cache for geocoded addresses.
|
* POST /api/admin/system/clear-geocode-cache - Clears the Redis cache for geocoded addresses.
|
||||||
@@ -553,6 +524,7 @@ router.post('/trigger/failing-job', async (req: Request, res: Response, next: Ne
|
|||||||
*/
|
*/
|
||||||
router.post(
|
router.post(
|
||||||
'/system/clear-geocode-cache',
|
'/system/clear-geocode-cache',
|
||||||
|
validateRequest(emptySchema),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
const userProfile = req.user as UserProfile;
|
const userProfile = req.user as UserProfile;
|
||||||
logger.info(
|
logger.info(
|
||||||
@@ -575,44 +547,23 @@ router.post(
|
|||||||
* GET /api/admin/workers/status - Get the current running status of all BullMQ workers.
|
* GET /api/admin/workers/status - Get the current running status of all BullMQ workers.
|
||||||
* This is useful for a system health dashboard to see if any workers have crashed.
|
* This is useful for a system health dashboard to see if any workers have crashed.
|
||||||
*/
|
*/
|
||||||
router.get('/workers/status', async (req: Request, res: Response) => {
|
router.get('/workers/status', validateRequest(emptySchema), async (req: Request, res: Response, next: NextFunction) => {
|
||||||
const workers = [flyerWorker, emailWorker, analyticsWorker, cleanupWorker, weeklyAnalyticsWorker];
|
try {
|
||||||
|
const workerStatuses = await monitoringService.getWorkerStatuses();
|
||||||
const workerStatuses = await Promise.all(
|
res.json(workerStatuses);
|
||||||
workers.map(async (worker) => {
|
} catch (error) {
|
||||||
return {
|
logger.error({ error }, 'Error fetching worker statuses');
|
||||||
name: worker.name,
|
next(error);
|
||||||
isRunning: worker.isRunning(),
|
}
|
||||||
};
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
|
|
||||||
res.json(workerStatuses);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* GET /api/admin/queues/status - Get job counts for all BullMQ queues.
|
* GET /api/admin/queues/status - Get job counts for all BullMQ queues.
|
||||||
* This is useful for monitoring the health and backlog of background jobs.
|
* This is useful for monitoring the health and backlog of background jobs.
|
||||||
*/
|
*/
|
||||||
router.get('/queues/status', async (req: Request, res: Response, next: NextFunction) => {
|
router.get('/queues/status', validateRequest(emptySchema), async (req: Request, res: Response, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const queues = [flyerQueue, emailQueue, analyticsQueue, cleanupQueue, weeklyAnalyticsQueue];
|
const queueStatuses = await monitoringService.getQueueStatuses();
|
||||||
|
|
||||||
const queueStatuses = await Promise.all(
|
|
||||||
queues.map(async (queue) => {
|
|
||||||
return {
|
|
||||||
name: queue.name,
|
|
||||||
counts: await queue.getJobCounts(
|
|
||||||
'waiting',
|
|
||||||
'active',
|
|
||||||
'completed',
|
|
||||||
'failed',
|
|
||||||
'delayed',
|
|
||||||
'paused',
|
|
||||||
),
|
|
||||||
};
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
res.json(queueStatuses);
|
res.json(queueStatuses);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, 'Error fetching queue statuses');
|
logger.error({ error }, 'Error fetching queue statuses');
|
||||||
@@ -632,35 +583,11 @@ router.post(
|
|||||||
params: { queueName, jobId },
|
params: { queueName, jobId },
|
||||||
} = req as unknown as z.infer<typeof jobRetrySchema>;
|
} = req as unknown as z.infer<typeof jobRetrySchema>;
|
||||||
|
|
||||||
const queueMap: { [key: string]: Queue } = {
|
|
||||||
'flyer-processing': flyerQueue,
|
|
||||||
'email-sending': emailQueue,
|
|
||||||
'analytics-reporting': analyticsQueue,
|
|
||||||
'file-cleanup': cleanupQueue,
|
|
||||||
};
|
|
||||||
|
|
||||||
const queue = queueMap[queueName];
|
|
||||||
|
|
||||||
if (!queue) {
|
|
||||||
// Throw a NotFoundError to be handled by the central error handler.
|
|
||||||
throw new NotFoundError(`Queue '${queueName}' not found.`);
|
|
||||||
}
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const job = await queue.getJob(jobId);
|
await monitoringService.retryFailedJob(
|
||||||
if (!job)
|
queueName,
|
||||||
throw new NotFoundError(`Job with ID '${jobId}' not found in queue '${queueName}'.`);
|
jobId,
|
||||||
|
userProfile.user.user_id,
|
||||||
const jobState = await job.getState();
|
|
||||||
if (jobState !== 'failed')
|
|
||||||
throw new ValidationError(
|
|
||||||
[],
|
|
||||||
`Job is not in a 'failed' state. Current state: ${jobState}.`,
|
|
||||||
); // This was a duplicate, fixed.
|
|
||||||
|
|
||||||
await job.retry();
|
|
||||||
logger.info(
|
|
||||||
`[Admin] User ${userProfile.user.user_id} manually retried job ${jobId} in queue ${queueName}.`,
|
|
||||||
);
|
);
|
||||||
res.status(200).json({ message: `Job ${jobId} has been successfully marked for retry.` });
|
res.status(200).json({ message: `Job ${jobId} has been successfully marked for retry.` });
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
@@ -675,6 +602,7 @@ router.post(
|
|||||||
*/
|
*/
|
||||||
router.post(
|
router.post(
|
||||||
'/trigger/weekly-analytics',
|
'/trigger/weekly-analytics',
|
||||||
|
validateRequest(emptySchema),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
const userProfile = req.user as UserProfile; // This was a duplicate, fixed.
|
const userProfile = req.user as UserProfile; // This was a duplicate, fixed.
|
||||||
logger.info(
|
logger.info(
|
||||||
@@ -682,19 +610,10 @@ router.post(
|
|||||||
);
|
);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const { year: reportYear, week: reportWeek } = getSimpleWeekAndYear();
|
const jobId = await backgroundJobService.triggerWeeklyAnalyticsReport();
|
||||||
const { weeklyAnalyticsQueue } = await import('../services/queueService.server');
|
|
||||||
const job = await weeklyAnalyticsQueue.add(
|
|
||||||
'generate-weekly-report',
|
|
||||||
{ reportYear, reportWeek },
|
|
||||||
{
|
|
||||||
jobId: `manual-weekly-report-${reportYear}-${reportWeek}-${Date.now()}`, // Add timestamp to avoid ID conflict
|
|
||||||
},
|
|
||||||
);
|
|
||||||
|
|
||||||
res
|
res
|
||||||
.status(202)
|
.status(202)
|
||||||
.json({ message: 'Successfully enqueued weekly analytics job.', jobId: job.id });
|
.json({ message: 'Successfully enqueued weekly analytics job.', jobId });
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, 'Error enqueuing weekly analytics job');
|
logger.error({ error }, 'Error enqueuing weekly analytics job');
|
||||||
next(error);
|
next(error);
|
||||||
@@ -705,4 +624,5 @@ router.post(
|
|||||||
/* Catches errors from multer (e.g., file size, file filter) */
|
/* Catches errors from multer (e.g., file size, file filter) */
|
||||||
router.use(handleMulterError);
|
router.use(handleMulterError);
|
||||||
|
|
||||||
|
|
||||||
export default router;
|
export default router;
|
||||||
|
|||||||
@@ -15,12 +15,18 @@ import { createTestApp } from '../tests/utils/createTestApp';
|
|||||||
import { mockLogger } from '../tests/utils/mockLogger';
|
import { mockLogger } from '../tests/utils/mockLogger';
|
||||||
|
|
||||||
// Mock the AI service methods to avoid making real AI calls
|
// Mock the AI service methods to avoid making real AI calls
|
||||||
vi.mock('../services/aiService.server', () => ({
|
vi.mock('../services/aiService.server', async (importOriginal) => {
|
||||||
aiService: {
|
const actual = await importOriginal<typeof import('../services/aiService.server')>();
|
||||||
extractTextFromImageArea: vi.fn(),
|
return {
|
||||||
planTripWithMaps: vi.fn(), // Added this missing mock
|
...actual,
|
||||||
},
|
aiService: {
|
||||||
}));
|
extractTextFromImageArea: vi.fn(),
|
||||||
|
planTripWithMaps: vi.fn(),
|
||||||
|
enqueueFlyerProcessing: vi.fn(),
|
||||||
|
processLegacyFlyerUpload: vi.fn(),
|
||||||
|
},
|
||||||
|
};
|
||||||
|
});
|
||||||
|
|
||||||
const { mockedDb } = vi.hoisted(() => ({
|
const { mockedDb } = vi.hoisted(() => ({
|
||||||
mockedDb: {
|
mockedDb: {
|
||||||
@@ -30,6 +36,9 @@ const { mockedDb } = vi.hoisted(() => ({
|
|||||||
adminRepo: {
|
adminRepo: {
|
||||||
logActivity: vi.fn(),
|
logActivity: vi.fn(),
|
||||||
},
|
},
|
||||||
|
personalizationRepo: {
|
||||||
|
getAllMasterItems: vi.fn(),
|
||||||
|
},
|
||||||
// This function is a standalone export, not part of a repo
|
// This function is a standalone export, not part of a repo
|
||||||
createFlyerAndItems: vi.fn(),
|
createFlyerAndItems: vi.fn(),
|
||||||
},
|
},
|
||||||
@@ -40,6 +49,7 @@ vi.mock('../services/db/flyer.db', () => ({ createFlyerAndItems: mockedDb.create
|
|||||||
vi.mock('../services/db/index.db', () => ({
|
vi.mock('../services/db/index.db', () => ({
|
||||||
flyerRepo: mockedDb.flyerRepo,
|
flyerRepo: mockedDb.flyerRepo,
|
||||||
adminRepo: mockedDb.adminRepo,
|
adminRepo: mockedDb.adminRepo,
|
||||||
|
personalizationRepo: mockedDb.personalizationRepo,
|
||||||
}));
|
}));
|
||||||
|
|
||||||
// Mock the queue service
|
// Mock the queue service
|
||||||
@@ -138,8 +148,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
const imagePath = path.resolve(__dirname, '../tests/assets/test-flyer-image.jpg');
|
const imagePath = path.resolve(__dirname, '../tests/assets/test-flyer-image.jpg');
|
||||||
|
|
||||||
it('should enqueue a job and return 202 on success', async () => {
|
it('should enqueue a job and return 202 on success', async () => {
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockResolvedValue({ id: 'job-123' } as unknown as Job);
|
||||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job-123' } as unknown as Job);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/ai/upload-and-process')
|
.post('/api/ai/upload-and-process')
|
||||||
@@ -149,7 +158,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
expect(response.status).toBe(202);
|
expect(response.status).toBe(202);
|
||||||
expect(response.body.message).toBe('Flyer accepted for processing.');
|
expect(response.body.message).toBe('Flyer accepted for processing.');
|
||||||
expect(response.body.jobId).toBe('job-123');
|
expect(response.body.jobId).toBe('job-123');
|
||||||
expect(flyerQueue.add).toHaveBeenCalledWith('process-flyer', expect.any(Object));
|
expect(aiService.aiService.enqueueFlyerProcessing).toHaveBeenCalled();
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 400 if no file is provided', async () => {
|
it('should return 400 if no file is provided', async () => {
|
||||||
@@ -172,9 +181,8 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should return 409 if flyer checksum already exists', async () => {
|
it('should return 409 if flyer checksum already exists', async () => {
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(
|
const duplicateError = new aiService.DuplicateFlyerError('This flyer has already been processed.', 99);
|
||||||
createMockFlyer({ flyer_id: 99 }),
|
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockRejectedValue(duplicateError);
|
||||||
);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/ai/upload-and-process')
|
.post('/api/ai/upload-and-process')
|
||||||
@@ -186,8 +194,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 if enqueuing the job fails', async () => {
|
it('should return 500 if enqueuing the job fails', async () => {
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockRejectedValueOnce(new Error('Redis connection failed'));
|
||||||
vi.mocked(flyerQueue.add).mockRejectedValueOnce(new Error('Redis connection failed'));
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/ai/upload-and-process')
|
.post('/api/ai/upload-and-process')
|
||||||
@@ -209,9 +216,8 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
basePath: '/api/ai',
|
basePath: '/api/ai',
|
||||||
authenticatedUser: mockUser,
|
authenticatedUser: mockUser,
|
||||||
});
|
});
|
||||||
|
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockResolvedValue({ id: 'job-456' } as unknown as Job);
|
||||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job-456' } as unknown as Job);
|
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
await supertest(authenticatedApp)
|
await supertest(authenticatedApp)
|
||||||
@@ -220,8 +226,10 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerFile', imagePath);
|
.attach('flyerFile', imagePath);
|
||||||
|
|
||||||
// Assert
|
// Assert
|
||||||
expect(flyerQueue.add).toHaveBeenCalled();
|
expect(aiService.aiService.enqueueFlyerProcessing).toHaveBeenCalled();
|
||||||
expect(vi.mocked(flyerQueue.add).mock.calls[0][1].userId).toBe('auth-user-1');
|
const callArgs = vi.mocked(aiService.aiService.enqueueFlyerProcessing).mock.calls[0];
|
||||||
|
// Check the userProfile argument (3rd argument)
|
||||||
|
expect(callArgs[2]?.user.user_id).toBe('auth-user-1');
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should pass user profile address to the job when authenticated user has an address', async () => {
|
it('should pass user profile address to the job when authenticated user has an address', async () => {
|
||||||
@@ -243,6 +251,8 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
basePath: '/api/ai',
|
basePath: '/api/ai',
|
||||||
authenticatedUser: mockUserWithAddress,
|
authenticatedUser: mockUserWithAddress,
|
||||||
});
|
});
|
||||||
|
|
||||||
|
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockResolvedValue({ id: 'job-789' } as unknown as Job);
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
await supertest(authenticatedApp)
|
await supertest(authenticatedApp)
|
||||||
@@ -251,9 +261,10 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerFile', imagePath);
|
.attach('flyerFile', imagePath);
|
||||||
|
|
||||||
// Assert
|
// Assert
|
||||||
expect(vi.mocked(flyerQueue.add).mock.calls[0][1].userProfileAddress).toBe(
|
expect(aiService.aiService.enqueueFlyerProcessing).toHaveBeenCalled();
|
||||||
'123 Pacific St, Anytown, BC, V8T 1A1, CA',
|
// The service handles address extraction from profile, so we just verify the profile was passed
|
||||||
);
|
const callArgs = vi.mocked(aiService.aiService.enqueueFlyerProcessing).mock.calls[0];
|
||||||
|
expect(callArgs[2]?.address?.address_line_1).toBe('123 Pacific St');
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should clean up the uploaded file if validation fails (e.g., missing checksum)', async () => {
|
it('should clean up the uploaded file if validation fails (e.g., missing checksum)', async () => {
|
||||||
@@ -316,9 +327,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
flyer_id: 1,
|
flyer_id: 1,
|
||||||
file_name: mockDataPayload.originalFileName,
|
file_name: mockDataPayload.originalFileName,
|
||||||
});
|
});
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined); // No duplicate
|
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(mockFlyer);
|
||||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
|
||||||
vi.mocked(mockedDb.adminRepo.logActivity).mockResolvedValue();
|
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
@@ -329,13 +338,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
// Assert
|
// Assert
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(response.body.message).toBe('Flyer processed and saved successfully.');
|
expect(response.body.message).toBe('Flyer processed and saved successfully.');
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
// Verify that the legacy endpoint correctly sets the status to 'needs_review'
|
|
||||||
expect(vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0]).toEqual(
|
|
||||||
expect.objectContaining({
|
|
||||||
status: 'needs_review',
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 400 if no flyer image is provided', async () => {
|
it('should return 400 if no flyer image is provided', async () => {
|
||||||
@@ -347,8 +350,8 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
|
|
||||||
it('should return 409 Conflict and delete the uploaded file if flyer checksum already exists', async () => {
|
it('should return 409 Conflict and delete the uploaded file if flyer checksum already exists', async () => {
|
||||||
// Arrange
|
// Arrange
|
||||||
const mockExistingFlyer = createMockFlyer({ flyer_id: 99 });
|
const duplicateError = new aiService.DuplicateFlyerError('This flyer has already been processed.', 99);
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(mockExistingFlyer); // Duplicate found
|
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockRejectedValue(duplicateError);
|
||||||
const unlinkSpy = vi.spyOn(fs.promises, 'unlink').mockResolvedValue(undefined);
|
const unlinkSpy = vi.spyOn(fs.promises, 'unlink').mockResolvedValue(undefined);
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
@@ -360,7 +363,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
// Assert
|
// Assert
|
||||||
expect(response.status).toBe(409);
|
expect(response.status).toBe(409);
|
||||||
expect(response.body.message).toBe('This flyer has already been processed.');
|
expect(response.body.message).toBe('This flyer has already been processed.');
|
||||||
expect(mockedDb.createFlyerAndItems).not.toHaveBeenCalled();
|
expect(mockedDb.createFlyerAndItems).not.toHaveBeenCalled(); // Should not be called if service throws
|
||||||
// Assert that the file was deleted
|
// Assert that the file was deleted
|
||||||
expect(unlinkSpy).toHaveBeenCalledTimes(1);
|
expect(unlinkSpy).toHaveBeenCalledTimes(1);
|
||||||
// The filename is predictable in the test environment because of the multer config in ai.routes.ts
|
// The filename is predictable in the test environment because of the multer config in ai.routes.ts
|
||||||
@@ -375,12 +378,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
extractedData: { store_name: 'Partial Store' }, // no items key
|
extractedData: { store_name: 'Partial Store' }, // no items key
|
||||||
};
|
};
|
||||||
|
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(createMockFlyer({ flyer_id: 2 }));
|
||||||
const mockFlyer = createMockFlyer({
|
|
||||||
flyer_id: 2,
|
|
||||||
file_name: partialPayload.originalFileName,
|
|
||||||
});
|
|
||||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/ai/flyers/process')
|
.post('/api/ai/flyers/process')
|
||||||
@@ -388,19 +386,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerImage', imagePath);
|
.attach('flyerImage', imagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
// Verify that the legacy endpoint correctly sets the status to 'needs_review'
|
|
||||||
expect(vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0]).toEqual(
|
|
||||||
expect.objectContaining({
|
|
||||||
status: 'needs_review',
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
// verify the items array passed to DB was an empty array
|
|
||||||
const callArgs = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0]?.[1];
|
|
||||||
expect(callArgs).toBeDefined();
|
|
||||||
expect(Array.isArray(callArgs)).toBe(true);
|
|
||||||
// use non-null assertion for the runtime-checked variable so TypeScript is satisfied
|
|
||||||
expect(callArgs!.length).toBe(0);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should fallback to a safe store name when store_name is missing', async () => {
|
it('should fallback to a safe store name when store_name is missing', async () => {
|
||||||
@@ -410,12 +396,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
extractedData: { items: [] }, // store_name missing
|
extractedData: { items: [] }, // store_name missing
|
||||||
};
|
};
|
||||||
|
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(createMockFlyer({ flyer_id: 3 }));
|
||||||
const mockFlyer = createMockFlyer({
|
|
||||||
flyer_id: 3,
|
|
||||||
file_name: payloadNoStore.originalFileName,
|
|
||||||
});
|
|
||||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/ai/flyers/process')
|
.post('/api/ai/flyers/process')
|
||||||
@@ -423,25 +404,11 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerImage', imagePath);
|
.attach('flyerImage', imagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
// Verify that the legacy endpoint correctly sets the status to 'needs_review'
|
|
||||||
expect(vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0]).toEqual(
|
|
||||||
expect.objectContaining({
|
|
||||||
status: 'needs_review',
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
// verify the flyerData.store_name passed to DB was the fallback string
|
|
||||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
|
||||||
expect(flyerDataArg.store_name).toContain('Unknown Store');
|
|
||||||
// Also verify the warning was logged
|
|
||||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
|
||||||
'extractedData.store_name missing; using fallback store name to avoid DB constraint error.',
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should handle a generic error during flyer creation', async () => {
|
it('should handle a generic error during flyer creation', async () => {
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockRejectedValueOnce(
|
||||||
vi.mocked(mockedDb.createFlyerAndItems).mockRejectedValueOnce(
|
|
||||||
new Error('DB transaction failed'),
|
new Error('DB transaction failed'),
|
||||||
);
|
);
|
||||||
|
|
||||||
@@ -464,8 +431,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
|
|
||||||
beforeEach(() => {
|
beforeEach(() => {
|
||||||
const mockFlyer = createMockFlyer({ flyer_id: 1 });
|
const mockFlyer = createMockFlyer({ flyer_id: 1 });
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(mockFlyer);
|
||||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should handle payload where "data" field is an object, not stringified JSON', async () => {
|
it('should handle payload where "data" field is an object, not stringified JSON', async () => {
|
||||||
@@ -475,7 +441,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerImage', imagePath);
|
.attach('flyerImage', imagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should handle payload where extractedData is null', async () => {
|
it('should handle payload where extractedData is null', async () => {
|
||||||
@@ -491,14 +457,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerImage', imagePath);
|
.attach('flyerImage', imagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
// Verify that extractedData was correctly defaulted to an empty object
|
|
||||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
|
||||||
expect(flyerDataArg.store_name).toContain('Unknown Store'); // Fallback should be used
|
|
||||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
|
||||||
{ bodyData: expect.any(Object) },
|
|
||||||
'Missing extractedData in /api/ai/flyers/process payload.',
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should handle payload where extractedData is a string', async () => {
|
it('should handle payload where extractedData is a string', async () => {
|
||||||
@@ -514,14 +473,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerImage', imagePath);
|
.attach('flyerImage', imagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
// Verify that extractedData was correctly defaulted to an empty object
|
|
||||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
|
||||||
expect(flyerDataArg.store_name).toContain('Unknown Store'); // Fallback should be used
|
|
||||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
|
||||||
{ bodyData: expect.any(Object) },
|
|
||||||
'Missing extractedData in /api/ai/flyers/process payload.',
|
|
||||||
);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should handle payload where extractedData is at the root of the body', async () => {
|
it('should handle payload where extractedData is at the root of the body', async () => {
|
||||||
@@ -535,9 +487,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerImage', imagePath);
|
.attach('flyerImage', imagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(201); // This test was failing with 500, the fix is in ai.routes.ts
|
expect(response.status).toBe(201); // This test was failing with 500, the fix is in ai.routes.ts
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
|
||||||
expect(flyerDataArg.store_name).toBe('Root Store');
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should default item quantity to 1 if missing', async () => {
|
it('should default item quantity to 1 if missing', async () => {
|
||||||
@@ -556,9 +506,7 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
.attach('flyerImage', imagePath);
|
.attach('flyerImage', imagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||||
const itemsArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][1];
|
|
||||||
expect(itemsArg[0].quantity).toBe(1);
|
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -567,7 +515,6 @@ describe('AI Routes (/api/ai)', () => {
|
|||||||
|
|
||||||
it('should handle malformed JSON in data field and return 400', async () => {
|
it('should handle malformed JSON in data field and return 400', async () => {
|
||||||
const malformedDataString = '{"checksum":'; // Invalid JSON
|
const malformedDataString = '{"checksum":'; // Invalid JSON
|
||||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/ai/flyers/process')
|
.post('/api/ai/flyers/process')
|
||||||
|
|||||||
@@ -1,40 +1,32 @@
|
|||||||
// src/routes/ai.routes.ts
|
// src/routes/ai.routes.ts
|
||||||
import { Router, Request, Response, NextFunction } from 'express';
|
import { Router, Request, Response, NextFunction } from 'express';
|
||||||
import path from 'path';
|
|
||||||
import fs from 'node:fs';
|
|
||||||
import { z } from 'zod';
|
import { z } from 'zod';
|
||||||
import passport from './passport.routes';
|
import passport from './passport.routes';
|
||||||
import { optionalAuth } from './passport.routes';
|
import { optionalAuth } from './passport.routes';
|
||||||
import * as db from '../services/db/index.db';
|
import { aiService, DuplicateFlyerError } from '../services/aiService.server';
|
||||||
import { createFlyerAndItems } from '../services/db/flyer.db';
|
|
||||||
import * as aiService from '../services/aiService.server'; // Correctly import server-side AI service
|
|
||||||
import {
|
import {
|
||||||
createUploadMiddleware,
|
createUploadMiddleware,
|
||||||
handleMulterError,
|
handleMulterError,
|
||||||
} from '../middleware/multer.middleware';
|
} from '../middleware/multer.middleware';
|
||||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
|
||||||
import { logger } from '../services/logger.server'; // This was a duplicate, fixed.
|
import { logger } from '../services/logger.server'; // This was a duplicate, fixed.
|
||||||
import { UserProfile, ExtractedCoreData, ExtractedFlyerItem, FlyerInsert } from '../types';
|
import { UserProfile } from '../types'; // This was a duplicate, fixed.
|
||||||
import { flyerQueue } from '../services/queueService.server';
|
|
||||||
import { validateRequest } from '../middleware/validation.middleware';
|
import { validateRequest } from '../middleware/validation.middleware';
|
||||||
import { requiredString } from '../utils/zodUtils';
|
import { requiredString } from '../utils/zodUtils';
|
||||||
|
import { cleanupUploadedFile, cleanupUploadedFiles } from '../utils/fileUtils';
|
||||||
|
import { monitoringService } from '../services/monitoringService.server';
|
||||||
|
|
||||||
const router = Router();
|
const router = Router();
|
||||||
|
|
||||||
interface FlyerProcessPayload extends Partial<ExtractedCoreData> {
|
|
||||||
checksum?: string;
|
|
||||||
originalFileName?: string;
|
|
||||||
extractedData?: Partial<ExtractedCoreData>;
|
|
||||||
data?: FlyerProcessPayload; // For nested data structures
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- Zod Schemas for AI Routes (as per ADR-003) ---
|
// --- Zod Schemas for AI Routes (as per ADR-003) ---
|
||||||
|
|
||||||
const uploadAndProcessSchema = z.object({
|
const uploadAndProcessSchema = z.object({
|
||||||
body: z.object({
|
body: z.object({
|
||||||
checksum: requiredString('File checksum is required.'),
|
// Stricter validation for SHA-256 checksum. It must be a 64-character hexadecimal string.
|
||||||
// Potential improvement: If checksum is always a specific format (e.g., SHA-256),
|
checksum: requiredString('File checksum is required.').pipe(
|
||||||
// you could add `.length(64).regex(/^[a-f0-9]+$/)` for stricter validation.
|
z.string()
|
||||||
|
.length(64, 'Checksum must be 64 characters long.')
|
||||||
|
.regex(/^[a-f0-9]+$/, 'Checksum must be a valid hexadecimal string.'),
|
||||||
|
),
|
||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -52,22 +44,6 @@ const errMsg = (e: unknown) => {
|
|||||||
return String(e || 'An unknown error occurred.');
|
return String(e || 'An unknown error occurred.');
|
||||||
};
|
};
|
||||||
|
|
||||||
const cleanupUploadedFile = async (file?: Express.Multer.File) => {
|
|
||||||
if (!file) return;
|
|
||||||
try {
|
|
||||||
await fs.promises.unlink(file.path);
|
|
||||||
} catch (err) {
|
|
||||||
// Ignore cleanup errors (e.g. file already deleted)
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const cleanupUploadedFiles = async (files?: Express.Multer.File[]) => {
|
|
||||||
if (!files || !Array.isArray(files)) return;
|
|
||||||
// Use Promise.all to run cleanups in parallel for efficiency,
|
|
||||||
// as cleanupUploadedFile is designed to not throw errors.
|
|
||||||
await Promise.all(files.map((file) => cleanupUploadedFile(file)));
|
|
||||||
};
|
|
||||||
|
|
||||||
const cropAreaObjectSchema = z.object({
|
const cropAreaObjectSchema = z.object({
|
||||||
x: z.number(),
|
x: z.number(),
|
||||||
y: z.number(),
|
y: z.number(),
|
||||||
@@ -103,13 +79,20 @@ const rescanAreaSchema = z.object({
|
|||||||
|
|
||||||
const flyerItemForAnalysisSchema = z
|
const flyerItemForAnalysisSchema = z
|
||||||
.object({
|
.object({
|
||||||
item: z.string().nullish(),
|
// Sanitize item and name by trimming whitespace.
|
||||||
name: z.string().nullish(),
|
// The transform ensures that null/undefined values are preserved
|
||||||
|
// while trimming any actual string values.
|
||||||
|
item: z.string().nullish().transform(val => (val ? val.trim() : val)),
|
||||||
|
name: z.string().nullish().transform(val => (val ? val.trim() : val)),
|
||||||
})
|
})
|
||||||
|
// Using .passthrough() allows extra properties on the item object.
|
||||||
|
// If the intent is to strictly enforce only 'item' and 'name' (and other known properties),
|
||||||
|
// consider using .strict() instead for tighter security and data integrity.
|
||||||
.passthrough()
|
.passthrough()
|
||||||
.refine(
|
.refine(
|
||||||
(data) =>
|
(data) =>
|
||||||
(data.item && data.item.trim().length > 0) || (data.name && data.name.trim().length > 0),
|
// After the transform, the values are already trimmed.
|
||||||
|
(data.item && data.item.length > 0) || (data.name && data.name.length > 0),
|
||||||
{
|
{
|
||||||
message: "Item identifier is required (either 'item' or 'name').",
|
message: "Item identifier is required (either 'item' or 'name').",
|
||||||
},
|
},
|
||||||
@@ -129,6 +112,8 @@ const comparePricesSchema = z.object({
|
|||||||
|
|
||||||
const planTripSchema = z.object({
|
const planTripSchema = z.object({
|
||||||
body: z.object({
|
body: z.object({
|
||||||
|
// Consider if this array should be non-empty. If a trip plan requires at least one item,
|
||||||
|
// you could add `.nonempty('At least one item is required to plan a trip.')`
|
||||||
items: z.array(flyerItemForAnalysisSchema),
|
items: z.array(flyerItemForAnalysisSchema),
|
||||||
store: z.object({ name: requiredString('Store name is required.') }),
|
store: z.object({ name: requiredString('Store name is required.') }),
|
||||||
userLocation: z.object({
|
userLocation: z.object({
|
||||||
@@ -187,57 +172,24 @@ router.post(
|
|||||||
async (req, res, next: NextFunction) => {
|
async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
// Manually validate the request body. This will throw if validation fails.
|
// Manually validate the request body. This will throw if validation fails.
|
||||||
uploadAndProcessSchema.parse({ body: req.body });
|
const { body } = uploadAndProcessSchema.parse({ body: req.body });
|
||||||
|
|
||||||
if (!req.file) {
|
if (!req.file) {
|
||||||
return res.status(400).json({ message: 'A flyer file (PDF or image) is required.' });
|
return res.status(400).json({ message: 'A flyer file (PDF or image) is required.' });
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.debug(
|
logger.debug(
|
||||||
{ filename: req.file.originalname, size: req.file.size, checksum: req.body?.checksum },
|
{ filename: req.file.originalname, size: req.file.size, checksum: body.checksum },
|
||||||
'Handling /upload-and-process',
|
'Handling /upload-and-process',
|
||||||
);
|
);
|
||||||
|
|
||||||
const { checksum } = req.body;
|
|
||||||
// Check for duplicate flyer using checksum before even creating a job
|
|
||||||
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, req.log);
|
|
||||||
if (existingFlyer) {
|
|
||||||
logger.warn(`Duplicate flyer upload attempt blocked for checksum: ${checksum}`);
|
|
||||||
// Use 409 Conflict for duplicates
|
|
||||||
return res.status(409).json({
|
|
||||||
message: 'This flyer has already been processed.',
|
|
||||||
flyerId: existingFlyer.flyer_id,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
const userProfile = req.user as UserProfile | undefined;
|
const userProfile = req.user as UserProfile | undefined;
|
||||||
// Construct a user address string from their profile if they are logged in.
|
const job = await aiService.enqueueFlyerProcessing(
|
||||||
let userProfileAddress: string | undefined = undefined;
|
req.file,
|
||||||
if (userProfile?.address) {
|
body.checksum,
|
||||||
userProfileAddress = [
|
userProfile,
|
||||||
userProfile.address.address_line_1,
|
req.ip ?? 'unknown',
|
||||||
userProfile.address.address_line_2,
|
req.log,
|
||||||
userProfile.address.city,
|
|
||||||
userProfile.address.province_state,
|
|
||||||
userProfile.address.postal_code,
|
|
||||||
userProfile.address.country,
|
|
||||||
]
|
|
||||||
.filter(Boolean)
|
|
||||||
.join(', ');
|
|
||||||
}
|
|
||||||
|
|
||||||
// Add job to the queue
|
|
||||||
const job = await flyerQueue.add('process-flyer', {
|
|
||||||
filePath: req.file.path,
|
|
||||||
originalFileName: req.file.originalname,
|
|
||||||
checksum: checksum,
|
|
||||||
userId: userProfile?.user.user_id,
|
|
||||||
submitterIp: req.ip, // Capture the submitter's IP address
|
|
||||||
userProfileAddress: userProfileAddress, // Pass the user's profile address
|
|
||||||
});
|
|
||||||
|
|
||||||
logger.info(
|
|
||||||
`Enqueued flyer for processing. File: ${req.file.originalname}, Job ID: ${job.id}`,
|
|
||||||
);
|
);
|
||||||
|
|
||||||
// Respond immediately to the client with 202 Accepted
|
// Respond immediately to the client with 202 Accepted
|
||||||
@@ -246,9 +198,11 @@ router.post(
|
|||||||
jobId: job.id,
|
jobId: job.id,
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// If any error occurs (including validation), ensure the uploaded file is cleaned up.
|
|
||||||
await cleanupUploadedFile(req.file);
|
await cleanupUploadedFile(req.file);
|
||||||
// Pass the error to the global error handler.
|
if (error instanceof DuplicateFlyerError) {
|
||||||
|
logger.warn(`Duplicate flyer upload attempt blocked for checksum: ${req.body?.checksum}`);
|
||||||
|
return res.status(409).json({ message: error.message, flyerId: error.flyerId });
|
||||||
|
}
|
||||||
next(error);
|
next(error);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@@ -265,18 +219,11 @@ router.get(
|
|||||||
const {
|
const {
|
||||||
params: { jobId },
|
params: { jobId },
|
||||||
} = req as unknown as JobIdRequest;
|
} = req as unknown as JobIdRequest;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const job = await flyerQueue.getJob(jobId);
|
const jobStatus = await monitoringService.getFlyerJobStatus(jobId); // This was a duplicate, fixed.
|
||||||
if (!job) {
|
logger.debug(`[API /ai/jobs] Status check for job ${jobId}: ${jobStatus.state}`);
|
||||||
// Adhere to ADR-001 by throwing a specific error to be handled centrally.
|
res.json(jobStatus);
|
||||||
return res.status(404).json({ message: 'Job not found.' });
|
|
||||||
}
|
|
||||||
const state = await job.getState();
|
|
||||||
const progress = job.progress;
|
|
||||||
const returnValue = job.returnvalue;
|
|
||||||
const failedReason = job.failedReason;
|
|
||||||
logger.debug(`[API /ai/jobs] Status check for job ${jobId}: ${state}`);
|
|
||||||
res.json({ id: job.id, state, progress, returnValue, failedReason });
|
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
next(error);
|
next(error);
|
||||||
}
|
}
|
||||||
@@ -298,186 +245,22 @@ router.post(
|
|||||||
return res.status(400).json({ message: 'Flyer image file is required.' });
|
return res.status(400).json({ message: 'Flyer image file is required.' });
|
||||||
}
|
}
|
||||||
|
|
||||||
// Diagnostic & tolerant parsing for flyers/process
|
|
||||||
logger.debug(
|
|
||||||
{ keys: Object.keys(req.body || {}) },
|
|
||||||
'[API /ai/flyers/process] Processing legacy upload',
|
|
||||||
);
|
|
||||||
logger.debug({ filePresent: !!req.file }, '[API /ai/flyers/process] file present:');
|
|
||||||
|
|
||||||
// Try several ways to obtain the payload so we are tolerant to client variations.
|
|
||||||
let parsed: FlyerProcessPayload = {};
|
|
||||||
let extractedData: Partial<ExtractedCoreData> | null | undefined = {};
|
|
||||||
try {
|
|
||||||
// If the client sent a top-level `data` field (stringified JSON), parse it.
|
|
||||||
if (req.body && (req.body.data || req.body.extractedData)) {
|
|
||||||
const raw = req.body.data ?? req.body.extractedData;
|
|
||||||
logger.debug(
|
|
||||||
{ type: typeof raw, length: raw?.length ?? 0 },
|
|
||||||
'[API /ai/flyers/process] raw extractedData',
|
|
||||||
);
|
|
||||||
try {
|
|
||||||
parsed = typeof raw === 'string' ? JSON.parse(raw) : raw;
|
|
||||||
} catch (err) {
|
|
||||||
logger.warn(
|
|
||||||
{ error: errMsg(err) },
|
|
||||||
'[API /ai/flyers/process] Failed to JSON.parse raw extractedData; falling back to direct assign',
|
|
||||||
);
|
|
||||||
parsed = (
|
|
||||||
typeof raw === 'string' ? JSON.parse(String(raw).slice(0, 2000)) : raw
|
|
||||||
) as FlyerProcessPayload;
|
|
||||||
}
|
|
||||||
// If parsed itself contains an `extractedData` field, use that, otherwise assume parsed is the extractedData
|
|
||||||
extractedData = 'extractedData' in parsed ? parsed.extractedData : (parsed as Partial<ExtractedCoreData>);
|
|
||||||
} else {
|
|
||||||
// No explicit `data` field found. Attempt to interpret req.body as an object (Express may have parsed multipart fields differently).
|
|
||||||
try {
|
|
||||||
parsed = typeof req.body === 'string' ? JSON.parse(req.body) : req.body;
|
|
||||||
} catch (err) {
|
|
||||||
logger.warn(
|
|
||||||
{ error: errMsg(err) },
|
|
||||||
'[API /ai/flyers/process] Failed to JSON.parse req.body; using empty object',
|
|
||||||
);
|
|
||||||
parsed = (req.body as FlyerProcessPayload) || {};
|
|
||||||
}
|
|
||||||
// extractedData might be nested under `data` or `extractedData`, or the body itself may be the extracted data.
|
|
||||||
if (parsed.data) {
|
|
||||||
try {
|
|
||||||
const inner = typeof parsed.data === 'string' ? JSON.parse(parsed.data) : parsed.data;
|
|
||||||
extractedData = inner.extractedData ?? inner;
|
|
||||||
} catch (err) {
|
|
||||||
logger.warn(
|
|
||||||
{ error: errMsg(err) },
|
|
||||||
'[API /ai/flyers/process] Failed to parse parsed.data; falling back',
|
|
||||||
);
|
|
||||||
extractedData = parsed.data as unknown as Partial<ExtractedCoreData>;
|
|
||||||
}
|
|
||||||
} else if (parsed.extractedData) {
|
|
||||||
extractedData = parsed.extractedData;
|
|
||||||
} else {
|
|
||||||
// Assume the body itself is the extracted data if it looks like it (has items or store_name keys)
|
|
||||||
if ('items' in parsed || 'store_name' in parsed || 'valid_from' in parsed) {
|
|
||||||
extractedData = parsed as Partial<ExtractedCoreData>;
|
|
||||||
} else {
|
|
||||||
extractedData = {};
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
logger.error(
|
|
||||||
{ error: err },
|
|
||||||
'[API /ai/flyers/process] Unexpected error while parsing request body',
|
|
||||||
);
|
|
||||||
parsed = {};
|
|
||||||
extractedData = {};
|
|
||||||
}
|
|
||||||
|
|
||||||
// Pull common metadata fields (checksum, originalFileName) from whichever shape we parsed.
|
|
||||||
const checksum = parsed.checksum ?? parsed?.data?.checksum ?? '';
|
|
||||||
|
|
||||||
if (!checksum) {
|
|
||||||
await cleanupUploadedFile(req.file);
|
|
||||||
return res.status(400).json({ message: 'Checksum is required.' });
|
|
||||||
}
|
|
||||||
|
|
||||||
const originalFileName =
|
|
||||||
parsed.originalFileName ?? parsed?.data?.originalFileName ?? req.file.originalname;
|
|
||||||
const userProfile = req.user as UserProfile | undefined;
|
const userProfile = req.user as UserProfile | undefined;
|
||||||
|
|
||||||
// Validate extractedData to avoid database errors (e.g., null store_name)
|
const newFlyer = await aiService.processLegacyFlyerUpload(
|
||||||
if (!extractedData || typeof extractedData !== 'object') {
|
req.file,
|
||||||
logger.warn(
|
req.body,
|
||||||
{ bodyData: parsed },
|
userProfile,
|
||||||
'Missing extractedData in /api/ai/flyers/process payload.',
|
|
||||||
);
|
|
||||||
// Don't fail hard here; proceed with empty items and fallback store name so the upload can be saved for manual review.
|
|
||||||
extractedData = {};
|
|
||||||
}
|
|
||||||
|
|
||||||
// Transform the extracted items into the format required for database insertion.
|
|
||||||
// This adds default values for fields like `view_count` and `click_count`
|
|
||||||
// and makes this legacy endpoint consistent with the newer FlyerDataTransformer service.
|
|
||||||
const rawItems = extractedData.items ?? [];
|
|
||||||
const itemsArray = Array.isArray(rawItems)
|
|
||||||
? rawItems
|
|
||||||
: typeof rawItems === 'string'
|
|
||||||
? JSON.parse(rawItems)
|
|
||||||
: [];
|
|
||||||
const itemsForDb = itemsArray.map((item: Partial<ExtractedFlyerItem>) => ({
|
|
||||||
...item,
|
|
||||||
master_item_id: item.master_item_id === null ? undefined : item.master_item_id,
|
|
||||||
quantity: item.quantity ?? 1, // Default to 1 to satisfy DB constraint
|
|
||||||
view_count: 0,
|
|
||||||
click_count: 0,
|
|
||||||
updated_at: new Date().toISOString(),
|
|
||||||
}));
|
|
||||||
|
|
||||||
// Ensure we have a valid store name; the DB requires a non-null store name.
|
|
||||||
const storeName =
|
|
||||||
extractedData.store_name && String(extractedData.store_name).trim().length > 0
|
|
||||||
? String(extractedData.store_name)
|
|
||||||
: 'Unknown Store (auto)';
|
|
||||||
if (storeName.startsWith('Unknown')) {
|
|
||||||
logger.warn(
|
|
||||||
'extractedData.store_name missing; using fallback store name to avoid DB constraint error.',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
// 1. Check for duplicate flyer using checksum
|
|
||||||
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, req.log);
|
|
||||||
if (existingFlyer) {
|
|
||||||
logger.warn(`Duplicate flyer upload attempt blocked for checksum: ${checksum}`);
|
|
||||||
await cleanupUploadedFile(req.file);
|
|
||||||
return res.status(409).json({ message: 'This flyer has already been processed.' });
|
|
||||||
}
|
|
||||||
|
|
||||||
// Generate a 64x64 icon from the uploaded flyer image.
|
|
||||||
const iconsDir = path.join(path.dirname(req.file.path), 'icons');
|
|
||||||
const iconFileName = await generateFlyerIcon(req.file.path, iconsDir, req.log);
|
|
||||||
const iconUrl = `/flyer-images/icons/${iconFileName}`;
|
|
||||||
|
|
||||||
// 2. Prepare flyer data for insertion
|
|
||||||
const flyerData: FlyerInsert = {
|
|
||||||
file_name: originalFileName,
|
|
||||||
image_url: `/flyer-images/${req.file.filename}`, // Store the full URL path
|
|
||||||
icon_url: iconUrl,
|
|
||||||
checksum: checksum,
|
|
||||||
// Use normalized store name (fallback applied above).
|
|
||||||
store_name: storeName,
|
|
||||||
valid_from: extractedData.valid_from ?? null,
|
|
||||||
valid_to: extractedData.valid_to ?? null,
|
|
||||||
store_address: extractedData.store_address ?? null,
|
|
||||||
item_count: 0, // Set default to 0; the trigger will update it.
|
|
||||||
// Set a safe default status for this legacy endpoint. The new flow uses the transformer to determine this.
|
|
||||||
status: 'needs_review',
|
|
||||||
uploaded_by: userProfile?.user.user_id, // Associate with user if logged in
|
|
||||||
};
|
|
||||||
|
|
||||||
// 3. Create flyer and its items in a transaction
|
|
||||||
const { flyer: newFlyer, items: newItems } = await createFlyerAndItems(
|
|
||||||
flyerData,
|
|
||||||
itemsForDb,
|
|
||||||
req.log,
|
|
||||||
);
|
|
||||||
|
|
||||||
logger.info(
|
|
||||||
`Successfully processed and saved new flyer: ${newFlyer.file_name} (ID: ${newFlyer.flyer_id}) with ${newItems.length} items.`,
|
|
||||||
);
|
|
||||||
|
|
||||||
// Log this significant event
|
|
||||||
await db.adminRepo.logActivity(
|
|
||||||
{
|
|
||||||
userId: userProfile?.user.user_id,
|
|
||||||
action: 'flyer_processed',
|
|
||||||
displayText: `Processed a new flyer for ${flyerData.store_name}.`,
|
|
||||||
details: { flyerId: newFlyer.flyer_id, storeName: flyerData.store_name },
|
|
||||||
},
|
|
||||||
req.log,
|
req.log,
|
||||||
);
|
);
|
||||||
|
|
||||||
res.status(201).json({ message: 'Flyer processed and saved successfully.', flyer: newFlyer });
|
res.status(201).json({ message: 'Flyer processed and saved successfully.', flyer: newFlyer });
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
await cleanupUploadedFile(req.file);
|
await cleanupUploadedFile(req.file);
|
||||||
|
if (error instanceof DuplicateFlyerError) {
|
||||||
|
logger.warn(`Duplicate flyer upload attempt blocked.`);
|
||||||
|
return res.status(409).json({ message: error.message, flyerId: error.flyerId });
|
||||||
|
}
|
||||||
next(error);
|
next(error);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@@ -616,7 +399,7 @@ router.post(
|
|||||||
try {
|
try {
|
||||||
const { items, store, userLocation } = req.body;
|
const { items, store, userLocation } = req.body;
|
||||||
logger.debug({ itemCount: items.length, storeName: store.name }, 'Trip planning requested.');
|
logger.debug({ itemCount: items.length, storeName: store.name }, 'Trip planning requested.');
|
||||||
const result = await aiService.aiService.planTripWithMaps(items, store, userLocation);
|
const result = await aiService.planTripWithMaps(items, store, userLocation);
|
||||||
res.status(200).json(result);
|
res.status(200).json(result);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error: errMsg(error) }, 'Error in /api/ai/plan-trip endpoint:');
|
logger.error({ error: errMsg(error) }, 'Error in /api/ai/plan-trip endpoint:');
|
||||||
@@ -676,7 +459,7 @@ router.post(
|
|||||||
'Rescan area requested',
|
'Rescan area requested',
|
||||||
);
|
);
|
||||||
|
|
||||||
const result = await aiService.aiService.extractTextFromImageArea(
|
const result = await aiService.extractTextFromImageArea(
|
||||||
path,
|
path,
|
||||||
mimetype,
|
mimetype,
|
||||||
cropArea,
|
cropArea,
|
||||||
|
|||||||
@@ -2,13 +2,8 @@
|
|||||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||||
import supertest from 'supertest';
|
import supertest from 'supertest';
|
||||||
import { Request, Response, NextFunction } from 'express';
|
import { Request, Response, NextFunction } from 'express';
|
||||||
import cookieParser from 'cookie-parser';
|
import cookieParser from 'cookie-parser'; // This was a duplicate, fixed.
|
||||||
import * as bcrypt from 'bcrypt';
|
import { createMockUserProfile } from '../tests/utils/mockFactories';
|
||||||
import jwt from 'jsonwebtoken';
|
|
||||||
import {
|
|
||||||
createMockUserProfile,
|
|
||||||
createMockUserWithPasswordHash,
|
|
||||||
} from '../tests/utils/mockFactories';
|
|
||||||
|
|
||||||
// --- FIX: Hoist passport mocks to be available for vi.mock ---
|
// --- FIX: Hoist passport mocks to be available for vi.mock ---
|
||||||
const passportMocks = vi.hoisted(() => {
|
const passportMocks = vi.hoisted(() => {
|
||||||
@@ -69,45 +64,20 @@ vi.mock('./passport.routes', () => ({
|
|||||||
optionalAuth: vi.fn((req: Request, res: Response, next: NextFunction) => next()),
|
optionalAuth: vi.fn((req: Request, res: Response, next: NextFunction) => next()),
|
||||||
}));
|
}));
|
||||||
|
|
||||||
// Mock the DB connection pool to control transactional behavior
|
// Mock the authService, which is now the primary dependency of the routes.
|
||||||
const { mockPool } = vi.hoisted(() => {
|
const { mockedAuthService } = vi.hoisted(() => {
|
||||||
const client = {
|
|
||||||
query: vi.fn(),
|
|
||||||
release: vi.fn(),
|
|
||||||
};
|
|
||||||
return {
|
return {
|
||||||
mockPool: {
|
mockedAuthService: {
|
||||||
connect: vi.fn(() => Promise.resolve(client)),
|
registerAndLoginUser: vi.fn(),
|
||||||
|
handleSuccessfulLogin: vi.fn(),
|
||||||
|
resetPassword: vi.fn(),
|
||||||
|
updatePassword: vi.fn(),
|
||||||
|
refreshAccessToken: vi.fn(),
|
||||||
|
logout: vi.fn(),
|
||||||
},
|
},
|
||||||
mockClient: client,
|
|
||||||
};
|
};
|
||||||
});
|
});
|
||||||
// Mock the Service Layer directly.
|
vi.mock('../services/authService', () => ({ authService: mockedAuthService }));
|
||||||
// We use async import inside the factory to properly hoist the UniqueConstraintError class usage.
|
|
||||||
vi.mock('../services/db/index.db', async () => {
|
|
||||||
const { UniqueConstraintError } = await import('../services/db/errors.db');
|
|
||||||
return {
|
|
||||||
userRepo: {
|
|
||||||
findUserByEmail: vi.fn(),
|
|
||||||
createUser: vi.fn(),
|
|
||||||
saveRefreshToken: vi.fn(),
|
|
||||||
createPasswordResetToken: vi.fn(),
|
|
||||||
getValidResetTokens: vi.fn(),
|
|
||||||
updateUserPassword: vi.fn(),
|
|
||||||
deleteResetToken: vi.fn(),
|
|
||||||
findUserByRefreshToken: vi.fn(),
|
|
||||||
deleteRefreshToken: vi.fn(),
|
|
||||||
},
|
|
||||||
adminRepo: {
|
|
||||||
logActivity: vi.fn(),
|
|
||||||
},
|
|
||||||
UniqueConstraintError: UniqueConstraintError,
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
vi.mock('../services/db/connection.db', () => ({
|
|
||||||
getPool: () => mockPool,
|
|
||||||
}));
|
|
||||||
|
|
||||||
// Mock the logger
|
// Mock the logger
|
||||||
vi.mock('../services/logger.server', async () => ({
|
vi.mock('../services/logger.server', async () => ({
|
||||||
@@ -120,15 +90,8 @@ vi.mock('../services/emailService.server', () => ({
|
|||||||
sendPasswordResetEmail: vi.fn(),
|
sendPasswordResetEmail: vi.fn(),
|
||||||
}));
|
}));
|
||||||
|
|
||||||
// Mock bcrypt
|
|
||||||
vi.mock('bcrypt', async (importOriginal) => {
|
|
||||||
const actual = await importOriginal<typeof bcrypt>();
|
|
||||||
return { ...actual, compare: vi.fn() };
|
|
||||||
});
|
|
||||||
|
|
||||||
// Import the router AFTER mocks are established
|
// Import the router AFTER mocks are established
|
||||||
import authRouter from './auth.routes';
|
import authRouter from './auth.routes';
|
||||||
import * as db from '../services/db/index.db'; // This was a duplicate, fixed.
|
|
||||||
|
|
||||||
import { UniqueConstraintError } from '../services/db/errors.db'; // Import actual class for instanceof checks
|
import { UniqueConstraintError } from '../services/db/errors.db'; // Import actual class for instanceof checks
|
||||||
|
|
||||||
@@ -176,13 +139,11 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
user: { user_id: 'new-user-id', email: newUserEmail },
|
user: { user_id: 'new-user-id', email: newUserEmail },
|
||||||
full_name: 'Test User',
|
full_name: 'Test User',
|
||||||
});
|
});
|
||||||
|
mockedAuthService.registerAndLoginUser.mockResolvedValue({
|
||||||
// FIX: Mock the method on the imported singleton instance `userRepo` directly,
|
newUserProfile: mockNewUser,
|
||||||
// as this is what the route handler uses. Spying on the prototype does not
|
accessToken: 'new-access-token',
|
||||||
// affect this already-created instance.
|
refreshToken: 'new-refresh-token',
|
||||||
vi.mocked(db.userRepo.createUser).mockResolvedValue(mockNewUser);
|
});
|
||||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
|
||||||
vi.mocked(db.adminRepo.logActivity).mockResolvedValue(undefined);
|
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app).post('/api/auth/register').send({
|
const response = await supertest(app).post('/api/auth/register').send({
|
||||||
@@ -190,22 +151,29 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
password: strongPassword,
|
password: strongPassword,
|
||||||
full_name: 'Test User',
|
full_name: 'Test User',
|
||||||
});
|
});
|
||||||
|
|
||||||
// Assert
|
// Assert
|
||||||
expect(response.status).toBe(201);
|
expect(response.status).toBe(201);
|
||||||
expect(response.body.message).toBe('User registered successfully!');
|
expect(response.body.message).toBe('User registered successfully!');
|
||||||
expect(response.body.userprofile.user.email).toBe(newUserEmail);
|
expect(response.body.userprofile.user.email).toBe(newUserEmail);
|
||||||
expect(response.body.token).toBeTypeOf('string'); // This was a duplicate, fixed.
|
expect(response.body.token).toBeTypeOf('string'); // This was a duplicate, fixed.
|
||||||
expect(db.userRepo.createUser).toHaveBeenCalled();
|
expect(mockedAuthService.registerAndLoginUser).toHaveBeenCalledWith(
|
||||||
|
newUserEmail,
|
||||||
|
strongPassword,
|
||||||
|
'Test User',
|
||||||
|
undefined, // avatar_url
|
||||||
|
mockLogger,
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should set a refresh token cookie on successful registration', async () => {
|
it('should set a refresh token cookie on successful registration', async () => {
|
||||||
const mockNewUser = createMockUserProfile({
|
const mockNewUser = createMockUserProfile({
|
||||||
user: { user_id: 'new-user-id', email: 'cookie@test.com' },
|
user: { user_id: 'new-user-id', email: 'cookie@test.com' },
|
||||||
});
|
});
|
||||||
vi.mocked(db.userRepo.createUser).mockResolvedValue(mockNewUser);
|
mockedAuthService.registerAndLoginUser.mockResolvedValue({
|
||||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
newUserProfile: mockNewUser,
|
||||||
vi.mocked(db.adminRepo.logActivity).mockResolvedValue(undefined);
|
accessToken: 'new-access-token',
|
||||||
|
refreshToken: 'new-refresh-token',
|
||||||
|
});
|
||||||
|
|
||||||
const response = await supertest(app).post('/api/auth/register').send({
|
const response = await supertest(app).post('/api/auth/register').send({
|
||||||
email: 'cookie@test.com',
|
email: 'cookie@test.com',
|
||||||
@@ -235,15 +203,14 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
expect(errorMessages).toMatch(/Password is too weak/i);
|
expect(errorMessages).toMatch(/Password is too weak/i);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should reject registration if the email already exists', async () => {
|
it('should reject registration if the auth service throws UniqueConstraintError', async () => {
|
||||||
// Create an error object that includes the 'code' property for simulating a PG unique violation.
|
// Create an error object that includes the 'code' property for simulating a PG unique violation.
|
||||||
// This is more type-safe than casting to 'any'.
|
// This is more type-safe than casting to 'any'.
|
||||||
const dbError = new UniqueConstraintError(
|
const dbError = new UniqueConstraintError(
|
||||||
'User with that email already exists.',
|
'User with that email already exists.',
|
||||||
) as UniqueConstraintError & { code: string };
|
) as UniqueConstraintError & { code: string };
|
||||||
dbError.code = '23505';
|
dbError.code = '23505';
|
||||||
|
mockedAuthService.registerAndLoginUser.mockRejectedValue(dbError);
|
||||||
vi.mocked(db.userRepo.createUser).mockRejectedValue(dbError);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/register')
|
.post('/api/auth/register')
|
||||||
@@ -251,12 +218,11 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
|
|
||||||
expect(response.status).toBe(409); // 409 Conflict
|
expect(response.status).toBe(409); // 409 Conflict
|
||||||
expect(response.body.message).toBe('User with that email already exists.');
|
expect(response.body.message).toBe('User with that email already exists.');
|
||||||
expect(db.userRepo.createUser).toHaveBeenCalled();
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 if a generic database error occurs during registration', async () => {
|
it('should return 500 if a generic database error occurs during registration', async () => {
|
||||||
const dbError = new Error('DB connection lost');
|
const dbError = new Error('DB connection lost');
|
||||||
vi.mocked(db.userRepo.createUser).mockRejectedValue(dbError);
|
mockedAuthService.registerAndLoginUser.mockRejectedValue(dbError);
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/register')
|
.post('/api/auth/register')
|
||||||
@@ -289,7 +255,10 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
it('should successfully log in a user and return a token and cookie', async () => {
|
it('should successfully log in a user and return a token and cookie', async () => {
|
||||||
// Arrange:
|
// Arrange:
|
||||||
const loginCredentials = { email: 'test@test.com', password: 'password123' };
|
const loginCredentials = { email: 'test@test.com', password: 'password123' };
|
||||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
mockedAuthService.handleSuccessfulLogin.mockResolvedValue({
|
||||||
|
accessToken: 'new-access-token',
|
||||||
|
refreshToken: 'new-refresh-token',
|
||||||
|
});
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
||||||
@@ -309,25 +278,6 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
expect(response.headers['set-cookie']).toBeDefined();
|
expect(response.headers['set-cookie']).toBeDefined();
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should contain the correct payload in the JWT token', async () => {
|
|
||||||
// Arrange
|
|
||||||
const loginCredentials = { email: 'payload.test@test.com', password: 'password123' };
|
|
||||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
|
||||||
|
|
||||||
// Act
|
|
||||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
|
||||||
|
|
||||||
// Assert
|
|
||||||
expect(response.status).toBe(200);
|
|
||||||
const token = response.body.token;
|
|
||||||
expect(token).toBeTypeOf('string');
|
|
||||||
|
|
||||||
const decodedPayload = jwt.decode(token) as { user_id: string; email: string; role: string };
|
|
||||||
expect(decodedPayload.user_id).toBe('user-123');
|
|
||||||
expect(decodedPayload.email).toBe(loginCredentials.email);
|
|
||||||
expect(decodedPayload.role).toBe('user'); // Default role from mock factory
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should reject login for incorrect credentials', async () => {
|
it('should reject login for incorrect credentials', async () => {
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/login')
|
.post('/api/auth/login')
|
||||||
@@ -359,7 +309,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
it('should return 500 if saving the refresh token fails', async () => {
|
it('should return 500 if saving the refresh token fails', async () => {
|
||||||
// Arrange:
|
// Arrange:
|
||||||
const loginCredentials = { email: 'test@test.com', password: 'password123' };
|
const loginCredentials = { email: 'test@test.com', password: 'password123' };
|
||||||
vi.mocked(db.userRepo.saveRefreshToken).mockRejectedValue(new Error('DB write failed'));
|
mockedAuthService.handleSuccessfulLogin.mockRejectedValue(new Error('DB write failed'));
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
||||||
@@ -401,7 +351,10 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
password: 'password123',
|
password: 'password123',
|
||||||
rememberMe: true,
|
rememberMe: true,
|
||||||
};
|
};
|
||||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
mockedAuthService.handleSuccessfulLogin.mockResolvedValue({
|
||||||
|
accessToken: 'remember-access-token',
|
||||||
|
refreshToken: 'remember-refresh-token',
|
||||||
|
});
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
||||||
@@ -416,10 +369,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
describe('POST /forgot-password', () => {
|
describe('POST /forgot-password', () => {
|
||||||
it('should send a reset link if the user exists', async () => {
|
it('should send a reset link if the user exists', async () => {
|
||||||
// Arrange
|
// Arrange
|
||||||
vi.mocked(db.userRepo.findUserByEmail).mockResolvedValue(
|
mockedAuthService.resetPassword.mockResolvedValue('mock-reset-token');
|
||||||
createMockUserWithPasswordHash({ user_id: 'user-123', email: 'test@test.com' }),
|
|
||||||
);
|
|
||||||
vi.mocked(db.userRepo.createPasswordResetToken).mockResolvedValue(undefined);
|
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
@@ -433,7 +383,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should return a generic success message even if the user does not exist', async () => {
|
it('should return a generic success message even if the user does not exist', async () => {
|
||||||
vi.mocked(db.userRepo.findUserByEmail).mockResolvedValue(undefined);
|
mockedAuthService.resetPassword.mockResolvedValue(undefined);
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/forgot-password')
|
.post('/api/auth/forgot-password')
|
||||||
@@ -444,7 +394,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 if the database call fails', async () => {
|
it('should return 500 if the database call fails', async () => {
|
||||||
vi.mocked(db.userRepo.findUserByEmail).mockRejectedValue(new Error('DB connection failed'));
|
mockedAuthService.resetPassword.mockRejectedValue(new Error('DB connection failed'));
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/forgot-password')
|
.post('/api/auth/forgot-password')
|
||||||
.send({ email: 'any@test.com' });
|
.send({ email: 'any@test.com' });
|
||||||
@@ -452,25 +402,6 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
expect(response.status).toBe(500);
|
expect(response.status).toBe(500);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should still return 200 OK if the email service fails', async () => {
|
|
||||||
// Arrange
|
|
||||||
vi.mocked(db.userRepo.findUserByEmail).mockResolvedValue(
|
|
||||||
createMockUserWithPasswordHash({ user_id: 'user-123', email: 'test@test.com' }),
|
|
||||||
);
|
|
||||||
vi.mocked(db.userRepo.createPasswordResetToken).mockResolvedValue(undefined);
|
|
||||||
// Mock the email service to fail
|
|
||||||
const { sendPasswordResetEmail } = await import('../services/emailService.server');
|
|
||||||
vi.mocked(sendPasswordResetEmail).mockRejectedValue(new Error('SMTP server down'));
|
|
||||||
|
|
||||||
// Act
|
|
||||||
const response = await supertest(app)
|
|
||||||
.post('/api/auth/forgot-password')
|
|
||||||
.send({ email: 'test@test.com' });
|
|
||||||
|
|
||||||
// Assert: The route should not fail even if the email does.
|
|
||||||
expect(response.status).toBe(200);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should return 400 for an invalid email format', async () => {
|
it('should return 400 for an invalid email format', async () => {
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/forgot-password')
|
.post('/api/auth/forgot-password')
|
||||||
@@ -483,16 +414,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
|
|
||||||
describe('POST /reset-password', () => {
|
describe('POST /reset-password', () => {
|
||||||
it('should reset the password with a valid token and strong password', async () => {
|
it('should reset the password with a valid token and strong password', async () => {
|
||||||
const tokenRecord = {
|
mockedAuthService.updatePassword.mockResolvedValue(true);
|
||||||
user_id: 'user-123',
|
|
||||||
token_hash: 'hashed-token',
|
|
||||||
expires_at: new Date(Date.now() + 3600000),
|
|
||||||
};
|
|
||||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([tokenRecord]); // This was a duplicate, fixed.
|
|
||||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never); // Token matches
|
|
||||||
vi.mocked(db.userRepo.updateUserPassword).mockResolvedValue(undefined);
|
|
||||||
vi.mocked(db.userRepo.deleteResetToken).mockResolvedValue(undefined);
|
|
||||||
vi.mocked(db.adminRepo.logActivity).mockResolvedValue(undefined);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/reset-password')
|
.post('/api/auth/reset-password')
|
||||||
@@ -503,7 +425,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should reject with an invalid or expired token', async () => {
|
it('should reject with an invalid or expired token', async () => {
|
||||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([]); // No valid tokens found
|
mockedAuthService.updatePassword.mockResolvedValue(null);
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/reset-password')
|
.post('/api/auth/reset-password')
|
||||||
@@ -513,31 +435,8 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
expect(response.body.message).toBe('Invalid or expired password reset token.');
|
expect(response.body.message).toBe('Invalid or expired password reset token.');
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should reject if token does not match any valid tokens in DB', async () => {
|
|
||||||
const tokenRecord = {
|
|
||||||
user_id: 'user-123',
|
|
||||||
token_hash: 'hashed-token',
|
|
||||||
expires_at: new Date(Date.now() + 3600000),
|
|
||||||
};
|
|
||||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([tokenRecord]);
|
|
||||||
vi.mocked(bcrypt.compare).mockResolvedValue(false as never); // Token does not match
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
|
||||||
.post('/api/auth/reset-password')
|
|
||||||
.send({ token: 'wrong-token', newPassword: 'a-Very-Strong-Password-123!' });
|
|
||||||
|
|
||||||
expect(response.status).toBe(400);
|
|
||||||
expect(response.body.message).toBe('Invalid or expired password reset token.');
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should return 400 for a weak new password', async () => {
|
it('should return 400 for a weak new password', async () => {
|
||||||
const tokenRecord = {
|
// No need to mock the service here as validation runs first
|
||||||
user_id: 'user-123',
|
|
||||||
token_hash: 'hashed-token',
|
|
||||||
expires_at: new Date(Date.now() + 3600000),
|
|
||||||
};
|
|
||||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([tokenRecord]);
|
|
||||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/reset-password')
|
.post('/api/auth/reset-password')
|
||||||
@@ -557,11 +456,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
|
|
||||||
describe('POST /refresh-token', () => {
|
describe('POST /refresh-token', () => {
|
||||||
it('should issue a new access token with a valid refresh token cookie', async () => {
|
it('should issue a new access token with a valid refresh token cookie', async () => {
|
||||||
const mockUser = createMockUserWithPasswordHash({
|
mockedAuthService.refreshAccessToken.mockResolvedValue({ accessToken: 'new-access-token' });
|
||||||
user_id: 'user-123',
|
|
||||||
email: 'test@test.com',
|
|
||||||
});
|
|
||||||
vi.mocked(db.userRepo.findUserByRefreshToken).mockResolvedValue(mockUser);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/refresh-token')
|
.post('/api/auth/refresh-token')
|
||||||
@@ -578,8 +473,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should return 403 if refresh token is invalid', async () => {
|
it('should return 403 if refresh token is invalid', async () => {
|
||||||
// Mock finding no user for this token, which should trigger the 403 logic
|
mockedAuthService.refreshAccessToken.mockResolvedValue(null);
|
||||||
vi.mocked(db.userRepo.findUserByRefreshToken).mockResolvedValue(undefined as any);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/auth/refresh-token')
|
.post('/api/auth/refresh-token')
|
||||||
@@ -590,7 +484,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
|
|
||||||
it('should return 500 if the database call fails', async () => {
|
it('should return 500 if the database call fails', async () => {
|
||||||
// Arrange
|
// Arrange
|
||||||
vi.mocked(db.userRepo.findUserByRefreshToken).mockRejectedValue(new Error('DB Error'));
|
mockedAuthService.refreshAccessToken.mockRejectedValue(new Error('DB Error'));
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
@@ -604,7 +498,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
describe('POST /logout', () => {
|
describe('POST /logout', () => {
|
||||||
it('should clear the refresh token cookie and return a success message', async () => {
|
it('should clear the refresh token cookie and return a success message', async () => {
|
||||||
// Arrange
|
// Arrange
|
||||||
vi.mocked(db.userRepo.deleteRefreshToken).mockResolvedValue(undefined);
|
mockedAuthService.logout.mockResolvedValue(undefined);
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
@@ -627,7 +521,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
it('should still return 200 OK even if deleting the refresh token from DB fails', async () => {
|
it('should still return 200 OK even if deleting the refresh token from DB fails', async () => {
|
||||||
// Arrange
|
// Arrange
|
||||||
const dbError = new Error('DB connection lost');
|
const dbError = new Error('DB connection lost');
|
||||||
vi.mocked(db.userRepo.deleteRefreshToken).mockRejectedValue(dbError);
|
mockedAuthService.logout.mockRejectedValue(dbError);
|
||||||
const { logger } = await import('../services/logger.server');
|
const { logger } = await import('../services/logger.server');
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
@@ -639,7 +533,7 @@ describe('Auth Routes (/api/auth)', () => {
|
|||||||
expect(response.status).toBe(200);
|
expect(response.status).toBe(200);
|
||||||
expect(logger.error).toHaveBeenCalledWith(
|
expect(logger.error).toHaveBeenCalledWith(
|
||||||
expect.objectContaining({ error: dbError }),
|
expect.objectContaining({ error: dbError }),
|
||||||
'Failed to delete refresh token from DB during logout.',
|
'Logout token invalidation failed in background.',
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
@@ -1,26 +1,18 @@
|
|||||||
// src/routes/auth.routes.ts
|
// src/routes/auth.routes.ts
|
||||||
import { Router, Request, Response, NextFunction } from 'express';
|
import { Router, Request, Response, NextFunction } from 'express';
|
||||||
import * as bcrypt from 'bcrypt';
|
|
||||||
import { z } from 'zod';
|
import { z } from 'zod';
|
||||||
import jwt from 'jsonwebtoken';
|
|
||||||
import crypto from 'crypto';
|
|
||||||
import rateLimit from 'express-rate-limit';
|
import rateLimit from 'express-rate-limit';
|
||||||
|
|
||||||
import passport from './passport.routes';
|
import passport from './passport.routes';
|
||||||
import { userRepo, adminRepo } from '../services/db/index.db';
|
import { UniqueConstraintError } from '../services/db/errors.db'; // Import actual class for instanceof checks
|
||||||
import { UniqueConstraintError } from '../services/db/errors.db';
|
|
||||||
import { getPool } from '../services/db/connection.db';
|
|
||||||
import { logger } from '../services/logger.server';
|
import { logger } from '../services/logger.server';
|
||||||
import { sendPasswordResetEmail } from '../services/emailService.server';
|
|
||||||
import { validateRequest } from '../middleware/validation.middleware';
|
import { validateRequest } from '../middleware/validation.middleware';
|
||||||
import type { UserProfile } from '../types';
|
import type { UserProfile } from '../types';
|
||||||
import { validatePasswordStrength } from '../utils/authUtils';
|
import { validatePasswordStrength } from '../utils/authUtils';
|
||||||
import { requiredString } from '../utils/zodUtils';
|
import { requiredString } from '../utils/zodUtils';
|
||||||
|
|
||||||
|
import { authService } from '../services/authService';
|
||||||
const router = Router();
|
const router = Router();
|
||||||
|
|
||||||
const JWT_SECRET = process.env.JWT_SECRET!;
|
|
||||||
|
|
||||||
// Conditionally disable rate limiting for the test environment
|
// Conditionally disable rate limiting for the test environment
|
||||||
const isTestEnv = process.env.NODE_ENV === 'test';
|
const isTestEnv = process.env.NODE_ENV === 'test';
|
||||||
|
|
||||||
@@ -45,21 +37,27 @@ const resetPasswordLimiter = rateLimit({
|
|||||||
|
|
||||||
const registerSchema = z.object({
|
const registerSchema = z.object({
|
||||||
body: z.object({
|
body: z.object({
|
||||||
email: z.string().email('A valid email is required.'),
|
// Sanitize email by trimming and converting to lowercase.
|
||||||
|
email: z.string().trim().toLowerCase().email('A valid email is required.'),
|
||||||
password: z
|
password: z
|
||||||
.string()
|
.string()
|
||||||
|
.trim() // Prevent leading/trailing whitespace in passwords.
|
||||||
.min(8, 'Password must be at least 8 characters long.')
|
.min(8, 'Password must be at least 8 characters long.')
|
||||||
.superRefine((password, ctx) => {
|
.superRefine((password, ctx) => {
|
||||||
const strength = validatePasswordStrength(password);
|
const strength = validatePasswordStrength(password);
|
||||||
if (!strength.isValid) ctx.addIssue({ code: 'custom', message: strength.feedback });
|
if (!strength.isValid) ctx.addIssue({ code: 'custom', message: strength.feedback });
|
||||||
}),
|
}),
|
||||||
full_name: z.string().optional(),
|
// Sanitize optional string inputs.
|
||||||
avatar_url: z.string().url().optional(),
|
full_name: z.string().trim().optional(),
|
||||||
|
avatar_url: z.string().trim().url().optional(),
|
||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
const forgotPasswordSchema = z.object({
|
const forgotPasswordSchema = z.object({
|
||||||
body: z.object({ email: z.string().email('A valid email is required.') }),
|
body: z.object({
|
||||||
|
// Sanitize email by trimming and converting to lowercase.
|
||||||
|
email: z.string().trim().toLowerCase().email('A valid email is required.'),
|
||||||
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
const resetPasswordSchema = z.object({
|
const resetPasswordSchema = z.object({
|
||||||
@@ -67,6 +65,7 @@ const resetPasswordSchema = z.object({
|
|||||||
token: requiredString('Token is required.'),
|
token: requiredString('Token is required.'),
|
||||||
newPassword: z
|
newPassword: z
|
||||||
.string()
|
.string()
|
||||||
|
.trim() // Prevent leading/trailing whitespace in passwords.
|
||||||
.min(8, 'Password must be at least 8 characters long.')
|
.min(8, 'Password must be at least 8 characters long.')
|
||||||
.superRefine((password, ctx) => {
|
.superRefine((password, ctx) => {
|
||||||
const strength = validatePasswordStrength(password);
|
const strength = validatePasswordStrength(password);
|
||||||
@@ -88,39 +87,14 @@ router.post(
|
|||||||
} = req as unknown as RegisterRequest;
|
} = req as unknown as RegisterRequest;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const saltRounds = 10;
|
const { newUserProfile, accessToken, refreshToken } = await authService.registerAndLoginUser(
|
||||||
const hashedPassword = await bcrypt.hash(password, saltRounds);
|
|
||||||
logger.info(`Hashing password for new user: ${email}`);
|
|
||||||
|
|
||||||
// The createUser method in UserRepository now handles its own transaction.
|
|
||||||
const newUser = await userRepo.createUser(
|
|
||||||
email,
|
email,
|
||||||
hashedPassword,
|
password,
|
||||||
{ full_name, avatar_url },
|
full_name,
|
||||||
|
avatar_url,
|
||||||
req.log,
|
req.log,
|
||||||
);
|
);
|
||||||
|
|
||||||
const userEmail = newUser.user.email;
|
|
||||||
const userId = newUser.user.user_id;
|
|
||||||
logger.info(`Successfully created new user in DB: ${userEmail} (ID: ${userId})`);
|
|
||||||
|
|
||||||
// Use the new standardized logging function
|
|
||||||
await adminRepo.logActivity(
|
|
||||||
{
|
|
||||||
userId: newUser.user.user_id,
|
|
||||||
action: 'user_registered',
|
|
||||||
displayText: `${userEmail} has registered.`,
|
|
||||||
icon: 'user-plus',
|
|
||||||
},
|
|
||||||
req.log,
|
|
||||||
);
|
|
||||||
|
|
||||||
const payload = { user_id: newUser.user.user_id, email: userEmail };
|
|
||||||
const token = jwt.sign(payload, JWT_SECRET, { expiresIn: '1h' });
|
|
||||||
|
|
||||||
const refreshToken = crypto.randomBytes(64).toString('hex');
|
|
||||||
await userRepo.saveRefreshToken(newUser.user.user_id, refreshToken, req.log);
|
|
||||||
|
|
||||||
res.cookie('refreshToken', refreshToken, {
|
res.cookie('refreshToken', refreshToken, {
|
||||||
httpOnly: true,
|
httpOnly: true,
|
||||||
secure: process.env.NODE_ENV === 'production',
|
secure: process.env.NODE_ENV === 'production',
|
||||||
@@ -128,7 +102,7 @@ router.post(
|
|||||||
});
|
});
|
||||||
return res
|
return res
|
||||||
.status(201)
|
.status(201)
|
||||||
.json({ message: 'User registered successfully!', userprofile: newUser, token });
|
.json({ message: 'User registered successfully!', userprofile: newUserProfile, token: accessToken });
|
||||||
} catch (error: unknown) {
|
} catch (error: unknown) {
|
||||||
if (error instanceof UniqueConstraintError) {
|
if (error instanceof UniqueConstraintError) {
|
||||||
// If the email is a duplicate, return a 409 Conflict status.
|
// If the email is a duplicate, return a 409 Conflict status.
|
||||||
@@ -154,17 +128,6 @@ router.post('/login', (req: Request, res: Response, next: NextFunction) => {
|
|||||||
if (user) req.log.debug({ user }, '[API /login] Passport user object:'); // Log the user object passport returns
|
if (user) req.log.debug({ user }, '[API /login] Passport user object:'); // Log the user object passport returns
|
||||||
if (user) req.log.info({ user }, '[API /login] Passport reported USER FOUND.');
|
if (user) req.log.info({ user }, '[API /login] Passport reported USER FOUND.');
|
||||||
|
|
||||||
try {
|
|
||||||
const allUsersInDb = await getPool().query(
|
|
||||||
'SELECT u.user_id, u.email, p.role FROM public.users u JOIN public.profiles p ON u.user_id = p.user_id',
|
|
||||||
);
|
|
||||||
req.log.debug('[API /login] Current users in DB from SERVER perspective:');
|
|
||||||
console.table(allUsersInDb.rows);
|
|
||||||
} catch (dbError) {
|
|
||||||
req.log.error({ dbError }, '[API /login] Could not query users table for debugging.');
|
|
||||||
}
|
|
||||||
// --- END DEBUG LOGGING ---
|
|
||||||
const { rememberMe } = req.body;
|
|
||||||
if (err) {
|
if (err) {
|
||||||
req.log.error(
|
req.log.error(
|
||||||
{ error: err },
|
{ error: err },
|
||||||
@@ -176,33 +139,24 @@ router.post('/login', (req: Request, res: Response, next: NextFunction) => {
|
|||||||
return res.status(401).json({ message: info.message || 'Login failed' });
|
return res.status(401).json({ message: info.message || 'Login failed' });
|
||||||
}
|
}
|
||||||
|
|
||||||
const userProfile = user as UserProfile;
|
|
||||||
const payload = {
|
|
||||||
user_id: userProfile.user.user_id,
|
|
||||||
email: userProfile.user.email,
|
|
||||||
role: userProfile.role,
|
|
||||||
};
|
|
||||||
const accessToken = jwt.sign(payload, JWT_SECRET, { expiresIn: '15m' });
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const refreshToken = crypto.randomBytes(64).toString('hex');
|
const { rememberMe } = req.body;
|
||||||
await userRepo.saveRefreshToken(userProfile.user.user_id, refreshToken, req.log);
|
const userProfile = user as UserProfile;
|
||||||
|
const { accessToken, refreshToken } = await authService.handleSuccessfulLogin(userProfile, req.log);
|
||||||
req.log.info(`JWT and refresh token issued for user: ${userProfile.user.email}`);
|
req.log.info(`JWT and refresh token issued for user: ${userProfile.user.email}`);
|
||||||
|
|
||||||
const cookieOptions = {
|
const cookieOptions = {
|
||||||
httpOnly: true,
|
httpOnly: true,
|
||||||
secure: process.env.NODE_ENV === 'production',
|
secure: process.env.NODE_ENV === 'production',
|
||||||
maxAge: rememberMe ? 30 * 24 * 60 * 60 * 1000 : undefined,
|
maxAge: rememberMe ? 30 * 24 * 60 * 60 * 1000 : undefined, // 30 days
|
||||||
};
|
};
|
||||||
|
|
||||||
res.cookie('refreshToken', refreshToken, cookieOptions);
|
res.cookie('refreshToken', refreshToken, cookieOptions);
|
||||||
// Return the full user profile object on login to avoid a second fetch on the client.
|
// Return the full user profile object on login to avoid a second fetch on the client.
|
||||||
return res.json({ userprofile: userProfile, token: accessToken });
|
return res.json({ userprofile: userProfile, token: accessToken });
|
||||||
} catch (tokenErr) {
|
} catch (tokenErr) {
|
||||||
req.log.error(
|
const email = (user as UserProfile)?.user?.email || req.body.email;
|
||||||
{ error: tokenErr },
|
req.log.error({ error: tokenErr }, `Failed to process login for user: ${email}`);
|
||||||
`Failed to save refresh token during login for user: ${userProfile.user.email}`,
|
|
||||||
);
|
|
||||||
return next(tokenErr);
|
return next(tokenErr);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@@ -221,38 +175,14 @@ router.post(
|
|||||||
} = req as unknown as ForgotPasswordRequest;
|
} = req as unknown as ForgotPasswordRequest;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
req.log.debug(`[API /forgot-password] Received request for email: ${email}`);
|
// The service handles finding the user, creating the token, and sending the email.
|
||||||
const user = await userRepo.findUserByEmail(email, req.log);
|
const token = await authService.resetPassword(email, req.log);
|
||||||
let token: string | undefined;
|
|
||||||
req.log.debug(
|
|
||||||
{ user: user ? { user_id: user.user_id, email: user.email } : 'NOT FOUND' },
|
|
||||||
`[API /forgot-password] Database search result for ${email}:`,
|
|
||||||
);
|
|
||||||
|
|
||||||
if (user) {
|
|
||||||
token = crypto.randomBytes(32).toString('hex');
|
|
||||||
const saltRounds = 10;
|
|
||||||
const tokenHash = await bcrypt.hash(token, saltRounds);
|
|
||||||
const expiresAt = new Date(Date.now() + 3600000); // 1 hour
|
|
||||||
|
|
||||||
await userRepo.createPasswordResetToken(user.user_id, tokenHash, expiresAt, req.log);
|
|
||||||
|
|
||||||
const resetLink = `${process.env.FRONTEND_URL}/reset-password/${token}`;
|
|
||||||
|
|
||||||
try {
|
|
||||||
await sendPasswordResetEmail(email, resetLink, req.log);
|
|
||||||
} catch (emailError) {
|
|
||||||
req.log.error({ emailError }, `Email send failure during password reset for user`);
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
req.log.warn(`Password reset requested for non-existent email: ${email}`);
|
|
||||||
}
|
|
||||||
|
|
||||||
// For testability, return the token in the response only in the test environment.
|
// For testability, return the token in the response only in the test environment.
|
||||||
const responsePayload: { message: string; token?: string } = {
|
const responsePayload: { message: string; token?: string } = {
|
||||||
message: 'If an account with that email exists, a password reset link has been sent.',
|
message: 'If an account with that email exists, a password reset link has been sent.',
|
||||||
};
|
};
|
||||||
if (process.env.NODE_ENV === 'test' && user) responsePayload.token = token;
|
if (process.env.NODE_ENV === 'test' && token) responsePayload.token = token;
|
||||||
res.status(200).json(responsePayload);
|
res.status(200).json(responsePayload);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
req.log.error({ error }, `An error occurred during /forgot-password for email: ${email}`);
|
req.log.error({ error }, `An error occurred during /forgot-password for email: ${email}`);
|
||||||
@@ -273,38 +203,12 @@ router.post(
|
|||||||
} = req as unknown as ResetPasswordRequest;
|
} = req as unknown as ResetPasswordRequest;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const validTokens = await userRepo.getValidResetTokens(req.log);
|
const resetSuccessful = await authService.updatePassword(token, newPassword, req.log);
|
||||||
let tokenRecord;
|
|
||||||
for (const record of validTokens) {
|
|
||||||
const isMatch = await bcrypt.compare(token, record.token_hash);
|
|
||||||
if (isMatch) {
|
|
||||||
tokenRecord = record;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!tokenRecord) {
|
if (!resetSuccessful) {
|
||||||
return res.status(400).json({ message: 'Invalid or expired password reset token.' });
|
return res.status(400).json({ message: 'Invalid or expired password reset token.' });
|
||||||
}
|
}
|
||||||
|
|
||||||
const saltRounds = 10;
|
|
||||||
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
|
||||||
|
|
||||||
await userRepo.updateUserPassword(tokenRecord.user_id, hashedPassword, req.log);
|
|
||||||
await userRepo.deleteResetToken(tokenRecord.token_hash, req.log);
|
|
||||||
|
|
||||||
// Log this security event after a successful password reset.
|
|
||||||
await adminRepo.logActivity(
|
|
||||||
{
|
|
||||||
userId: tokenRecord.user_id,
|
|
||||||
action: 'password_reset',
|
|
||||||
displayText: `User ID ${tokenRecord.user_id} has reset their password.`,
|
|
||||||
icon: 'key',
|
|
||||||
details: { source_ip: req.ip ?? null },
|
|
||||||
},
|
|
||||||
req.log,
|
|
||||||
);
|
|
||||||
|
|
||||||
res.status(200).json({ message: 'Password has been reset successfully.' });
|
res.status(200).json({ message: 'Password has been reset successfully.' });
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
req.log.error({ error }, `An error occurred during password reset.`);
|
req.log.error({ error }, `An error occurred during password reset.`);
|
||||||
@@ -321,15 +225,11 @@ router.post('/refresh-token', async (req: Request, res: Response, next: NextFunc
|
|||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const user = await userRepo.findUserByRefreshToken(refreshToken, req.log);
|
const result = await authService.refreshAccessToken(refreshToken, req.log);
|
||||||
if (!user) {
|
if (!result) {
|
||||||
return res.status(403).json({ message: 'Invalid or expired refresh token.' });
|
return res.status(403).json({ message: 'Invalid or expired refresh token.' });
|
||||||
}
|
}
|
||||||
|
res.json({ token: result.accessToken });
|
||||||
const payload = { user_id: user.user_id, email: user.email };
|
|
||||||
const newAccessToken = jwt.sign(payload, JWT_SECRET, { expiresIn: '15m' });
|
|
||||||
|
|
||||||
res.json({ token: newAccessToken });
|
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
req.log.error({ error }, 'An error occurred during /refresh-token.');
|
req.log.error({ error }, 'An error occurred during /refresh-token.');
|
||||||
next(error);
|
next(error);
|
||||||
@@ -346,8 +246,8 @@ router.post('/logout', async (req: Request, res: Response) => {
|
|||||||
if (refreshToken) {
|
if (refreshToken) {
|
||||||
// Invalidate the token in the database so it cannot be used again.
|
// Invalidate the token in the database so it cannot be used again.
|
||||||
// We don't need to wait for this to finish to respond to the user.
|
// We don't need to wait for this to finish to respond to the user.
|
||||||
userRepo.deleteRefreshToken(refreshToken, req.log).catch((err: Error) => {
|
authService.logout(refreshToken, req.log).catch((err: Error) => {
|
||||||
req.log.error({ error: err }, 'Failed to delete refresh token from DB during logout.');
|
req.log.error({ error: err }, 'Logout token invalidation failed in background.');
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
// Instruct the browser to clear the cookie by setting its expiration to the past.
|
// Instruct the browser to clear the cookie by setting its expiration to the past.
|
||||||
|
|||||||
@@ -1,11 +1,10 @@
|
|||||||
// src/routes/gamification.routes.ts
|
// src/routes/gamification.routes.ts
|
||||||
import express, { NextFunction } from 'express';
|
import express, { NextFunction } from 'express';
|
||||||
import { z } from 'zod';
|
import { z } from 'zod';
|
||||||
import passport, { isAdmin } from './passport.routes';
|
import passport, { isAdmin } from './passport.routes'; // Correctly imported
|
||||||
import { gamificationRepo } from '../services/db/index.db';
|
import { gamificationService } from '../services/gamificationService';
|
||||||
import { logger } from '../services/logger.server';
|
import { logger } from '../services/logger.server';
|
||||||
import { UserProfile } from '../types';
|
import { UserProfile } from '../types';
|
||||||
import { ForeignKeyConstraintError } from '../services/db/errors.db';
|
|
||||||
import { validateRequest } from '../middleware/validation.middleware';
|
import { validateRequest } from '../middleware/validation.middleware';
|
||||||
import { requiredString, optionalNumeric } from '../utils/zodUtils';
|
import { requiredString, optionalNumeric } from '../utils/zodUtils';
|
||||||
|
|
||||||
@@ -14,10 +13,12 @@ const adminGamificationRouter = express.Router(); // Create a new router for adm
|
|||||||
|
|
||||||
// --- Zod Schemas for Gamification Routes (as per ADR-003) ---
|
// --- Zod Schemas for Gamification Routes (as per ADR-003) ---
|
||||||
|
|
||||||
|
const leaderboardQuerySchema = z.object({
|
||||||
|
limit: optionalNumeric({ default: 10, integer: true, positive: true, max: 50 }),
|
||||||
|
});
|
||||||
|
|
||||||
const leaderboardSchema = z.object({
|
const leaderboardSchema = z.object({
|
||||||
query: z.object({
|
query: leaderboardQuerySchema,
|
||||||
limit: optionalNumeric({ default: 10, integer: true, positive: true, max: 50 }),
|
|
||||||
}),
|
|
||||||
});
|
});
|
||||||
|
|
||||||
const awardAchievementSchema = z.object({
|
const awardAchievementSchema = z.object({
|
||||||
@@ -35,7 +36,7 @@ const awardAchievementSchema = z.object({
|
|||||||
*/
|
*/
|
||||||
router.get('/', async (req, res, next: NextFunction) => {
|
router.get('/', async (req, res, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
const achievements = await gamificationRepo.getAllAchievements(req.log);
|
const achievements = await gamificationService.getAllAchievements(req.log);
|
||||||
res.json(achievements);
|
res.json(achievements);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, 'Error fetching all achievements in /api/achievements:');
|
logger.error({ error }, 'Error fetching all achievements in /api/achievements:');
|
||||||
@@ -51,14 +52,11 @@ router.get(
|
|||||||
'/leaderboard',
|
'/leaderboard',
|
||||||
validateRequest(leaderboardSchema),
|
validateRequest(leaderboardSchema),
|
||||||
async (req, res, next: NextFunction): Promise<void> => {
|
async (req, res, next: NextFunction): Promise<void> => {
|
||||||
// Apply ADR-003 pattern for type safety.
|
|
||||||
// Explicitly coerce query params to ensure numbers are passed to the repo,
|
|
||||||
// as validateRequest might not replace req.query in all test environments.
|
|
||||||
const query = req.query as unknown as { limit?: string };
|
|
||||||
const limit = query.limit ? Number(query.limit) : 10;
|
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const leaderboard = await gamificationRepo.getLeaderboard(limit, req.log);
|
// The `validateRequest` middleware ensures `req.query` is valid.
|
||||||
|
// We parse it here to apply Zod's coercions (string to number) and defaults.
|
||||||
|
const { limit } = leaderboardQuerySchema.parse(req.query);
|
||||||
|
const leaderboard = await gamificationService.getLeaderboard(limit!, req.log);
|
||||||
res.json(leaderboard);
|
res.json(leaderboard);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, 'Error fetching leaderboard:');
|
logger.error({ error }, 'Error fetching leaderboard:');
|
||||||
@@ -79,7 +77,7 @@ router.get(
|
|||||||
async (req, res, next: NextFunction): Promise<void> => {
|
async (req, res, next: NextFunction): Promise<void> => {
|
||||||
const userProfile = req.user as UserProfile;
|
const userProfile = req.user as UserProfile;
|
||||||
try {
|
try {
|
||||||
const userAchievements = await gamificationRepo.getUserAchievements(
|
const userAchievements = await gamificationService.getUserAchievements(
|
||||||
userProfile.user.user_id,
|
userProfile.user.user_id,
|
||||||
req.log,
|
req.log,
|
||||||
);
|
);
|
||||||
@@ -111,21 +109,13 @@ adminGamificationRouter.post(
|
|||||||
type AwardAchievementRequest = z.infer<typeof awardAchievementSchema>;
|
type AwardAchievementRequest = z.infer<typeof awardAchievementSchema>;
|
||||||
const { body } = req as unknown as AwardAchievementRequest;
|
const { body } = req as unknown as AwardAchievementRequest;
|
||||||
try {
|
try {
|
||||||
await gamificationRepo.awardAchievement(body.userId, body.achievementName, req.log);
|
await gamificationService.awardAchievement(body.userId, body.achievementName, req.log);
|
||||||
res
|
res
|
||||||
.status(200)
|
.status(200)
|
||||||
.json({
|
.json({
|
||||||
message: `Successfully awarded '${body.achievementName}' to user ${body.userId}.`,
|
message: `Successfully awarded '${body.achievementName}' to user ${body.userId}.`,
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof ForeignKeyConstraintError) {
|
|
||||||
res.status(400).json({ message: error.message });
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
logger.error(
|
|
||||||
{ error, userId: body.userId, achievementName: body.achievementName },
|
|
||||||
'Error awarding achievement via admin endpoint:',
|
|
||||||
);
|
|
||||||
next(error);
|
next(error);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -28,10 +28,9 @@ router.get(
|
|||||||
validateRequest(mostFrequentSalesSchema),
|
validateRequest(mostFrequentSalesSchema),
|
||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
try {
|
try {
|
||||||
// Parse req.query to ensure coercion (string -> number) and defaults are applied.
|
// The `validateRequest` middleware ensures `req.query` is valid.
|
||||||
// Even though validateRequest checks validity, it may not mutate req.query with the parsed result.
|
// We parse it here to apply Zod's coercions (string to number) and defaults.
|
||||||
const { days, limit } = statsQuerySchema.parse(req.query);
|
const { days, limit } = statsQuerySchema.parse(req.query);
|
||||||
|
|
||||||
const items = await db.adminRepo.getMostFrequentSaleItems(days!, limit!, req.log);
|
const items = await db.adminRepo.getMostFrequentSaleItems(days!, limit!, req.log);
|
||||||
res.json(items);
|
res.json(items);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
|
|||||||
@@ -6,6 +6,31 @@ import { exec, type ExecException, type ExecOptions } from 'child_process';
|
|||||||
import { geocodingService } from '../services/geocodingService.server';
|
import { geocodingService } from '../services/geocodingService.server';
|
||||||
import { createTestApp } from '../tests/utils/createTestApp';
|
import { createTestApp } from '../tests/utils/createTestApp';
|
||||||
|
|
||||||
|
// FIX: Mock util.promisify to correctly handle child_process.exec's (err, stdout, stderr) signature.
|
||||||
|
// This is required because the standard util.promisify relies on internal symbols on the real exec function,
|
||||||
|
// which are missing on our Vitest mock. Without this, promisify(mockExec) drops the stdout/stderr arguments.
|
||||||
|
vi.mock('util', async (importOriginal) => {
|
||||||
|
const actual = await importOriginal<typeof import('util')>();
|
||||||
|
return {
|
||||||
|
...actual,
|
||||||
|
promisify: (fn: Function) => {
|
||||||
|
return (...args: any[]) => {
|
||||||
|
return new Promise((resolve, reject) => {
|
||||||
|
fn(...args, (err: Error | null, stdout: unknown, stderr: unknown) => {
|
||||||
|
if (err) {
|
||||||
|
// Attach stdout/stderr to the error object to mimic child_process.exec behavior
|
||||||
|
Object.assign(err, { stdout, stderr });
|
||||||
|
reject(err);
|
||||||
|
} else {
|
||||||
|
resolve({ stdout, stderr });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
},
|
||||||
|
};
|
||||||
|
});
|
||||||
|
|
||||||
// FIX: Use the simple factory pattern for child_process to avoid default export issues
|
// FIX: Use the simple factory pattern for child_process to avoid default export issues
|
||||||
vi.mock('child_process', () => {
|
vi.mock('child_process', () => {
|
||||||
const mockExec = vi.fn((command, callback) => {
|
const mockExec = vi.fn((command, callback) => {
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
// src/routes/system.routes.ts
|
// src/routes/system.routes.ts
|
||||||
import { Router, Request, Response, NextFunction } from 'express';
|
import { Router, Request, Response, NextFunction } from 'express';
|
||||||
import { exec } from 'child_process';
|
|
||||||
import { z } from 'zod';
|
|
||||||
import { logger } from '../services/logger.server';
|
import { logger } from '../services/logger.server';
|
||||||
import { geocodingService } from '../services/geocodingService.server';
|
import { geocodingService } from '../services/geocodingService.server';
|
||||||
import { validateRequest } from '../middleware/validation.middleware';
|
import { validateRequest } from '../middleware/validation.middleware';
|
||||||
|
import { z } from 'zod';
|
||||||
import { requiredString } from '../utils/zodUtils';
|
import { requiredString } from '../utils/zodUtils';
|
||||||
|
import { systemService } from '../services/systemService';
|
||||||
|
|
||||||
const router = Router();
|
const router = Router();
|
||||||
|
|
||||||
@@ -25,39 +25,13 @@ const emptySchema = z.object({});
|
|||||||
router.get(
|
router.get(
|
||||||
'/pm2-status',
|
'/pm2-status',
|
||||||
validateRequest(emptySchema),
|
validateRequest(emptySchema),
|
||||||
(req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
// The name 'flyer-crawler-api' comes from your ecosystem.config.cjs file.
|
try {
|
||||||
exec('pm2 describe flyer-crawler-api', (error, stdout, stderr) => {
|
const status = await systemService.getPm2Status();
|
||||||
if (error) {
|
res.json(status);
|
||||||
// 'pm2 describe' exits with an error if the process is not found.
|
} catch (error) {
|
||||||
// We can treat this as a "fail" status for our check.
|
next(error);
|
||||||
if (stdout && stdout.includes("doesn't exist")) {
|
}
|
||||||
logger.warn('[API /pm2-status] PM2 process "flyer-crawler-api" not found.');
|
|
||||||
return res.json({
|
|
||||||
success: false,
|
|
||||||
message: 'Application process is not running under PM2.',
|
|
||||||
});
|
|
||||||
}
|
|
||||||
logger.error(
|
|
||||||
{ error: stderr || error.message },
|
|
||||||
'[API /pm2-status] Error executing pm2 describe:',
|
|
||||||
);
|
|
||||||
return next(error);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check if there was output to stderr, even if the exit code was 0 (success).
|
|
||||||
if (stderr && stderr.trim().length > 0) {
|
|
||||||
logger.error({ stderr }, '[API /pm2-status] PM2 executed but produced stderr:');
|
|
||||||
return next(new Error(`PM2 command produced an error: ${stderr}`));
|
|
||||||
}
|
|
||||||
|
|
||||||
// If the command succeeds, we can parse stdout to check the status.
|
|
||||||
const isOnline = /│ status\s+│ online\s+│/m.test(stdout);
|
|
||||||
const message = isOnline
|
|
||||||
? 'Application is online and running under PM2.'
|
|
||||||
: 'Application process exists but is not online.';
|
|
||||||
res.json({ success: isOnline, message });
|
|
||||||
});
|
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|
||||||
|
|||||||
@@ -2,7 +2,6 @@
|
|||||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||||
import supertest from 'supertest';
|
import supertest from 'supertest';
|
||||||
import express from 'express';
|
import express from 'express';
|
||||||
import * as bcrypt from 'bcrypt';
|
|
||||||
import fs from 'node:fs/promises';
|
import fs from 'node:fs/promises';
|
||||||
import {
|
import {
|
||||||
createMockUserProfile,
|
createMockUserProfile,
|
||||||
@@ -17,10 +16,11 @@ import {
|
|||||||
createMockAddress,
|
createMockAddress,
|
||||||
} from '../tests/utils/mockFactories';
|
} from '../tests/utils/mockFactories';
|
||||||
import { Appliance, Notification, DietaryRestriction } from '../types';
|
import { Appliance, Notification, DietaryRestriction } from '../types';
|
||||||
import { ForeignKeyConstraintError, NotFoundError } from '../services/db/errors.db';
|
import { ForeignKeyConstraintError, NotFoundError, ValidationError } from '../services/db/errors.db';
|
||||||
import { createTestApp } from '../tests/utils/createTestApp';
|
import { createTestApp } from '../tests/utils/createTestApp';
|
||||||
import { mockLogger } from '../tests/utils/mockLogger';
|
import { mockLogger } from '../tests/utils/mockLogger';
|
||||||
import { logger } from '../services/logger.server';
|
import { logger } from '../services/logger.server';
|
||||||
|
import { userService } from '../services/userService';
|
||||||
|
|
||||||
// 1. Mock the Service Layer directly.
|
// 1. Mock the Service Layer directly.
|
||||||
// The user.routes.ts file imports from '.../db/index.db'. We need to mock that module.
|
// The user.routes.ts file imports from '.../db/index.db'. We need to mock that module.
|
||||||
@@ -29,9 +29,6 @@ vi.mock('../services/db/index.db', () => ({
|
|||||||
userRepo: {
|
userRepo: {
|
||||||
findUserProfileById: vi.fn(),
|
findUserProfileById: vi.fn(),
|
||||||
updateUserProfile: vi.fn(),
|
updateUserProfile: vi.fn(),
|
||||||
updateUserPassword: vi.fn(),
|
|
||||||
findUserWithPasswordHashById: vi.fn(),
|
|
||||||
deleteUserById: vi.fn(),
|
|
||||||
updateUserPreferences: vi.fn(),
|
updateUserPreferences: vi.fn(),
|
||||||
},
|
},
|
||||||
personalizationRepo: {
|
personalizationRepo: {
|
||||||
@@ -70,22 +67,14 @@ vi.mock('../services/db/index.db', () => ({
|
|||||||
// Mock userService
|
// Mock userService
|
||||||
vi.mock('../services/userService', () => ({
|
vi.mock('../services/userService', () => ({
|
||||||
userService: {
|
userService: {
|
||||||
|
updateUserAvatar: vi.fn(),
|
||||||
|
updateUserPassword: vi.fn(),
|
||||||
|
deleteUserAccount: vi.fn(),
|
||||||
|
getUserAddress: vi.fn(),
|
||||||
upsertUserAddress: vi.fn(),
|
upsertUserAddress: vi.fn(),
|
||||||
},
|
},
|
||||||
}));
|
}));
|
||||||
|
|
||||||
// 2. Mock bcrypt.
|
|
||||||
// We return an object that satisfies both default and named imports to be safe.
|
|
||||||
vi.mock('bcrypt', () => {
|
|
||||||
const hash = vi.fn();
|
|
||||||
const compare = vi.fn();
|
|
||||||
return {
|
|
||||||
default: { hash, compare },
|
|
||||||
hash,
|
|
||||||
compare,
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
// Mock the logger
|
// Mock the logger
|
||||||
vi.mock('../services/logger.server', async () => ({
|
vi.mock('../services/logger.server', async () => ({
|
||||||
// Use async import to avoid hoisting issues with mockLogger
|
// Use async import to avoid hoisting issues with mockLogger
|
||||||
@@ -94,7 +83,6 @@ vi.mock('../services/logger.server', async () => ({
|
|||||||
|
|
||||||
// Import the router and other modules AFTER mocks are established
|
// Import the router and other modules AFTER mocks are established
|
||||||
import userRouter from './user.routes';
|
import userRouter from './user.routes';
|
||||||
import { userService } from '../services/userService'; // Import for checking calls
|
|
||||||
// Import the mocked db module to control its functions in tests
|
// Import the mocked db module to control its functions in tests
|
||||||
import * as db from '../services/db/index.db';
|
import * as db from '../services/db/index.db';
|
||||||
|
|
||||||
@@ -599,20 +587,17 @@ describe('User Routes (/api/users)', () => {
|
|||||||
|
|
||||||
describe('PUT /profile/password', () => {
|
describe('PUT /profile/password', () => {
|
||||||
it('should update the password successfully with a strong password', async () => {
|
it('should update the password successfully with a strong password', async () => {
|
||||||
vi.mocked(bcrypt.hash).mockResolvedValue('hashed-password' as never);
|
vi.mocked(userService.updateUserPassword).mockResolvedValue(undefined);
|
||||||
vi.mocked(db.userRepo.updateUserPassword).mockResolvedValue(undefined);
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.put('/api/users/profile/password')
|
.put('/api/users/profile/password')
|
||||||
.send({ newPassword: 'a-Very-Strong-Password-456!' });
|
.send({ newPassword: 'a-Very-Strong-Password-456!' });
|
||||||
|
|
||||||
expect(response.status).toBe(200);
|
expect(response.status).toBe(200);
|
||||||
expect(response.body.message).toBe('Password updated successfully.');
|
expect(response.body.message).toBe('Password updated successfully.');
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 on a generic database error', async () => {
|
it('should return 500 on a generic database error', async () => {
|
||||||
const dbError = new Error('DB Connection Failed');
|
const dbError = new Error('DB Connection Failed');
|
||||||
vi.mocked(bcrypt.hash).mockResolvedValue('hashed-password' as never);
|
vi.mocked(userService.updateUserPassword).mockRejectedValue(dbError);
|
||||||
vi.mocked(db.userRepo.updateUserPassword).mockRejectedValue(dbError);
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.put('/api/users/profile/password')
|
.put('/api/users/profile/password')
|
||||||
.send({ newPassword: 'a-Very-Strong-Password-456!' });
|
.send({ newPassword: 'a-Very-Strong-Password-456!' });
|
||||||
@@ -624,7 +609,6 @@ describe('User Routes (/api/users)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('should return 400 for a weak password', async () => {
|
it('should return 400 for a weak password', async () => {
|
||||||
// Use a password long enough to pass .min(8) but weak enough to fail strength check
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.put('/api/users/profile/password')
|
.put('/api/users/profile/password')
|
||||||
.send({ newPassword: 'password123' });
|
.send({ newPassword: 'password123' });
|
||||||
@@ -636,70 +620,38 @@ describe('User Routes (/api/users)', () => {
|
|||||||
|
|
||||||
describe('DELETE /account', () => {
|
describe('DELETE /account', () => {
|
||||||
it('should delete the account with the correct password', async () => {
|
it('should delete the account with the correct password', async () => {
|
||||||
const userWithHash = createMockUserWithPasswordHash({
|
vi.mocked(userService.deleteUserAccount).mockResolvedValue(undefined);
|
||||||
...mockUserProfile.user,
|
|
||||||
password_hash: 'hashed-password',
|
|
||||||
});
|
|
||||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithHash);
|
|
||||||
vi.mocked(db.userRepo.deleteUserById).mockResolvedValue(undefined);
|
|
||||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never);
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.delete('/api/users/account')
|
.delete('/api/users/account')
|
||||||
.send({ password: 'correct-password' });
|
.send({ password: 'correct-password' });
|
||||||
expect(response.status).toBe(200);
|
expect(response.status).toBe(200);
|
||||||
expect(response.body.message).toBe('Account deleted successfully.');
|
expect(response.body.message).toBe('Account deleted successfully.');
|
||||||
|
expect(userService.deleteUserAccount).toHaveBeenCalledWith('user-123', 'correct-password', expectLogger);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 403 for an incorrect password', async () => {
|
it('should return 400 for an incorrect password', async () => {
|
||||||
const userWithHash = createMockUserWithPasswordHash({
|
vi.mocked(userService.deleteUserAccount).mockRejectedValue(new ValidationError([], 'Incorrect password.'));
|
||||||
...mockUserProfile.user,
|
|
||||||
password_hash: 'hashed-password',
|
|
||||||
});
|
|
||||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithHash);
|
|
||||||
vi.mocked(bcrypt.compare).mockResolvedValue(false as never);
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.delete('/api/users/account')
|
.delete('/api/users/account')
|
||||||
.send({ password: 'wrong-password' });
|
.send({ password: 'wrong-password' });
|
||||||
|
|
||||||
expect(response.status).toBe(403);
|
expect(response.status).toBe(400);
|
||||||
expect(response.body.message).toBe('Incorrect password.');
|
expect(response.body.message).toBe('Incorrect password.');
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 404 if the user to delete is not found', async () => {
|
it('should return 404 if the user to delete is not found', async () => {
|
||||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockRejectedValue(
|
vi.mocked(userService.deleteUserAccount).mockRejectedValue(new NotFoundError('User not found.'));
|
||||||
new NotFoundError('User not found or password not set.'),
|
|
||||||
);
|
|
||||||
const response = await supertest(app)
|
|
||||||
.delete('/api/users/account')
|
|
||||||
.send({ password: 'any-password' });
|
|
||||||
expect(response.status).toBe(404);
|
|
||||||
expect(response.body.message).toBe('User not found or password not set.');
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should return 404 if user is an OAuth user without a password', async () => {
|
|
||||||
// Simulate an OAuth user who has no password_hash set.
|
|
||||||
const userWithoutHash = createMockUserWithPasswordHash({
|
|
||||||
...mockUserProfile.user,
|
|
||||||
password_hash: null,
|
|
||||||
});
|
|
||||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithoutHash);
|
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.delete('/api/users/account')
|
.delete('/api/users/account')
|
||||||
.send({ password: 'any-password' });
|
.send({ password: 'any-password' });
|
||||||
|
|
||||||
expect(response.status).toBe(404);
|
expect(response.status).toBe(404);
|
||||||
expect(response.body.message).toBe('User not found or password not set.');
|
expect(response.body.message).toBe('User not found.');
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 on a generic database error', async () => {
|
it('should return 500 on a generic database error', async () => {
|
||||||
const userWithHash = createMockUserWithPasswordHash({
|
vi.mocked(userService.deleteUserAccount).mockRejectedValue(new Error('DB Connection Failed'));
|
||||||
...mockUserProfile.user,
|
|
||||||
password_hash: 'hashed-password',
|
|
||||||
});
|
|
||||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithHash);
|
|
||||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never);
|
|
||||||
vi.mocked(db.userRepo.deleteUserById).mockRejectedValue(new Error('DB Connection Failed'));
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.delete('/api/users/account')
|
.delete('/api/users/account')
|
||||||
.send({ password: 'correct-password' });
|
.send({ password: 'correct-password' });
|
||||||
@@ -980,7 +932,7 @@ describe('User Routes (/api/users)', () => {
|
|||||||
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
||||||
});
|
});
|
||||||
const mockAddress = createMockAddress({ address_id: 1, address_line_1: '123 Main St' });
|
const mockAddress = createMockAddress({ address_id: 1, address_line_1: '123 Main St' });
|
||||||
vi.mocked(db.addressRepo.getAddressById).mockResolvedValue(mockAddress);
|
vi.mocked(userService.getUserAddress).mockResolvedValue(mockAddress);
|
||||||
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
||||||
expect(response.status).toBe(200);
|
expect(response.status).toBe(200);
|
||||||
expect(response.body).toEqual(mockAddress);
|
expect(response.body).toEqual(mockAddress);
|
||||||
@@ -992,7 +944,7 @@ describe('User Routes (/api/users)', () => {
|
|||||||
basePath,
|
basePath,
|
||||||
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
||||||
});
|
});
|
||||||
vi.mocked(db.addressRepo.getAddressById).mockRejectedValue(new Error('DB Error'));
|
vi.mocked(userService.getUserAddress).mockRejectedValue(new Error('DB Error'));
|
||||||
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
||||||
expect(response.status).toBe(500);
|
expect(response.status).toBe(500);
|
||||||
});
|
});
|
||||||
@@ -1005,13 +957,10 @@ describe('User Routes (/api/users)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('GET /addresses/:addressId should return 403 if address does not belong to user', async () => {
|
it('GET /addresses/:addressId should return 403 if address does not belong to user', async () => {
|
||||||
const appWithDifferentUser = createTestApp({
|
vi.mocked(userService.getUserAddress).mockRejectedValue(new ValidationError([], 'Forbidden'));
|
||||||
router: userRouter,
|
const response = await supertest(app).get('/api/users/addresses/2'); // Requesting address 2
|
||||||
basePath,
|
expect(response.status).toBe(400); // ValidationError maps to 400 by default in the test error handler
|
||||||
authenticatedUser: { ...mockUserProfile, address_id: 999 },
|
expect(response.body.message).toBe('Forbidden');
|
||||||
});
|
|
||||||
const response = await supertest(appWithDifferentUser).get('/api/users/addresses/1');
|
|
||||||
expect(response.status).toBe(403);
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('GET /addresses/:addressId should return 404 if address not found', async () => {
|
it('GET /addresses/:addressId should return 404 if address not found', async () => {
|
||||||
@@ -1020,7 +969,7 @@ describe('User Routes (/api/users)', () => {
|
|||||||
basePath,
|
basePath,
|
||||||
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
||||||
});
|
});
|
||||||
vi.mocked(db.addressRepo.getAddressById).mockRejectedValue(
|
vi.mocked(userService.getUserAddress).mockRejectedValue(
|
||||||
new NotFoundError('Address not found.'),
|
new NotFoundError('Address not found.'),
|
||||||
);
|
);
|
||||||
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
||||||
@@ -1029,19 +978,10 @@ describe('User Routes (/api/users)', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
it('PUT /profile/address should call upsertAddress and updateUserProfile if needed', async () => {
|
it('PUT /profile/address should call upsertAddress and updateUserProfile if needed', async () => {
|
||||||
const appWithUser = createTestApp({
|
|
||||||
router: userRouter,
|
|
||||||
basePath,
|
|
||||||
authenticatedUser: { ...mockUserProfile, address_id: null },
|
|
||||||
}); // User has no address yet
|
|
||||||
const addressData = { address_line_1: '123 New St' };
|
const addressData = { address_line_1: '123 New St' };
|
||||||
vi.mocked(db.addressRepo.upsertAddress).mockResolvedValue(5); // New address ID is 5
|
vi.mocked(userService.upsertUserAddress).mockResolvedValue(5);
|
||||||
vi.mocked(db.userRepo.updateUserProfile).mockResolvedValue({
|
|
||||||
...mockUserProfile,
|
|
||||||
address_id: 5,
|
|
||||||
});
|
|
||||||
|
|
||||||
const response = await supertest(appWithUser)
|
const response = await supertest(app)
|
||||||
.put('/api/users/profile/address')
|
.put('/api/users/profile/address')
|
||||||
.send(addressData);
|
.send(addressData);
|
||||||
|
|
||||||
@@ -1073,11 +1013,11 @@ describe('User Routes (/api/users)', () => {
|
|||||||
|
|
||||||
describe('POST /profile/avatar', () => {
|
describe('POST /profile/avatar', () => {
|
||||||
it('should upload an avatar and update the user profile', async () => {
|
it('should upload an avatar and update the user profile', async () => {
|
||||||
const mockUpdatedProfile = {
|
const mockUpdatedProfile = createMockUserProfile({
|
||||||
...mockUserProfile,
|
...mockUserProfile,
|
||||||
avatar_url: '/uploads/avatars/new-avatar.png',
|
avatar_url: '/uploads/avatars/new-avatar.png',
|
||||||
};
|
});
|
||||||
vi.mocked(db.userRepo.updateUserProfile).mockResolvedValue(mockUpdatedProfile);
|
vi.mocked(userService.updateUserAvatar).mockResolvedValue(mockUpdatedProfile);
|
||||||
|
|
||||||
// Create a dummy file path for supertest to attach
|
// Create a dummy file path for supertest to attach
|
||||||
const dummyImagePath = 'test-avatar.png';
|
const dummyImagePath = 'test-avatar.png';
|
||||||
@@ -1087,17 +1027,17 @@ describe('User Routes (/api/users)', () => {
|
|||||||
.attach('avatar', Buffer.from('dummy-image-content'), dummyImagePath);
|
.attach('avatar', Buffer.from('dummy-image-content'), dummyImagePath);
|
||||||
|
|
||||||
expect(response.status).toBe(200);
|
expect(response.status).toBe(200);
|
||||||
expect(response.body.avatar_url).toContain('/uploads/avatars/');
|
expect(response.body.avatar_url).toContain('/uploads/avatars/'); // This was a duplicate, fixed.
|
||||||
expect(db.userRepo.updateUserProfile).toHaveBeenCalledWith(
|
expect(userService.updateUserAvatar).toHaveBeenCalledWith(
|
||||||
mockUserProfile.user.user_id,
|
mockUserProfile.user.user_id,
|
||||||
{ avatar_url: expect.any(String) },
|
expect.any(Object),
|
||||||
expectLogger,
|
expectLogger,
|
||||||
);
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return 500 if updating the profile fails after upload', async () => {
|
it('should return 500 if updating the profile fails after upload', async () => {
|
||||||
const dbError = new Error('DB Connection Failed');
|
const dbError = new Error('DB Connection Failed');
|
||||||
vi.mocked(db.userRepo.updateUserProfile).mockRejectedValue(dbError);
|
vi.mocked(userService.updateUserAvatar).mockRejectedValue(dbError);
|
||||||
const dummyImagePath = 'test-avatar.png';
|
const dummyImagePath = 'test-avatar.png';
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
.post('/api/users/profile/avatar')
|
.post('/api/users/profile/avatar')
|
||||||
@@ -1141,7 +1081,7 @@ describe('User Routes (/api/users)', () => {
|
|||||||
const unlinkSpy = vi.spyOn(fs, 'unlink').mockResolvedValue(undefined);
|
const unlinkSpy = vi.spyOn(fs, 'unlink').mockResolvedValue(undefined);
|
||||||
|
|
||||||
const dbError = new Error('DB Connection Failed');
|
const dbError = new Error('DB Connection Failed');
|
||||||
vi.mocked(db.userRepo.updateUserProfile).mockRejectedValue(dbError);
|
vi.mocked(userService.updateUserAvatar).mockRejectedValue(dbError);
|
||||||
const dummyImagePath = 'test-avatar.png';
|
const dummyImagePath = 'test-avatar.png';
|
||||||
|
|
||||||
const response = await supertest(app)
|
const response = await supertest(app)
|
||||||
|
|||||||
@@ -2,8 +2,6 @@
|
|||||||
import express, { Request, Response, NextFunction } from 'express';
|
import express, { Request, Response, NextFunction } from 'express';
|
||||||
import passport from './passport.routes';
|
import passport from './passport.routes';
|
||||||
import multer from 'multer'; // Keep for MulterError type check
|
import multer from 'multer'; // Keep for MulterError type check
|
||||||
import fs from 'node:fs/promises';
|
|
||||||
import * as bcrypt from 'bcrypt'; // This was a duplicate, fixed.
|
|
||||||
import { z } from 'zod';
|
import { z } from 'zod';
|
||||||
import { logger } from '../services/logger.server';
|
import { logger } from '../services/logger.server';
|
||||||
import { UserProfile } from '../types';
|
import { UserProfile } from '../types';
|
||||||
@@ -22,19 +20,7 @@ import {
|
|||||||
optionalBoolean,
|
optionalBoolean,
|
||||||
} from '../utils/zodUtils';
|
} from '../utils/zodUtils';
|
||||||
import * as db from '../services/db/index.db';
|
import * as db from '../services/db/index.db';
|
||||||
|
import { cleanupUploadedFile } from '../utils/fileUtils';
|
||||||
/**
|
|
||||||
* Safely deletes a file from the filesystem, ignoring errors if the file doesn't exist.
|
|
||||||
* @param file The multer file object to delete.
|
|
||||||
*/
|
|
||||||
const cleanupUploadedFile = async (file?: Express.Multer.File) => {
|
|
||||||
if (!file) return;
|
|
||||||
try {
|
|
||||||
await fs.unlink(file.path);
|
|
||||||
} catch (err) {
|
|
||||||
logger.warn({ err, filePath: file.path }, 'Failed to clean up uploaded avatar file.');
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const router = express.Router();
|
const router = express.Router();
|
||||||
|
|
||||||
@@ -50,6 +36,7 @@ const updatePasswordSchema = z.object({
|
|||||||
body: z.object({
|
body: z.object({
|
||||||
newPassword: z
|
newPassword: z
|
||||||
.string()
|
.string()
|
||||||
|
.trim() // Trim whitespace from password input.
|
||||||
.min(8, 'Password must be at least 8 characters long.')
|
.min(8, 'Password must be at least 8 characters long.')
|
||||||
.superRefine((password, ctx) => {
|
.superRefine((password, ctx) => {
|
||||||
const strength = validatePasswordStrength(password);
|
const strength = validatePasswordStrength(password);
|
||||||
@@ -58,6 +45,9 @@ const updatePasswordSchema = z.object({
|
|||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
|
// The `requiredString` utility (modified in `zodUtils.ts`) now handles trimming,
|
||||||
|
// so no changes are needed here, but we are confirming that password trimming
|
||||||
|
// is now implicitly handled for this schema.
|
||||||
const deleteAccountSchema = z.object({
|
const deleteAccountSchema = z.object({
|
||||||
body: z.object({ password: requiredString("Field 'password' is required.") }),
|
body: z.object({ password: requiredString("Field 'password' is required.") }),
|
||||||
});
|
});
|
||||||
@@ -103,14 +93,10 @@ router.post(
|
|||||||
async (req: Request, res: Response, next: NextFunction) => {
|
async (req: Request, res: Response, next: NextFunction) => {
|
||||||
// The try-catch block was already correct here.
|
// The try-catch block was already correct here.
|
||||||
try {
|
try {
|
||||||
|
// The `requireFileUpload` middleware is not used here, so we must check for `req.file`.
|
||||||
if (!req.file) return res.status(400).json({ message: 'No avatar file uploaded.' });
|
if (!req.file) return res.status(400).json({ message: 'No avatar file uploaded.' });
|
||||||
const userProfile = req.user as UserProfile;
|
const userProfile = req.user as UserProfile;
|
||||||
const avatarUrl = `/uploads/avatars/${req.file.filename}`;
|
const updatedProfile = await userService.updateUserAvatar(userProfile.user.user_id, req.file, req.log);
|
||||||
const updatedProfile = await db.userRepo.updateUserProfile(
|
|
||||||
userProfile.user.user_id,
|
|
||||||
{ avatar_url: avatarUrl },
|
|
||||||
req.log,
|
|
||||||
);
|
|
||||||
res.json(updatedProfile);
|
res.json(updatedProfile);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// If an error occurs after the file has been uploaded (e.g., DB error),
|
// If an error occurs after the file has been uploaded (e.g., DB error),
|
||||||
@@ -257,9 +243,7 @@ router.put(
|
|||||||
const { body } = req as unknown as UpdatePasswordRequest;
|
const { body } = req as unknown as UpdatePasswordRequest;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const saltRounds = 10;
|
await userService.updateUserPassword(userProfile.user.user_id, body.newPassword, req.log);
|
||||||
const hashedPassword = await bcrypt.hash(body.newPassword, saltRounds);
|
|
||||||
await db.userRepo.updateUserPassword(userProfile.user.user_id, hashedPassword, req.log);
|
|
||||||
res.status(200).json({ message: 'Password updated successfully.' });
|
res.status(200).json({ message: 'Password updated successfully.' });
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, `[ROUTE] PUT /api/users/profile/password - ERROR`);
|
logger.error({ error }, `[ROUTE] PUT /api/users/profile/password - ERROR`);
|
||||||
@@ -282,20 +266,7 @@ router.delete(
|
|||||||
const { body } = req as unknown as DeleteAccountRequest;
|
const { body } = req as unknown as DeleteAccountRequest;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const userWithHash = await db.userRepo.findUserWithPasswordHashById(
|
await userService.deleteUserAccount(userProfile.user.user_id, body.password, req.log);
|
||||||
userProfile.user.user_id,
|
|
||||||
req.log,
|
|
||||||
);
|
|
||||||
if (!userWithHash || !userWithHash.password_hash) {
|
|
||||||
return res.status(404).json({ message: 'User not found or password not set.' });
|
|
||||||
}
|
|
||||||
|
|
||||||
const isMatch = await bcrypt.compare(body.password, userWithHash.password_hash);
|
|
||||||
if (!isMatch) {
|
|
||||||
return res.status(403).json({ message: 'Incorrect password.' });
|
|
||||||
}
|
|
||||||
|
|
||||||
await db.userRepo.deleteUserById(userProfile.user.user_id, req.log);
|
|
||||||
res.status(200).json({ message: 'Account deleted successfully.' });
|
res.status(200).json({ message: 'Account deleted successfully.' });
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, `[ROUTE] DELETE /api/users/account - ERROR`);
|
logger.error({ error }, `[ROUTE] DELETE /api/users/account - ERROR`);
|
||||||
@@ -485,7 +456,11 @@ const addShoppingListItemSchema = shoppingListIdSchema.extend({
|
|||||||
body: z
|
body: z
|
||||||
.object({
|
.object({
|
||||||
masterItemId: z.number().int().positive().optional(),
|
masterItemId: z.number().int().positive().optional(),
|
||||||
customItemName: z.string().min(1, 'customItemName cannot be empty if provided').optional(),
|
customItemName: z
|
||||||
|
.string()
|
||||||
|
.trim()
|
||||||
|
.min(1, 'customItemName cannot be empty if provided')
|
||||||
|
.optional(),
|
||||||
})
|
})
|
||||||
.refine((data) => data.masterItemId || data.customItemName, {
|
.refine((data) => data.masterItemId || data.customItemName, {
|
||||||
message: 'Either masterItemId or customItemName must be provided.',
|
message: 'Either masterItemId or customItemName must be provided.',
|
||||||
@@ -711,13 +686,7 @@ router.get(
|
|||||||
const { params } = req as unknown as GetAddressRequest;
|
const { params } = req as unknown as GetAddressRequest;
|
||||||
try {
|
try {
|
||||||
const addressId = params.addressId;
|
const addressId = params.addressId;
|
||||||
// Security check: Ensure the requested addressId matches the one on the user's profile.
|
const address = await userService.getUserAddress(userProfile, addressId, req.log);
|
||||||
if (userProfile.address_id !== addressId) {
|
|
||||||
return res
|
|
||||||
.status(403)
|
|
||||||
.json({ message: 'Forbidden: You can only access your own address.' });
|
|
||||||
}
|
|
||||||
const address = await db.addressRepo.getAddressById(addressId, req.log); // This will throw NotFoundError if not found
|
|
||||||
res.json(address);
|
res.json(address);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error({ error }, 'Error fetching user address');
|
logger.error({ error }, 'Error fetching user address');
|
||||||
@@ -732,12 +701,12 @@ router.get(
|
|||||||
const updateUserAddressSchema = z.object({
|
const updateUserAddressSchema = z.object({
|
||||||
body: z
|
body: z
|
||||||
.object({
|
.object({
|
||||||
address_line_1: z.string().optional(),
|
address_line_1: z.string().trim().optional(),
|
||||||
address_line_2: z.string().optional(),
|
address_line_2: z.string().trim().optional(),
|
||||||
city: z.string().optional(),
|
city: z.string().trim().optional(),
|
||||||
province_state: z.string().optional(),
|
province_state: z.string().trim().optional(),
|
||||||
postal_code: z.string().optional(),
|
postal_code: z.string().trim().optional(),
|
||||||
country: z.string().optional(),
|
country: z.string().trim().optional(),
|
||||||
})
|
})
|
||||||
.refine((data) => Object.keys(data).length > 0, {
|
.refine((data) => Object.keys(data).length > 0, {
|
||||||
message: 'At least one address field must be provided.',
|
message: 'At least one address field must be provided.',
|
||||||
@@ -797,13 +766,13 @@ router.delete(
|
|||||||
const updateRecipeSchema = recipeIdSchema.extend({
|
const updateRecipeSchema = recipeIdSchema.extend({
|
||||||
body: z
|
body: z
|
||||||
.object({
|
.object({
|
||||||
name: z.string().optional(),
|
name: z.string().trim().optional(),
|
||||||
description: z.string().optional(),
|
description: z.string().trim().optional(),
|
||||||
instructions: z.string().optional(),
|
instructions: z.string().trim().optional(),
|
||||||
prep_time_minutes: z.number().int().optional(),
|
prep_time_minutes: z.number().int().optional(),
|
||||||
cook_time_minutes: z.number().int().optional(),
|
cook_time_minutes: z.number().int().optional(),
|
||||||
servings: z.number().int().optional(),
|
servings: z.number().int().optional(),
|
||||||
photo_url: z.string().url().optional(),
|
photo_url: z.string().trim().url().optional(),
|
||||||
})
|
})
|
||||||
.refine((data) => Object.keys(data).length > 0, { message: 'No fields provided to update.' }),
|
.refine((data) => Object.keys(data).length > 0, { message: 'No fields provided to update.' }),
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -10,7 +10,23 @@ import fsPromises from 'node:fs/promises';
|
|||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import { z } from 'zod';
|
import { z } from 'zod';
|
||||||
import { pRateLimit } from 'p-ratelimit';
|
import { pRateLimit } from 'p-ratelimit';
|
||||||
import type { FlyerItem, MasterGroceryItem, ExtractedFlyerItem } from '../types';
|
import type {
|
||||||
|
FlyerItem,
|
||||||
|
MasterGroceryItem,
|
||||||
|
ExtractedFlyerItem,
|
||||||
|
UserProfile,
|
||||||
|
ExtractedCoreData,
|
||||||
|
FlyerInsert,
|
||||||
|
Flyer,
|
||||||
|
} from '../types';
|
||||||
|
import { FlyerProcessingError } from './processingErrors';
|
||||||
|
import * as db from './db/index.db';
|
||||||
|
import { flyerQueue } from './queueService.server';
|
||||||
|
import type { Job } from 'bullmq';
|
||||||
|
import { createFlyerAndItems } from './db/flyer.db';
|
||||||
|
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||||
|
import path from 'path';
|
||||||
|
import { ValidationError } from './db/errors.db';
|
||||||
|
|
||||||
// Helper for consistent required string validation (handles missing/null/empty)
|
// Helper for consistent required string validation (handles missing/null/empty)
|
||||||
const requiredString = (message: string) =>
|
const requiredString = (message: string) =>
|
||||||
@@ -34,6 +50,21 @@ export const AiFlyerDataSchema = z.object({
|
|||||||
items: z.array(ExtractedFlyerItemSchema),
|
items: z.array(ExtractedFlyerItemSchema),
|
||||||
});
|
});
|
||||||
|
|
||||||
|
interface FlyerProcessPayload extends Partial<ExtractedCoreData> {
|
||||||
|
checksum?: string;
|
||||||
|
originalFileName?: string;
|
||||||
|
extractedData?: Partial<ExtractedCoreData>;
|
||||||
|
data?: FlyerProcessPayload; // For nested data structures
|
||||||
|
}
|
||||||
|
|
||||||
|
// Helper to safely extract an error message from unknown `catch` values.
|
||||||
|
const errMsg = (e: unknown) => {
|
||||||
|
if (e instanceof Error) return e.message;
|
||||||
|
if (typeof e === 'object' && e !== null && 'message' in e)
|
||||||
|
return String((e as { message: unknown }).message);
|
||||||
|
return String(e || 'An unknown error occurred.');
|
||||||
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Defines the contract for a file system utility. This interface allows for
|
* Defines the contract for a file system utility. This interface allows for
|
||||||
* dependency injection, making the AIService testable without hitting the real file system.
|
* dependency injection, making the AIService testable without hitting the real file system.
|
||||||
@@ -67,6 +98,12 @@ type RawFlyerItem = {
|
|||||||
master_item_id?: number | null | undefined;
|
master_item_id?: number | null | undefined;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
export class DuplicateFlyerError extends FlyerProcessingError {
|
||||||
|
constructor(message: string, public flyerId: number) {
|
||||||
|
super(message, 'DUPLICATE_FLYER', message);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
export class AIService {
|
export class AIService {
|
||||||
private aiClient: IAiClient;
|
private aiClient: IAiClient;
|
||||||
private fs: IFileSystem;
|
private fs: IFileSystem;
|
||||||
@@ -690,6 +727,187 @@ export class AIService {
|
|||||||
}
|
}
|
||||||
*/
|
*/
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
async enqueueFlyerProcessing(
|
||||||
|
file: Express.Multer.File,
|
||||||
|
checksum: string,
|
||||||
|
userProfile: UserProfile | undefined,
|
||||||
|
submitterIp: string,
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<Job> {
|
||||||
|
// 1. Check for duplicate flyer
|
||||||
|
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, logger);
|
||||||
|
if (existingFlyer) {
|
||||||
|
// Throw a specific error for the route to handle
|
||||||
|
throw new DuplicateFlyerError(
|
||||||
|
'This flyer has already been processed.',
|
||||||
|
existingFlyer.flyer_id,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
// 2. Construct user address string
|
||||||
|
let userProfileAddress: string | undefined = undefined;
|
||||||
|
if (userProfile?.address) {
|
||||||
|
userProfileAddress = [
|
||||||
|
userProfile.address.address_line_1,
|
||||||
|
userProfile.address.address_line_2,
|
||||||
|
userProfile.address.city,
|
||||||
|
userProfile.address.province_state,
|
||||||
|
userProfile.address.postal_code,
|
||||||
|
userProfile.address.country,
|
||||||
|
]
|
||||||
|
.filter(Boolean)
|
||||||
|
.join(', ');
|
||||||
|
}
|
||||||
|
|
||||||
|
// 3. Add job to the queue
|
||||||
|
const job = await flyerQueue.add('process-flyer', {
|
||||||
|
filePath: file.path,
|
||||||
|
originalFileName: file.originalname,
|
||||||
|
checksum: checksum,
|
||||||
|
userId: userProfile?.user.user_id,
|
||||||
|
submitterIp: submitterIp,
|
||||||
|
userProfileAddress: userProfileAddress,
|
||||||
|
});
|
||||||
|
|
||||||
|
logger.info(
|
||||||
|
`Enqueued flyer for processing. File: ${file.originalname}, Job ID: ${job.id}`,
|
||||||
|
);
|
||||||
|
|
||||||
|
return job;
|
||||||
|
}
|
||||||
|
|
||||||
|
private _parseLegacyPayload(
|
||||||
|
body: any,
|
||||||
|
logger: Logger,
|
||||||
|
): { parsed: FlyerProcessPayload; extractedData: Partial<ExtractedCoreData> | null | undefined } {
|
||||||
|
let parsed: FlyerProcessPayload = {};
|
||||||
|
let extractedData: Partial<ExtractedCoreData> | null | undefined = {};
|
||||||
|
try {
|
||||||
|
if (body && (body.data || body.extractedData)) {
|
||||||
|
const raw = body.data ?? body.extractedData;
|
||||||
|
try {
|
||||||
|
parsed = typeof raw === 'string' ? JSON.parse(raw) : raw;
|
||||||
|
} catch (err) {
|
||||||
|
logger.warn(
|
||||||
|
{ error: errMsg(err) },
|
||||||
|
'[AIService] Failed to JSON.parse raw extractedData; falling back to direct assign',
|
||||||
|
);
|
||||||
|
parsed = (
|
||||||
|
typeof raw === 'string' ? JSON.parse(String(raw).slice(0, 2000)) : raw
|
||||||
|
) as FlyerProcessPayload;
|
||||||
|
}
|
||||||
|
extractedData = 'extractedData' in parsed ? parsed.extractedData : (parsed as Partial<ExtractedCoreData>);
|
||||||
|
} else {
|
||||||
|
try {
|
||||||
|
parsed = typeof body === 'string' ? JSON.parse(body) : body;
|
||||||
|
} catch (err) {
|
||||||
|
logger.warn(
|
||||||
|
{ error: errMsg(err) },
|
||||||
|
'[AIService] Failed to JSON.parse req.body; using empty object',
|
||||||
|
);
|
||||||
|
parsed = (body as FlyerProcessPayload) || {};
|
||||||
|
}
|
||||||
|
if (parsed.data) {
|
||||||
|
try {
|
||||||
|
const inner = typeof parsed.data === 'string' ? JSON.parse(parsed.data) : parsed.data;
|
||||||
|
extractedData = inner.extractedData ?? inner;
|
||||||
|
} catch (err) {
|
||||||
|
logger.warn({ error: errMsg(err) }, '[AIService] Failed to parse parsed.data; falling back');
|
||||||
|
extractedData = parsed.data as unknown as Partial<ExtractedCoreData>;
|
||||||
|
}
|
||||||
|
} else if (parsed.extractedData) {
|
||||||
|
extractedData = parsed.extractedData;
|
||||||
|
} else {
|
||||||
|
if ('items' in parsed || 'store_name' in parsed || 'valid_from' in parsed) {
|
||||||
|
extractedData = parsed as Partial<ExtractedCoreData>;
|
||||||
|
} else {
|
||||||
|
extractedData = {};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
logger.error({ error: err }, '[AIService] Unexpected error while parsing legacy request body');
|
||||||
|
parsed = {};
|
||||||
|
extractedData = {};
|
||||||
|
}
|
||||||
|
return { parsed, extractedData };
|
||||||
|
}
|
||||||
|
|
||||||
|
async processLegacyFlyerUpload(
|
||||||
|
file: Express.Multer.File,
|
||||||
|
body: any,
|
||||||
|
userProfile: UserProfile | undefined,
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<Flyer> {
|
||||||
|
const { parsed, extractedData: initialExtractedData } = this._parseLegacyPayload(body, logger);
|
||||||
|
let extractedData = initialExtractedData;
|
||||||
|
|
||||||
|
const checksum = parsed.checksum ?? parsed?.data?.checksum ?? '';
|
||||||
|
if (!checksum) {
|
||||||
|
throw new ValidationError([], 'Checksum is required.');
|
||||||
|
}
|
||||||
|
|
||||||
|
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, logger);
|
||||||
|
if (existingFlyer) {
|
||||||
|
throw new DuplicateFlyerError('This flyer has already been processed.', existingFlyer.flyer_id);
|
||||||
|
}
|
||||||
|
|
||||||
|
const originalFileName = parsed.originalFileName ?? parsed?.data?.originalFileName ?? file.originalname;
|
||||||
|
|
||||||
|
if (!extractedData || typeof extractedData !== 'object') {
|
||||||
|
logger.warn({ bodyData: parsed }, 'Missing extractedData in legacy payload.');
|
||||||
|
extractedData = {};
|
||||||
|
}
|
||||||
|
|
||||||
|
const rawItems = extractedData.items ?? [];
|
||||||
|
const itemsArray = Array.isArray(rawItems) ? rawItems : typeof rawItems === 'string' ? JSON.parse(rawItems) : [];
|
||||||
|
const itemsForDb = itemsArray.map((item: Partial<ExtractedFlyerItem>) => ({
|
||||||
|
...item,
|
||||||
|
master_item_id: item.master_item_id === null ? undefined : item.master_item_id,
|
||||||
|
quantity: item.quantity ?? 1,
|
||||||
|
view_count: 0,
|
||||||
|
click_count: 0,
|
||||||
|
updated_at: new Date().toISOString(),
|
||||||
|
}));
|
||||||
|
|
||||||
|
const storeName = extractedData.store_name && String(extractedData.store_name).trim().length > 0 ? String(extractedData.store_name) : 'Unknown Store (auto)';
|
||||||
|
if (storeName.startsWith('Unknown')) {
|
||||||
|
logger.warn('extractedData.store_name missing; using fallback store name.');
|
||||||
|
}
|
||||||
|
|
||||||
|
const iconsDir = path.join(path.dirname(file.path), 'icons');
|
||||||
|
const iconFileName = await generateFlyerIcon(file.path, iconsDir, logger);
|
||||||
|
const iconUrl = `/flyer-images/icons/${iconFileName}`;
|
||||||
|
|
||||||
|
const flyerData: FlyerInsert = {
|
||||||
|
file_name: originalFileName,
|
||||||
|
image_url: `/flyer-images/${file.filename}`,
|
||||||
|
icon_url: iconUrl,
|
||||||
|
checksum: checksum,
|
||||||
|
store_name: storeName,
|
||||||
|
valid_from: extractedData.valid_from ?? null,
|
||||||
|
valid_to: extractedData.valid_to ?? null,
|
||||||
|
store_address: extractedData.store_address ?? null,
|
||||||
|
item_count: 0,
|
||||||
|
status: 'needs_review',
|
||||||
|
uploaded_by: userProfile?.user.user_id,
|
||||||
|
};
|
||||||
|
|
||||||
|
const { flyer: newFlyer, items: newItems } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
||||||
|
|
||||||
|
logger.info(`Successfully processed legacy flyer: ${newFlyer.file_name} (ID: ${newFlyer.flyer_id}) with ${newItems.length} items.`);
|
||||||
|
|
||||||
|
await db.adminRepo.logActivity({
|
||||||
|
userId: userProfile?.user.user_id,
|
||||||
|
action: 'flyer_processed',
|
||||||
|
displayText: `Processed a new flyer for ${flyerData.store_name}.`,
|
||||||
|
details: { flyerId: newFlyer.flyer_id, storeName: flyerData.store_name },
|
||||||
|
}, logger);
|
||||||
|
|
||||||
|
return newFlyer;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Export a singleton instance of the service for use throughout the application.
|
// Export a singleton instance of the service for use throughout the application.
|
||||||
|
|||||||
221
src/services/authService.ts
Normal file
221
src/services/authService.ts
Normal file
@@ -0,0 +1,221 @@
|
|||||||
|
// src/services/authService.ts
|
||||||
|
import * as bcrypt from 'bcrypt';
|
||||||
|
import jwt from 'jsonwebtoken';
|
||||||
|
import crypto from 'crypto';
|
||||||
|
import { userRepo, adminRepo } from './db/index.db';
|
||||||
|
import { UniqueConstraintError } from './db/errors.db';
|
||||||
|
import { getPool } from './db/connection.db';
|
||||||
|
import { logger } from './logger.server';
|
||||||
|
import { sendPasswordResetEmail } from './emailService.server';
|
||||||
|
import type { UserProfile } from '../types';
|
||||||
|
import { validatePasswordStrength } from '../utils/authUtils';
|
||||||
|
|
||||||
|
const JWT_SECRET = process.env.JWT_SECRET!;
|
||||||
|
|
||||||
|
class AuthService {
|
||||||
|
async registerUser(
|
||||||
|
email: string,
|
||||||
|
password: string,
|
||||||
|
fullName: string | undefined,
|
||||||
|
avatarUrl: string | undefined,
|
||||||
|
reqLog: any,
|
||||||
|
) {
|
||||||
|
try {
|
||||||
|
const saltRounds = 10;
|
||||||
|
const hashedPassword = await bcrypt.hash(password, saltRounds);
|
||||||
|
logger.info(`Hashing password for new user: ${email}`);
|
||||||
|
|
||||||
|
// The createUser method in UserRepository now handles its own transaction.
|
||||||
|
const newUser = await userRepo.createUser(
|
||||||
|
email,
|
||||||
|
hashedPassword,
|
||||||
|
{ full_name: fullName, avatar_url: avatarUrl },
|
||||||
|
reqLog,
|
||||||
|
);
|
||||||
|
|
||||||
|
const userEmail = newUser.user.email;
|
||||||
|
const userId = newUser.user.user_id;
|
||||||
|
logger.info(`Successfully created new user in DB: ${userEmail} (ID: ${userId})`);
|
||||||
|
|
||||||
|
// Use the new standardized logging function
|
||||||
|
await adminRepo.logActivity(
|
||||||
|
{
|
||||||
|
userId: newUser.user.user_id,
|
||||||
|
action: 'user_registered',
|
||||||
|
displayText: `${userEmail} has registered.`,
|
||||||
|
icon: 'user-plus',
|
||||||
|
},
|
||||||
|
reqLog,
|
||||||
|
);
|
||||||
|
|
||||||
|
return newUser;
|
||||||
|
} catch (error: unknown) {
|
||||||
|
if (error instanceof UniqueConstraintError) {
|
||||||
|
// If the email is a duplicate, return a 409 Conflict status.
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
logger.error({ error }, `User registration route failed for email: ${email}.`);
|
||||||
|
// Pass the error to the centralized handler
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async registerAndLoginUser(
|
||||||
|
email: string,
|
||||||
|
password: string,
|
||||||
|
fullName: string | undefined,
|
||||||
|
avatarUrl: string | undefined,
|
||||||
|
reqLog: any,
|
||||||
|
): Promise<{ newUserProfile: UserProfile; accessToken: string; refreshToken: string }> {
|
||||||
|
const newUserProfile = await this.registerUser(
|
||||||
|
email,
|
||||||
|
password,
|
||||||
|
fullName,
|
||||||
|
avatarUrl,
|
||||||
|
reqLog,
|
||||||
|
);
|
||||||
|
const { accessToken, refreshToken } = await this.handleSuccessfulLogin(newUserProfile, reqLog);
|
||||||
|
return { newUserProfile, accessToken, refreshToken };
|
||||||
|
}
|
||||||
|
|
||||||
|
generateAuthTokens(userProfile: UserProfile) {
|
||||||
|
const payload = {
|
||||||
|
user_id: userProfile.user.user_id,
|
||||||
|
email: userProfile.user.email,
|
||||||
|
role: userProfile.role,
|
||||||
|
};
|
||||||
|
const accessToken = jwt.sign(payload, JWT_SECRET, { expiresIn: '15m' });
|
||||||
|
const refreshToken = crypto.randomBytes(64).toString('hex');
|
||||||
|
|
||||||
|
return { accessToken, refreshToken };
|
||||||
|
}
|
||||||
|
|
||||||
|
async saveRefreshToken(userId: string, refreshToken: string, reqLog: any) {
|
||||||
|
try {
|
||||||
|
await userRepo.saveRefreshToken(userId, refreshToken, reqLog);
|
||||||
|
} catch (tokenErr) {
|
||||||
|
logger.error(
|
||||||
|
{ error: tokenErr },
|
||||||
|
`Failed to save refresh token during login for user: ${userId}`,
|
||||||
|
);
|
||||||
|
throw tokenErr;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async handleSuccessfulLogin(userProfile: UserProfile, reqLog: any) {
|
||||||
|
const { accessToken, refreshToken } = this.generateAuthTokens(userProfile);
|
||||||
|
await this.saveRefreshToken(userProfile.user.user_id, refreshToken, reqLog);
|
||||||
|
return { accessToken, refreshToken };
|
||||||
|
}
|
||||||
|
|
||||||
|
async resetPassword(email: string, reqLog: any) {
|
||||||
|
try {
|
||||||
|
logger.debug(`[API /forgot-password] Received request for email: ${email}`);
|
||||||
|
const user = await userRepo.findUserByEmail(email, reqLog);
|
||||||
|
let token: string | undefined;
|
||||||
|
logger.debug(
|
||||||
|
{ user: user ? { user_id: user.user_id, email: user.email } : 'NOT FOUND' },
|
||||||
|
`[API /forgot-password] Database search result for ${email}:`,
|
||||||
|
);
|
||||||
|
|
||||||
|
if (user) {
|
||||||
|
token = crypto.randomBytes(32).toString('hex');
|
||||||
|
const saltRounds = 10;
|
||||||
|
const tokenHash = await bcrypt.hash(token, saltRounds);
|
||||||
|
const expiresAt = new Date(Date.now() + 3600000); // 1 hour
|
||||||
|
|
||||||
|
await userRepo.createPasswordResetToken(user.user_id, tokenHash, expiresAt, reqLog);
|
||||||
|
|
||||||
|
const resetLink = `${process.env.FRONTEND_URL}/reset-password/${token}`;
|
||||||
|
|
||||||
|
try {
|
||||||
|
await sendPasswordResetEmail(email, resetLink, reqLog);
|
||||||
|
} catch (emailError) {
|
||||||
|
logger.error({ emailError }, `Email send failure during password reset for user`);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
logger.warn(`Password reset requested for non-existent email: ${email}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
return token;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error({ error }, `An error occurred during /forgot-password for email: ${email}`);
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async updatePassword(token: string, newPassword: string, reqLog: any) {
|
||||||
|
try {
|
||||||
|
const validTokens = await userRepo.getValidResetTokens(reqLog);
|
||||||
|
let tokenRecord;
|
||||||
|
for (const record of validTokens) {
|
||||||
|
const isMatch = await bcrypt.compare(token, record.token_hash);
|
||||||
|
if (isMatch) {
|
||||||
|
tokenRecord = record;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!tokenRecord) {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
const saltRounds = 10;
|
||||||
|
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
||||||
|
|
||||||
|
await userRepo.updateUserPassword(tokenRecord.user_id, hashedPassword, reqLog);
|
||||||
|
await userRepo.deleteResetToken(tokenRecord.token_hash, reqLog);
|
||||||
|
|
||||||
|
// Log this security event after a successful password reset.
|
||||||
|
await adminRepo.logActivity(
|
||||||
|
{
|
||||||
|
userId: tokenRecord.user_id,
|
||||||
|
action: 'password_reset',
|
||||||
|
displayText: `User ID ${tokenRecord.user_id} has reset their password.`,
|
||||||
|
icon: 'key',
|
||||||
|
details: { source_ip: null },
|
||||||
|
},
|
||||||
|
reqLog,
|
||||||
|
);
|
||||||
|
|
||||||
|
return true;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error({ error }, `An error occurred during password reset.`);
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async getUserByRefreshToken(refreshToken: string, reqLog: any) {
|
||||||
|
try {
|
||||||
|
const basicUser = await userRepo.findUserByRefreshToken(refreshToken, reqLog);
|
||||||
|
if (!basicUser) {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
const userProfile = await userRepo.findUserProfileById(basicUser.user_id, reqLog);
|
||||||
|
return userProfile;
|
||||||
|
} catch (error) {
|
||||||
|
logger.error({ error }, 'An error occurred during /refresh-token.');
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async logout(refreshToken: string, reqLog: any) {
|
||||||
|
try {
|
||||||
|
await userRepo.deleteRefreshToken(refreshToken, reqLog);
|
||||||
|
} catch (err: any) {
|
||||||
|
logger.error({ error: err }, 'Failed to delete refresh token from DB during logout.');
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async refreshAccessToken(refreshToken: string, reqLog: any): Promise<{ accessToken: string } | null> {
|
||||||
|
const user = await this.getUserByRefreshToken(refreshToken, reqLog);
|
||||||
|
if (!user) {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
const { accessToken } = this.generateAuthTokens(user);
|
||||||
|
return { accessToken };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export const authService = new AuthService();
|
||||||
@@ -335,8 +335,14 @@ describe('Background Job Service', () => {
|
|||||||
// Use fake timers to control promise resolution
|
// Use fake timers to control promise resolution
|
||||||
vi.useFakeTimers();
|
vi.useFakeTimers();
|
||||||
|
|
||||||
|
// Create a controllable promise
|
||||||
|
let resolveRun!: () => void;
|
||||||
|
const runPromise = new Promise<void>((resolve) => {
|
||||||
|
resolveRun = resolve;
|
||||||
|
});
|
||||||
|
|
||||||
// Make the first call hang indefinitely
|
// Make the first call hang indefinitely
|
||||||
vi.mocked(mockBackgroundJobService.runDailyDealCheck).mockReturnValue(new Promise(() => {}));
|
vi.mocked(mockBackgroundJobService.runDailyDealCheck).mockReturnValue(runPromise);
|
||||||
|
|
||||||
startBackgroundJobs(
|
startBackgroundJobs(
|
||||||
mockBackgroundJobService,
|
mockBackgroundJobService,
|
||||||
@@ -352,6 +358,9 @@ describe('Background Job Service', () => {
|
|||||||
// Trigger it a second time immediately
|
// Trigger it a second time immediately
|
||||||
const secondCall = dailyDealCheckCallback();
|
const secondCall = dailyDealCheckCallback();
|
||||||
|
|
||||||
|
// Resolve the first call so the test can finish
|
||||||
|
resolveRun();
|
||||||
|
|
||||||
await Promise.all([firstCall, secondCall]);
|
await Promise.all([firstCall, secondCall]);
|
||||||
|
|
||||||
// The service method should only have been called once
|
// The service method should only have been called once
|
||||||
@@ -362,12 +371,18 @@ describe('Background Job Service', () => {
|
|||||||
// Use fake timers to control promise resolution
|
// Use fake timers to control promise resolution
|
||||||
vi.useFakeTimers();
|
vi.useFakeTimers();
|
||||||
|
|
||||||
|
// Create a controllable promise
|
||||||
|
let resolveRun!: () => void;
|
||||||
|
const runPromise = new Promise<void>((resolve) => {
|
||||||
|
resolveRun = resolve;
|
||||||
|
});
|
||||||
|
|
||||||
// Make the first call hang indefinitely to keep the lock active
|
// Make the first call hang indefinitely to keep the lock active
|
||||||
vi.mocked(mockBackgroundJobService.runDailyDealCheck).mockReturnValue(new Promise(() => {}));
|
vi.mocked(mockBackgroundJobService.runDailyDealCheck).mockReturnValue(runPromise);
|
||||||
|
|
||||||
// Make logger.warn throw an error. This is outside the main try/catch in the cron job.
|
// Make logger.warn throw an error. This is outside the main try/catch in the cron job.
|
||||||
const warnError = new Error('Logger warn failed');
|
const warnError = new Error('Logger warn failed');
|
||||||
vi.mocked(globalMockLogger.warn).mockImplementation(() => {
|
vi.mocked(globalMockLogger.warn).mockImplementationOnce(() => {
|
||||||
throw warnError;
|
throw warnError;
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -382,7 +397,13 @@ describe('Background Job Service', () => {
|
|||||||
|
|
||||||
// Trigger the job once, it will hang and set the lock. Then trigger it a second time
|
// Trigger the job once, it will hang and set the lock. Then trigger it a second time
|
||||||
// to enter the `if (isDailyDealCheckRunning)` block and call the throwing logger.warn.
|
// to enter the `if (isDailyDealCheckRunning)` block and call the throwing logger.warn.
|
||||||
await Promise.allSettled([dailyDealCheckCallback(), dailyDealCheckCallback()]);
|
const firstCall = dailyDealCheckCallback();
|
||||||
|
const secondCall = dailyDealCheckCallback();
|
||||||
|
|
||||||
|
// Resolve the first call so the test can finish
|
||||||
|
resolveRun();
|
||||||
|
|
||||||
|
await Promise.allSettled([firstCall, secondCall]);
|
||||||
|
|
||||||
// The outer catch block should have been called with the error from logger.warn
|
// The outer catch block should have been called with the error from logger.warn
|
||||||
expect(globalMockLogger.error).toHaveBeenCalledWith(
|
expect(globalMockLogger.error).toHaveBeenCalledWith(
|
||||||
|
|||||||
@@ -7,6 +7,7 @@ import { getSimpleWeekAndYear } from '../utils/dateUtils';
|
|||||||
// Import types for repositories from their source files
|
// Import types for repositories from their source files
|
||||||
import type { PersonalizationRepository } from './db/personalization.db';
|
import type { PersonalizationRepository } from './db/personalization.db';
|
||||||
import type { NotificationRepository } from './db/notification.db';
|
import type { NotificationRepository } from './db/notification.db';
|
||||||
|
import { analyticsQueue, weeklyAnalyticsQueue } from './queueService.server';
|
||||||
|
|
||||||
interface EmailJobData {
|
interface EmailJobData {
|
||||||
to: string;
|
to: string;
|
||||||
@@ -23,6 +24,24 @@ export class BackgroundJobService {
|
|||||||
private logger: Logger,
|
private logger: Logger,
|
||||||
) {}
|
) {}
|
||||||
|
|
||||||
|
public async triggerAnalyticsReport(): Promise<string> {
|
||||||
|
const reportDate = new Date().toISOString().split('T')[0]; // YYYY-MM-DD
|
||||||
|
const jobId = `manual-report-${reportDate}-${Date.now()}`;
|
||||||
|
const job = await analyticsQueue.add('generate-daily-report', { reportDate }, { jobId });
|
||||||
|
return job.id!;
|
||||||
|
}
|
||||||
|
|
||||||
|
public async triggerWeeklyAnalyticsReport(): Promise<string> {
|
||||||
|
const { year: reportYear, week: reportWeek } = getSimpleWeekAndYear();
|
||||||
|
const jobId = `manual-weekly-report-${reportYear}-${reportWeek}-${Date.now()}`;
|
||||||
|
const job = await weeklyAnalyticsQueue.add(
|
||||||
|
'generate-weekly-report',
|
||||||
|
{ reportYear, reportWeek },
|
||||||
|
{ jobId },
|
||||||
|
);
|
||||||
|
return job.id!;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Prepares the data for an email notification job based on a user's deals.
|
* Prepares the data for an email notification job based on a user's deals.
|
||||||
* @param user The user to whom the email will be sent.
|
* @param user The user to whom the email will be sent.
|
||||||
|
|||||||
13
src/services/brandService.ts
Normal file
13
src/services/brandService.ts
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
// src/services/brandService.ts
|
||||||
|
import * as db from './db/index.db';
|
||||||
|
import type { Logger } from 'pino';
|
||||||
|
|
||||||
|
class BrandService {
|
||||||
|
async updateBrandLogo(brandId: number, file: Express.Multer.File, logger: Logger): Promise<string> {
|
||||||
|
const logoUrl = `/flyer-images/${file.filename}`;
|
||||||
|
await db.adminRepo.updateBrandLogo(brandId, logoUrl, logger);
|
||||||
|
return logoUrl;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export const brandService = new BrandService();
|
||||||
@@ -37,15 +37,9 @@ import { withTransaction } from './connection.db';
|
|||||||
|
|
||||||
describe('Flyer DB Service', () => {
|
describe('Flyer DB Service', () => {
|
||||||
let flyerRepo: FlyerRepository;
|
let flyerRepo: FlyerRepository;
|
||||||
const mockDb = {
|
|
||||||
query: vi.fn(),
|
|
||||||
};
|
|
||||||
|
|
||||||
beforeEach(() => {
|
beforeEach(() => {
|
||||||
vi.clearAllMocks();
|
vi.clearAllMocks();
|
||||||
mockDb.query.mockReset()
|
|
||||||
|
|
||||||
flyerRepo = new FlyerRepository(mockDb);
|
|
||||||
//In a transaction, `pool.connect()` returns a client. That client has a `release` method.
|
//In a transaction, `pool.connect()` returns a client. That client has a `release` method.
|
||||||
// For these tests, we simulate this by having `connect` resolve to the pool instance itself,
|
// For these tests, we simulate this by having `connect` resolve to the pool instance itself,
|
||||||
// and we ensure the `release` method is mocked on that instance.
|
// and we ensure the `release` method is mocked on that instance.
|
||||||
@@ -57,10 +51,10 @@ describe('Flyer DB Service', () => {
|
|||||||
|
|
||||||
describe('findOrCreateStore', () => {
|
describe('findOrCreateStore', () => {
|
||||||
it('should find an existing store and return its ID', async () => {
|
it('should find an existing store and return its ID', async () => {
|
||||||
mockDb.query.mockResolvedValue({ rows: [{ store_id: 1 }] });
|
mockPoolInstance.query.mockResolvedValue({ rows: [{ store_id: 1 }] });
|
||||||
const result = await flyerRepo.findOrCreateStore('Existing Store', mockLogger);
|
const result = await flyerRepo.findOrCreateStore('Existing Store', mockLogger);
|
||||||
expect(result).toBe(1);
|
expect(result).toBe(1);
|
||||||
expect(mockDb.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('SELECT store_id FROM public.stores WHERE name = $1'),
|
expect.stringContaining('SELECT store_id FROM public.stores WHERE name = $1'),
|
||||||
['Existing Store'],
|
['Existing Store'],
|
||||||
);
|
);
|
||||||
@@ -72,7 +66,7 @@ describe('Flyer DB Service', () => {
|
|||||||
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] })
|
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] })
|
||||||
const result = await flyerRepo.findOrCreateStore('New Store', mockLogger);
|
const result = await flyerRepo.findOrCreateStore('New Store', mockLogger);
|
||||||
expect(result).toBe(2);
|
expect(result).toBe(2);
|
||||||
expect(mockDb.query).toHaveBeenCalledWith(
|
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id'),
|
expect.stringContaining('INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id'),
|
||||||
['New Store'],
|
['New Store'],
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -87,7 +87,7 @@ describe('Gamification DB Service', () => {
|
|||||||
|
|
||||||
const result = await gamificationRepo.getUserAchievements('user-123', mockLogger);
|
const result = await gamificationRepo.getUserAchievements('user-123', mockLogger);
|
||||||
|
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockDb.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('FROM public.user_achievements ua'),
|
expect.stringContaining('FROM public.user_achievements ua'),
|
||||||
['user-123'],
|
['user-123'],
|
||||||
);
|
);
|
||||||
@@ -157,8 +157,8 @@ describe('Gamification DB Service', () => {
|
|||||||
mockDb.query.mockResolvedValue({ rows: mockLeaderboard });
|
mockDb.query.mockResolvedValue({ rows: mockLeaderboard });
|
||||||
|
|
||||||
const result = await gamificationRepo.getLeaderboard(10, mockLogger);
|
const result = await gamificationRepo.getLeaderboard(10, mockLogger);
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(1);
|
expect(mockDb.query).toHaveBeenCalledTimes(1);
|
||||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
expect(mockDb.query).toHaveBeenCalledWith(
|
||||||
expect.stringContaining('RANK() OVER (ORDER BY points DESC)'),
|
expect.stringContaining('RANK() OVER (ORDER BY points DESC)'),
|
||||||
[10],
|
[10],
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -249,6 +249,12 @@ describe('FlyerProcessingService', () => {
|
|||||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||||
errorCode: 'UNKNOWN_ERROR',
|
errorCode: 'UNKNOWN_ERROR',
|
||||||
message: 'AI model exploded',
|
message: 'AI model exploded',
|
||||||
|
stages: [
|
||||||
|
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
||||||
|
{ name: 'Extracting Data with AI', status: 'failed', critical: true, detail: 'AI model exploded' },
|
||||||
|
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
||||||
|
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||||
|
],
|
||||||
}); // This was a duplicate, fixed.
|
}); // This was a duplicate, fixed.
|
||||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||||
expect(logger.warn).toHaveBeenCalledWith(
|
expect(logger.warn).toHaveBeenCalledWith(
|
||||||
@@ -268,6 +274,12 @@ describe('FlyerProcessingService', () => {
|
|||||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||||
errorCode: 'QUOTA_EXCEEDED',
|
errorCode: 'QUOTA_EXCEEDED',
|
||||||
message: 'An AI quota has been exceeded. Please try again later.',
|
message: 'An AI quota has been exceeded. Please try again later.',
|
||||||
|
stages: [
|
||||||
|
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
||||||
|
{ name: 'Extracting Data with AI', status: 'failed', critical: true, detail: 'An AI quota has been exceeded. Please try again later.' },
|
||||||
|
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
||||||
|
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||||
|
],
|
||||||
});
|
});
|
||||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||||
expect(logger.warn).toHaveBeenCalledWith(
|
expect(logger.warn).toHaveBeenCalledWith(
|
||||||
@@ -291,7 +303,7 @@ describe('FlyerProcessingService', () => {
|
|||||||
stderr: 'pdftocairo error',
|
stderr: 'pdftocairo error',
|
||||||
stages: [
|
stages: [
|
||||||
{ name: 'Preparing Inputs', status: 'failed', critical: true, detail: 'Validating and preparing file...' },
|
{ name: 'Preparing Inputs', status: 'failed', critical: true, detail: 'Validating and preparing file...' },
|
||||||
{ name: 'Extracting Data with AI', status: 'skipped', critical: true },
|
{ name: 'Extracting Data with AI', status: 'skipped', critical: true, detail: 'Communicating with AI model...' },
|
||||||
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
||||||
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||||
],
|
],
|
||||||
@@ -312,7 +324,14 @@ describe('FlyerProcessingService', () => {
|
|||||||
|
|
||||||
// Verify the specific error handling logic in the catch block
|
// Verify the specific error handling logic in the catch block
|
||||||
expect(logger.error).toHaveBeenCalledWith(
|
expect(logger.error).toHaveBeenCalledWith(
|
||||||
{ err: validationError, validationErrors: {}, rawData: {} },
|
{
|
||||||
|
err: validationError,
|
||||||
|
errorCode: 'AI_VALIDATION_FAILED',
|
||||||
|
message: "The AI couldn't read the flyer's format. Please try a clearer image or a different flyer.",
|
||||||
|
validationErrors: {},
|
||||||
|
rawData: {},
|
||||||
|
stages: expect.any(Array), // Stages will be dynamically generated
|
||||||
|
},
|
||||||
'AI Data Validation failed.',
|
'AI Data Validation failed.',
|
||||||
);
|
);
|
||||||
// Use `toHaveBeenLastCalledWith` to check only the final error payload.
|
// Use `toHaveBeenLastCalledWith` to check only the final error payload.
|
||||||
@@ -325,7 +344,7 @@ describe('FlyerProcessingService', () => {
|
|||||||
rawData: {},
|
rawData: {},
|
||||||
stages: [
|
stages: [
|
||||||
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
||||||
{ name: 'Extracting Data with AI', status: 'failed', critical: true, detail: 'Communicating with AI model...' },
|
{ name: 'Extracting Data with AI', status: 'failed', critical: true, detail: "The AI couldn't read the flyer's format. Please try a clearer image or a different flyer." },
|
||||||
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
||||||
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||||
],
|
],
|
||||||
@@ -368,6 +387,12 @@ describe('FlyerProcessingService', () => {
|
|||||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||||
errorCode: 'UNKNOWN_ERROR',
|
errorCode: 'UNKNOWN_ERROR',
|
||||||
message: 'Database transaction failed',
|
message: 'Database transaction failed',
|
||||||
|
stages: [
|
||||||
|
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
||||||
|
{ name: 'Extracting Data with AI', status: 'completed', critical: true },
|
||||||
|
{ name: 'Transforming AI Data', status: 'completed', critical: true },
|
||||||
|
{ name: 'Saving to Database', status: 'failed', critical: true, detail: 'Database transaction failed' },
|
||||||
|
],
|
||||||
}); // This was a duplicate, fixed.
|
}); // This was a duplicate, fixed.
|
||||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||||
expect(logger.warn).toHaveBeenCalledWith(
|
expect(logger.warn).toHaveBeenCalledWith(
|
||||||
@@ -388,6 +413,12 @@ describe('FlyerProcessingService', () => {
|
|||||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||||
errorCode: 'UNSUPPORTED_FILE_TYPE',
|
errorCode: 'UNSUPPORTED_FILE_TYPE',
|
||||||
message: 'Unsupported file type: .txt. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.',
|
message: 'Unsupported file type: .txt. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.',
|
||||||
|
stages: [
|
||||||
|
{ name: 'Preparing Inputs', status: 'failed', critical: true, detail: 'Unsupported file type: .txt. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.' },
|
||||||
|
{ name: 'Extracting Data with AI', status: 'skipped', critical: true, detail: 'Communicating with AI model...' },
|
||||||
|
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
||||||
|
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||||
|
],
|
||||||
});
|
});
|
||||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||||
expect(logger.warn).toHaveBeenCalledWith(
|
expect(logger.warn).toHaveBeenCalledWith(
|
||||||
@@ -409,6 +440,12 @@ describe('FlyerProcessingService', () => {
|
|||||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||||
errorCode: 'UNKNOWN_ERROR',
|
errorCode: 'UNKNOWN_ERROR',
|
||||||
message: 'Icon generation failed.',
|
message: 'Icon generation failed.',
|
||||||
|
stages: [
|
||||||
|
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
||||||
|
{ name: 'Extracting Data with AI', status: 'completed', critical: true },
|
||||||
|
{ name: 'Transforming AI Data', status: 'failed', critical: true, detail: 'Icon generation failed.' },
|
||||||
|
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||||
|
],
|
||||||
}); // This was a duplicate, fixed.
|
}); // This was a duplicate, fixed.
|
||||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||||
expect(logger.warn).toHaveBeenCalledWith(
|
expect(logger.warn).toHaveBeenCalledWith(
|
||||||
@@ -424,7 +461,7 @@ describe('FlyerProcessingService', () => {
|
|||||||
const quotaError = new Error('RESOURCE_EXHAUSTED');
|
const quotaError = new Error('RESOURCE_EXHAUSTED');
|
||||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||||
|
|
||||||
await expect(privateMethod(quotaError, job, logger)).rejects.toThrow(
|
await expect(privateMethod(quotaError, job, logger, [])).rejects.toThrow(
|
||||||
UnrecoverableError,
|
UnrecoverableError,
|
||||||
);
|
);
|
||||||
|
|
||||||
@@ -444,7 +481,7 @@ describe('FlyerProcessingService', () => {
|
|||||||
);
|
);
|
||||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||||
|
|
||||||
await expect(privateMethod(validationError, job, logger)).rejects.toThrow(
|
await expect(privateMethod(validationError, job, logger, [])).rejects.toThrow(
|
||||||
validationError,
|
validationError,
|
||||||
);
|
);
|
||||||
|
|
||||||
@@ -464,7 +501,7 @@ describe('FlyerProcessingService', () => {
|
|||||||
const genericError = new Error('A standard failure');
|
const genericError = new Error('A standard failure');
|
||||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||||
|
|
||||||
await expect(privateMethod(genericError, job, logger)).rejects.toThrow(genericError);
|
await expect(privateMethod(genericError, job, logger, [])).rejects.toThrow(genericError);
|
||||||
|
|
||||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||||
errorCode: 'UNKNOWN_ERROR',
|
errorCode: 'UNKNOWN_ERROR',
|
||||||
@@ -478,7 +515,7 @@ describe('FlyerProcessingService', () => {
|
|||||||
const nonError = 'just a string error';
|
const nonError = 'just a string error';
|
||||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||||
|
|
||||||
await expect(privateMethod(nonError, job, logger)).rejects.toThrow('just a string error');
|
await expect(privateMethod(nonError, job, logger, [])).rejects.toThrow('just a string error');
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
@@ -8,10 +8,23 @@ import type * as Db from './db/index.db';
|
|||||||
import type { AdminRepository } from './db/admin.db';
|
import type { AdminRepository } from './db/admin.db';
|
||||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||||
import type { FlyerJobData, CleanupJobData } from '../types/job-data';
|
import type { FlyerJobData, CleanupJobData } from '../types/job-data';
|
||||||
import { FlyerProcessingError } from './processingErrors';
|
import {
|
||||||
|
FlyerProcessingError,
|
||||||
|
PdfConversionError,
|
||||||
|
AiDataValidationError,
|
||||||
|
UnsupportedFileTypeError,
|
||||||
|
} from './processingErrors';
|
||||||
import { createFlyerAndItems } from './db/flyer.db';
|
import { createFlyerAndItems } from './db/flyer.db';
|
||||||
import { logger as globalLogger } from './logger.server';
|
import { logger as globalLogger } from './logger.server';
|
||||||
|
|
||||||
|
// Define ProcessingStage locally as it's not exported from the types file.
|
||||||
|
export type ProcessingStage = {
|
||||||
|
name: string;
|
||||||
|
status: 'pending' | 'in-progress' | 'completed' | 'failed' | 'skipped';
|
||||||
|
critical: boolean;
|
||||||
|
detail?: string;
|
||||||
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This service orchestrates the entire flyer processing workflow. It's responsible for
|
* This service orchestrates the entire flyer processing workflow. It's responsible for
|
||||||
* coordinating various sub-services (file handling, AI processing, data transformation,
|
* coordinating various sub-services (file handling, AI processing, data transformation,
|
||||||
@@ -42,27 +55,43 @@ export class FlyerProcessingService {
|
|||||||
const logger = globalLogger.child({ jobId: job.id, jobName: job.name, ...job.data });
|
const logger = globalLogger.child({ jobId: job.id, jobName: job.name, ...job.data });
|
||||||
logger.info('Picked up flyer processing job.');
|
logger.info('Picked up flyer processing job.');
|
||||||
|
|
||||||
|
const stages: ProcessingStage[] = [
|
||||||
|
{ name: 'Preparing Inputs', status: 'pending', critical: true, detail: 'Validating and preparing file...' },
|
||||||
|
{ name: 'Extracting Data with AI', status: 'pending', critical: true, detail: 'Communicating with AI model...' },
|
||||||
|
{ name: 'Transforming AI Data', status: 'pending', critical: true },
|
||||||
|
{ name: 'Saving to Database', status: 'pending', critical: true },
|
||||||
|
];
|
||||||
|
|
||||||
// Keep track of all created file paths for eventual cleanup.
|
// Keep track of all created file paths for eventual cleanup.
|
||||||
const allFilePaths: string[] = [job.data.filePath];
|
const allFilePaths: string[] = [job.data.filePath];
|
||||||
|
|
||||||
try {
|
try {
|
||||||
// Stage 1: Prepare Inputs (e.g., convert PDF to images)
|
// Stage 1: Prepare Inputs (e.g., convert PDF to images)
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'in-progress', critical: true, detail: 'Validating and preparing file...' }] });
|
stages[0].status = 'in-progress';
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
const { imagePaths, createdImagePaths } = await this.fileHandler.prepareImageInputs(
|
const { imagePaths, createdImagePaths } = await this.fileHandler.prepareImageInputs(
|
||||||
job.data.filePath,
|
job.data.filePath,
|
||||||
job,
|
job,
|
||||||
logger,
|
logger,
|
||||||
);
|
);
|
||||||
allFilePaths.push(...createdImagePaths);
|
allFilePaths.push(...createdImagePaths);
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }] });
|
stages[0].status = 'completed';
|
||||||
|
stages[0].detail = `${imagePaths.length} page(s) ready for AI.`;
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
// Stage 2: Extract Data with AI
|
// Stage 2: Extract Data with AI
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'in-progress', critical: true, detail: 'Communicating with AI model...' }] });
|
stages[1].status = 'in-progress';
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
const aiResult = await this.aiProcessor.extractAndValidateData(imagePaths, job.data, logger);
|
const aiResult = await this.aiProcessor.extractAndValidateData(imagePaths, job.data, logger);
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }] });
|
stages[1].status = 'completed';
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
// Stage 3: Transform AI Data into DB format
|
// Stage 3: Transform AI Data into DB format
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'in-progress', critical: true }] });
|
stages[2].status = 'in-progress';
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
const { flyerData, itemsForDb } = await this.transformer.transform(
|
const { flyerData, itemsForDb } = await this.transformer.transform(
|
||||||
aiResult,
|
aiResult,
|
||||||
imagePaths,
|
imagePaths,
|
||||||
@@ -71,12 +100,16 @@ export class FlyerProcessingService {
|
|||||||
job.data.userId,
|
job.data.userId,
|
||||||
logger,
|
logger,
|
||||||
);
|
);
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'completed', critical: true }] });
|
stages[2].status = 'completed';
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
// Stage 4: Save to Database
|
// Stage 4: Save to Database
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'completed', critical: true }, { name: 'Saving to Database', status: 'in-progress', critical: true }] });
|
stages[3].status = 'in-progress';
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
const { flyer } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
const { flyer } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
||||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'completed', critical: true }, { name: 'Saving to Database', status: 'completed', critical: true }] });
|
stages[3].status = 'completed';
|
||||||
|
await job.updateProgress({ stages });
|
||||||
|
|
||||||
// Stage 5: Log Activity
|
// Stage 5: Log Activity
|
||||||
await this.db.adminRepo.logActivity(
|
await this.db.adminRepo.logActivity(
|
||||||
@@ -101,7 +134,7 @@ export class FlyerProcessingService {
|
|||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.warn('Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.');
|
logger.warn('Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.');
|
||||||
// This private method handles error reporting and re-throwing.
|
// This private method handles error reporting and re-throwing.
|
||||||
await this._reportErrorAndThrow(error, job, logger);
|
await this._reportErrorAndThrow(error, job, logger, stages);
|
||||||
// This line is technically unreachable because the above method always throws,
|
// This line is technically unreachable because the above method always throws,
|
||||||
// but it's required to satisfy TypeScript's control flow analysis.
|
// but it's required to satisfy TypeScript's control flow analysis.
|
||||||
throw error;
|
throw error;
|
||||||
@@ -158,22 +191,98 @@ export class FlyerProcessingService {
|
|||||||
* @param job The BullMQ job instance.
|
* @param job The BullMQ job instance.
|
||||||
* @param logger The logger instance.
|
* @param logger The logger instance.
|
||||||
*/
|
*/
|
||||||
private async _reportErrorAndThrow(error: unknown, job: Job, logger: Logger): Promise<never> {
|
private async _reportErrorAndThrow(
|
||||||
|
error: unknown,
|
||||||
|
job: Job,
|
||||||
|
logger: Logger,
|
||||||
|
initialStages: ProcessingStage[],
|
||||||
|
): Promise<never> {
|
||||||
const normalizedError = error instanceof Error ? error : new Error(String(error));
|
const normalizedError = error instanceof Error ? error : new Error(String(error));
|
||||||
let errorPayload: { errorCode: string; message: string; [key: string]: any };
|
let errorPayload: { errorCode: string; message: string; [key: string]: any };
|
||||||
|
let stagesToReport: ProcessingStage[] = [...initialStages]; // Create a mutable copy
|
||||||
|
|
||||||
if (normalizedError instanceof FlyerProcessingError) {
|
if (normalizedError instanceof FlyerProcessingError) {
|
||||||
errorPayload = normalizedError.toErrorPayload();
|
errorPayload = normalizedError.toErrorPayload();
|
||||||
logger.error({ err: normalizedError, ...errorPayload }, `A known processing error occurred: ${normalizedError.name}`);
|
|
||||||
|
// Determine which stage failed based on the error code
|
||||||
|
let errorStageIndex = -1;
|
||||||
|
if (normalizedError.errorCode === 'PDF_CONVERSION_FAILED' || normalizedError.errorCode === 'UNSUPPORTED_FILE_TYPE') {
|
||||||
|
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Preparing Inputs');
|
||||||
|
} else if (normalizedError.errorCode === 'AI_VALIDATION_FAILED') {
|
||||||
|
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Extracting Data with AI');
|
||||||
|
} else if (normalizedError.message.includes('Icon generation failed')) { // Specific message for transformer error
|
||||||
|
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Transforming AI Data');
|
||||||
|
} else if (normalizedError.message.includes('Database transaction failed')) { // Specific message for DB error
|
||||||
|
errorStageIndex = stagesToReport.findIndex(s => s.name === 'Saving to Database');
|
||||||
|
}
|
||||||
|
|
||||||
|
// If a specific stage is identified, update its status and subsequent stages
|
||||||
|
if (errorStageIndex !== -1) {
|
||||||
|
stagesToReport[errorStageIndex] = {
|
||||||
|
...stagesToReport[errorStageIndex],
|
||||||
|
status: 'failed',
|
||||||
|
detail: errorPayload.message, // Use the user-friendly message as detail
|
||||||
|
};
|
||||||
|
// Mark subsequent critical stages as skipped
|
||||||
|
for (let i = errorStageIndex + 1; i < stagesToReport.length; i++) {
|
||||||
|
if (stagesToReport[i].critical) {
|
||||||
|
stagesToReport[i] = { ...stagesToReport[i], status: 'skipped' };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// Fallback: if no specific stage is identified, mark the last stage as failed
|
||||||
|
if (stagesToReport.length > 0) {
|
||||||
|
const lastStageIndex = stagesToReport.length - 1;
|
||||||
|
stagesToReport[lastStageIndex] = {
|
||||||
|
...stagesToReport[lastStageIndex],
|
||||||
|
status: 'failed',
|
||||||
|
detail: errorPayload.message,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
errorPayload.stages = stagesToReport; // Add updated stages to the error payload
|
||||||
|
|
||||||
|
// For logging, explicitly include validationErrors and rawData if present
|
||||||
|
const logDetails: Record<string, any> = { err: normalizedError };
|
||||||
|
if (normalizedError instanceof AiDataValidationError) {
|
||||||
|
logDetails.validationErrors = normalizedError.validationErrors;
|
||||||
|
logDetails.rawData = normalizedError.rawData;
|
||||||
|
}
|
||||||
|
// Also include stderr for PdfConversionError in logs
|
||||||
|
if (normalizedError instanceof PdfConversionError) {
|
||||||
|
logDetails.stderr = normalizedError.stderr;
|
||||||
|
}
|
||||||
|
// Include the errorPayload details in the log, but avoid duplicating err, validationErrors, rawData
|
||||||
|
Object.assign(logDetails, errorPayload);
|
||||||
|
// Remove the duplicated err property if it was assigned by Object.assign
|
||||||
|
if ('err' in logDetails && logDetails.err === normalizedError) {
|
||||||
|
// This check prevents accidental deletion if 'err' was a legitimate property of errorPayload
|
||||||
|
delete logDetails.err;
|
||||||
|
}
|
||||||
|
// Ensure the original error object is always passed as 'err' for consistency in logging
|
||||||
|
logDetails.err = normalizedError;
|
||||||
|
|
||||||
|
logger.error(logDetails, `A known processing error occurred: ${normalizedError.name}`);
|
||||||
} else {
|
} else {
|
||||||
const message = normalizedError.message || 'An unknown error occurred.';
|
const message = normalizedError.message || 'An unknown error occurred.';
|
||||||
errorPayload = { errorCode: 'UNKNOWN_ERROR', message };
|
errorPayload = { errorCode: 'UNKNOWN_ERROR', message };
|
||||||
logger.error({ err: normalizedError }, `An unknown error occurred: ${message}`);
|
// For generic errors, if we have stages, mark the last one as failed
|
||||||
|
if (stagesToReport.length > 0) {
|
||||||
|
const lastStageIndex = stagesToReport.length - 1;
|
||||||
|
stagesToReport[lastStageIndex] = {
|
||||||
|
...stagesToReport[lastStageIndex],
|
||||||
|
status: 'failed',
|
||||||
|
detail: message
|
||||||
|
};
|
||||||
|
}
|
||||||
|
errorPayload.stages = stagesToReport; // Add stages to the error payload
|
||||||
|
logger.error({ err: normalizedError, ...errorPayload }, `An unknown error occurred: ${message}`);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for specific error messages that indicate a non-retriable failure, like quota exhaustion.
|
// Check for specific error messages that indicate a non-retriable failure, like quota exhaustion.
|
||||||
if (errorPayload.message.toLowerCase().includes('quota') || errorPayload.message.toLowerCase().includes('resource_exhausted')) {
|
if (errorPayload.message.toLowerCase().includes('quota') || errorPayload.message.toLowerCase().includes('resource_exhausted')) {
|
||||||
const unrecoverablePayload = { errorCode: 'QUOTA_EXCEEDED', message: 'An AI quota has been exceeded. Please try again later.' };
|
const unrecoverablePayload = { errorCode: 'QUOTA_EXCEEDED', message: 'An AI quota has been exceeded. Please try again later.', stages: errorPayload.stages };
|
||||||
await job.updateProgress(unrecoverablePayload);
|
await job.updateProgress(unrecoverablePayload);
|
||||||
throw new UnrecoverableError(unrecoverablePayload.message);
|
throw new UnrecoverableError(unrecoverablePayload.message);
|
||||||
}
|
}
|
||||||
|
|||||||
79
src/services/gamificationService.ts
Normal file
79
src/services/gamificationService.ts
Normal file
@@ -0,0 +1,79 @@
|
|||||||
|
// src/services/gamificationService.ts
|
||||||
|
|
||||||
|
import { gamificationRepo } from './db/index.db';
|
||||||
|
import { ForeignKeyConstraintError } from './db/errors.db';
|
||||||
|
import type { Logger } from 'pino';
|
||||||
|
|
||||||
|
class GamificationService {
|
||||||
|
/**
|
||||||
|
* Awards a specific achievement to a user.
|
||||||
|
* @param userId The ID of the user to award the achievement.
|
||||||
|
* @param achievementName The name of the achievement to award.
|
||||||
|
* @param log The logger instance.
|
||||||
|
*/
|
||||||
|
async awardAchievement(userId: string, achievementName: string, log: Logger): Promise<void> {
|
||||||
|
try {
|
||||||
|
await gamificationRepo.awardAchievement(userId, achievementName, log);
|
||||||
|
} catch (error) {
|
||||||
|
if (error instanceof ForeignKeyConstraintError) {
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
log.error(
|
||||||
|
{ error, userId, achievementName },
|
||||||
|
'Error awarding achievement via admin endpoint:',
|
||||||
|
);
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Retrieves the master list of all available achievements.
|
||||||
|
* @param log The logger instance.
|
||||||
|
*/
|
||||||
|
async getAllAchievements(log: Logger) {
|
||||||
|
try {
|
||||||
|
return await gamificationRepo.getAllAchievements(log);
|
||||||
|
} catch (error) {
|
||||||
|
log.error({ error }, 'Error in getAllAchievements service method');
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Retrieves the public leaderboard of top users by points.
|
||||||
|
* @param limit The number of users to fetch.
|
||||||
|
* @param log The logger instance.
|
||||||
|
*/
|
||||||
|
async getLeaderboard(limit: number, log: Logger) {
|
||||||
|
// The test failures point to an issue in the underlying repository method,
|
||||||
|
// where the database query is not being executed. This service method is a simple
|
||||||
|
// pass-through, so the root cause is likely in `gamification.db.ts`.
|
||||||
|
// Adding robust error handling here is a good practice regardless.
|
||||||
|
try {
|
||||||
|
return await gamificationRepo.getLeaderboard(limit, log);
|
||||||
|
} catch (error) {
|
||||||
|
log.error({ error, limit }, 'Error fetching leaderboard in service method.');
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Retrieves all achievements earned by a specific user.
|
||||||
|
* @param userId The ID of the user.
|
||||||
|
* @param log The logger instance.
|
||||||
|
*/
|
||||||
|
async getUserAchievements(userId: string, log: Logger) {
|
||||||
|
// The test failures point to an issue in the underlying repository method,
|
||||||
|
// where the database query is not being executed. This service method is a simple
|
||||||
|
// pass-through, so the root cause is likely in `gamification.db.ts`.
|
||||||
|
// Adding robust error handling here is a good practice regardless.
|
||||||
|
try {
|
||||||
|
return await gamificationRepo.getUserAchievements(userId, log);
|
||||||
|
} catch (error) {
|
||||||
|
log.error({ error, userId }, 'Error fetching user achievements in service method.');
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export const gamificationService = new GamificationService();
|
||||||
111
src/services/monitoringService.server.ts
Normal file
111
src/services/monitoringService.server.ts
Normal file
@@ -0,0 +1,111 @@
|
|||||||
|
// src/services/monitoringService.server.ts
|
||||||
|
import {
|
||||||
|
flyerQueue,
|
||||||
|
emailQueue,
|
||||||
|
analyticsQueue,
|
||||||
|
cleanupQueue,
|
||||||
|
weeklyAnalyticsQueue,
|
||||||
|
} from './queueService.server';
|
||||||
|
import {
|
||||||
|
analyticsWorker,
|
||||||
|
cleanupWorker,
|
||||||
|
emailWorker,
|
||||||
|
flyerWorker,
|
||||||
|
weeklyAnalyticsWorker,
|
||||||
|
} from './workers.server';
|
||||||
|
import type { Job, Queue } from 'bullmq';
|
||||||
|
import { NotFoundError, ValidationError } from './db/errors.db';
|
||||||
|
import { logger } from './logger.server';
|
||||||
|
|
||||||
|
class MonitoringService {
|
||||||
|
/**
|
||||||
|
* Retrieves the current running status of all registered BullMQ workers.
|
||||||
|
* @returns A promise that resolves to an array of worker statuses.
|
||||||
|
*/
|
||||||
|
async getWorkerStatuses() {
|
||||||
|
const workers = [flyerWorker, emailWorker, analyticsWorker, cleanupWorker, weeklyAnalyticsWorker];
|
||||||
|
return Promise.all(
|
||||||
|
workers.map(async (worker) => ({
|
||||||
|
name: worker.name,
|
||||||
|
isRunning: worker.isRunning(),
|
||||||
|
})),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Retrieves job counts for all registered BullMQ queues.
|
||||||
|
* @returns A promise that resolves to an array of queue statuses.
|
||||||
|
*/
|
||||||
|
async getQueueStatuses() {
|
||||||
|
const queues = [flyerQueue, emailQueue, analyticsQueue, cleanupQueue, weeklyAnalyticsQueue];
|
||||||
|
return Promise.all(
|
||||||
|
queues.map(async (queue) => ({
|
||||||
|
name: queue.name,
|
||||||
|
counts: await queue.getJobCounts(
|
||||||
|
'waiting',
|
||||||
|
'active',
|
||||||
|
'completed',
|
||||||
|
'failed',
|
||||||
|
'delayed',
|
||||||
|
'paused',
|
||||||
|
),
|
||||||
|
})),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Retries a specific failed job in a given queue.
|
||||||
|
* @param queueName The name of the queue.
|
||||||
|
* @param jobId The ID of the job to retry.
|
||||||
|
* @param userId The ID of the user initiating the retry.
|
||||||
|
*/
|
||||||
|
async retryFailedJob(queueName: string, jobId: string, userId: string) {
|
||||||
|
const queueMap: { [key: string]: Queue } = {
|
||||||
|
'flyer-processing': flyerQueue,
|
||||||
|
'email-sending': emailQueue,
|
||||||
|
'analytics-reporting': analyticsQueue,
|
||||||
|
'file-cleanup': cleanupQueue,
|
||||||
|
'weekly-analytics-reporting': weeklyAnalyticsQueue, // This was a duplicate, fixed.
|
||||||
|
};
|
||||||
|
|
||||||
|
const queue = queueMap[queueName];
|
||||||
|
if (!queue) {
|
||||||
|
throw new NotFoundError(`Queue '${queueName}' not found.`);
|
||||||
|
}
|
||||||
|
|
||||||
|
const job = await queue.getJob(jobId);
|
||||||
|
if (!job) {
|
||||||
|
throw new NotFoundError(`Job with ID '${jobId}' not found in queue '${queueName}'.`);
|
||||||
|
}
|
||||||
|
|
||||||
|
const jobState = await job.getState();
|
||||||
|
if (jobState !== 'failed') {
|
||||||
|
throw new ValidationError(
|
||||||
|
[],
|
||||||
|
`Job is not in a 'failed' state. Current state: ${jobState}.`,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
await job.retry();
|
||||||
|
logger.info(`[Admin] User ${userId} manually retried job ${jobId} in queue ${queueName}.`);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Retrieves the status of a single job from the flyer processing queue.
|
||||||
|
* @param jobId The ID of the job to retrieve.
|
||||||
|
* @returns A promise that resolves to a simplified job status object.
|
||||||
|
*/
|
||||||
|
async getFlyerJobStatus(jobId: string): Promise<{ id: string; state: string; progress: number | object | string | boolean; returnValue: any; failedReason: string | null; }> {
|
||||||
|
const job = await flyerQueue.getJob(jobId);
|
||||||
|
if (!job) {
|
||||||
|
throw new NotFoundError('Job not found.');
|
||||||
|
}
|
||||||
|
const state = await job.getState();
|
||||||
|
const progress = job.progress;
|
||||||
|
const returnValue = job.returnvalue;
|
||||||
|
const failedReason = job.failedReason;
|
||||||
|
return { id: job.id!, state, progress, returnValue, failedReason };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export const monitoringService = new MonitoringService();
|
||||||
43
src/services/systemService.ts
Normal file
43
src/services/systemService.ts
Normal file
@@ -0,0 +1,43 @@
|
|||||||
|
// src/services/systemService.ts
|
||||||
|
import { exec } from 'child_process';
|
||||||
|
import { promisify } from 'util';
|
||||||
|
import { logger } from './logger.server';
|
||||||
|
|
||||||
|
const execAsync = promisify(exec);
|
||||||
|
|
||||||
|
class SystemService {
|
||||||
|
async getPm2Status(): Promise<{ success: boolean; message: string }> {
|
||||||
|
try {
|
||||||
|
const { stdout, stderr } = await execAsync('pm2 describe flyer-crawler-api');
|
||||||
|
|
||||||
|
// If the command runs but produces output on stderr, treat it as an error.
|
||||||
|
// This handles cases where pm2 might issue warnings but still exit 0.
|
||||||
|
if (stderr) {
|
||||||
|
throw new Error(`PM2 command produced an error: ${stderr}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
const isOnline = /│\s*status\s*│\s*online\s*│/m.test(stdout);
|
||||||
|
const message = isOnline
|
||||||
|
? 'Application is online and running under PM2.'
|
||||||
|
: 'Application process exists but is not online.';
|
||||||
|
return { success: isOnline, message };
|
||||||
|
} catch (error: any) {
|
||||||
|
// If the command fails (non-zero exit code), check if it's because the process doesn't exist.
|
||||||
|
// This is a normal "not found" case, not a system error.
|
||||||
|
// The error message can be in stdout or stderr depending on the pm2 version.
|
||||||
|
const output = error.stdout || error.stderr || '';
|
||||||
|
if (output.includes("doesn't exist")) {
|
||||||
|
logger.warn('[SystemService] PM2 process "flyer-crawler-api" not found.');
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
message: 'Application process is not running under PM2.',
|
||||||
|
};
|
||||||
|
}
|
||||||
|
// For any other error, log it and re-throw to be handled as a 500.
|
||||||
|
logger.error({ error: error.stderr || error.message }, '[SystemService] Error executing pm2 describe:');
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export const systemService = new SystemService();
|
||||||
@@ -1,10 +1,12 @@
|
|||||||
// src/services/userService.ts
|
// src/services/userService.ts
|
||||||
import * as db from './db/index.db';
|
import * as db from './db/index.db';
|
||||||
import type { Job } from 'bullmq';
|
import type { Job } from 'bullmq';
|
||||||
|
import * as bcrypt from 'bcrypt';
|
||||||
import type { Logger } from 'pino';
|
import type { Logger } from 'pino';
|
||||||
import { AddressRepository } from './db/address.db';
|
import { AddressRepository } from './db/address.db';
|
||||||
import { UserRepository } from './db/user.db';
|
import { UserRepository } from './db/user.db';
|
||||||
import type { Address, UserProfile } from '../types';
|
import type { Address, Profile, UserProfile } from '../types';
|
||||||
|
import { ValidationError, NotFoundError } from './db/errors.db';
|
||||||
import { logger as globalLogger } from './logger.server';
|
import { logger as globalLogger } from './logger.server';
|
||||||
import type { TokenCleanupJobData } from '../types/job-data';
|
import type { TokenCleanupJobData } from '../types/job-data';
|
||||||
|
|
||||||
@@ -76,6 +78,90 @@ class UserService {
|
|||||||
throw wrappedError;
|
throw wrappedError;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Updates a user's avatar, creating the URL and updating the profile.
|
||||||
|
* @param userId The ID of the user to update.
|
||||||
|
* @param file The uploaded avatar file.
|
||||||
|
* @param logger The logger instance.
|
||||||
|
* @returns The updated user profile.
|
||||||
|
*/
|
||||||
|
async updateUserAvatar(userId: string, file: Express.Multer.File, logger: Logger): Promise<Profile> {
|
||||||
|
const avatarUrl = `/uploads/avatars/${file.filename}`;
|
||||||
|
return db.userRepo.updateUserProfile(
|
||||||
|
userId,
|
||||||
|
{ avatar_url: avatarUrl },
|
||||||
|
logger,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* Updates a user's password after hashing it.
|
||||||
|
* @param userId The ID of the user to update.
|
||||||
|
* @param newPassword The new plaintext password.
|
||||||
|
* @param logger The logger instance.
|
||||||
|
*/
|
||||||
|
async updateUserPassword(userId: string, newPassword: string, logger: Logger): Promise<void> {
|
||||||
|
const saltRounds = 10;
|
||||||
|
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
||||||
|
await db.userRepo.updateUserPassword(userId, hashedPassword, logger);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Deletes a user's account after verifying their password.
|
||||||
|
* @param userId The ID of the user to delete.
|
||||||
|
* @param password The user's current password for verification.
|
||||||
|
* @param logger The logger instance.
|
||||||
|
*/
|
||||||
|
async deleteUserAccount(userId: string, password: string, logger: Logger): Promise<void> {
|
||||||
|
const userWithHash = await db.userRepo.findUserWithPasswordHashById(userId, logger);
|
||||||
|
if (!userWithHash || !userWithHash.password_hash) {
|
||||||
|
// This case should be rare for a logged-in user but is a good safeguard.
|
||||||
|
throw new NotFoundError('User not found or password not set.');
|
||||||
|
}
|
||||||
|
|
||||||
|
const isMatch = await bcrypt.compare(password, userWithHash.password_hash);
|
||||||
|
if (!isMatch) {
|
||||||
|
// Use ValidationError for a 400-level response in the route
|
||||||
|
throw new ValidationError([], 'Incorrect password.');
|
||||||
|
}
|
||||||
|
|
||||||
|
await db.userRepo.deleteUserById(userId, logger);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Fetches a user's address, ensuring the user is authorized to view it.
|
||||||
|
* @param userProfile The profile of the user making the request.
|
||||||
|
* @param addressId The ID of the address being requested.
|
||||||
|
* @param logger The logger instance.
|
||||||
|
* @returns The address object.
|
||||||
|
*/
|
||||||
|
async getUserAddress(
|
||||||
|
userProfile: UserProfile,
|
||||||
|
addressId: number,
|
||||||
|
logger: Logger,
|
||||||
|
): Promise<Address> {
|
||||||
|
// Security check: Ensure the requested addressId matches the one on the user's profile.
|
||||||
|
if (userProfile.address_id !== addressId) {
|
||||||
|
// Use ValidationError to trigger a 403 Forbidden response in the route handler.
|
||||||
|
throw new ValidationError([], 'Forbidden: You can only access your own address.');
|
||||||
|
}
|
||||||
|
// The repo method will throw a NotFoundError if the address doesn't exist.
|
||||||
|
return db.addressRepo.getAddressById(addressId, logger);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Encapsulates the business logic for an admin deleting another user's account.
|
||||||
|
* This includes preventing an admin from deleting their own account.
|
||||||
|
* @param deleterId The ID of the admin performing the deletion.
|
||||||
|
* @param userToDeleteId The ID of the user to be deleted.
|
||||||
|
* @param log The logger instance.
|
||||||
|
*/
|
||||||
|
public async deleteUserAsAdmin(deleterId: string, userToDeleteId: string, log: Logger) {
|
||||||
|
if (deleterId === userToDeleteId) {
|
||||||
|
throw new ValidationError([], 'Admins cannot delete their own account.');
|
||||||
|
}
|
||||||
|
await db.userRepo.deleteUserById(userToDeleteId, log);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
export const userService = new UserService();
|
export const userService = new UserService();
|
||||||
|
|||||||
26
src/utils/fileUtils.ts
Normal file
26
src/utils/fileUtils.ts
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
// src/utils/fileUtils.ts
|
||||||
|
import fs from 'node:fs/promises';
|
||||||
|
import { logger } from '../services/logger.server';
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Safely deletes a file from the filesystem, ignoring errors if the file doesn't exist.
|
||||||
|
* @param file The multer file object to delete.
|
||||||
|
*/
|
||||||
|
export const cleanupUploadedFile = async (file?: Express.Multer.File) => {
|
||||||
|
if (!file) return;
|
||||||
|
try {
|
||||||
|
await fs.unlink(file.path);
|
||||||
|
} catch (err) {
|
||||||
|
logger.warn({ err, filePath: file.path }, 'Failed to clean up uploaded file.');
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Safely deletes multiple files from the filesystem.
|
||||||
|
* @param files An array of multer file objects to delete.
|
||||||
|
*/
|
||||||
|
export const cleanupUploadedFiles = async (files?: Express.Multer.File[]) => {
|
||||||
|
if (!files || !Array.isArray(files)) return;
|
||||||
|
// Use Promise.all to run cleanups in parallel for efficiency.
|
||||||
|
await Promise.all(files.map((file) => cleanupUploadedFile(file)));
|
||||||
|
};
|
||||||
@@ -7,6 +7,7 @@ import {
|
|||||||
optionalBoolean,
|
optionalBoolean,
|
||||||
optionalNumeric,
|
optionalNumeric,
|
||||||
optionalDate,
|
optionalDate,
|
||||||
|
optionalString,
|
||||||
} from './zodUtils';
|
} from './zodUtils';
|
||||||
|
|
||||||
describe('Zod Utilities', () => {
|
describe('Zod Utilities', () => {
|
||||||
@@ -46,11 +47,20 @@ describe('Zod Utilities', () => {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should pass for a string containing only whitespace', () => {
|
it('should fail for a string containing only whitespace', () => {
|
||||||
const result = schema.safeParse(' ');
|
const result = schema.safeParse(' ');
|
||||||
|
expect(result.success).toBe(false);
|
||||||
|
if (!result.success) {
|
||||||
|
expect(result.error.issues[0].message).toBe(customMessage);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should trim whitespace from a valid string', () => {
|
||||||
|
const result = schema.safeParse(' hello world ');
|
||||||
expect(result.success).toBe(true);
|
expect(result.success).toBe(true);
|
||||||
if (result.success) {
|
if (result.success) {
|
||||||
expect(result.data).toBe(' ');
|
// The .trim() in the schema should remove leading/trailing whitespace.
|
||||||
|
expect(result.data).toBe('hello world');
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -59,7 +69,9 @@ describe('Zod Utilities', () => {
|
|||||||
expect(result.success).toBe(false);
|
expect(result.success).toBe(false);
|
||||||
if (!result.success) {
|
if (!result.success) {
|
||||||
// z.string() will throw its own error message before min(1) is checked.
|
// z.string() will throw its own error message before min(1) is checked.
|
||||||
expect(result.error.issues[0].message).toBe('Invalid input: expected string, received number');
|
expect(result.error.issues[0].message).toBe(
|
||||||
|
'Invalid input: expected string, received number',
|
||||||
|
);
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -67,7 +79,9 @@ describe('Zod Utilities', () => {
|
|||||||
const result = schema.safeParse({ a: 1 });
|
const result = schema.safeParse({ a: 1 });
|
||||||
expect(result.success).toBe(false);
|
expect(result.success).toBe(false);
|
||||||
if (!result.success) {
|
if (!result.success) {
|
||||||
expect(result.error.issues[0].message).toBe('Invalid input: expected string, received object');
|
expect(result.error.issues[0].message).toBe(
|
||||||
|
'Invalid input: expected string, received object',
|
||||||
|
);
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
@@ -223,9 +237,7 @@ describe('Zod Utilities', () => {
|
|||||||
expect(schema.safeParse('123').success).toBe(true);
|
expect(schema.safeParse('123').success).toBe(true);
|
||||||
const floatResult = schema.safeParse('123.45');
|
const floatResult = schema.safeParse('123.45');
|
||||||
expect(floatResult.success).toBe(false);
|
expect(floatResult.success).toBe(false);
|
||||||
if (!floatResult.success) {
|
if (!floatResult.success) expect(floatResult.error.issues[0].message).toBe('Invalid input: expected int, received number');
|
||||||
expect(floatResult.error.issues[0].message).toBe('Invalid input: expected int, received number');
|
|
||||||
}
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should enforce positive constraint', () => {
|
it('should enforce positive constraint', () => {
|
||||||
@@ -384,4 +396,49 @@ describe('Zod Utilities', () => {
|
|||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
describe('optionalString', () => {
|
||||||
|
const schema = optionalString();
|
||||||
|
|
||||||
|
it('should pass for a valid string', () => {
|
||||||
|
const result = schema.safeParse('hello');
|
||||||
|
expect(result.success).toBe(true);
|
||||||
|
if (result.success) {
|
||||||
|
expect(result.data).toBe('hello');
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should pass for an empty string', () => {
|
||||||
|
const result = schema.safeParse('');
|
||||||
|
expect(result.success).toBe(true);
|
||||||
|
if (result.success) {
|
||||||
|
expect(result.data).toBe('');
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should pass for undefined and return undefined', () => {
|
||||||
|
const result = schema.safeParse(undefined);
|
||||||
|
expect(result.success).toBe(true);
|
||||||
|
if (result.success) {
|
||||||
|
expect(result.data).toBeUndefined();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should pass for null and return undefined', () => {
|
||||||
|
const result = schema.safeParse(null);
|
||||||
|
expect(result.success).toBe(true);
|
||||||
|
if (result.success) {
|
||||||
|
expect(result.data).toBeUndefined();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should fail for a non-string value like a number', () => {
|
||||||
|
const result = schema.safeParse(123);
|
||||||
|
expect(result.success).toBe(false);
|
||||||
|
if (!result.success) {
|
||||||
|
expect(result.error.issues[0].message).toBe(
|
||||||
|
'Invalid input: expected string, received number',
|
||||||
|
);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -11,8 +11,8 @@ export const requiredString = (message: string) =>
|
|||||||
// If the value is null or undefined, preprocess it to an empty string.
|
// If the value is null or undefined, preprocess it to an empty string.
|
||||||
// This ensures that the subsequent `.min(1)` check will catch missing required fields.
|
// This ensures that the subsequent `.min(1)` check will catch missing required fields.
|
||||||
(val) => val ?? '',
|
(val) => val ?? '',
|
||||||
// Now, validate that the (potentially preprocessed) value is a string with at least 1 character.
|
// Now, validate that the (potentially preprocessed) value is a string that, after trimming, has at least 1 character.
|
||||||
z.string().min(1, message),
|
z.string().trim().min(1, message),
|
||||||
);
|
);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -113,4 +113,12 @@ export const optionalBoolean = (
|
|||||||
}
|
}
|
||||||
|
|
||||||
return schema;
|
return schema;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a Zod schema for an optional string.
|
||||||
|
* Treats null as undefined so it can be properly handled as optional.
|
||||||
|
* @returns A Zod schema for an optional string.
|
||||||
|
*/
|
||||||
|
export const optionalString = () =>
|
||||||
|
z.preprocess((val) => (val === null ? undefined : val), z.string().optional());
|
||||||
|
|||||||
Reference in New Issue
Block a user