Compare commits
22 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
28d03f4e21 | ||
| 2e72ee81dd | |||
|
|
ba67ace190 | ||
|
|
50782c30e5 | ||
| 4a2ff8afc5 | |||
|
|
7a1c14ce89 | ||
| 6fafc3d089 | |||
|
|
4316866bce | ||
| 356c1a1894 | |||
|
|
2a310648ca | ||
| 8592633c22 | |||
|
|
0a9cdb8709 | ||
| 0d21e098f8 | |||
| b6799ed167 | |||
|
|
be5bda169e | ||
| 4ede403356 | |||
| 5d31605b80 | |||
| ddd4ad024e | |||
|
|
4e927f48bd | ||
| af5644d17a | |||
|
|
016c0a883a | ||
| c6a5f889b4 |
@@ -158,7 +158,7 @@ jobs:
|
||||
else
|
||||
echo "Version mismatch (Running: $RUNNING_VERSION -> Deployed: $NEW_VERSION) or app not running. Reloading PM2..."
|
||||
fi
|
||||
pm2 startOrReload ecosystem.config.cjs --env production && pm2 save
|
||||
pm2 startOrReload ecosystem.config.cjs --env production --update-env && pm2 save
|
||||
echo "Production backend server reloaded successfully."
|
||||
else
|
||||
echo "Version $NEW_VERSION is already running. Skipping PM2 reload."
|
||||
|
||||
@@ -90,10 +90,11 @@ jobs:
|
||||
# integration test suite can launch its own, fresh server instance.
|
||||
# '|| true' ensures the workflow doesn't fail if the process isn't running.
|
||||
run: |
|
||||
pm2 stop flyer-crawler-api-test || true
|
||||
pm2 stop flyer-crawler-worker-test || true
|
||||
pm2 delete flyer-crawler-api-test || true
|
||||
pm2 delete flyer-crawler-worker-test || true
|
||||
echo "--- Stopping and deleting all test processes ---"
|
||||
# Use a script to parse pm2's JSON output and delete any process whose name ends with '-test'.
|
||||
# This is safer than 'pm2 delete all' and more robust than naming each process individually.
|
||||
# It prevents the accumulation of duplicate processes from previous test runs.
|
||||
node -e "const exec = require('child_process').execSync; try { const list = JSON.parse(exec('pm2 jlist').toString()); list.forEach(p => { if (p.name && p.name.endsWith('-test')) { console.log('Deleting test process: ' + p.name + ' (' + p.pm2_env.pm_id + ')'); try { exec('pm2 delete ' + p.pm2_env.pm_id); } catch(e) { console.error('Failed to delete ' + p.pm2_env.pm_id, e.message); } } }); console.log('✅ Test process cleanup complete.'); } catch (e) { if (e.stdout.toString().includes('No process found')) { console.log('No PM2 processes running, cleanup not needed.'); } else { console.error('Error cleaning up test processes:', e.message); } }" || true
|
||||
|
||||
- name: Run All Tests and Generate Merged Coverage Report
|
||||
# This single step runs both unit and integration tests, then merges their
|
||||
@@ -389,8 +390,15 @@ jobs:
|
||||
|
||||
run: |
|
||||
# Fail-fast check to ensure secrets are configured in Gitea.
|
||||
if [ -z "$DB_HOST" ] || [ -z "$DB_USER" ] || [ -z "$DB_PASSWORD" ] || [ -z "$DB_NAME" ]; then
|
||||
echo "ERROR: One or more test database secrets (DB_HOST, DB_USER, DB_PASSWORD, DB_DATABASE_TEST) are not set in Gitea repository settings."
|
||||
MISSING_SECRETS=""
|
||||
if [ -z "$DB_HOST" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_HOST"; fi
|
||||
if [ -z "$DB_USER" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_USER"; fi
|
||||
if [ -z "$DB_PASSWORD" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_PASSWORD"; fi
|
||||
if [ -z "$DB_NAME" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_NAME"; fi
|
||||
if [ -z "$JWT_SECRET" ]; then MISSING_SECRETS="${MISSING_SECRETS} JWT_SECRET"; fi
|
||||
|
||||
if [ ! -z "$MISSING_SECRETS" ]; then
|
||||
echo "ERROR: The following required secrets are missing in Gitea:${MISSING_SECRETS}"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
@@ -405,7 +413,7 @@ jobs:
|
||||
# Use `startOrReload` with the ecosystem file. This is the standard, idempotent way to deploy.
|
||||
# It will START the process if it's not running, or RELOAD it if it is.
|
||||
# We also add `&& pm2 save` to persist the process list across server reboots.
|
||||
pm2 startOrReload ecosystem.config.cjs --env test && pm2 save
|
||||
pm2 startOrReload ecosystem.config.cjs --env test --update-env && pm2 save
|
||||
echo "Test backend server reloaded successfully."
|
||||
|
||||
# After a successful deployment, update the schema hash in the database.
|
||||
|
||||
@@ -157,7 +157,7 @@ jobs:
|
||||
else
|
||||
echo "Version mismatch (Running: $RUNNING_VERSION -> Deployed: $NEW_VERSION) or app not running. Reloading PM2..."
|
||||
fi
|
||||
pm2 startOrReload ecosystem.config.cjs --env production && pm2 save
|
||||
pm2 startOrReload ecosystem.config.cjs --env production --update-env && pm2 save
|
||||
echo "Production backend server reloaded successfully."
|
||||
else
|
||||
echo "Version $NEW_VERSION is already running. Skipping PM2 reload."
|
||||
|
||||
@@ -3,23 +3,38 @@
|
||||
// It allows us to define all the settings for our application in one place.
|
||||
// The .cjs extension is required because the project's package.json has "type": "module".
|
||||
|
||||
// --- Environment Variable Validation ---
|
||||
const requiredSecrets = ['DB_HOST', 'JWT_SECRET', 'GEMINI_API_KEY'];
|
||||
const missingSecrets = requiredSecrets.filter(key => !process.env[key]);
|
||||
|
||||
if (missingSecrets.length > 0) {
|
||||
console.warn('\n[ecosystem.config.cjs] ⚠️ WARNING: The following environment variables are MISSING in the shell:');
|
||||
missingSecrets.forEach(key => console.warn(` - ${key}`));
|
||||
console.warn('[ecosystem.config.cjs] The application may crash if these are required for startup.\n');
|
||||
process.exit(1); // Fail fast so PM2 doesn't attempt to start a broken app
|
||||
} else {
|
||||
console.log('[ecosystem.config.cjs] ✅ Critical environment variables are present.');
|
||||
}
|
||||
|
||||
module.exports = {
|
||||
apps: [
|
||||
{
|
||||
// --- API Server ---
|
||||
// The name is now dynamically set based on the environment.
|
||||
// This is a common pattern but requires you to call pm2 with the correct name.
|
||||
// The deploy script handles this by using 'flyer-crawler-api' for prod and 'flyer-crawler-api-test' for test.
|
||||
name: 'flyer-crawler-api',
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'server.ts', // tsx will execute this file
|
||||
max_memory_restart: '500M', // Restart if memory usage exceeds 500MB
|
||||
args: 'server.ts',
|
||||
max_memory_restart: '500M',
|
||||
|
||||
// Restart Logic
|
||||
max_restarts: 40,
|
||||
exp_backoff_restart_delay: 100,
|
||||
min_uptime: '10s',
|
||||
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production', // Set the Node.js environment to production
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-api',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -39,10 +54,9 @@ module.exports = {
|
||||
},
|
||||
// Test Environment Settings
|
||||
env_test: {
|
||||
NODE_ENV: 'test', // Set to 'test' to match the environment purpose and disable pino-pretty
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-api-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -66,7 +80,6 @@ module.exports = {
|
||||
name: 'flyer-crawler-api-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -89,14 +102,19 @@ module.exports = {
|
||||
// --- General Worker ---
|
||||
name: 'flyer-crawler-worker',
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'src/services/worker.ts', // tsx will execute this file
|
||||
max_memory_restart: '1G', // Restart if memory usage exceeds 1GB
|
||||
args: 'src/services/worker.ts',
|
||||
max_memory_restart: '1G',
|
||||
|
||||
// Restart Logic
|
||||
max_restarts: 40,
|
||||
exp_backoff_restart_delay: 100,
|
||||
min_uptime: '10s',
|
||||
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-worker',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -119,7 +137,6 @@ module.exports = {
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-worker-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -143,7 +160,6 @@ module.exports = {
|
||||
name: 'flyer-crawler-worker-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -166,14 +182,19 @@ module.exports = {
|
||||
// --- Analytics Worker ---
|
||||
name: 'flyer-crawler-analytics-worker',
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'src/services/worker.ts', // tsx will execute this file
|
||||
max_memory_restart: '1G', // Restart if memory usage exceeds 1GB
|
||||
args: 'src/services/worker.ts',
|
||||
max_memory_restart: '1G',
|
||||
|
||||
// Restart Logic
|
||||
max_restarts: 40,
|
||||
exp_backoff_restart_delay: 100,
|
||||
min_uptime: '10s',
|
||||
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-analytics-worker',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -196,7 +217,6 @@ module.exports = {
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-analytics-worker-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -220,7 +240,6 @@ module.exports = {
|
||||
name: 'flyer-crawler-analytics-worker-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
|
||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.2.8",
|
||||
"version": "0.2.19",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.2.8",
|
||||
"version": "0.2.19",
|
||||
"dependencies": {
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"private": true,
|
||||
"version": "0.2.8",
|
||||
"version": "0.2.19",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||
|
||||
@@ -115,6 +115,7 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
||||
valid_from DATE,
|
||||
valid_to DATE,
|
||||
store_address TEXT,
|
||||
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
||||
item_count INTEGER DEFAULT 0 NOT NULL,
|
||||
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
@@ -130,11 +131,13 @@ COMMENT ON COLUMN public.flyers.store_id IS 'Foreign key linking this flyer to a
|
||||
COMMENT ON COLUMN public.flyers.valid_from IS 'The start date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.valid_to IS 'The end date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.store_address IS 'The physical store address if it was successfully extracted from the flyer image.';
|
||||
COMMENT ON COLUMN public.flyers.status IS 'The processing status of the flyer, e.g., if it needs manual review.';
|
||||
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
||||
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
||||
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||
-- 7. The 'master_grocery_items' table. This is the master dictionary.
|
||||
CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
||||
master_grocery_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
|
||||
@@ -131,6 +131,7 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
||||
valid_from DATE,
|
||||
valid_to DATE,
|
||||
store_address TEXT,
|
||||
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
||||
item_count INTEGER DEFAULT 0 NOT NULL,
|
||||
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
@@ -146,11 +147,13 @@ COMMENT ON COLUMN public.flyers.store_id IS 'Foreign key linking this flyer to a
|
||||
COMMENT ON COLUMN public.flyers.valid_from IS 'The start date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.valid_to IS 'The end date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.store_address IS 'The physical store address if it was successfully extracted from the flyer image.';
|
||||
COMMENT ON COLUMN public.flyers.status IS 'The processing status of the flyer, e.g., if it needs manual review.';
|
||||
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
||||
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
||||
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||
-- 7. The 'master_grocery_items' table. This is the master dictionary.
|
||||
CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
||||
master_grocery_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
|
||||
@@ -21,7 +21,6 @@ import {
|
||||
mockUseFlyerItems,
|
||||
} from './tests/setup/mockHooks';
|
||||
import { useAppInitialization } from './hooks/useAppInitialization';
|
||||
import { useModal } from './hooks/useModal';
|
||||
|
||||
// Mock top-level components rendered by App's routes
|
||||
|
||||
@@ -57,10 +56,6 @@ vi.mock('./hooks/useFlyerItems', async () => {
|
||||
vi.mock('./hooks/useAppInitialization');
|
||||
const mockedUseAppInitialization = vi.mocked(useAppInitialization);
|
||||
|
||||
// Mock useModal directly in this file to avoid dependency on mockHooks.ts
|
||||
vi.mock('./hooks/useModal');
|
||||
const mockedUseModal = vi.mocked(useModal);
|
||||
|
||||
vi.mock('./hooks/useAuth', async () => {
|
||||
const hooks = await import('./tests/setup/mockHooks');
|
||||
return { useAuth: hooks.mockUseAuth };
|
||||
@@ -131,11 +126,21 @@ vi.mock('./layouts/MainLayout', async () => {
|
||||
return { MainLayout: MockMainLayout };
|
||||
});
|
||||
|
||||
vi.mock('./components/AppGuard', () => ({
|
||||
AppGuard: ({ children }: { children: React.ReactNode }) => (
|
||||
<div data-testid="app-guard-mock">{children}</div>
|
||||
),
|
||||
}));
|
||||
vi.mock('./components/AppGuard', async () => {
|
||||
// We need to use the real useModal hook inside our mock AppGuard
|
||||
const { useModal } = await vi.importActual<typeof import('./hooks/useModal')>('./hooks/useModal');
|
||||
return {
|
||||
AppGuard: ({ children }: { children: React.ReactNode }) => {
|
||||
const { isModalOpen } = useModal();
|
||||
return (
|
||||
<div data-testid="app-guard-mock">
|
||||
{children}
|
||||
{isModalOpen('whatsNew') && <div data-testid="whats-new-modal-mock" />}
|
||||
</div>
|
||||
);
|
||||
},
|
||||
};
|
||||
});
|
||||
|
||||
const mockedAiApiClient = vi.mocked(aiApiClient);
|
||||
const mockedApiClient = vi.mocked(apiClient);
|
||||
@@ -196,11 +201,6 @@ describe('App Component', () => {
|
||||
error: null,
|
||||
});
|
||||
mockedUseAppInitialization.mockReturnValue({ isDarkMode: false, unitSystem: 'imperial' });
|
||||
mockedUseModal.mockReturnValue({
|
||||
isModalOpen: vi.fn(),
|
||||
openModal: vi.fn(),
|
||||
closeModal: vi.fn(),
|
||||
});
|
||||
|
||||
// Default mocks for API calls
|
||||
// Use mockImplementation to create a new Response object for each call,
|
||||
@@ -391,6 +391,7 @@ describe('App Component', () => {
|
||||
|
||||
describe('Modal Interactions', () => {
|
||||
it('should open and close the ProfileManager modal', async () => {
|
||||
console.log('[TEST DEBUG] Test Start: should open and close the ProfileManager modal');
|
||||
renderApp();
|
||||
expect(screen.queryByTestId('profile-manager-mock')).not.toBeInTheDocument();
|
||||
|
||||
@@ -398,11 +399,13 @@ describe('App Component', () => {
|
||||
fireEvent.click(screen.getByText('Open Profile'));
|
||||
expect(await screen.findByTestId('profile-manager-mock')).toBeInTheDocument();
|
||||
|
||||
console.log('[TEST DEBUG] ProfileManager modal opened. Now closing...');
|
||||
// Close modal
|
||||
fireEvent.click(screen.getByText('Close Profile'));
|
||||
await waitFor(() => {
|
||||
expect(screen.queryByTestId('profile-manager-mock')).not.toBeInTheDocument();
|
||||
});
|
||||
console.log('[TEST DEBUG] ProfileManager modal closed.');
|
||||
});
|
||||
|
||||
it('should open and close the VoiceAssistant modal for authenticated users', async () => {
|
||||
@@ -427,7 +430,7 @@ describe('App Component', () => {
|
||||
fireEvent.click(screen.getByText('Open Voice Assistant'));
|
||||
|
||||
console.log('[TEST DEBUG] Waiting for voice-assistant-mock');
|
||||
expect(await screen.findByTestId('voice-assistant-mock')).toBeInTheDocument();
|
||||
expect(await screen.findByTestId('voice-assistant-mock', {}, { timeout: 3000 })).toBeInTheDocument();
|
||||
|
||||
// Close modal
|
||||
fireEvent.click(screen.getByText('Close Voice Assistant'));
|
||||
@@ -586,7 +589,7 @@ describe('App Component', () => {
|
||||
renderApp();
|
||||
console.log('[TEST DEBUG] Opening Profile');
|
||||
fireEvent.click(screen.getByText('Open Profile'));
|
||||
const loginButton = await screen.findByText('Login');
|
||||
const loginButton = await screen.findByRole('button', { name: 'Login' });
|
||||
console.log('[TEST DEBUG] Clicking Login');
|
||||
fireEvent.click(loginButton);
|
||||
|
||||
@@ -622,7 +625,8 @@ describe('App Component', () => {
|
||||
renderApp();
|
||||
const openButton = await screen.findByTitle("Show what's new in this version");
|
||||
fireEvent.click(openButton);
|
||||
expect(mockedUseModal().openModal).toHaveBeenCalledWith('whatsNew');
|
||||
// The mock AppGuard now renders the modal when it's open
|
||||
expect(await screen.findByTestId('whats-new-modal-mock')).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -13,6 +13,7 @@ import { AdminPage } from './pages/admin/AdminPage';
|
||||
import { AdminRoute } from './components/AdminRoute';
|
||||
import { CorrectionsPage } from './pages/admin/CorrectionsPage';
|
||||
import { AdminStatsPage } from './pages/admin/AdminStatsPage';
|
||||
import { FlyerReviewPage } from './pages/admin/FlyerReviewPage';
|
||||
import { ResetPasswordPage } from './pages/ResetPasswordPage';
|
||||
import { VoiceLabPage } from './pages/VoiceLabPage';
|
||||
import { FlyerCorrectionTool } from './components/FlyerCorrectionTool';
|
||||
@@ -228,6 +229,7 @@ function App() {
|
||||
<Route path="/admin" element={<AdminPage />} />
|
||||
<Route path="/admin/corrections" element={<CorrectionsPage />} />
|
||||
<Route path="/admin/stats" element={<AdminStatsPage />} />
|
||||
<Route path="/admin/flyer-review" element={<FlyerReviewPage />} />
|
||||
<Route path="/admin/voice-lab" element={<VoiceLabPage />} />
|
||||
</Route>
|
||||
<Route path="/reset-password/:token" element={<ResetPasswordPage />} />
|
||||
|
||||
18
src/components/icons/DocumentMagnifyingGlassIcon.tsx
Normal file
18
src/components/icons/DocumentMagnifyingGlassIcon.tsx
Normal file
@@ -0,0 +1,18 @@
|
||||
import React from 'react';
|
||||
|
||||
export const DocumentMagnifyingGlassIcon: React.FC<React.SVGProps<SVGSVGElement>> = (props) => (
|
||||
<svg
|
||||
xmlns="http://www.w3.org/2000/svg"
|
||||
fill="none"
|
||||
viewBox="0 0 24 24"
|
||||
strokeWidth={1.5}
|
||||
stroke="currentColor"
|
||||
{...props}
|
||||
>
|
||||
<path
|
||||
strokeLinecap="round"
|
||||
strokeLinejoin="round"
|
||||
d="M19.5 14.25v-2.625a3.375 3.375 0 0 0-3.375-3.375h-1.5A1.125 1.125 0 0 1 13.5 7.125v-1.5a3.375 3.375 0 0 0-3.375-3.375H8.25m5.231 13.481L15 17.25m-4.5 4.5L6.75 21.75m0 0L2.25 17.25m4.5 4.5v-4.5m13.5-3V9A2.25 2.25 0 0 0 16.5 6.75h-9A2.25 2.25 0 0 0 5.25 9v9.75m14.25-10.5a2.25 2.25 0 0 0-2.25-2.25H5.25a2.25 2.25 0 0 0-2.25 2.25v10.5a2.25 2.25 0 0 0 2.25 225h5.25"
|
||||
/>
|
||||
</svg>
|
||||
);
|
||||
@@ -28,7 +28,7 @@ const mockedUseAuth = vi.mocked(useAuth);
|
||||
const mockedUseModal = vi.mocked(useModal);
|
||||
const mockedUseNavigate = vi.mocked(useNavigate);
|
||||
|
||||
const mockLogin = vi.fn();
|
||||
const mockLogin = vi.fn().mockResolvedValue(undefined);
|
||||
const mockNavigate = vi.fn();
|
||||
const mockOpenModal = vi.fn();
|
||||
|
||||
@@ -61,7 +61,7 @@ describe('useAppInitialization Hook', () => {
|
||||
// Mock localStorage
|
||||
Object.defineProperty(window, 'localStorage', {
|
||||
value: {
|
||||
getItem: vi.fn(),
|
||||
getItem: vi.fn().mockReturnValue(null),
|
||||
setItem: vi.fn(),
|
||||
removeItem: vi.fn(),
|
||||
clear: vi.fn(),
|
||||
@@ -74,6 +74,7 @@ describe('useAppInitialization Hook', () => {
|
||||
matches: false, // default to light mode
|
||||
})),
|
||||
writable: true,
|
||||
configurable: true,
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
@@ -69,7 +69,7 @@ export const useAppInitialization = () => {
|
||||
if (userProfile && userProfile.preferences?.darkMode !== undefined) {
|
||||
localStorage.setItem('darkMode', String(userProfile.preferences.darkMode));
|
||||
}
|
||||
}, [userProfile?.preferences?.darkMode, userProfile?.user.user_id]);
|
||||
}, [userProfile]);
|
||||
|
||||
// Effect to set initial unit system based on user profile or local storage
|
||||
useEffect(() => {
|
||||
|
||||
@@ -79,7 +79,7 @@ vi.mock('../pages/admin/ActivityLog', async () => {
|
||||
),
|
||||
};
|
||||
});
|
||||
vi.mock('../pages/admin/components/AnonymousUserBanner', () => ({
|
||||
vi.mock('../components/AnonymousUserBanner', () => ({
|
||||
AnonymousUserBanner: () => <div data-testid="anonymous-banner" />,
|
||||
}));
|
||||
vi.mock('../components/ErrorDisplay', () => ({
|
||||
|
||||
@@ -4,6 +4,7 @@ import { SystemCheck } from './components/SystemCheck';
|
||||
import { Link } from 'react-router-dom';
|
||||
import { ShieldExclamationIcon } from '../../components/icons/ShieldExclamationIcon';
|
||||
import { ChartBarIcon } from '../../components/icons/ChartBarIcon';
|
||||
import { DocumentMagnifyingGlassIcon } from '../../components/icons/DocumentMagnifyingGlassIcon';
|
||||
|
||||
export const AdminPage: React.FC = () => {
|
||||
// The onReady prop for SystemCheck is present to allow for future UI changes,
|
||||
@@ -39,6 +40,13 @@ export const AdminPage: React.FC = () => {
|
||||
<ChartBarIcon className="w-6 h-6 mr-3 text-brand-primary" />
|
||||
<span className="font-semibold">View Statistics</span>
|
||||
</Link>
|
||||
<Link
|
||||
to="/admin/flyer-review"
|
||||
className="flex items-center p-3 rounded-lg hover:bg-gray-100 dark:hover:bg-gray-700/50 transition-colors"
|
||||
>
|
||||
<DocumentMagnifyingGlassIcon className="w-6 h-6 mr-3 text-brand-primary" />
|
||||
<span className="font-semibold">Flyer Review Queue</span>
|
||||
</Link>
|
||||
</div>
|
||||
</div>
|
||||
<SystemCheck />
|
||||
|
||||
93
src/pages/admin/FlyerReviewPage.tsx
Normal file
93
src/pages/admin/FlyerReviewPage.tsx
Normal file
@@ -0,0 +1,93 @@
|
||||
// src/pages/admin/FlyerReviewPage.tsx
|
||||
import React, { useEffect, useState } from 'react';
|
||||
import { Link } from 'react-router-dom';
|
||||
import { getFlyersForReview } from '../../services/apiClient';
|
||||
import { logger } from '../../services/logger.client';
|
||||
import type { Flyer } from '../../types';
|
||||
import { LoadingSpinner } from '../../components/LoadingSpinner';
|
||||
import { format } from 'date-fns';
|
||||
|
||||
export const FlyerReviewPage: React.FC = () => {
|
||||
const [flyers, setFlyers] = useState<Flyer[]>([]);
|
||||
const [isLoading, setIsLoading] = useState(true);
|
||||
const [error, setError] = useState<string | null>(null);
|
||||
|
||||
useEffect(() => {
|
||||
const fetchFlyers = async () => {
|
||||
setIsLoading(true);
|
||||
setError(null);
|
||||
try {
|
||||
const response = await getFlyersForReview();
|
||||
if (!response.ok) {
|
||||
throw new Error((await response.json()).message || 'Failed to fetch flyers for review.');
|
||||
}
|
||||
setFlyers(await response.json());
|
||||
} catch (err) {
|
||||
const errorMessage =
|
||||
err instanceof Error ? err.message : 'An unknown error occurred while fetching data.';
|
||||
logger.error({ err }, 'Failed to fetch flyers for review');
|
||||
setError(errorMessage);
|
||||
} finally {
|
||||
setIsLoading(false);
|
||||
}
|
||||
};
|
||||
|
||||
fetchFlyers();
|
||||
}, []);
|
||||
|
||||
return (
|
||||
<div className="max-w-7xl mx-auto py-8 px-4">
|
||||
<div className="mb-8">
|
||||
<Link to="/admin" className="text-brand-primary hover:underline">
|
||||
← Back to Admin Dashboard
|
||||
</Link>
|
||||
<h1 className="text-3xl font-bold text-gray-800 dark:text-white mt-2">
|
||||
Flyer Review Queue
|
||||
</h1>
|
||||
<p className="text-gray-500 dark:text-gray-400">
|
||||
Review flyers that were processed with low confidence by the AI.
|
||||
</p>
|
||||
</div>
|
||||
|
||||
{isLoading && (
|
||||
<div
|
||||
role="status"
|
||||
aria-label="Loading flyers for review"
|
||||
className="flex justify-center items-center h-64"
|
||||
>
|
||||
<LoadingSpinner />
|
||||
</div>
|
||||
)}
|
||||
{error && (
|
||||
<div className="text-red-500 bg-red-100 dark:bg-red-900/20 p-4 rounded-lg">{error}</div>
|
||||
)}
|
||||
|
||||
{!isLoading && !error && (
|
||||
<div className="bg-white dark:bg-gray-800 rounded-lg border border-gray-200 dark:border-gray-700 overflow-hidden">
|
||||
<ul className="divide-y divide-gray-200 dark:divide-gray-700">
|
||||
{flyers.length === 0 ? (
|
||||
<li className="p-6 text-center text-gray-500">
|
||||
The review queue is empty. Great job!
|
||||
</li>
|
||||
) : (
|
||||
flyers.map((flyer) => (
|
||||
<li key={flyer.flyer_id} className="p-4 hover:bg-gray-50 dark:hover:bg-gray-700/50">
|
||||
<Link to={`/flyers/${flyer.flyer_id}`} className="flex items-center space-x-4">
|
||||
<img src={flyer.icon_url || ''} alt={flyer.store?.name} className="w-12 h-12 rounded-md object-cover" />
|
||||
<div className="flex-1">
|
||||
<p className="font-semibold text-gray-800 dark:text-white">{flyer.store?.name || 'Unknown Store'}</p>
|
||||
<p className="text-sm text-gray-500 dark:text-gray-400">{flyer.file_name}</p>
|
||||
</div>
|
||||
<div className="text-right text-sm text-gray-500 dark:text-gray-400">
|
||||
<p>Uploaded: {format(new Date(flyer.created_at), 'MMM d, yyyy')}</p>
|
||||
</div>
|
||||
</Link>
|
||||
</li>
|
||||
))
|
||||
)}
|
||||
</ul>
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
);
|
||||
};
|
||||
@@ -374,7 +374,11 @@ describe('AuthView', () => {
|
||||
fireEvent.submit(screen.getByTestId('auth-form'));
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.getByRole('button', { name: 'Register' })).toBeDisabled();
|
||||
const submitButton = screen.getByTestId('auth-form').querySelector('button[type="submit"]');
|
||||
expect(submitButton).toBeInTheDocument();
|
||||
expect(submitButton).toBeDisabled();
|
||||
// Verify the text 'Register' is gone from any button
|
||||
expect(screen.queryByRole('button', { name: 'Register' })).not.toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -148,6 +148,18 @@ router.get('/corrections', async (req, res, next: NextFunction) => {
|
||||
}
|
||||
});
|
||||
|
||||
router.get('/review/flyers', async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
req.log.debug('Fetching flyers for review via adminRepo');
|
||||
const flyers = await db.adminRepo.getFlyersForReview(req.log);
|
||||
req.log.info({ count: Array.isArray(flyers) ? flyers.length : 'unknown' }, 'Successfully fetched flyers for review');
|
||||
res.json(flyers);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching flyers for review');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
|
||||
router.get('/brands', async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const brands = await db.flyerRepo.getAllBrands(req.log);
|
||||
|
||||
@@ -330,6 +330,12 @@ describe('AI Routes (/api/ai)', () => {
|
||||
expect(response.status).toBe(201);
|
||||
expect(response.body.message).toBe('Flyer processed and saved successfully.');
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// Verify that the legacy endpoint correctly sets the status to 'needs_review'
|
||||
expect(vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0]).toEqual(
|
||||
expect.objectContaining({
|
||||
status: 'needs_review',
|
||||
}),
|
||||
);
|
||||
});
|
||||
|
||||
it('should return 400 if no flyer image is provided', async () => {
|
||||
@@ -383,6 +389,12 @@ describe('AI Routes (/api/ai)', () => {
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// Verify that the legacy endpoint correctly sets the status to 'needs_review'
|
||||
expect(vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0]).toEqual(
|
||||
expect.objectContaining({
|
||||
status: 'needs_review',
|
||||
}),
|
||||
);
|
||||
// verify the items array passed to DB was an empty array
|
||||
const callArgs = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0]?.[1];
|
||||
expect(callArgs).toBeDefined();
|
||||
@@ -412,6 +424,12 @@ describe('AI Routes (/api/ai)', () => {
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// Verify that the legacy endpoint correctly sets the status to 'needs_review'
|
||||
expect(vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0]).toEqual(
|
||||
expect.objectContaining({
|
||||
status: 'needs_review',
|
||||
}),
|
||||
);
|
||||
// verify the flyerData.store_name passed to DB was the fallback string
|
||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
||||
expect(flyerDataArg.store_name).toContain('Unknown Store');
|
||||
|
||||
@@ -13,8 +13,8 @@ import {
|
||||
handleMulterError,
|
||||
} from '../middleware/multer.middleware';
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { UserProfile, ExtractedCoreData, ExtractedFlyerItem } from '../types';
|
||||
import { logger } from '../services/logger.server'; // This was a duplicate, fixed.
|
||||
import { UserProfile, ExtractedCoreData, ExtractedFlyerItem, FlyerInsert } from '../types';
|
||||
import { flyerQueue } from '../services/queueService.server';
|
||||
import { validateRequest } from '../middleware/validation.middleware';
|
||||
import { requiredString } from '../utils/zodUtils';
|
||||
@@ -437,7 +437,7 @@ router.post(
|
||||
const iconUrl = `/flyer-images/icons/${iconFileName}`;
|
||||
|
||||
// 2. Prepare flyer data for insertion
|
||||
const flyerData = {
|
||||
const flyerData: FlyerInsert = {
|
||||
file_name: originalFileName,
|
||||
image_url: `/flyer-images/${req.file.filename}`, // Store the full URL path
|
||||
icon_url: iconUrl,
|
||||
@@ -448,6 +448,8 @@ router.post(
|
||||
valid_to: extractedData.valid_to ?? null,
|
||||
store_address: extractedData.store_address ?? null,
|
||||
item_count: 0, // Set default to 0; the trigger will update it.
|
||||
// Set a safe default status for this legacy endpoint. The new flow uses the transformer to determine this.
|
||||
status: 'needs_review',
|
||||
uploaded_by: userProfile?.user.user_id, // Associate with user if logged in
|
||||
};
|
||||
|
||||
|
||||
@@ -260,6 +260,13 @@ const jwtOptions = {
|
||||
secretOrKey: JWT_SECRET,
|
||||
};
|
||||
|
||||
// --- DEBUG LOGGING FOR JWT SECRET ---
|
||||
if (!JWT_SECRET) {
|
||||
logger.fatal('[Passport] CRITICAL: JWT_SECRET is missing or empty in environment variables! JwtStrategy will fail.');
|
||||
} else {
|
||||
logger.info(`[Passport] JWT_SECRET loaded successfully (length: ${JWT_SECRET.length}).`);
|
||||
}
|
||||
|
||||
passport.use(
|
||||
new JwtStrategy(jwtOptions, async (jwt_payload, done) => {
|
||||
logger.debug(
|
||||
|
||||
@@ -19,13 +19,15 @@ vi.mock('./logger.client', () => ({
|
||||
debug: vi.fn(),
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// 2. Mock ./apiClient to simply pass calls through to the global fetch.
|
||||
vi.mock('./apiClient', async (importOriginal) => {
|
||||
return {
|
||||
apiFetch: (
|
||||
// This is the core logic we want to preserve: it calls the global fetch
|
||||
// which is then intercepted by MSW.
|
||||
const apiFetch = (
|
||||
url: string,
|
||||
options: RequestInit = {},
|
||||
apiOptions: import('./apiClient').ApiOptions = {},
|
||||
@@ -59,6 +61,26 @@ vi.mock('./apiClient', async (importOriginal) => {
|
||||
const request = new Request(fullUrl, options);
|
||||
console.log(`[apiFetch MOCK] Executing fetch for URL: ${request.url}.`);
|
||||
return fetch(request);
|
||||
};
|
||||
|
||||
return {
|
||||
// The original mock only had apiFetch. We need to add the helpers.
|
||||
apiFetch,
|
||||
|
||||
// These helpers are what aiApiClient.ts actually calls.
|
||||
// Their mock implementation should just call our mocked apiFetch.
|
||||
authedGet: (endpoint: string, options: import('./apiClient').ApiOptions = {}) => {
|
||||
return apiFetch(endpoint, { method: 'GET' }, options);
|
||||
},
|
||||
authedPost: <T>(endpoint: string, body: T, options: import('./apiClient').ApiOptions = {}) => {
|
||||
return apiFetch(
|
||||
endpoint,
|
||||
{ method: 'POST', headers: { 'Content-Type': 'application/json' }, body: JSON.stringify(body) },
|
||||
options,
|
||||
);
|
||||
},
|
||||
authedPostForm: (endpoint: string, formData: FormData, options: import('./apiClient').ApiOptions = {}) => {
|
||||
return apiFetch(endpoint, { method: 'POST', body: formData }, options);
|
||||
},
|
||||
// Add a mock for ApiOptions to satisfy the compiler
|
||||
ApiOptions: vi.fn(),
|
||||
@@ -285,9 +307,25 @@ describe('AI API Client (Network Mocking with MSW)', () => {
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow('Job not found');
|
||||
});
|
||||
|
||||
it('should throw a generic error if the API response is not valid JSON', async () => {
|
||||
server.use(http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => HttpResponse.text('Invalid JSON')));
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow(expect.any(SyntaxError));
|
||||
it('should throw a specific error if a 200 OK response is not valid JSON', async () => {
|
||||
server.use(
|
||||
http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => {
|
||||
// A 200 OK response that is not JSON is a server-side contract violation.
|
||||
return HttpResponse.text('This should have been JSON', { status: 200 });
|
||||
}),
|
||||
);
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow(
|
||||
'Failed to parse job status from a successful API response.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error with status text if the non-ok API response is not valid JSON', async () => {
|
||||
server.use(
|
||||
http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => {
|
||||
return HttpResponse.text('Gateway Timeout', { status: 504, statusText: 'Gateway Timeout' });
|
||||
}),
|
||||
);
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow('API Error: 504 Gateway Timeout');
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
@@ -12,7 +12,7 @@ import type {
|
||||
GroundedResponse,
|
||||
} from '../types';
|
||||
import { logger } from './logger.client';
|
||||
import { apiFetch } from './apiClient';
|
||||
import { apiFetch, authedGet, authedPost, authedPostForm } from './apiClient';
|
||||
|
||||
/**
|
||||
* Uploads a flyer file to the backend to be processed asynchronously.
|
||||
@@ -33,14 +33,7 @@ export const uploadAndProcessFlyer = async (
|
||||
|
||||
logger.info(`[aiApiClient] Starting background processing for file: ${file.name}`);
|
||||
|
||||
const response = await apiFetch(
|
||||
'/ai/upload-and-process',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
const response = await authedPostForm('/ai/upload-and-process', formData, { tokenOverride });
|
||||
|
||||
if (!response.ok) {
|
||||
let errorBody;
|
||||
@@ -101,18 +94,29 @@ export const getJobStatus = async (
|
||||
jobId: string,
|
||||
tokenOverride?: string,
|
||||
): Promise<JobStatus> => {
|
||||
const response = await apiFetch(`/ai/jobs/${jobId}/status`, {}, { tokenOverride });
|
||||
const response = await authedGet(`/ai/jobs/${jobId}/status`, { tokenOverride });
|
||||
|
||||
// Handle non-OK responses first, as they might not have a JSON body.
|
||||
if (!response.ok) {
|
||||
let errorMessage = `API Error: ${response.status} ${response.statusText}`;
|
||||
try {
|
||||
// Try to get a more specific message from the body.
|
||||
const errorData = await response.json();
|
||||
if (errorData.message) {
|
||||
errorMessage = errorData.message;
|
||||
}
|
||||
} catch (e) {
|
||||
// The body was not JSON, which is fine for a server error page.
|
||||
// The default message is sufficient.
|
||||
logger.warn('getJobStatus received a non-JSON error response.', { status: response.status });
|
||||
}
|
||||
throw new Error(errorMessage);
|
||||
}
|
||||
|
||||
// If we get here, the response is OK (2xx). Now parse the body.
|
||||
try {
|
||||
const statusData: JobStatus = await response.json();
|
||||
|
||||
if (!response.ok) {
|
||||
// If the HTTP response itself is an error (e.g., 404, 500), throw an error.
|
||||
// Use the message from the JSON body if available.
|
||||
const errorMessage = (statusData as any).message || `API Error: ${response.status}`;
|
||||
throw new Error(errorMessage);
|
||||
}
|
||||
|
||||
// If the job itself has failed, we should treat this as an error condition
|
||||
// for the polling logic by rejecting the promise. This will stop the polling loop.
|
||||
if (statusData.state === 'failed') {
|
||||
@@ -130,9 +134,13 @@ export const getJobStatus = async (
|
||||
|
||||
return statusData;
|
||||
} catch (error) {
|
||||
// This block catches errors from `response.json()` (if the body is not valid JSON)
|
||||
// and also re-throws the errors we created above.
|
||||
throw error;
|
||||
// If it's the specific error we threw, just re-throw it.
|
||||
if (error instanceof JobFailedError) {
|
||||
throw error;
|
||||
}
|
||||
// This now primarily catches JSON parsing errors on an OK response, which is unexpected.
|
||||
logger.error('getJobStatus failed to parse a successful API response.', { error });
|
||||
throw new Error('Failed to parse job status from a successful API response.');
|
||||
}
|
||||
};
|
||||
|
||||
@@ -145,14 +153,7 @@ export const isImageAFlyer = (
|
||||
|
||||
// Use apiFetchWithAuth for FormData to let the browser set the correct Content-Type.
|
||||
// The URL must be relative, as the helper constructs the full path.
|
||||
return apiFetch(
|
||||
'/ai/check-flyer',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/check-flyer', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
export const extractAddressFromImage = (
|
||||
@@ -162,14 +163,7 @@ export const extractAddressFromImage = (
|
||||
const formData = new FormData();
|
||||
formData.append('image', imageFile);
|
||||
|
||||
return apiFetch(
|
||||
'/ai/extract-address',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/extract-address', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
export const extractLogoFromImage = (
|
||||
@@ -181,14 +175,7 @@ export const extractLogoFromImage = (
|
||||
formData.append('images', file);
|
||||
});
|
||||
|
||||
return apiFetch(
|
||||
'/ai/extract-logo',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/extract-logo', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
export const getQuickInsights = (
|
||||
@@ -196,16 +183,7 @@ export const getQuickInsights = (
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
return apiFetch(
|
||||
'/ai/quick-insights',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/quick-insights', { items }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
export const getDeepDiveAnalysis = (
|
||||
@@ -213,16 +191,7 @@ export const getDeepDiveAnalysis = (
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
return apiFetch(
|
||||
'/ai/deep-dive',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/deep-dive', { items }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
export const searchWeb = (
|
||||
@@ -230,16 +199,7 @@ export const searchWeb = (
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
return apiFetch(
|
||||
'/ai/search-web',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ query }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/search-web', { query }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
// ============================================================================
|
||||
@@ -254,15 +214,7 @@ export const planTripWithMaps = async (
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
logger.debug('Stub: planTripWithMaps called with location:', { userLocation });
|
||||
return apiFetch(
|
||||
'/ai/plan-trip',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items, store, userLocation }),
|
||||
},
|
||||
{ signal, tokenOverride },
|
||||
);
|
||||
return authedPost('/ai/plan-trip', { items, store, userLocation }, { signal, tokenOverride });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -276,16 +228,7 @@ export const generateImageFromText = (
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
logger.debug('Stub: generateImageFromText called with prompt:', { prompt });
|
||||
return apiFetch(
|
||||
'/ai/generate-image',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ prompt }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/generate-image', { prompt }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -299,16 +242,7 @@ export const generateSpeechFromText = (
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
logger.debug('Stub: generateSpeechFromText called with text:', { text });
|
||||
return apiFetch(
|
||||
'/ai/generate-speech',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ text }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/generate-speech', { text }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -361,11 +295,7 @@ export const rescanImageArea = (
|
||||
formData.append('cropArea', JSON.stringify(cropArea));
|
||||
formData.append('extractionType', extractionType);
|
||||
|
||||
return apiFetch(
|
||||
'/ai/rescan-area',
|
||||
{ method: 'POST', body: formData },
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/rescan-area', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -379,12 +309,5 @@ export const compareWatchedItemPrices = (
|
||||
): Promise<Response> => {
|
||||
// Use the apiFetch wrapper for consistency with other API calls in this file.
|
||||
// This centralizes token handling and base URL logic.
|
||||
return apiFetch(
|
||||
'/ai/compare-prices',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items: watchedItems }),
|
||||
},
|
||||
{ signal },
|
||||
)};
|
||||
return authedPost('/ai/compare-prices', { items: watchedItems }, { signal });
|
||||
};
|
||||
|
||||
@@ -596,40 +596,6 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('_normalizeExtractedItems (private method)', () => {
|
||||
it('should replace null or undefined fields with default values', () => {
|
||||
const rawItems: {
|
||||
item: string;
|
||||
price_display: null;
|
||||
quantity: undefined;
|
||||
category_name: null;
|
||||
master_item_id: null;
|
||||
}[] = [
|
||||
{
|
||||
item: 'Test',
|
||||
price_display: null,
|
||||
quantity: undefined,
|
||||
category_name: null,
|
||||
master_item_id: null,
|
||||
},
|
||||
];
|
||||
const [normalized] = (
|
||||
aiServiceInstance as unknown as {
|
||||
_normalizeExtractedItems: (items: typeof rawItems) => {
|
||||
price_display: string;
|
||||
quantity: string;
|
||||
category_name: string;
|
||||
master_item_id: undefined;
|
||||
}[];
|
||||
}
|
||||
)._normalizeExtractedItems(rawItems);
|
||||
expect(normalized.price_display).toBe('');
|
||||
expect(normalized.quantity).toBe('');
|
||||
expect(normalized.category_name).toBe('Other/Miscellaneous');
|
||||
expect(normalized.master_item_id).toBeUndefined();
|
||||
});
|
||||
});
|
||||
|
||||
describe('extractTextFromImageArea', () => {
|
||||
it('should call sharp to crop the image and call the AI with the correct prompt', async () => {
|
||||
console.log("TEST START: 'should call sharp to crop...'");
|
||||
|
||||
79
src/services/analyticsService.server.ts
Normal file
79
src/services/analyticsService.server.ts
Normal file
@@ -0,0 +1,79 @@
|
||||
// src/services/analyticsService.server.ts
|
||||
import type { Job } from 'bullmq';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
import type { AnalyticsJobData, WeeklyAnalyticsJobData } from './queues.server';
|
||||
|
||||
/**
|
||||
* A service class to encapsulate business logic for analytics-related background jobs.
|
||||
*/
|
||||
export class AnalyticsService {
|
||||
/**
|
||||
* Processes a job to generate a daily analytics report.
|
||||
* This is currently a mock implementation.
|
||||
* @param job The BullMQ job object.
|
||||
*/
|
||||
async processDailyReportJob(job: Job<AnalyticsJobData>) {
|
||||
const { reportDate } = job.data;
|
||||
const logger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
reportDate,
|
||||
});
|
||||
|
||||
logger.info(`Picked up daily analytics job.`);
|
||||
|
||||
try {
|
||||
// This is mock logic, but we keep it in the service
|
||||
if (reportDate === 'FAIL') {
|
||||
throw new Error('This is a test failure for the analytics job.');
|
||||
}
|
||||
// Simulate work
|
||||
await new Promise((resolve) => setTimeout(resolve, 10000));
|
||||
logger.info(`Successfully generated report for ${reportDate}.`);
|
||||
return { status: 'success', reportDate };
|
||||
} catch (error) {
|
||||
const wrappedError = error instanceof Error ? error : new Error(String(error));
|
||||
logger.error(
|
||||
{
|
||||
err: wrappedError,
|
||||
attemptsMade: job.attemptsMade,
|
||||
},
|
||||
`Daily analytics job failed.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Processes a job to generate a weekly analytics report.
|
||||
* This is currently a mock implementation.
|
||||
* @param job The BullMQ job object.
|
||||
*/
|
||||
async processWeeklyReportJob(job: Job<WeeklyAnalyticsJobData>) {
|
||||
const { reportYear, reportWeek } = job.data;
|
||||
const logger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
reportYear,
|
||||
reportWeek,
|
||||
});
|
||||
|
||||
logger.info(`Picked up weekly analytics job.`);
|
||||
|
||||
try {
|
||||
// Mock logic
|
||||
await new Promise((resolve) => setTimeout(resolve, 30000));
|
||||
logger.info(`Successfully generated weekly report for week ${reportWeek}, ${reportYear}.`);
|
||||
return { status: 'success', reportYear, reportWeek };
|
||||
} catch (error) {
|
||||
const wrappedError = error instanceof Error ? error : new Error(String(error));
|
||||
logger.error(
|
||||
{ err: wrappedError, attemptsMade: job.attemptsMade },
|
||||
`Weekly analytics job failed.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
export const analyticsService = new AnalyticsService();
|
||||
@@ -7,6 +7,17 @@ import { http, HttpResponse } from 'msw';
|
||||
vi.unmock('./apiClient');
|
||||
|
||||
import * as apiClient from './apiClient';
|
||||
import {
|
||||
createMockAddressPayload,
|
||||
createMockBudget,
|
||||
createMockLoginPayload,
|
||||
createMockProfileUpdatePayload,
|
||||
createMockRecipeCommentPayload,
|
||||
createMockRegisterUserPayload,
|
||||
createMockSearchQueryPayload,
|
||||
createMockShoppingListItemPayload,
|
||||
createMockWatchedItemPayload,
|
||||
} from '../tests/utils/mockFactories';
|
||||
|
||||
// Mock the logger to keep test output clean and verifiable.
|
||||
vi.mock('./logger', () => ({
|
||||
@@ -229,33 +240,6 @@ describe('API Client', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('Analytics API Functions', () => {
|
||||
it('trackFlyerItemInteraction should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
|
||||
// We can now await this properly because we added 'return' in apiClient.ts
|
||||
await apiClient.trackFlyerItemInteraction(123, 'click');
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
});
|
||||
|
||||
it('logSearchQuery should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
|
||||
await apiClient.logSearchQuery({
|
||||
query_text: 'test',
|
||||
result_count: 0,
|
||||
was_successful: false,
|
||||
});
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
});
|
||||
});
|
||||
|
||||
describe('apiFetch (with FormData)', () => {
|
||||
it('should handle FormData correctly by not setting Content-Type', async () => {
|
||||
localStorage.setItem('authToken', 'form-data-token');
|
||||
@@ -317,10 +301,11 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('addWatchedItem should send a POST request with the correct body', async () => {
|
||||
await apiClient.addWatchedItem('Apples', 'Produce');
|
||||
const watchedItemData = createMockWatchedItemPayload({ itemName: 'Apples', category: 'Produce' });
|
||||
await apiClient.addWatchedItem(watchedItemData.itemName, watchedItemData.category);
|
||||
|
||||
expect(capturedUrl?.pathname).toBe('/api/users/watched-items');
|
||||
expect(capturedBody).toEqual({ itemName: 'Apples', category: 'Produce' });
|
||||
expect(capturedBody).toEqual(watchedItemData);
|
||||
});
|
||||
|
||||
it('removeWatchedItem should send a DELETE request to the correct URL', async () => {
|
||||
@@ -337,12 +322,12 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('createBudget should send a POST request with budget data', async () => {
|
||||
const budgetData = {
|
||||
const budgetData = createMockBudget({
|
||||
name: 'Groceries',
|
||||
amount_cents: 50000,
|
||||
period: 'monthly' as const,
|
||||
period: 'monthly',
|
||||
start_date: '2024-01-01',
|
||||
};
|
||||
});
|
||||
await apiClient.createBudget(budgetData);
|
||||
|
||||
expect(capturedUrl?.pathname).toBe('/api/budgets');
|
||||
@@ -461,7 +446,7 @@ describe('API Client', () => {
|
||||
|
||||
it('addShoppingListItem should send a POST request with item data', async () => {
|
||||
const listId = 42;
|
||||
const itemData = { customItemName: 'Paper Towels' };
|
||||
const itemData = createMockShoppingListItemPayload({ customItemName: 'Paper Towels' });
|
||||
await apiClient.addShoppingListItem(listId, itemData);
|
||||
|
||||
expect(capturedUrl?.pathname).toBe(`/api/users/shopping-lists/${listId}/items`);
|
||||
@@ -547,7 +532,7 @@ describe('API Client', () => {
|
||||
|
||||
it('addRecipeComment should send a POST request with content and optional parentId', async () => {
|
||||
const recipeId = 456;
|
||||
const commentData = { content: 'This is a reply', parentCommentId: 789 };
|
||||
const commentData = createMockRecipeCommentPayload({ content: 'This is a reply', parentCommentId: 789 });
|
||||
await apiClient.addRecipeComment(recipeId, commentData.content, commentData.parentCommentId);
|
||||
expect(capturedUrl?.pathname).toBe(`/api/recipes/${recipeId}/comments`);
|
||||
expect(capturedBody).toEqual(commentData);
|
||||
@@ -563,7 +548,7 @@ describe('API Client', () => {
|
||||
describe('User Profile and Settings API Functions', () => {
|
||||
it('updateUserProfile should send a PUT request with profile data', async () => {
|
||||
localStorage.setItem('authToken', 'user-settings-token');
|
||||
const profileData = { full_name: 'John Doe' };
|
||||
const profileData = createMockProfileUpdatePayload({ full_name: 'John Doe' });
|
||||
await apiClient.updateUserProfile(profileData, { tokenOverride: 'override-token' });
|
||||
expect(capturedUrl?.pathname).toBe('/api/users/profile');
|
||||
expect(capturedBody).toEqual(profileData);
|
||||
@@ -619,14 +604,14 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('registerUser should send a POST request with user data', async () => {
|
||||
await apiClient.registerUser('test@example.com', 'password123', 'Test User');
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/register');
|
||||
expect(capturedBody).toEqual({
|
||||
const userData = createMockRegisterUserPayload({
|
||||
email: 'test@example.com',
|
||||
password: 'password123',
|
||||
full_name: 'Test User',
|
||||
avatar_url: undefined,
|
||||
});
|
||||
await apiClient.registerUser(userData.email, userData.password, userData.full_name);
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/register');
|
||||
expect(capturedBody).toEqual(userData);
|
||||
});
|
||||
|
||||
it('deleteUserAccount should send a DELETE request with the confirmation password', async () => {
|
||||
@@ -654,7 +639,7 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('updateUserAddress should send a PUT request with address data', async () => {
|
||||
const addressData = { address_line_1: '123 Main St', city: 'Anytown' };
|
||||
const addressData = createMockAddressPayload({ address_line_1: '123 Main St', city: 'Anytown' });
|
||||
await apiClient.updateUserAddress(addressData);
|
||||
expect(capturedUrl?.pathname).toBe('/api/users/profile/address');
|
||||
expect(capturedBody).toEqual(addressData);
|
||||
@@ -890,6 +875,11 @@ describe('API Client', () => {
|
||||
expect(capturedUrl?.pathname).toBe('/api/admin/corrections');
|
||||
});
|
||||
|
||||
it('getFlyersForReview should call the correct endpoint', async () => {
|
||||
await apiClient.getFlyersForReview();
|
||||
expect(capturedUrl?.pathname).toBe('/api/admin/review/flyers');
|
||||
});
|
||||
|
||||
it('rejectCorrection should send a POST request to the correct URL', async () => {
|
||||
const correctionId = 46;
|
||||
await apiClient.rejectCorrection(correctionId);
|
||||
@@ -942,53 +932,49 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('logSearchQuery should send a POST request with query data', async () => {
|
||||
const queryData = { query_text: 'apples', result_count: 10, was_successful: true };
|
||||
const queryData = createMockSearchQueryPayload({ query_text: 'apples', result_count: 10, was_successful: true });
|
||||
await apiClient.logSearchQuery(queryData);
|
||||
expect(capturedUrl?.pathname).toBe('/api/search/log');
|
||||
expect(capturedBody).toEqual(queryData);
|
||||
});
|
||||
|
||||
it('trackFlyerItemInteraction should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
const { logger } = await import('./logger.client');
|
||||
|
||||
// We can now await this properly because we added 'return' in apiClient.ts
|
||||
await apiClient.trackFlyerItemInteraction(123, 'click');
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
});
|
||||
|
||||
it('logSearchQuery should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
const { logger } = await import('./logger.client');
|
||||
|
||||
await apiClient.logSearchQuery({
|
||||
const queryData = createMockSearchQueryPayload({
|
||||
query_text: 'test',
|
||||
result_count: 0,
|
||||
was_successful: false,
|
||||
});
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
|
||||
await apiClient.logSearchQuery(queryData);
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
});
|
||||
});
|
||||
|
||||
describe('Authentication API Functions', () => {
|
||||
it('loginUser should send a POST request with credentials', async () => {
|
||||
await apiClient.loginUser('test@example.com', 'password123', true);
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/login');
|
||||
expect(capturedBody).toEqual({
|
||||
const loginData = createMockLoginPayload({
|
||||
email: 'test@example.com',
|
||||
password: 'password123',
|
||||
rememberMe: true,
|
||||
});
|
||||
await apiClient.loginUser(loginData.email, loginData.password, loginData.rememberMe);
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/login');
|
||||
expect(capturedBody).toEqual(loginData);
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
@@ -1,14 +1,9 @@
|
||||
// src/services/db/address.db.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import type { Pool } from 'pg';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import { AddressRepository } from './address.db';
|
||||
import type { Address } from '../../types';
|
||||
import { UniqueConstraintError, NotFoundError } from './errors.db';
|
||||
|
||||
// Un-mock the module we are testing
|
||||
vi.unmock('./address.db');
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('../logger.server', () => ({
|
||||
logger: { info: vi.fn(), warn: vi.fn(), error: vi.fn(), debug: vi.fn() },
|
||||
@@ -17,10 +12,13 @@ import { logger as mockLogger } from '../logger.server';
|
||||
|
||||
describe('Address DB Service', () => {
|
||||
let addressRepo: AddressRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
addressRepo = new AddressRepository(mockPoolInstance as unknown as Pool);
|
||||
addressRepo = new AddressRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getAddressById', () => {
|
||||
@@ -35,19 +33,19 @@ describe('Address DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
};
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockAddress] });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockAddress], rowCount: 1 });
|
||||
|
||||
const result = await addressRepo.getAddressById(1, mockLogger);
|
||||
|
||||
expect(result).toEqual(mockAddress);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.addresses WHERE address_id = $1',
|
||||
[1],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if no address is found', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(addressRepo.getAddressById(999, mockLogger)).rejects.toThrow(NotFoundError);
|
||||
await expect(addressRepo.getAddressById(999, mockLogger)).rejects.toThrow(
|
||||
'Address with ID 999 not found.',
|
||||
@@ -56,7 +54,7 @@ describe('Address DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.getAddressById(1, mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve address.',
|
||||
@@ -71,12 +69,12 @@ describe('Address DB Service', () => {
|
||||
describe('upsertAddress', () => {
|
||||
it('should INSERT a new address when no address_id is provided', async () => {
|
||||
const newAddressData = { address_line_1: '456 New Ave', city: 'Newville' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ address_id: 2 }] });
|
||||
mockDb.query.mockResolvedValue({ rows: [{ address_id: 2 }] });
|
||||
|
||||
const result = await addressRepo.upsertAddress(newAddressData, mockLogger);
|
||||
|
||||
expect(result).toBe(2);
|
||||
const [query, values] = mockPoolInstance.query.mock.calls[0];
|
||||
const [query, values] = mockDb.query.mock.calls[0];
|
||||
expect(query).toContain('INSERT INTO public.addresses');
|
||||
expect(query).toContain('ON CONFLICT (address_id) DO UPDATE');
|
||||
expect(values).toEqual(['456 New Ave', 'Newville']);
|
||||
@@ -84,64 +82,15 @@ describe('Address DB Service', () => {
|
||||
|
||||
it('should UPDATE an existing address when an address_id is provided', async () => {
|
||||
const existingAddressData = { address_id: 1, address_line_1: '789 Old Rd', city: 'Oldtown' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ address_id: 1 }] });
|
||||
mockDb.query.mockResolvedValue({ rows: [{ address_id: 1 }] });
|
||||
|
||||
const result = await addressRepo.upsertAddress(existingAddressData, mockLogger);
|
||||
|
||||
expect(result).toBe(1);
|
||||
const [query, values] = mockPoolInstance.query.mock.calls[0];
|
||||
const [query, values] = mockDb.query.mock.calls[0];
|
||||
expect(query).toContain('INSERT INTO public.addresses');
|
||||
expect(query).toContain('ON CONFLICT (address_id) DO UPDATE');
|
||||
// The values array should now include the address_id at the beginning
|
||||
expect(values).toEqual([1, '789 Old Rd', 'Oldtown']);
|
||||
});
|
||||
|
||||
it('should throw a generic error on INSERT failure', async () => {
|
||||
const newAddressData = { address_line_1: '456 New Ave', city: 'Newville' };
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.upsertAddress(newAddressData, mockLogger)).rejects.toThrow(
|
||||
'Failed to upsert address.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: newAddressData },
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error on UPDATE failure', async () => {
|
||||
const existingAddressData = { address_id: 1, address_line_1: '789 Old Rd', city: 'Oldtown' };
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.upsertAddress(existingAddressData, mockLogger)).rejects.toThrow(
|
||||
'Failed to upsert address.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: existingAddressData },
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw UniqueConstraintError on duplicate address insert', async () => {
|
||||
const newAddressData = { address_line_1: '123 Main St', city: 'Anytown' };
|
||||
const dbError = new Error('duplicate key value violates unique constraint') as Error & {
|
||||
code: string;
|
||||
};
|
||||
dbError.code = '23505';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.upsertAddress(newAddressData, mockLogger)).rejects.toThrow(
|
||||
UniqueConstraintError,
|
||||
);
|
||||
await expect(addressRepo.upsertAddress(newAddressData, mockLogger)).rejects.toThrow(
|
||||
'An identical address already exists.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: newAddressData },
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -6,9 +6,11 @@ import { UniqueConstraintError, NotFoundError } from './errors.db';
|
||||
import { Address } from '../../types';
|
||||
|
||||
export class AddressRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -1,6 +1,5 @@
|
||||
// src/services/db/admin.db.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, Mock } from 'vitest';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import type { Pool, PoolClient } from 'pg';
|
||||
import { ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
||||
import { AdminRepository } from './admin.db';
|
||||
@@ -33,6 +32,9 @@ import { withTransaction } from './connection.db';
|
||||
|
||||
describe('Admin DB Service', () => {
|
||||
let adminRepo: AdminRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset the global mock's call history before each test.
|
||||
@@ -43,8 +45,8 @@ describe('Admin DB Service', () => {
|
||||
const mockClient = { query: vi.fn() };
|
||||
return callback(mockClient as unknown as PoolClient);
|
||||
});
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
adminRepo = new AdminRepository(mockPoolInstance as unknown as Pool);
|
||||
// Instantiate the repository with the minimal mock db for each test
|
||||
adminRepo = new AdminRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getSuggestedCorrections', () => {
|
||||
@@ -52,11 +54,11 @@ describe('Admin DB Service', () => {
|
||||
const mockCorrections: SuggestedCorrection[] = [
|
||||
createMockSuggestedCorrection({ suggested_correction_id: 1 }),
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockCorrections });
|
||||
mockDb.query.mockResolvedValue({ rows: mockCorrections });
|
||||
|
||||
const result = await adminRepo.getSuggestedCorrections(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.suggested_corrections sc'),
|
||||
);
|
||||
expect(result).toEqual(mockCorrections);
|
||||
@@ -64,7 +66,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getSuggestedCorrections(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve suggested corrections.',
|
||||
);
|
||||
@@ -77,10 +79,10 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('approveCorrection', () => {
|
||||
it('should call the approve_correction database function', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] }); // Mock the function call
|
||||
mockDb.query.mockResolvedValue({ rows: [] }); // Mock the function call
|
||||
await adminRepo.approveCorrection(123, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT public.approve_correction($1)',
|
||||
[123],
|
||||
);
|
||||
@@ -88,7 +90,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database function fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.approveCorrection(123, mockLogger)).rejects.toThrow(
|
||||
'Failed to approve correction.',
|
||||
);
|
||||
@@ -101,17 +103,17 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('rejectCorrection', () => {
|
||||
it('should update the correction status to rejected', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 1 });
|
||||
await adminRepo.rejectCorrection(123, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining("UPDATE public.suggested_corrections SET status = 'rejected'"),
|
||||
[123],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if the correction is not found or not pending', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0 });
|
||||
await expect(adminRepo.rejectCorrection(123, mockLogger)).rejects.toThrow(NotFoundError);
|
||||
await expect(adminRepo.rejectCorrection(123, mockLogger)).rejects.toThrow(
|
||||
"Correction with ID 123 not found or not in 'pending' state.",
|
||||
@@ -119,7 +121,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
mockDb.query.mockRejectedValue(new Error('DB Error'));
|
||||
await expect(adminRepo.rejectCorrection(123, mockLogger)).rejects.toThrow(
|
||||
'Failed to reject correction.',
|
||||
);
|
||||
@@ -136,11 +138,11 @@ describe('Admin DB Service', () => {
|
||||
suggested_correction_id: 1,
|
||||
suggested_value: '300',
|
||||
});
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockCorrection], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockCorrection], rowCount: 1 });
|
||||
|
||||
const result = await adminRepo.updateSuggestedCorrection(1, '300', mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.suggested_corrections SET suggested_value = $1'),
|
||||
['300', 1],
|
||||
);
|
||||
@@ -148,7 +150,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the correction is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(
|
||||
adminRepo.updateSuggestedCorrection(999, 'new value', mockLogger),
|
||||
).rejects.toThrow(NotFoundError);
|
||||
@@ -158,7 +160,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
mockDb.query.mockRejectedValue(new Error('DB Error'));
|
||||
await expect(adminRepo.updateSuggestedCorrection(1, 'new value', mockLogger)).rejects.toThrow(
|
||||
'Failed to update suggested correction.',
|
||||
);
|
||||
@@ -172,7 +174,7 @@ describe('Admin DB Service', () => {
|
||||
describe('getApplicationStats', () => {
|
||||
it('should execute 5 parallel count queries and return the aggregated stats', async () => {
|
||||
// Mock responses for each of the 5 parallel queries
|
||||
mockPoolInstance.query
|
||||
mockDb.query
|
||||
.mockResolvedValueOnce({ rows: [{ count: '10' }] }) // flyerCount
|
||||
.mockResolvedValueOnce({ rows: [{ count: '20' }] }) // userCount
|
||||
.mockResolvedValueOnce({ rows: [{ count: '300' }] }) // flyerItemCount
|
||||
@@ -182,7 +184,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
const stats = await adminRepo.getApplicationStats(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(6);
|
||||
expect(mockDb.query).toHaveBeenCalledTimes(6);
|
||||
expect(stats).toEqual({
|
||||
flyerCount: 10,
|
||||
userCount: 20,
|
||||
@@ -195,7 +197,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if one of the parallel queries fails', async () => {
|
||||
// Mock one query to succeed and another to fail
|
||||
mockPoolInstance.query
|
||||
mockDb.query
|
||||
.mockResolvedValueOnce({ rows: [{ count: '10' }] })
|
||||
.mockRejectedValueOnce(new Error('DB Read Error'));
|
||||
|
||||
@@ -211,11 +213,11 @@ describe('Admin DB Service', () => {
|
||||
describe('getDailyStatsForLast30Days', () => {
|
||||
it('should execute the correct query to get daily stats', async () => {
|
||||
const mockStats = [{ date: '2023-01-01', new_users: 5, new_flyers: 2 }];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockStats });
|
||||
mockDb.query.mockResolvedValue({ rows: mockStats });
|
||||
|
||||
const result = await adminRepo.getDailyStatsForLast30Days(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('WITH date_series AS'),
|
||||
);
|
||||
expect(result).toEqual(mockStats);
|
||||
@@ -223,7 +225,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getDailyStatsForLast30Days(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve daily statistics.',
|
||||
);
|
||||
@@ -236,18 +238,18 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('logActivity', () => {
|
||||
it('should insert a new activity log entry', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
const logData = { userId: 'user-123', action: 'test_action', displayText: 'Test activity' };
|
||||
await adminRepo.logActivity(logData, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO public.activity_log'),
|
||||
[logData.userId, logData.action, logData.displayText, null, null],
|
||||
);
|
||||
});
|
||||
|
||||
it('should not throw an error if the database query fails (non-critical)', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
mockDb.query.mockRejectedValue(new Error('DB Error'));
|
||||
const logData = { action: 'test_action', displayText: 'Test activity' };
|
||||
await expect(adminRepo.logActivity(logData, mockLogger)).resolves.toBeUndefined();
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
@@ -259,9 +261,9 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('getMostFrequentSaleItems', () => {
|
||||
it('should call the correct database function', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.getMostFrequentSaleItems(30, 10, mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.flyer_items fi'),
|
||||
[30, 10],
|
||||
);
|
||||
@@ -269,7 +271,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getMostFrequentSaleItems(30, 10, mockLogger)).rejects.toThrow(
|
||||
'Failed to get most frequent sale items.',
|
||||
);
|
||||
@@ -283,9 +285,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateRecipeCommentStatus', () => {
|
||||
it('should update the comment status and return the updated comment', async () => {
|
||||
const mockComment = { comment_id: 1, status: 'hidden' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockComment], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockComment], rowCount: 1 });
|
||||
const result = await adminRepo.updateRecipeCommentStatus(1, 'hidden', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.recipe_comments'),
|
||||
['hidden', 1],
|
||||
);
|
||||
@@ -293,7 +295,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the comment is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateRecipeCommentStatus(999, 'hidden', mockLogger)).rejects.toThrow(
|
||||
'Recipe comment with ID 999 not found.',
|
||||
);
|
||||
@@ -301,7 +303,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateRecipeCommentStatus(1, 'hidden', mockLogger)).rejects.toThrow(
|
||||
'Failed to update recipe comment status.',
|
||||
);
|
||||
@@ -314,16 +316,16 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('getUnmatchedFlyerItems', () => {
|
||||
it('should execute the correct query to get unmatched items', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.getUnmatchedFlyerItems(mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.unmatched_flyer_items ufi'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getUnmatchedFlyerItems(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve unmatched flyer items.',
|
||||
);
|
||||
@@ -337,9 +339,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateRecipeStatus', () => {
|
||||
it('should update the recipe status and return the updated recipe', async () => {
|
||||
const mockRecipe = { recipe_id: 1, status: 'public' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockRecipe], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockRecipe], rowCount: 1 });
|
||||
const result = await adminRepo.updateRecipeStatus(1, 'public', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.recipes'),
|
||||
['public', 1],
|
||||
);
|
||||
@@ -347,7 +349,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the recipe is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateRecipeStatus(999, 'public', mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -358,7 +360,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateRecipeStatus(1, 'public', mockLogger)).rejects.toThrow(
|
||||
'Failed to update recipe status.',
|
||||
);
|
||||
@@ -437,16 +439,16 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('ignoreUnmatchedFlyerItem', () => {
|
||||
it('should update the status of an unmatched item to "ignored"', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 1 });
|
||||
await adminRepo.ignoreUnmatchedFlyerItem(1, mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
"UPDATE public.unmatched_flyer_items SET status = 'ignored' WHERE unmatched_flyer_item_id = $1 AND status = 'pending'",
|
||||
[1],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if the unmatched item is not found or not pending', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0 });
|
||||
await expect(adminRepo.ignoreUnmatchedFlyerItem(999, mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -457,11 +459,11 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.ignoreUnmatchedFlyerItem(1, mockLogger)).rejects.toThrow(
|
||||
'Failed to ignore unmatched flyer item.',
|
||||
);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining("UPDATE public.unmatched_flyer_items SET status = 'ignored'"),
|
||||
[1],
|
||||
);
|
||||
@@ -474,7 +476,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('resetFailedLoginAttempts', () => {
|
||||
it('should execute a specific UPDATE query to reset attempts and log login details', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.resetFailedLoginAttempts('user-123', '127.0.0.1', mockLogger);
|
||||
|
||||
// Use a regular expression to match the SQL query while ignoring whitespace differences.
|
||||
@@ -482,7 +484,7 @@ describe('Admin DB Service', () => {
|
||||
const expectedQueryRegex =
|
||||
/UPDATE\s+public\.users\s+SET\s+failed_login_attempts\s*=\s*0,\s*last_failed_login\s*=\s*NULL,\s*last_login_ip\s*=\s*\$2,\s*last_login_at\s*=\s*NOW\(\)\s+WHERE\s+user_id\s*=\s*\$1\s+AND\s+failed_login_attempts\s*>\s*0/;
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
// The test now verifies the full structure of the query.
|
||||
expect.stringMatching(expectedQueryRegex),
|
||||
['user-123', '127.0.0.1'],
|
||||
@@ -491,7 +493,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should not throw an error if the database query fails (non-critical)', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
adminRepo.resetFailedLoginAttempts('user-123', '127.0.0.1', mockLogger),
|
||||
).resolves.toBeUndefined();
|
||||
@@ -506,21 +508,21 @@ describe('Admin DB Service', () => {
|
||||
describe('incrementFailedLoginAttempts', () => {
|
||||
it('should execute an UPDATE query and return the new attempt count', async () => {
|
||||
// Mock the DB to return the new count
|
||||
mockPoolInstance.query.mockResolvedValue({
|
||||
mockDb.query.mockResolvedValue({
|
||||
rows: [{ failed_login_attempts: 3 }],
|
||||
rowCount: 1,
|
||||
});
|
||||
const newCount = await adminRepo.incrementFailedLoginAttempts('user-123', mockLogger);
|
||||
|
||||
expect(newCount).toBe(3);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('RETURNING failed_login_attempts'),
|
||||
['user-123'],
|
||||
);
|
||||
});
|
||||
|
||||
it('should return 0 if the user is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
const newCount = await adminRepo.incrementFailedLoginAttempts('user-not-found', mockLogger);
|
||||
expect(newCount).toBe(0);
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
@@ -531,7 +533,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should return -1 if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
const newCount = await adminRepo.incrementFailedLoginAttempts('user-123', mockLogger);
|
||||
|
||||
expect(newCount).toBe(-1);
|
||||
@@ -544,16 +546,16 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('updateBrandLogo', () => {
|
||||
it('should execute an UPDATE query for the brand logo', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.updateBrandLogo(1, '/logo.png', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'UPDATE public.brands SET logo_url = $1 WHERE brand_id = $2',
|
||||
['/logo.png', 1],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if the brand is not found', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0 });
|
||||
await expect(adminRepo.updateBrandLogo(999, '/logo.png', mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -564,11 +566,11 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateBrandLogo(1, '/logo.png', mockLogger)).rejects.toThrow(
|
||||
'Failed to update brand logo in database.',
|
||||
);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.brands SET logo_url'),
|
||||
['/logo.png', 1],
|
||||
);
|
||||
@@ -582,9 +584,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateReceiptStatus', () => {
|
||||
it('should update the receipt status and return the updated receipt', async () => {
|
||||
const mockReceipt = { receipt_id: 1, status: 'completed' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockReceipt], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockReceipt], rowCount: 1 });
|
||||
const result = await adminRepo.updateReceiptStatus(1, 'completed', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.receipts'),
|
||||
['completed', 1],
|
||||
);
|
||||
@@ -592,7 +594,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the receipt is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateReceiptStatus(999, 'completed', mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -603,7 +605,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateReceiptStatus(1, 'completed', mockLogger)).rejects.toThrow(
|
||||
'Failed to update receipt status.',
|
||||
);
|
||||
@@ -616,9 +618,9 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('getActivityLog', () => {
|
||||
it('should call the get_activity_log database function', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.getActivityLog(50, 0, mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.get_activity_log($1, $2)',
|
||||
[50, 0],
|
||||
);
|
||||
@@ -626,7 +628,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getActivityLog(50, 0, mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve activity log.',
|
||||
);
|
||||
@@ -642,9 +644,9 @@ describe('Admin DB Service', () => {
|
||||
const mockUsers: AdminUserView[] = [
|
||||
createMockAdminUserView({ user_id: '1', email: 'test@test.com' }),
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockUsers });
|
||||
mockDb.query.mockResolvedValue({ rows: mockUsers });
|
||||
const result = await adminRepo.getAllUsers(mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.users u JOIN public.profiles p'),
|
||||
);
|
||||
expect(result).toEqual(mockUsers);
|
||||
@@ -652,7 +654,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getAllUsers(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve all users.',
|
||||
);
|
||||
@@ -666,9 +668,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateUserRole', () => {
|
||||
it('should update the user role and return the updated user', async () => {
|
||||
const mockProfile: Profile = createMockProfile({ role: 'admin' });
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockProfile], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockProfile], rowCount: 1 });
|
||||
const result = await adminRepo.updateUserRole('1', 'admin', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'UPDATE public.profiles SET role = $1 WHERE user_id = $2 RETURNING *',
|
||||
['admin', '1'],
|
||||
);
|
||||
@@ -676,7 +678,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the user is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateUserRole('999', 'admin', mockLogger)).rejects.toThrow(
|
||||
'User with ID 999 not found.',
|
||||
);
|
||||
@@ -684,7 +686,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should re-throw a generic error if the database query fails for other reasons', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateUserRole('1', 'admin', mockLogger)).rejects.toThrow('DB Error');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: '1', role: 'admin' },
|
||||
@@ -697,7 +699,7 @@ describe('Admin DB Service', () => {
|
||||
const dbError = new Error('violates foreign key constraint');
|
||||
// Create a more specific type for the error object to avoid using 'any'
|
||||
(dbError as Error & { code: string }).code = '23503';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(
|
||||
adminRepo.updateUserRole('non-existent-user', 'admin', mockLogger),
|
||||
|
||||
@@ -13,12 +13,15 @@ import {
|
||||
Receipt,
|
||||
AdminUserView,
|
||||
Profile,
|
||||
Flyer,
|
||||
} from '../../types';
|
||||
|
||||
export class AdminRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
@@ -612,4 +615,32 @@ export class AdminRepository {
|
||||
throw error; // Re-throw to be handled by the route
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Retrieves all flyers that have been flagged with a 'needs_review' status.
|
||||
* @param logger The logger instance.
|
||||
* @returns A promise that resolves to an array of Flyer objects.
|
||||
*/
|
||||
async getFlyersForReview(logger: Logger): Promise<Flyer[]> {
|
||||
try {
|
||||
const query = `
|
||||
SELECT
|
||||
f.*,
|
||||
json_build_object(
|
||||
'store_id', s.store_id,
|
||||
'name', s.name,
|
||||
'logo_url', s.logo_url
|
||||
) as store
|
||||
FROM public.flyers f
|
||||
LEFT JOIN public.stores s ON f.store_id = s.store_id
|
||||
WHERE f.status = 'needs_review'
|
||||
ORDER BY f.created_at DESC;
|
||||
`;
|
||||
const res = await this.db.query<Flyer>(query);
|
||||
return res.rows;
|
||||
} catch (error) {
|
||||
logger.error({ err: error }, 'Database error in getFlyersForReview');
|
||||
throw new Error('Failed to retrieve flyers for review.');
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -7,7 +7,6 @@ vi.unmock('./budget.db');
|
||||
|
||||
import { BudgetRepository } from './budget.db';
|
||||
import type { Pool, PoolClient } from 'pg';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import type { Budget, SpendingByCategory } from '../../types';
|
||||
|
||||
// Mock the logger to prevent console output during tests
|
||||
@@ -42,11 +41,14 @@ import { withTransaction } from './connection.db';
|
||||
|
||||
describe('Budget DB Service', () => {
|
||||
let budgetRepo: BudgetRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
budgetRepo = new BudgetRepository(mockPoolInstance as unknown as Pool);
|
||||
// Instantiate the repository with the minimal mock db for each test
|
||||
budgetRepo = new BudgetRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getBudgetsForUser', () => {
|
||||
@@ -63,11 +65,11 @@ describe('Budget DB Service', () => {
|
||||
updated_at: new Date().toISOString(),
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockBudgets });
|
||||
mockDb.query.mockResolvedValue({ rows: mockBudgets });
|
||||
|
||||
const result = await budgetRepo.getBudgetsForUser('user-123', mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.budgets WHERE user_id = $1 ORDER BY start_date DESC',
|
||||
['user-123'],
|
||||
);
|
||||
@@ -75,15 +77,15 @@ describe('Budget DB Service', () => {
|
||||
});
|
||||
|
||||
it('should return an empty array if the user has no budgets', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
const result = await budgetRepo.getBudgetsForUser('user-123', mockLogger);
|
||||
expect(result).toEqual([]);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(expect.any(String), ['user-123']);
|
||||
expect(mockDb.query).toHaveBeenCalledWith(expect.any(String), ['user-123']);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(budgetRepo.getBudgetsForUser('user-123', mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve budgets.',
|
||||
);
|
||||
@@ -236,11 +238,11 @@ describe('Budget DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
};
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockUpdatedBudget], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockUpdatedBudget], rowCount: 1 });
|
||||
|
||||
const result = await budgetRepo.updateBudget(1, 'user-123', budgetUpdates, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.budgets SET'),
|
||||
[budgetUpdates.name, budgetUpdates.amount_cents, undefined, undefined, 1, 'user-123'],
|
||||
);
|
||||
@@ -249,7 +251,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if no rows are updated', async () => {
|
||||
// Arrange: Mock the query to return 0 rows affected
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
|
||||
await expect(
|
||||
budgetRepo.updateBudget(999, 'user-123', { name: 'Fail' }, mockLogger),
|
||||
@@ -258,7 +260,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
budgetRepo.updateBudget(1, 'user-123', { name: 'Fail' }, mockLogger),
|
||||
).rejects.toThrow('Failed to update budget.');
|
||||
@@ -271,9 +273,9 @@ describe('Budget DB Service', () => {
|
||||
|
||||
describe('deleteBudget', () => {
|
||||
it('should execute a DELETE query with user ownership check', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 1, command: 'DELETE', rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 1, command: 'DELETE', rows: [] });
|
||||
await budgetRepo.deleteBudget(1, 'user-123', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'DELETE FROM public.budgets WHERE budget_id = $1 AND user_id = $2',
|
||||
[1, 'user-123'],
|
||||
);
|
||||
@@ -281,7 +283,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if no rows are deleted', async () => {
|
||||
// Arrange: Mock the query to return 0 rows affected
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
|
||||
await expect(budgetRepo.deleteBudget(999, 'user-123', mockLogger)).rejects.toThrow(
|
||||
'Budget not found or user does not have permission to delete.',
|
||||
@@ -290,7 +292,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(budgetRepo.deleteBudget(1, 'user-123', mockLogger)).rejects.toThrow(
|
||||
'Failed to delete budget.',
|
||||
);
|
||||
@@ -306,7 +308,7 @@ describe('Budget DB Service', () => {
|
||||
const mockSpendingData: SpendingByCategory[] = [
|
||||
{ category_id: 1, category_name: 'Produce', total_spent_cents: 12345 },
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockSpendingData });
|
||||
mockDb.query.mockResolvedValue({ rows: mockSpendingData });
|
||||
|
||||
const result = await budgetRepo.getSpendingByCategory(
|
||||
'user-123',
|
||||
@@ -315,7 +317,7 @@ describe('Budget DB Service', () => {
|
||||
mockLogger,
|
||||
);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.get_spending_by_category($1, $2, $3)',
|
||||
['user-123', '2024-01-01', '2024-01-31'],
|
||||
);
|
||||
@@ -323,7 +325,7 @@ describe('Budget DB Service', () => {
|
||||
});
|
||||
|
||||
it('should return an empty array if there is no spending data', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
const result = await budgetRepo.getSpendingByCategory(
|
||||
'user-123',
|
||||
'2024-01-01',
|
||||
@@ -335,7 +337,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
budgetRepo.getSpendingByCategory('user-123', '2024-01-01', '2024-01-31', mockLogger),
|
||||
).rejects.toThrow('Failed to get spending analysis.');
|
||||
|
||||
@@ -7,9 +7,11 @@ import type { Budget, SpendingByCategory } from '../../types';
|
||||
import { GamificationRepository } from './gamification.db';
|
||||
|
||||
export class BudgetRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -1,9 +1,7 @@
|
||||
// src/services/db/deals.db.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import { DealsRepository } from './deals.db';
|
||||
import type { WatchedItemDeal } from '../../types';
|
||||
import type { Pool } from 'pg';
|
||||
|
||||
// Un-mock the module we are testing to ensure we use the real implementation.
|
||||
vi.unmock('./deals.db');
|
||||
@@ -22,11 +20,17 @@ import { logger as mockLogger } from '../logger.server';
|
||||
describe('Deals DB Service', () => {
|
||||
// Import the Pool type to use for casting the mock instance.
|
||||
let dealsRepo: DealsRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn()
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
dealsRepo = new DealsRepository(mockPoolInstance as unknown as Pool);
|
||||
|
||||
mockDb.query.mockReset()
|
||||
|
||||
// Instantiate the repository with the minimal mock db for each test
|
||||
dealsRepo = new DealsRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('findBestPricesForWatchedItems', () => {
|
||||
@@ -50,14 +54,14 @@ describe('Deals DB Service', () => {
|
||||
valid_to: '2025-12-24',
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockDeals });
|
||||
mockDb.query.mockResolvedValue({ rows: mockDeals });
|
||||
|
||||
// Act
|
||||
const result = await dealsRepo.findBestPricesForWatchedItems('user-123', mockLogger);
|
||||
|
||||
// Assert
|
||||
expect(result).toEqual(mockDeals);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM flyer_items fi'),
|
||||
['user-123'],
|
||||
);
|
||||
@@ -68,7 +72,7 @@ describe('Deals DB Service', () => {
|
||||
});
|
||||
|
||||
it('should return an empty array if no deals are found', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
|
||||
const result = await dealsRepo.findBestPricesForWatchedItems(
|
||||
'user-with-no-deals',
|
||||
@@ -80,7 +84,7 @@ describe('Deals DB Service', () => {
|
||||
|
||||
it('should re-throw the error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Connection Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(dealsRepo.findBestPricesForWatchedItems('user-1', mockLogger)).rejects.toThrow(
|
||||
dbError,
|
||||
|
||||
@@ -6,9 +6,11 @@ import type { Logger } from 'pino';
|
||||
import { logger as globalLogger } from '../logger.server';
|
||||
|
||||
export class DealsRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -37,11 +37,16 @@ import { withTransaction } from './connection.db';
|
||||
|
||||
describe('Flyer DB Service', () => {
|
||||
let flyerRepo: FlyerRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
mockDb.query.mockReset()
|
||||
|
||||
// In a transaction, `pool.connect()` returns a client. That client has a `release` method.
|
||||
flyerRepo = new FlyerRepository(mockDb);
|
||||
//In a transaction, `pool.connect()` returns a client. That client has a `release` method.
|
||||
// For these tests, we simulate this by having `connect` resolve to the pool instance itself,
|
||||
// and we ensure the `release` method is mocked on that instance.
|
||||
const mockClient = { ...mockPoolInstance, release: vi.fn() } as unknown as PoolClient;
|
||||
@@ -52,11 +57,11 @@ describe('Flyer DB Service', () => {
|
||||
|
||||
describe('findOrCreateStore', () => {
|
||||
it('should find an existing store and return its ID', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ store_id: 1 }] });
|
||||
mockDb.query.mockResolvedValue({ rows: [{ store_id: 1 }] });
|
||||
const result = await flyerRepo.findOrCreateStore('Existing Store', mockLogger);
|
||||
expect(result).toBe(1);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'SELECT store_id FROM public.stores WHERE name = $1',
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('SELECT store_id FROM public.stores WHERE name = $1'),
|
||||
['Existing Store'],
|
||||
);
|
||||
});
|
||||
@@ -64,11 +69,11 @@ describe('Flyer DB Service', () => {
|
||||
it('should create a new store if it does not exist', async () => {
|
||||
mockPoolInstance.query
|
||||
.mockResolvedValueOnce({ rows: [] }) // First SELECT finds nothing
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] }); // INSERT returns new ID
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] })
|
||||
const result = await flyerRepo.findOrCreateStore('New Store', mockLogger);
|
||||
expect(result).toBe(2);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id',
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id'),
|
||||
['New Store'],
|
||||
);
|
||||
});
|
||||
@@ -83,11 +88,11 @@ describe('Flyer DB Service', () => {
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 3 }] }); // Second SELECT finds the store
|
||||
|
||||
const result = await flyerRepo.findOrCreateStore('Racy Store', mockLogger);
|
||||
expect(result).toBe(3);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(3);
|
||||
});
|
||||
expect(result).toBe(3);
|
||||
//expect(mockDb.query).toHaveBeenCalledTimes(3);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
await expect(flyerRepo.findOrCreateStore('Any Store', mockLogger)).rejects.toThrow(
|
||||
@@ -129,6 +134,7 @@ describe('Flyer DB Service', () => {
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
status: 'processed',
|
||||
item_count: 10,
|
||||
uploaded_by: 'user-1',
|
||||
};
|
||||
@@ -139,7 +145,7 @@ describe('Flyer DB Service', () => {
|
||||
|
||||
expect(result).toEqual(mockFlyer);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(1);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO flyers'),
|
||||
[
|
||||
'test.jpg',
|
||||
@@ -150,6 +156,7 @@ describe('Flyer DB Service', () => {
|
||||
'2024-01-01',
|
||||
'2024-01-07',
|
||||
'123 Test St',
|
||||
'processed',
|
||||
10,
|
||||
'user-1',
|
||||
],
|
||||
|
||||
@@ -13,9 +13,11 @@ import type {
|
||||
} from '../../types';
|
||||
|
||||
export class FlyerRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
@@ -78,10 +80,10 @@ export class FlyerRepository {
|
||||
try {
|
||||
const query = `
|
||||
INSERT INTO flyers (
|
||||
file_name, image_url, icon_url, checksum, store_id, valid_from, valid_to,
|
||||
store_address, item_count, uploaded_by
|
||||
file_name, image_url, icon_url, checksum, store_id, valid_from, valid_to, store_address,
|
||||
status, item_count, uploaded_by
|
||||
)
|
||||
VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10)
|
||||
VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11)
|
||||
RETURNING *;
|
||||
`;
|
||||
const values = [
|
||||
@@ -93,8 +95,9 @@ export class FlyerRepository {
|
||||
flyerData.valid_from, // $6
|
||||
flyerData.valid_to, // $7
|
||||
flyerData.store_address, // $8
|
||||
flyerData.item_count, // $9
|
||||
flyerData.uploaded_by, // $10
|
||||
flyerData.status, // $9
|
||||
flyerData.item_count, // $10
|
||||
flyerData.uploaded_by, // $11
|
||||
];
|
||||
|
||||
const result = await this.db.query<Flyer>(query, values);
|
||||
|
||||
@@ -22,14 +22,18 @@ import { logger as mockLogger } from '../logger.server';
|
||||
|
||||
describe('Gamification DB Service', () => {
|
||||
let gamificationRepo: GamificationRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset the global mock's call history before each test.
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
gamificationRepo = new GamificationRepository(mockPoolInstance as unknown as Pool);
|
||||
});
|
||||
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
gamificationRepo = new GamificationRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getAllAchievements', () => {
|
||||
it('should execute the correct SELECT query and return achievements', async () => {
|
||||
const mockAchievements: Achievement[] = [
|
||||
@@ -42,11 +46,11 @@ describe('Gamification DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockAchievements });
|
||||
mockDb.query.mockResolvedValue({ rows: mockAchievements });
|
||||
|
||||
const result = await gamificationRepo.getAllAchievements(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.achievements ORDER BY points_value ASC, name ASC',
|
||||
);
|
||||
expect(result).toEqual(mockAchievements);
|
||||
@@ -54,7 +58,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(gamificationRepo.getAllAchievements(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve achievements.',
|
||||
);
|
||||
@@ -79,10 +83,10 @@ describe('Gamification DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockUserAchievements });
|
||||
mockDb.query.mockResolvedValue({ rows: mockUserAchievements });
|
||||
|
||||
const result = await gamificationRepo.getUserAchievements('user-123', mockLogger);
|
||||
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.user_achievements ua'),
|
||||
['user-123'],
|
||||
@@ -92,7 +96,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(gamificationRepo.getUserAchievements('user-123', mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve user achievements.',
|
||||
);
|
||||
@@ -105,10 +109,10 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
describe('awardAchievement', () => {
|
||||
it('should call the award_achievement database function with the correct parameters', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] }); // The function returns void
|
||||
mockDb.query.mockResolvedValue({ rows: [] }); // The function returns void
|
||||
await gamificationRepo.awardAchievement('user-123', 'Test Achievement', mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT public.award_achievement($1, $2)',
|
||||
['user-123', 'Test Achievement'],
|
||||
);
|
||||
@@ -117,7 +121,7 @@ describe('Gamification DB Service', () => {
|
||||
it('should throw ForeignKeyConstraintError if user or achievement does not exist', async () => {
|
||||
const dbError = new Error('violates foreign key constraint');
|
||||
(dbError as Error & { code: string }).code = '23503';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
gamificationRepo.awardAchievement(
|
||||
'non-existent-user',
|
||||
@@ -133,7 +137,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
gamificationRepo.awardAchievement('user-123', 'Test Achievement', mockLogger),
|
||||
).rejects.toThrow('Failed to award achievement.');
|
||||
@@ -147,13 +151,12 @@ describe('Gamification DB Service', () => {
|
||||
describe('getLeaderboard', () => {
|
||||
it('should execute the correct SELECT query with a LIMIT and return leaderboard users', async () => {
|
||||
const mockLeaderboard: LeaderboardUser[] = [
|
||||
{ user_id: 'user-1', full_name: 'User One', avatar_url: null, points: 500, rank: '1' },
|
||||
{ user_id: 'user-2', full_name: 'User Two', avatar_url: null, points: 450, rank: '2' },
|
||||
{ user_id: 'user-1', full_name: 'User One', avatar_url: null, points: 500, rank: '1' },
|
||||
{ user_id: 'user-2', full_name: 'User Two', avatar_url: null, points: 450, rank: '2' }
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockLeaderboard });
|
||||
mockDb.query.mockResolvedValue({ rows: mockLeaderboard });
|
||||
|
||||
const result = await gamificationRepo.getLeaderboard(10, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(1);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('RANK() OVER (ORDER BY points DESC)'),
|
||||
@@ -164,7 +167,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(gamificationRepo.getLeaderboard(10, mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve leaderboard.',
|
||||
);
|
||||
|
||||
@@ -6,9 +6,11 @@ import type { Logger } from 'pino';
|
||||
import { Achievement, UserAchievement, LeaderboardUser } from '../../types';
|
||||
|
||||
export class GamificationRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -2,7 +2,6 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import type { Pool } from 'pg';
|
||||
|
||||
// Un-mock the module we are testing to ensure we use the real implementation.
|
||||
vi.unmock('./notification.db');
|
||||
|
||||
import { NotificationRepository } from './notification.db';
|
||||
@@ -11,6 +10,7 @@ import { ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
||||
import type { Notification } from '../../types';
|
||||
import { createMockNotification } from '../../tests/utils/mockFactories';
|
||||
|
||||
|
||||
// Mock the logger to prevent console output during tests
|
||||
vi.mock('../logger.server', () => ({
|
||||
logger: {
|
||||
@@ -24,10 +24,14 @@ import { logger as mockLogger } from '../logger.server';
|
||||
|
||||
describe('Notification DB Service', () => {
|
||||
let notificationRepo: NotificationRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
|
||||
notificationRepo = new NotificationRepository(mockPoolInstance as unknown as Pool);
|
||||
});
|
||||
|
||||
|
||||
@@ -6,9 +6,11 @@ import type { Logger } from 'pino';
|
||||
import type { Notification } from '../../types';
|
||||
|
||||
export class NotificationRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -16,9 +16,11 @@ import {
|
||||
} from '../../types';
|
||||
|
||||
export class PersonalizationRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -4,8 +4,11 @@
|
||||
* It is configured via environment variables and should only be used on the server.
|
||||
*/
|
||||
import nodemailer from 'nodemailer';
|
||||
import type { Job } from 'bullmq';
|
||||
import type { Logger } from 'pino';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
import { WatchedItemDeal } from '../types';
|
||||
import type { EmailJobData } from './queues.server';
|
||||
|
||||
// 1. Create a Nodemailer transporter using SMTP configuration from environment variables.
|
||||
// For development, you can use a service like Ethereal (https://ethereal.email/)
|
||||
@@ -20,18 +23,11 @@ const transporter = nodemailer.createTransport({
|
||||
},
|
||||
});
|
||||
|
||||
interface EmailOptions {
|
||||
to: string;
|
||||
subject: string;
|
||||
text: string;
|
||||
html: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Sends an email using the pre-configured transporter.
|
||||
* @param options The email options, including recipient, subject, and body.
|
||||
*/
|
||||
export const sendEmail = async (options: EmailOptions, logger: Logger) => {
|
||||
export const sendEmail = async (options: EmailJobData, logger: Logger) => {
|
||||
const mailOptions = {
|
||||
from: `"Flyer Crawler" <${process.env.SMTP_FROM_EMAIL}>`, // sender address
|
||||
to: options.to,
|
||||
@@ -40,16 +36,37 @@ export const sendEmail = async (options: EmailOptions, logger: Logger) => {
|
||||
html: options.html,
|
||||
};
|
||||
|
||||
const info = await transporter.sendMail(mailOptions);
|
||||
logger.info(
|
||||
{ to: options.to, subject: options.subject, messageId: info.messageId },
|
||||
`Email sent successfully.`,
|
||||
);
|
||||
};
|
||||
|
||||
/**
|
||||
* Processes an email sending job from the queue.
|
||||
* This is the entry point for the email worker.
|
||||
* It encapsulates logging and error handling for the job.
|
||||
* @param job The BullMQ job object.
|
||||
*/
|
||||
export const processEmailJob = async (job: Job<EmailJobData>) => {
|
||||
const jobLogger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
recipient: job.data.to,
|
||||
});
|
||||
|
||||
jobLogger.info(`Picked up email job.`);
|
||||
|
||||
try {
|
||||
const info = await transporter.sendMail(mailOptions);
|
||||
logger.info(
|
||||
{ to: options.to, subject: options.subject, messageId: info.messageId },
|
||||
`Email sent successfully.`,
|
||||
);
|
||||
await sendEmail(job.data, jobLogger);
|
||||
} catch (error) {
|
||||
logger.error({ err: error, to: options.to, subject: options.subject }, 'Failed to send email.');
|
||||
// Re-throwing the error is important so the background job knows it failed.
|
||||
throw error;
|
||||
const wrappedError = error instanceof Error ? error : new Error(String(error));
|
||||
jobLogger.error(
|
||||
{ err: wrappedError, jobData: job.data, attemptsMade: job.attemptsMade },
|
||||
`Email job failed.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
};
|
||||
|
||||
@@ -92,16 +109,22 @@ export const sendDealNotificationEmail = async (
|
||||
|
||||
const text = `Hi ${recipientName},\n\nWe found some great deals on items you're watching. Visit the deals page on the site to learn more.\n\nFlyer Crawler`;
|
||||
|
||||
// Use the generic sendEmail function to send the composed email
|
||||
await sendEmail(
|
||||
{
|
||||
to,
|
||||
subject,
|
||||
text,
|
||||
html,
|
||||
},
|
||||
logger,
|
||||
);
|
||||
try {
|
||||
// Use the generic sendEmail function to send the composed email
|
||||
await sendEmail(
|
||||
{
|
||||
to,
|
||||
subject,
|
||||
text,
|
||||
html,
|
||||
},
|
||||
logger,
|
||||
);
|
||||
} catch (err) {
|
||||
const error = err instanceof Error ? err : new Error(String(err));
|
||||
logger.error({ err: error, to, subject }, 'Failed to send email.');
|
||||
throw error;
|
||||
}
|
||||
};
|
||||
|
||||
/**
|
||||
|
||||
130
src/services/flyerAiProcessor.server.test.ts
Normal file
130
src/services/flyerAiProcessor.server.test.ts
Normal file
@@ -0,0 +1,130 @@
|
||||
// src/services/flyerAiProcessor.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||
import { AiDataValidationError } from './processingErrors';
|
||||
import { logger } from './logger.server';
|
||||
import type { AIService } from './aiService.server';
|
||||
import type { PersonalizationRepository } from './db/personalization.db';
|
||||
import type { FlyerJobData } from './queues.server';
|
||||
|
||||
vi.mock('./logger.server', () => ({
|
||||
logger: {
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
child: vi.fn().mockReturnThis(),
|
||||
},
|
||||
}));
|
||||
|
||||
const createMockJobData = (data: Partial<FlyerJobData>): FlyerJobData => ({
|
||||
filePath: '/tmp/flyer.jpg',
|
||||
originalFileName: 'flyer.jpg',
|
||||
checksum: 'checksum-123',
|
||||
...data,
|
||||
});
|
||||
|
||||
describe('FlyerAiProcessor', () => {
|
||||
let service: FlyerAiProcessor;
|
||||
let mockAiService: AIService;
|
||||
let mockPersonalizationRepo: PersonalizationRepository;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
mockAiService = {
|
||||
extractCoreDataFromFlyerImage: vi.fn(),
|
||||
} as unknown as AIService;
|
||||
mockPersonalizationRepo = {
|
||||
getAllMasterItems: vi.fn().mockResolvedValue([]),
|
||||
} as unknown as PersonalizationRepository;
|
||||
|
||||
service = new FlyerAiProcessor(mockAiService, mockPersonalizationRepo);
|
||||
});
|
||||
|
||||
it('should call AI service and return validated data on success', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: 'AI Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 AI St',
|
||||
// FIX: Add an item to pass the new "must have items" quality check.
|
||||
items: [
|
||||
{
|
||||
item: 'Test Item',
|
||||
price_display: '$1.99',
|
||||
price_in_cents: 199,
|
||||
// ADDED to satisfy ExtractedFlyerItem type
|
||||
quantity: 'each',
|
||||
category_name: 'Grocery',
|
||||
},
|
||||
],
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
|
||||
const result = await service.extractAndValidateData([], jobData, logger);
|
||||
|
||||
expect(mockAiService.extractCoreDataFromFlyerImage).toHaveBeenCalledTimes(1);
|
||||
expect(mockPersonalizationRepo.getAllMasterItems).toHaveBeenCalledTimes(1);
|
||||
expect(result.data).toEqual(mockAiResponse);
|
||||
expect(result.needsReview).toBe(false);
|
||||
});
|
||||
|
||||
it('should throw AiDataValidationError if AI response has incorrect data structure', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
// Mock AI to return a structurally invalid response (e.g., items is not an array)
|
||||
const invalidResponse = {
|
||||
store_name: 'Invalid Store',
|
||||
items: 'not-an-array',
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(invalidResponse as any);
|
||||
|
||||
await expect(service.extractAndValidateData([], jobData, logger)).rejects.toThrow(
|
||||
AiDataValidationError,
|
||||
);
|
||||
});
|
||||
|
||||
it('should pass validation even if store_name is missing', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: null, // Missing store name
|
||||
items: [{ item: 'Test Item', price_display: '$1.99', price_in_cents: 199, quantity: 'each', category_name: 'Grocery' }],
|
||||
// ADDED to satisfy AiFlyerDataSchema
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse as any);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const result = await service.extractAndValidateData([], jobData, logger);
|
||||
|
||||
// It should not throw, but return the data and log a warning.
|
||||
expect(result.data).toEqual(mockAiResponse);
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(expect.any(Object), expect.stringContaining('missing a store name. The transformer will use a fallback. Flagging for review.'));
|
||||
});
|
||||
|
||||
it('should pass validation even if items array is empty', async () => {
|
||||
const jobData = createMockJobData({});
|
||||
const mockAiResponse = {
|
||||
store_name: 'Test Store',
|
||||
items: [], // Empty items array
|
||||
// ADDED to satisfy AiFlyerDataSchema
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
};
|
||||
vi.mocked(mockAiService.extractCoreDataFromFlyerImage).mockResolvedValue(mockAiResponse);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
const result = await service.extractAndValidateData([], jobData, logger);
|
||||
expect(result.data).toEqual(mockAiResponse);
|
||||
expect(result.needsReview).toBe(true);
|
||||
expect(logger.warn).toHaveBeenCalledWith(expect.any(Object), expect.stringContaining('contains no items. The flyer will be saved with an item_count of 0. Flagging for review.'));
|
||||
});
|
||||
});
|
||||
112
src/services/flyerAiProcessor.server.ts
Normal file
112
src/services/flyerAiProcessor.server.ts
Normal file
@@ -0,0 +1,112 @@
|
||||
// src/services/flyerAiProcessor.server.ts
|
||||
import { z } from 'zod';
|
||||
import type { Logger } from 'pino';
|
||||
import type { AIService } from './aiService.server';
|
||||
import type { PersonalizationRepository } from './db/personalization.db';
|
||||
import { AiDataValidationError } from './processingErrors';
|
||||
import type { FlyerJobData } from '../types/job-data';
|
||||
|
||||
// Helper for consistent required string validation (handles missing/null/empty)
|
||||
const requiredString = (message: string) =>
|
||||
z.preprocess((val) => val ?? '', z.string().min(1, message));
|
||||
|
||||
// --- Zod Schemas for AI Response Validation ---
|
||||
const ExtractedFlyerItemSchema = z.object({
|
||||
item: z.string().nullable(),
|
||||
price_display: z.string().nullable(),
|
||||
price_in_cents: z.number().nullable(),
|
||||
quantity: z.string().nullable(),
|
||||
category_name: z.string().nullable(),
|
||||
master_item_id: z.number().nullish(),
|
||||
});
|
||||
|
||||
export const AiFlyerDataSchema = z.object({
|
||||
store_name: z.string().nullable(),
|
||||
valid_from: z.string().nullable(),
|
||||
valid_to: z.string().nullable(),
|
||||
store_address: z.string().nullable(),
|
||||
items: z.array(ExtractedFlyerItemSchema),
|
||||
});
|
||||
|
||||
export type ValidatedAiDataType = z.infer<typeof AiFlyerDataSchema>;
|
||||
|
||||
export interface AiProcessorResult {
|
||||
data: ValidatedAiDataType;
|
||||
needsReview: boolean;
|
||||
}
|
||||
|
||||
/**
|
||||
* This class encapsulates the logic for interacting with the AI service
|
||||
* to extract and validate data from flyer images.
|
||||
*/
|
||||
export class FlyerAiProcessor {
|
||||
constructor(
|
||||
private ai: AIService,
|
||||
private personalizationRepo: PersonalizationRepository,
|
||||
) {}
|
||||
|
||||
/**
|
||||
* Validates the raw data from the AI against the Zod schema.
|
||||
*/
|
||||
private _validateAiData(
|
||||
extractedData: unknown,
|
||||
logger: Logger,
|
||||
): AiProcessorResult {
|
||||
const validationResult = AiFlyerDataSchema.safeParse(extractedData);
|
||||
if (!validationResult.success) {
|
||||
const errors = validationResult.error.flatten();
|
||||
logger.error({ errors, rawData: extractedData }, 'AI response failed validation.');
|
||||
throw new AiDataValidationError(
|
||||
'AI response validation failed. The returned data structure is incorrect.',
|
||||
errors,
|
||||
extractedData,
|
||||
);
|
||||
}
|
||||
|
||||
// --- NEW QUALITY CHECK ---
|
||||
// After structural validation, perform semantic quality checks.
|
||||
const { store_name, items } = validationResult.data;
|
||||
let needsReview = false;
|
||||
|
||||
// 1. Check for a valid store name, but don't fail the job.
|
||||
// The data transformer will handle this by assigning a fallback name.
|
||||
if (!store_name || store_name.trim() === '') {
|
||||
logger.warn({ rawData: extractedData }, 'AI response is missing a store name. The transformer will use a fallback. Flagging for review.');
|
||||
needsReview = true;
|
||||
}
|
||||
|
||||
// 2. Check that at least one item was extracted, but don't fail the job.
|
||||
// An admin can review a flyer with 0 items.
|
||||
if (!items || items.length === 0) {
|
||||
logger.warn({ rawData: extractedData }, 'AI response contains no items. The flyer will be saved with an item_count of 0. Flagging for review.');
|
||||
needsReview = true;
|
||||
}
|
||||
|
||||
logger.info(`AI extracted ${validationResult.data.items.length} items.`);
|
||||
return { data: validationResult.data, needsReview };
|
||||
}
|
||||
|
||||
/**
|
||||
* Calls the AI service to extract structured data from the flyer images and validates the response.
|
||||
*/
|
||||
public async extractAndValidateData(
|
||||
imagePaths: { path: string; mimetype: string }[],
|
||||
jobData: FlyerJobData,
|
||||
logger: Logger,
|
||||
): Promise<AiProcessorResult> {
|
||||
logger.info(`Starting AI data extraction.`);
|
||||
const { submitterIp, userProfileAddress } = jobData;
|
||||
const masterItems = await this.personalizationRepo.getAllMasterItems(logger);
|
||||
logger.debug(`Retrieved ${masterItems.length} master items for AI matching.`);
|
||||
|
||||
const extractedData = await this.ai.extractCoreDataFromFlyerImage(
|
||||
imagePaths,
|
||||
masterItems,
|
||||
submitterIp,
|
||||
userProfileAddress,
|
||||
logger,
|
||||
);
|
||||
|
||||
return this._validateAiData(extractedData, logger);
|
||||
}
|
||||
}
|
||||
@@ -3,8 +3,7 @@ import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
import { logger as mockLogger } from './logger.server';
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
import type { z } from 'zod';
|
||||
import type { AiFlyerDataSchema } from './flyerProcessingService.server';
|
||||
import type { AiProcessorResult } from './flyerAiProcessor.server';
|
||||
import type { FlyerItemInsert } from '../types';
|
||||
|
||||
// Mock the dependencies
|
||||
@@ -29,29 +28,32 @@ describe('FlyerDataTransformer', () => {
|
||||
|
||||
it('should transform AI data into database-ready format with a user ID', async () => {
|
||||
// Arrange
|
||||
const extractedData: z.infer<typeof AiFlyerDataSchema> = {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{
|
||||
item: 'Milk',
|
||||
price_display: '$3.99',
|
||||
price_in_cents: 399,
|
||||
quantity: '1L',
|
||||
category_name: 'Dairy',
|
||||
master_item_id: 10,
|
||||
},
|
||||
{
|
||||
item: 'Bread',
|
||||
price_display: '$2.49',
|
||||
price_in_cents: 249,
|
||||
quantity: '1 loaf',
|
||||
category_name: 'Bakery',
|
||||
master_item_id: null,
|
||||
},
|
||||
],
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
{
|
||||
item: 'Milk',
|
||||
price_display: '$3.99',
|
||||
price_in_cents: 399,
|
||||
quantity: '1L',
|
||||
category_name: 'Dairy',
|
||||
master_item_id: 10,
|
||||
},
|
||||
{
|
||||
item: 'Bread',
|
||||
price_display: '$2.49',
|
||||
price_in_cents: 249,
|
||||
quantity: '1 loaf',
|
||||
category_name: 'Bakery',
|
||||
master_item_id: null,
|
||||
},
|
||||
],
|
||||
},
|
||||
needsReview: false,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
const originalFileName = 'my-flyer.pdf';
|
||||
@@ -60,7 +62,7 @@ describe('FlyerDataTransformer', () => {
|
||||
|
||||
// Act
|
||||
const { flyerData, itemsForDb } = await transformer.transform(
|
||||
extractedData,
|
||||
aiResult,
|
||||
imagePaths,
|
||||
originalFileName,
|
||||
checksum,
|
||||
@@ -89,6 +91,7 @@ describe('FlyerDataTransformer', () => {
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
item_count: 2,
|
||||
status: 'processed',
|
||||
uploaded_by: userId,
|
||||
});
|
||||
|
||||
@@ -109,9 +112,6 @@ describe('FlyerDataTransformer', () => {
|
||||
view_count: 0,
|
||||
click_count: 0,
|
||||
}),
|
||||
); // Use a more specific type assertion to check for the added property.
|
||||
expect((itemsForDb[0] as FlyerItemInsert & { updated_at: string }).updated_at).toBeTypeOf(
|
||||
'string',
|
||||
);
|
||||
|
||||
// 3. Check that generateFlyerIcon was called correctly
|
||||
@@ -124,12 +124,15 @@ describe('FlyerDataTransformer', () => {
|
||||
|
||||
it('should handle missing optional data gracefully', async () => {
|
||||
// Arrange
|
||||
const extractedData: z.infer<typeof AiFlyerDataSchema> = {
|
||||
store_name: '', // Empty store name
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
items: [], // No items
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: '', // Empty store name
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
items: [], // No items
|
||||
},
|
||||
needsReview: true,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/another.png', mimetype: 'image/png' }];
|
||||
const originalFileName = 'another.png';
|
||||
@@ -140,7 +143,7 @@ describe('FlyerDataTransformer', () => {
|
||||
|
||||
// Act
|
||||
const { flyerData, itemsForDb } = await transformer.transform(
|
||||
extractedData,
|
||||
aiResult,
|
||||
imagePaths,
|
||||
originalFileName,
|
||||
checksum,
|
||||
@@ -153,6 +156,9 @@ describe('FlyerDataTransformer', () => {
|
||||
expect(mockLogger.info).toHaveBeenCalledWith(
|
||||
'Starting data transformation from AI output to database format.',
|
||||
);
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
'AI did not return a store name. Using fallback "Unknown Store (auto)".',
|
||||
);
|
||||
expect(mockLogger.info).toHaveBeenCalledWith(
|
||||
{ itemCount: 0, storeName: 'Unknown Store (auto)' },
|
||||
'Data transformation complete.',
|
||||
@@ -169,7 +175,69 @@ describe('FlyerDataTransformer', () => {
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
item_count: 0,
|
||||
status: 'needs_review',
|
||||
uploaded_by: undefined, // Should be undefined
|
||||
});
|
||||
});
|
||||
|
||||
it('should correctly normalize item fields with null, undefined, or empty values', async () => {
|
||||
// Arrange
|
||||
const aiResult: AiProcessorResult = {
|
||||
data: {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
items: [
|
||||
// Case 1: All fields are null or undefined
|
||||
{
|
||||
item: null,
|
||||
price_display: null,
|
||||
price_in_cents: null,
|
||||
quantity: null,
|
||||
category_name: null,
|
||||
master_item_id: null,
|
||||
},
|
||||
// Case 2: Fields are empty strings
|
||||
{
|
||||
item: ' ', // whitespace only
|
||||
price_display: '',
|
||||
price_in_cents: 200,
|
||||
quantity: '',
|
||||
category_name: '',
|
||||
master_item_id: 20,
|
||||
},
|
||||
],
|
||||
},
|
||||
needsReview: false,
|
||||
};
|
||||
const imagePaths = [{ path: '/uploads/flyer-page-1.jpg', mimetype: 'image/jpeg' }];
|
||||
|
||||
// Act
|
||||
const { itemsForDb } = await transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
'file.pdf',
|
||||
'checksum',
|
||||
'user-1',
|
||||
mockLogger,
|
||||
);
|
||||
|
||||
// Assert
|
||||
expect(itemsForDb).toHaveLength(2);
|
||||
|
||||
// Check Case 1 (null/undefined values)
|
||||
expect(itemsForDb[0]).toEqual(
|
||||
expect.objectContaining({
|
||||
item: 'Unknown Item', price_display: '', quantity: '', category_name: 'Other/Miscellaneous', master_item_id: undefined,
|
||||
}),
|
||||
);
|
||||
|
||||
// Check Case 2 (empty string values)
|
||||
expect(itemsForDb[1]).toEqual(
|
||||
expect.objectContaining({
|
||||
item: 'Unknown Item', price_display: '', quantity: '', category_name: 'Other/Miscellaneous', master_item_id: 20,
|
||||
}),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -2,8 +2,8 @@
|
||||
import path from 'path';
|
||||
import type { z } from 'zod';
|
||||
import type { Logger } from 'pino';
|
||||
import type { FlyerInsert, FlyerItemInsert } from '../types';
|
||||
import type { AiFlyerDataSchema } from './flyerProcessingService.server';
|
||||
import type { FlyerInsert, FlyerItemInsert, FlyerStatus } from '../types';
|
||||
import type { AiFlyerDataSchema, AiProcessorResult } from './flyerAiProcessor.server';
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
|
||||
/**
|
||||
@@ -11,6 +11,31 @@ import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
* into the structured format required for database insertion (FlyerInsert and FlyerItemInsert).
|
||||
*/
|
||||
export class FlyerDataTransformer {
|
||||
/**
|
||||
* Normalizes a single raw item from the AI, providing default values for nullable fields.
|
||||
* @param item The raw item object from the AI.
|
||||
* @returns A normalized item object ready for database insertion.
|
||||
*/
|
||||
private _normalizeItem(
|
||||
item: z.infer<typeof AiFlyerDataSchema>['items'][number],
|
||||
): FlyerItemInsert {
|
||||
return {
|
||||
...item,
|
||||
// Use logical OR to default falsy values (null, undefined, '') to a fallback.
|
||||
// The trim is important for cases where the AI returns only whitespace.
|
||||
item: String(item.item || '').trim() || 'Unknown Item',
|
||||
// Use nullish coalescing to default only null/undefined to an empty string.
|
||||
price_display: String(item.price_display ?? ''),
|
||||
quantity: String(item.quantity ?? ''),
|
||||
// Use logical OR to default falsy category names (null, undefined, '') to a fallback.
|
||||
category_name: String(item.category_name || 'Other/Miscellaneous'),
|
||||
// Use nullish coalescing to convert null to undefined for the database.
|
||||
master_item_id: item.master_item_id ?? undefined,
|
||||
view_count: 0,
|
||||
click_count: 0,
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Transforms AI-extracted data into database-ready flyer and item records.
|
||||
* @param extractedData The validated data from the AI.
|
||||
@@ -22,7 +47,7 @@ export class FlyerDataTransformer {
|
||||
* @returns A promise that resolves to an object containing the prepared flyer and item data.
|
||||
*/
|
||||
async transform(
|
||||
extractedData: z.infer<typeof AiFlyerDataSchema>,
|
||||
aiResult: AiProcessorResult,
|
||||
imagePaths: { path: string; mimetype: string }[],
|
||||
originalFileName: string,
|
||||
checksum: string,
|
||||
@@ -31,6 +56,8 @@ export class FlyerDataTransformer {
|
||||
): Promise<{ flyerData: FlyerInsert; itemsForDb: FlyerItemInsert[] }> {
|
||||
logger.info('Starting data transformation from AI output to database format.');
|
||||
|
||||
const { data: extractedData, needsReview } = aiResult;
|
||||
|
||||
const firstImage = imagePaths[0].path;
|
||||
const iconFileName = await generateFlyerIcon(
|
||||
firstImage,
|
||||
@@ -38,39 +65,25 @@ export class FlyerDataTransformer {
|
||||
logger,
|
||||
);
|
||||
|
||||
const itemsForDb: FlyerItemInsert[] = extractedData.items.map((item) => ({
|
||||
...item,
|
||||
// Ensure 'item' is always a string, defaulting to 'Unknown Item' if null/undefined/empty.
|
||||
item:
|
||||
item.item === null || item.item === undefined || String(item.item).trim() === ''
|
||||
? 'Unknown Item'
|
||||
: String(item.item),
|
||||
// Ensure 'price_display' is always a string, defaulting to empty if null/undefined.
|
||||
price_display:
|
||||
item.price_display === null || item.price_display === undefined
|
||||
? ''
|
||||
: String(item.price_display),
|
||||
// Ensure 'quantity' is always a string, defaulting to empty if null/undefined.
|
||||
quantity: item.quantity === null || item.quantity === undefined ? '' : String(item.quantity),
|
||||
// Ensure 'category_name' is always a string, defaulting to 'Other/Miscellaneous' if null/undefined.
|
||||
category_name: item.category_name === null || item.category_name === undefined ? 'Other/Miscellaneous' : String(item.category_name),
|
||||
master_item_id: item.master_item_id === null ? undefined : item.master_item_id, // Convert null to undefined
|
||||
view_count: 0,
|
||||
click_count: 0,
|
||||
updated_at: new Date().toISOString(),
|
||||
}));
|
||||
const itemsForDb: FlyerItemInsert[] = extractedData.items.map((item) => this._normalizeItem(item));
|
||||
|
||||
const storeName = extractedData.store_name || 'Unknown Store (auto)';
|
||||
if (!extractedData.store_name) {
|
||||
logger.warn('AI did not return a store name. Using fallback "Unknown Store (auto)".');
|
||||
}
|
||||
|
||||
const flyerData: FlyerInsert = {
|
||||
file_name: originalFileName,
|
||||
image_url: `/flyer-images/${path.basename(firstImage)}`,
|
||||
icon_url: `/flyer-images/icons/${iconFileName}`,
|
||||
checksum,
|
||||
store_name: extractedData.store_name || 'Unknown Store (auto)',
|
||||
store_name: storeName,
|
||||
valid_from: extractedData.valid_from,
|
||||
valid_to: extractedData.valid_to,
|
||||
store_address: extractedData.store_address, // The number of items is now calculated directly from the transformed data.
|
||||
item_count: itemsForDb.length,
|
||||
uploaded_by: userId,
|
||||
status: needsReview ? 'needs_review' : 'processed',
|
||||
};
|
||||
|
||||
logger.info(
|
||||
|
||||
129
src/services/flyerFileHandler.server.test.ts
Normal file
129
src/services/flyerFileHandler.server.test.ts
Normal file
@@ -0,0 +1,129 @@
|
||||
// src/services/flyerFileHandler.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { Job } from 'bullmq';
|
||||
import type { Dirent } from 'node:fs';
|
||||
import sharp from 'sharp';
|
||||
import { FlyerFileHandler, ICommandExecutor, IFileSystem } from './flyerFileHandler.server';
|
||||
import { PdfConversionError, UnsupportedFileTypeError } from './processingErrors';
|
||||
import { logger } from './logger.server';
|
||||
import type { FlyerJobData } from './queues.server';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('sharp', () => {
|
||||
const mockSharpInstance = {
|
||||
png: vi.fn().mockReturnThis(),
|
||||
toFile: vi.fn().mockResolvedValue({}),
|
||||
};
|
||||
return {
|
||||
__esModule: true,
|
||||
default: vi.fn(() => mockSharpInstance),
|
||||
};
|
||||
});
|
||||
|
||||
vi.mock('./logger.server', () => ({
|
||||
logger: {
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
child: vi.fn().mockReturnThis(),
|
||||
},
|
||||
}));
|
||||
|
||||
const createMockJob = (data: Partial<FlyerJobData>): Job<FlyerJobData> => {
|
||||
return {
|
||||
id: 'job-1',
|
||||
data: {
|
||||
filePath: '/tmp/flyer.jpg',
|
||||
originalFileName: 'flyer.jpg',
|
||||
checksum: 'checksum-123',
|
||||
...data,
|
||||
},
|
||||
updateProgress: vi.fn(),
|
||||
} as unknown as Job<FlyerJobData>;
|
||||
};
|
||||
|
||||
describe('FlyerFileHandler', () => {
|
||||
let service: FlyerFileHandler;
|
||||
let mockFs: IFileSystem;
|
||||
let mockExec: ICommandExecutor;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
mockFs = {
|
||||
readdir: vi.fn().mockResolvedValue([]),
|
||||
unlink: vi.fn(),
|
||||
};
|
||||
mockExec = vi.fn().mockResolvedValue({ stdout: 'success', stderr: '' });
|
||||
|
||||
service = new FlyerFileHandler(mockFs, mockExec);
|
||||
});
|
||||
|
||||
it('should convert a PDF and return image paths', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.pdf' });
|
||||
vi.mocked(mockFs.readdir).mockResolvedValue([
|
||||
{ name: 'flyer-1.jpg' },
|
||||
{ name: 'flyer-2.jpg' },
|
||||
] as Dirent[]);
|
||||
|
||||
const { imagePaths, createdImagePaths } = await service.prepareImageInputs(
|
||||
'/tmp/flyer.pdf',
|
||||
job,
|
||||
logger,
|
||||
);
|
||||
|
||||
expect(mockExec).toHaveBeenCalledWith('pdftocairo -jpeg -r 150 "/tmp/flyer.pdf" "/tmp/flyer"');
|
||||
expect(imagePaths).toHaveLength(2);
|
||||
expect(imagePaths[0].path).toContain('flyer-1.jpg');
|
||||
expect(createdImagePaths).toHaveLength(2);
|
||||
});
|
||||
|
||||
it('should throw PdfConversionError if PDF conversion yields no images', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.pdf' });
|
||||
vi.mocked(mockFs.readdir).mockResolvedValue([]); // No images found
|
||||
|
||||
await expect(service.prepareImageInputs('/tmp/flyer.pdf', job, logger)).rejects.toThrow(
|
||||
PdfConversionError,
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle supported image types directly', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.jpg' });
|
||||
const { imagePaths, createdImagePaths } = await service.prepareImageInputs(
|
||||
'/tmp/flyer.jpg',
|
||||
job,
|
||||
logger,
|
||||
);
|
||||
|
||||
expect(imagePaths).toEqual([{ path: '/tmp/flyer.jpg', mimetype: 'image/jpeg' }]);
|
||||
expect(createdImagePaths).toEqual([]);
|
||||
expect(mockExec).not.toHaveBeenCalled();
|
||||
expect(sharp).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should convert convertible image types to PNG', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.gif' });
|
||||
const mockSharpInstance = sharp('/tmp/flyer.gif');
|
||||
vi.mocked(mockSharpInstance.toFile).mockResolvedValue({} as any);
|
||||
|
||||
const { imagePaths, createdImagePaths } = await service.prepareImageInputs(
|
||||
'/tmp/flyer.gif',
|
||||
job,
|
||||
logger,
|
||||
);
|
||||
|
||||
expect(sharp).toHaveBeenCalledWith('/tmp/flyer.gif');
|
||||
expect(mockSharpInstance.png).toHaveBeenCalled();
|
||||
expect(mockSharpInstance.toFile).toHaveBeenCalledWith('/tmp/flyer-converted.png');
|
||||
expect(imagePaths).toEqual([{ path: '/tmp/flyer-converted.png', mimetype: 'image/png' }]);
|
||||
expect(createdImagePaths).toEqual(['/tmp/flyer-converted.png']);
|
||||
});
|
||||
|
||||
it('should throw UnsupportedFileTypeError for unsupported types', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/document.txt' });
|
||||
await expect(service.prepareImageInputs('/tmp/document.txt', job, logger)).rejects.toThrow(
|
||||
UnsupportedFileTypeError,
|
||||
);
|
||||
});
|
||||
});
|
||||
206
src/services/flyerFileHandler.server.ts
Normal file
206
src/services/flyerFileHandler.server.ts
Normal file
@@ -0,0 +1,206 @@
|
||||
// src/services/flyerFileHandler.server.ts
|
||||
import path from 'path';
|
||||
import sharp from 'sharp';
|
||||
import type { Dirent } from 'node:fs';
|
||||
import type { Job } from 'bullmq';
|
||||
import type { Logger } from 'pino';
|
||||
import { ImageConversionError, PdfConversionError, UnsupportedFileTypeError } from './processingErrors';
|
||||
import type { FlyerJobData } from './queues.server';
|
||||
|
||||
// Define the image formats supported by the AI model
|
||||
const SUPPORTED_IMAGE_EXTENSIONS = ['.jpg', '.jpeg', '.png', '.webp', '.heic', '.heif'];
|
||||
// Define image formats that are not directly supported but can be converted to PNG.
|
||||
const CONVERTIBLE_IMAGE_EXTENSIONS = ['.gif', '.tiff', '.svg', '.bmp'];
|
||||
|
||||
export interface IFileSystem {
|
||||
readdir(path: string, options: { withFileTypes: true }): Promise<Dirent[]>;
|
||||
unlink(path: string): Promise<void>;
|
||||
}
|
||||
|
||||
export interface ICommandExecutor {
|
||||
(command: string): Promise<{ stdout: string; stderr: string }>;
|
||||
}
|
||||
|
||||
/**
|
||||
* This class encapsulates the logic for handling different file types (PDF, images)
|
||||
* and preparing them for AI processing.
|
||||
*/
|
||||
export class FlyerFileHandler {
|
||||
constructor(
|
||||
private fs: IFileSystem,
|
||||
private exec: ICommandExecutor,
|
||||
) {}
|
||||
|
||||
/**
|
||||
* Executes the pdftocairo command to convert the PDF.
|
||||
*/
|
||||
private async _executePdfConversion(
|
||||
filePath: string,
|
||||
outputFilePrefix: string,
|
||||
logger: Logger,
|
||||
): Promise<{ stdout: string; stderr: string }> {
|
||||
const command = `pdftocairo -jpeg -r 150 "${filePath}" "${outputFilePrefix}"`;
|
||||
logger.info(`Executing PDF conversion command`);
|
||||
logger.debug({ command });
|
||||
try {
|
||||
const { stdout, stderr } = await this.exec(command);
|
||||
if (stdout) logger.debug({ stdout }, `[Worker] pdftocairo stdout for ${filePath}:`);
|
||||
if (stderr) logger.warn({ stderr }, `[Worker] pdftocairo stderr for ${filePath}:`);
|
||||
return { stdout, stderr };
|
||||
} catch (error) {
|
||||
const execError = error as Error & { stderr?: string };
|
||||
const errorMessage = `The pdftocairo command failed for file: ${filePath}.`;
|
||||
logger.error({ err: execError, stderr: execError.stderr }, errorMessage);
|
||||
throw new PdfConversionError(errorMessage, execError.stderr);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Scans the output directory for generated JPEG images and returns their paths.
|
||||
*/
|
||||
private async _collectGeneratedImages(
|
||||
outputDir: string,
|
||||
outputFilePrefix: string,
|
||||
logger: Logger,
|
||||
): Promise<string[]> {
|
||||
logger.debug(`[Worker] Reading contents of output directory: ${outputDir}`);
|
||||
const filesInDir = await this.fs.readdir(outputDir, { withFileTypes: true });
|
||||
logger.debug(`[Worker] Found ${filesInDir.length} total entries in output directory.`);
|
||||
|
||||
const generatedImages = filesInDir
|
||||
.filter((f) => f.name.startsWith(path.basename(outputFilePrefix)) && f.name.endsWith('.jpg'))
|
||||
.sort((a, b) => a.name.localeCompare(b.name, undefined, { numeric: true }));
|
||||
|
||||
logger.debug(
|
||||
{ imageNames: generatedImages.map((f) => f.name) },
|
||||
`Filtered down to ${generatedImages.length} generated JPGs.`,
|
||||
);
|
||||
|
||||
return generatedImages.map((img) => path.join(outputDir, img.name));
|
||||
}
|
||||
|
||||
/**
|
||||
* Converts a PDF file to a series of JPEG images using an external tool.
|
||||
*/
|
||||
private async _convertPdfToImages(
|
||||
filePath: string,
|
||||
job: Job<FlyerJobData>,
|
||||
logger: Logger,
|
||||
): Promise<string[]> {
|
||||
logger.info(`Starting PDF conversion for: ${filePath}`);
|
||||
|
||||
const outputDir = path.dirname(filePath);
|
||||
const outputFilePrefix = path.join(outputDir, path.basename(filePath, '.pdf'));
|
||||
logger.debug({ outputDir, outputFilePrefix }, `PDF output details`);
|
||||
|
||||
const { stderr } = await this._executePdfConversion(filePath, outputFilePrefix, logger);
|
||||
|
||||
const imagePaths = await this._collectGeneratedImages(outputDir, outputFilePrefix, logger);
|
||||
|
||||
if (imagePaths.length === 0) {
|
||||
const errorMessage = `PDF conversion resulted in 0 images for file: ${filePath}. The PDF might be blank or corrupt.`;
|
||||
logger.error({ stderr }, `PdfConversionError: ${errorMessage}`);
|
||||
throw new PdfConversionError(errorMessage, stderr);
|
||||
}
|
||||
|
||||
return imagePaths;
|
||||
}
|
||||
|
||||
/**
|
||||
* Converts an image file (e.g., GIF, TIFF) to a PNG format that the AI can process.
|
||||
*/
|
||||
private async _convertImageToPng(filePath: string, logger: Logger): Promise<string> {
|
||||
const outputDir = path.dirname(filePath);
|
||||
const originalFileName = path.parse(path.basename(filePath)).name;
|
||||
const newFileName = `${originalFileName}-converted.png`;
|
||||
const outputPath = path.join(outputDir, newFileName);
|
||||
|
||||
logger.info({ from: filePath, to: outputPath }, 'Converting unsupported image format to PNG.');
|
||||
|
||||
try {
|
||||
await sharp(filePath).png().toFile(outputPath);
|
||||
return outputPath;
|
||||
} catch (error) {
|
||||
logger.error({ err: error, filePath }, 'Failed to convert image to PNG using sharp.');
|
||||
throw new ImageConversionError(`Image conversion to PNG failed for ${path.basename(filePath)}.`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Handles PDF files by converting them to a series of JPEG images.
|
||||
*/
|
||||
private async _handlePdfInput(
|
||||
filePath: string,
|
||||
job: Job<FlyerJobData>,
|
||||
logger: Logger,
|
||||
): Promise<{ imagePaths: { path: string; mimetype: string }[]; createdImagePaths: string[] }> {
|
||||
const createdImagePaths = await this._convertPdfToImages(filePath, job, logger);
|
||||
const imagePaths = createdImagePaths.map((p) => ({ path: p, mimetype: 'image/jpeg' }));
|
||||
logger.info(`Converted PDF to ${imagePaths.length} images.`);
|
||||
return { imagePaths, createdImagePaths };
|
||||
}
|
||||
|
||||
/**
|
||||
* Handles image files that are directly supported by the AI.
|
||||
*/
|
||||
private async _handleSupportedImageInput(
|
||||
filePath: string,
|
||||
fileExt: string,
|
||||
logger: Logger,
|
||||
): Promise<{ imagePaths: { path: string; mimetype: string }[]; createdImagePaths: string[] }> {
|
||||
logger.info(`Processing as a single image file: ${filePath}`);
|
||||
const mimetype =
|
||||
fileExt === '.jpg' || fileExt === '.jpeg' ? 'image/jpeg' : `image/${fileExt.slice(1)}`;
|
||||
const imagePaths = [{ path: filePath, mimetype }];
|
||||
return { imagePaths, createdImagePaths: [] };
|
||||
}
|
||||
|
||||
/**
|
||||
* Handles image files that need to be converted to PNG before AI processing.
|
||||
*/
|
||||
private async _handleConvertibleImageInput(
|
||||
filePath: string,
|
||||
logger: Logger,
|
||||
): Promise<{ imagePaths: { path: string; mimetype: string }[]; createdImagePaths: string[] }> {
|
||||
const createdPngPath = await this._convertImageToPng(filePath, logger);
|
||||
const imagePaths = [{ path: createdPngPath, mimetype: 'image/png' }];
|
||||
const createdImagePaths = [createdPngPath];
|
||||
return { imagePaths, createdImagePaths };
|
||||
}
|
||||
|
||||
/**
|
||||
* Throws an error for unsupported file types.
|
||||
*/
|
||||
private _handleUnsupportedInput(
|
||||
fileExt: string,
|
||||
originalFileName: string,
|
||||
logger: Logger,
|
||||
): never {
|
||||
const errorMessage = `Unsupported file type: ${fileExt}. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.`;
|
||||
logger.error({ originalFileName, fileExt }, errorMessage);
|
||||
throw new UnsupportedFileTypeError(errorMessage);
|
||||
}
|
||||
|
||||
/**
|
||||
* Prepares the input images for the AI service. If the input is a PDF, it's converted to images.
|
||||
*/
|
||||
public async prepareImageInputs(
|
||||
filePath: string,
|
||||
job: Job<FlyerJobData>,
|
||||
logger: Logger,
|
||||
): Promise<{ imagePaths: { path: string; mimetype: string }[]; createdImagePaths: string[] }> {
|
||||
const fileExt = path.extname(filePath).toLowerCase();
|
||||
|
||||
if (fileExt === '.pdf') {
|
||||
return this._handlePdfInput(filePath, job, logger);
|
||||
}
|
||||
if (SUPPORTED_IMAGE_EXTENSIONS.includes(fileExt)) {
|
||||
return this._handleSupportedImageInput(filePath, fileExt, logger);
|
||||
}
|
||||
if (CONVERTIBLE_IMAGE_EXTENSIONS.includes(fileExt)) {
|
||||
return this._handleConvertibleImageInput(filePath, logger);
|
||||
}
|
||||
|
||||
return this._handleUnsupportedInput(fileExt, job.data.originalFileName, logger);
|
||||
}
|
||||
}
|
||||
@@ -1,21 +1,13 @@
|
||||
// src/services/flyerProcessingService.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, type Mocked } from 'vitest';
|
||||
import sharp from 'sharp';
|
||||
import { Job } from 'bullmq';
|
||||
import { Job, UnrecoverableError } from 'bullmq';
|
||||
import type { Dirent } from 'node:fs';
|
||||
import type { Logger } from 'pino';
|
||||
import { z } from 'zod';
|
||||
import { AiFlyerDataSchema } from './flyerProcessingService.server';
|
||||
import type { Flyer, FlyerInsert } from '../types';
|
||||
|
||||
export interface FlyerJobData {
|
||||
filePath: string;
|
||||
originalFileName: string;
|
||||
checksum: string;
|
||||
userId?: string;
|
||||
submitterIp?: string;
|
||||
userProfileAddress?: string;
|
||||
}
|
||||
import { AiFlyerDataSchema } from './flyerAiProcessor.server';
|
||||
import type { Flyer, FlyerInsert, FlyerItemInsert } from '../types';
|
||||
import type { CleanupJobData, FlyerJobData } from '../types/job-data';
|
||||
|
||||
// 1. Create hoisted mocks FIRST
|
||||
const mocks = vi.hoisted(() => ({
|
||||
@@ -36,22 +28,10 @@ vi.mock('node:fs/promises', async (importOriginal) => {
|
||||
};
|
||||
});
|
||||
|
||||
// Mock sharp for the new image conversion logic
|
||||
const mockSharpInstance = {
|
||||
png: vi.fn(() => mockSharpInstance),
|
||||
toFile: vi.fn().mockResolvedValue({}),
|
||||
};
|
||||
vi.mock('sharp', () => ({
|
||||
__esModule: true,
|
||||
default: vi.fn(() => mockSharpInstance),
|
||||
}));
|
||||
|
||||
// Import service and dependencies (FlyerJobData already imported from types above)
|
||||
import { FlyerProcessingService } from './flyerProcessingService.server';
|
||||
import * as aiService from './aiService.server';
|
||||
import * as db from './db/index.db';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import * as imageProcessor from '../utils/imageProcessor';
|
||||
import { createMockFlyer } from '../tests/utils/mockFactories';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
import {
|
||||
@@ -59,6 +39,10 @@ import {
|
||||
PdfConversionError,
|
||||
UnsupportedFileTypeError,
|
||||
} from './processingErrors';
|
||||
import { FlyerFileHandler } from './flyerFileHandler.server';
|
||||
import { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||
import type { IFileSystem, ICommandExecutor } from './flyerFileHandler.server';
|
||||
import type { AIService } from './aiService.server';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('./aiService.server', () => ({
|
||||
@@ -73,9 +57,6 @@ vi.mock('./db/index.db', () => ({
|
||||
personalizationRepo: { getAllMasterItems: vi.fn() },
|
||||
adminRepo: { logActivity: vi.fn() },
|
||||
}));
|
||||
vi.mock('../utils/imageProcessor', () => ({
|
||||
generateFlyerIcon: vi.fn().mockResolvedValue('icon-test.webp'),
|
||||
}));
|
||||
vi.mock('./logger.server', () => ({
|
||||
logger: {
|
||||
info: vi.fn(),
|
||||
@@ -85,13 +66,15 @@ vi.mock('./logger.server', () => ({
|
||||
child: vi.fn().mockReturnThis(),
|
||||
},
|
||||
}));
|
||||
vi.mock('./flyerFileHandler.server');
|
||||
vi.mock('./flyerAiProcessor.server');
|
||||
|
||||
const mockedAiService = aiService as Mocked<typeof aiService>;
|
||||
const mockedDb = db as Mocked<typeof db>;
|
||||
const mockedImageProcessor = imageProcessor as Mocked<typeof imageProcessor>;
|
||||
|
||||
describe('FlyerProcessingService', () => {
|
||||
let service: FlyerProcessingService;
|
||||
let mockFileHandler: Mocked<FlyerFileHandler>;
|
||||
let mockAiProcessor: Mocked<FlyerAiProcessor>;
|
||||
const mockCleanupQueue = {
|
||||
add: vi.fn(),
|
||||
};
|
||||
@@ -108,49 +91,66 @@ describe('FlyerProcessingService', () => {
|
||||
icon_url: 'icon.webp',
|
||||
checksum: 'checksum-123',
|
||||
store_name: 'Mock Store',
|
||||
} as FlyerInsert,
|
||||
// Add required fields for FlyerInsert type
|
||||
status: 'processed',
|
||||
item_count: 0,
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Mock St',
|
||||
} as FlyerInsert, // Cast is okay here as it's a mock value
|
||||
itemsForDb: [],
|
||||
});
|
||||
|
||||
// Default mock implementation for the promisified exec
|
||||
mocks.execAsync.mockResolvedValue({ stdout: 'success', stderr: '' });
|
||||
|
||||
// Default mock for readdir returns an empty array of Dirent-like objects.
|
||||
mocks.readdir.mockResolvedValue([]);
|
||||
|
||||
// Mock the file system adapter that will be passed to the service
|
||||
const mockFs = {
|
||||
const mockFs: IFileSystem = {
|
||||
readdir: mocks.readdir,
|
||||
unlink: mocks.unlink,
|
||||
};
|
||||
|
||||
mockFileHandler = new FlyerFileHandler(mockFs, vi.fn()) as Mocked<FlyerFileHandler>;
|
||||
mockAiProcessor = new FlyerAiProcessor(
|
||||
{} as AIService,
|
||||
mockedDb.personalizationRepo,
|
||||
) as Mocked<FlyerAiProcessor>;
|
||||
|
||||
// Instantiate the service with all its dependencies mocked
|
||||
service = new FlyerProcessingService(
|
||||
mockedAiService.aiService,
|
||||
mockFileHandler,
|
||||
mockAiProcessor,
|
||||
mockedDb,
|
||||
mockFs,
|
||||
mocks.execAsync,
|
||||
mockCleanupQueue,
|
||||
new FlyerDataTransformer(),
|
||||
);
|
||||
|
||||
// Provide default successful mock implementations for dependencies
|
||||
vi.mocked(mockedAiService.aiService.extractCoreDataFromFlyerImage).mockResolvedValue({
|
||||
store_name: 'Mock Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Mock St',
|
||||
items: [
|
||||
{
|
||||
item: 'Test Item',
|
||||
price_display: '$1.99',
|
||||
price_in_cents: 199,
|
||||
quantity: 'each',
|
||||
category_name: 'Test Category',
|
||||
master_item_id: 1,
|
||||
},
|
||||
],
|
||||
mockAiProcessor.extractAndValidateData.mockResolvedValue({
|
||||
data: {
|
||||
store_name: 'Mock Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Mock St',
|
||||
items: [
|
||||
{
|
||||
item: 'Test Item',
|
||||
price_display: '$1.99',
|
||||
price_in_cents: 199,
|
||||
quantity: 'each',
|
||||
category_name: 'Test Category',
|
||||
master_item_id: 1,
|
||||
},
|
||||
],
|
||||
},
|
||||
needsReview: false,
|
||||
});
|
||||
mockFileHandler.prepareImageInputs.mockResolvedValue({
|
||||
imagePaths: [{ path: '/tmp/flyer.jpg', mimetype: 'image/jpeg' }],
|
||||
createdImagePaths: [],
|
||||
});
|
||||
|
||||
vi.mocked(createFlyerAndItems).mockResolvedValue({
|
||||
flyer: createMockFlyer({
|
||||
flyer_id: 1,
|
||||
@@ -160,7 +160,6 @@ describe('FlyerProcessingService', () => {
|
||||
}),
|
||||
items: [],
|
||||
});
|
||||
mockedImageProcessor.generateFlyerIcon.mockResolvedValue('icon-test.jpg');
|
||||
vi.mocked(mockedDb.adminRepo.logActivity).mockResolvedValue();
|
||||
// FIX: Provide a default mock for getAllMasterItems to prevent a TypeError on `.length`.
|
||||
vi.mocked(mockedDb.personalizationRepo.getAllMasterItems).mockResolvedValue([]);
|
||||
@@ -181,6 +180,16 @@ describe('FlyerProcessingService', () => {
|
||||
} as unknown as Job<FlyerJobData>;
|
||||
};
|
||||
|
||||
const createMockCleanupJob = (data: CleanupJobData): Job<CleanupJobData> => {
|
||||
return {
|
||||
id: `cleanup-job-${data.flyerId}`,
|
||||
data,
|
||||
opts: { attempts: 3 },
|
||||
attemptsMade: 1,
|
||||
updateProgress: vi.fn(),
|
||||
} as unknown as Job<CleanupJobData>;
|
||||
};
|
||||
|
||||
describe('processJob (Orchestrator)', () => {
|
||||
it('should process an image file successfully and enqueue a cleanup job', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.jpg', originalFileName: 'flyer.jpg' });
|
||||
@@ -188,10 +197,10 @@ describe('FlyerProcessingService', () => {
|
||||
const result = await service.processJob(job);
|
||||
|
||||
expect(result).toEqual({ flyerId: 1 });
|
||||
expect(mockedAiService.aiService.extractCoreDataFromFlyerImage).toHaveBeenCalledTimes(1);
|
||||
expect(mockFileHandler.prepareImageInputs).toHaveBeenCalledWith(job.data.filePath, job, expect.any(Object));
|
||||
expect(mockAiProcessor.extractAndValidateData).toHaveBeenCalledTimes(1);
|
||||
expect(createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
expect(mockedDb.adminRepo.logActivity).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.execAsync).not.toHaveBeenCalled();
|
||||
expect(mockCleanupQueue.add).toHaveBeenCalledWith(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId: 1, paths: ['/tmp/flyer.jpg'] },
|
||||
@@ -202,29 +211,17 @@ describe('FlyerProcessingService', () => {
|
||||
it('should convert a PDF, process its images, and enqueue a cleanup job for all files', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.pdf', originalFileName: 'flyer.pdf' });
|
||||
|
||||
// Mock readdir to return Dirent-like objects for the converted files
|
||||
mocks.readdir.mockResolvedValue([
|
||||
{ name: 'flyer-1.jpg' },
|
||||
{ name: 'flyer-2.jpg' },
|
||||
] as Dirent[]);
|
||||
// Mock the file handler to return multiple created paths
|
||||
const createdPaths = ['/tmp/flyer-1.jpg', '/tmp/flyer-2.jpg'];
|
||||
mockFileHandler.prepareImageInputs.mockResolvedValue({
|
||||
imagePaths: createdPaths.map(p => ({ path: p, mimetype: 'image/jpeg' })),
|
||||
createdImagePaths: createdPaths,
|
||||
});
|
||||
|
||||
await service.processJob(job);
|
||||
|
||||
// Verify that pdftocairo was called
|
||||
expect(mocks.execAsync).toHaveBeenCalledWith(
|
||||
expect.stringContaining('pdftocairo -jpeg -r 150'),
|
||||
);
|
||||
// Verify AI service was called with the converted images
|
||||
expect(mockedAiService.aiService.extractCoreDataFromFlyerImage).toHaveBeenCalledWith(
|
||||
expect.arrayContaining([
|
||||
expect.objectContaining({ path: expect.stringContaining('flyer-1.jpg') }),
|
||||
expect.objectContaining({ path: expect.stringContaining('flyer-2.jpg') }),
|
||||
]),
|
||||
expect.any(Array),
|
||||
undefined, // submitterIp
|
||||
undefined, // userProfileAddress
|
||||
expect.any(Object), // The job-specific logger
|
||||
);
|
||||
expect(mockFileHandler.prepareImageInputs).toHaveBeenCalledWith('/tmp/flyer.pdf', job, expect.any(Object));
|
||||
expect(mockAiProcessor.extractAndValidateData).toHaveBeenCalledTimes(1);
|
||||
expect(createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// Verify cleanup job includes original PDF and both generated images
|
||||
expect(mockCleanupQueue.add).toHaveBeenCalledWith(
|
||||
@@ -233,8 +230,8 @@ describe('FlyerProcessingService', () => {
|
||||
flyerId: 1,
|
||||
paths: [
|
||||
'/tmp/flyer.pdf',
|
||||
expect.stringContaining('flyer-1.jpg'),
|
||||
expect.stringContaining('flyer-2.jpg'),
|
||||
'/tmp/flyer-1.jpg',
|
||||
'/tmp/flyer-2.jpg',
|
||||
],
|
||||
},
|
||||
expect.any(Object),
|
||||
@@ -243,42 +240,73 @@ describe('FlyerProcessingService', () => {
|
||||
|
||||
it('should throw an error and not enqueue cleanup if the AI service fails', async () => {
|
||||
const job = createMockJob({});
|
||||
const { logger } = await import('./logger.server');
|
||||
const aiError = new Error('AI model exploded');
|
||||
vi.mocked(mockedAiService.aiService.extractCoreDataFromFlyerImage).mockRejectedValue(aiError);
|
||||
mockAiProcessor.extractAndValidateData.mockRejectedValue(aiError);
|
||||
|
||||
await expect(service.processJob(job)).rejects.toThrow('AI model exploded');
|
||||
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'UNKNOWN_ERROR',
|
||||
message: 'AI model exploded',
|
||||
}); // This was a duplicate, fixed.
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw UnrecoverableError for quota issues and not enqueue cleanup', async () => {
|
||||
const job = createMockJob({});
|
||||
// Simulate an AI error that contains a keyword for unrecoverable errors
|
||||
const quotaError = new Error('AI model quota exceeded');
|
||||
const { logger } = await import('./logger.server');
|
||||
mockAiProcessor.extractAndValidateData.mockRejectedValue(quotaError);
|
||||
|
||||
await expect(service.processJob(job)).rejects.toThrow(UnrecoverableError);
|
||||
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'QUOTA_EXCEEDED',
|
||||
message: 'An AI quota has been exceeded. Please try again later.',
|
||||
});
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw PdfConversionError and not enqueue cleanup if PDF conversion fails', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/bad.pdf', originalFileName: 'bad.pdf' });
|
||||
const { logger } = await import('./logger.server');
|
||||
const conversionError = new PdfConversionError('Conversion failed', 'pdftocairo error');
|
||||
// Make the conversion step fail
|
||||
mocks.execAsync.mockRejectedValue(conversionError);
|
||||
mockFileHandler.prepareImageInputs.mockRejectedValue(conversionError);
|
||||
|
||||
await expect(service.processJob(job)).rejects.toThrow(conversionError);
|
||||
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
// Use `toHaveBeenLastCalledWith` to check only the final error payload, ignoring earlier progress updates.
|
||||
expect(job.updateProgress).toHaveBeenLastCalledWith({
|
||||
errorCode: 'PDF_CONVERSION_FAILED',
|
||||
message:
|
||||
'The uploaded PDF could not be processed. It might be blank, corrupt, or password-protected.',
|
||||
'The uploaded PDF could not be processed. It might be blank, corrupt, or password-protected.', // This was a duplicate, fixed.
|
||||
stderr: 'pdftocairo error',
|
||||
stages: [
|
||||
{ name: 'Preparing Inputs', status: 'failed', critical: true, detail: 'Validating and preparing file...' },
|
||||
{ name: 'Extracting Data with AI', status: 'skipped', critical: true },
|
||||
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
||||
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||
],
|
||||
});
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw AiDataValidationError and not enqueue cleanup if AI validation fails', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({});
|
||||
const validationError = new AiDataValidationError('Validation failed', {}, {});
|
||||
// Make the AI extraction step fail with a validation error
|
||||
vi.mocked(mockedAiService.aiService.extractCoreDataFromFlyerImage).mockRejectedValue(
|
||||
validationError,
|
||||
);
|
||||
mockAiProcessor.extractAndValidateData.mockRejectedValue(validationError);
|
||||
|
||||
await expect(service.processJob(job)).rejects.toThrow(validationError);
|
||||
|
||||
@@ -287,77 +315,51 @@ describe('FlyerProcessingService', () => {
|
||||
{ err: validationError, validationErrors: {}, rawData: {} },
|
||||
'AI Data Validation failed.',
|
||||
);
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
// Use `toHaveBeenLastCalledWith` to check only the final error payload.
|
||||
// FIX: The payload from AiDataValidationError includes validationErrors and rawData.
|
||||
expect(job.updateProgress).toHaveBeenLastCalledWith({
|
||||
errorCode: 'AI_VALIDATION_FAILED',
|
||||
message:
|
||||
"The AI couldn't read the flyer's format. Please try a clearer image or a different flyer.",
|
||||
"The AI couldn't read the flyer's format. Please try a clearer image or a different flyer.", // This was a duplicate, fixed.
|
||||
validationErrors: {},
|
||||
rawData: {},
|
||||
stages: [
|
||||
{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: '1 page(s) ready for AI.' },
|
||||
{ name: 'Extracting Data with AI', status: 'failed', critical: true, detail: 'Communicating with AI model...' },
|
||||
{ name: 'Transforming AI Data', status: 'skipped', critical: true },
|
||||
{ name: 'Saving to Database', status: 'skipped', critical: true },
|
||||
],
|
||||
});
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
);
|
||||
});
|
||||
|
||||
// FIX: This test was incorrect. The service *does* support GIF conversion.
|
||||
// It is now a success case, verifying that conversion works as intended.
|
||||
it('should convert a GIF image to PNG and then process it', async () => {
|
||||
console.log('\n--- [TEST LOG] ---: Starting GIF conversion success test...');
|
||||
it('should handle convertible image types and include original and converted files in cleanup', async () => {
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.gif', originalFileName: 'flyer.gif' });
|
||||
const convertedPath = '/tmp/flyer-converted.png';
|
||||
|
||||
// Mock the file handler to return the converted path
|
||||
mockFileHandler.prepareImageInputs.mockResolvedValue({
|
||||
imagePaths: [{ path: convertedPath, mimetype: 'image/png' }],
|
||||
createdImagePaths: [convertedPath],
|
||||
});
|
||||
|
||||
await service.processJob(job);
|
||||
|
||||
console.log('--- [TEST LOG] ---: Verifying sharp conversion for GIF...');
|
||||
expect(sharp).toHaveBeenCalledWith('/tmp/flyer.gif');
|
||||
expect(mockSharpInstance.toFile).toHaveBeenCalledWith('/tmp/flyer-converted.png');
|
||||
|
||||
console.log('--- [TEST LOG] ---: Verifying AI service call and cleanup for GIF...');
|
||||
expect(mockedAiService.aiService.extractCoreDataFromFlyerImage).toHaveBeenCalledWith(
|
||||
[{ path: '/tmp/flyer-converted.png', mimetype: 'image/png' }],
|
||||
[],
|
||||
undefined,
|
||||
undefined,
|
||||
expect.any(Object),
|
||||
);
|
||||
expect(mockFileHandler.prepareImageInputs).toHaveBeenCalledWith('/tmp/flyer.gif', job, expect.any(Object));
|
||||
expect(mockAiProcessor.extractAndValidateData).toHaveBeenCalledTimes(1);
|
||||
expect(mockCleanupQueue.add).toHaveBeenCalledWith(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId: 1, paths: ['/tmp/flyer.gif', '/tmp/flyer-converted.png'] },
|
||||
expect.any(Object),
|
||||
);
|
||||
});
|
||||
|
||||
it('should convert a TIFF image to PNG and then process it', async () => {
|
||||
console.log('\n--- [TEST LOG] ---: Starting TIFF conversion success test...');
|
||||
const job = createMockJob({ filePath: '/tmp/flyer.tiff', originalFileName: 'flyer.tiff' });
|
||||
|
||||
await service.processJob(job);
|
||||
|
||||
expect(sharp).toHaveBeenCalledWith('/tmp/flyer.tiff');
|
||||
expect(mockSharpInstance.png).toHaveBeenCalled();
|
||||
expect(mockSharpInstance.toFile).toHaveBeenCalledWith('/tmp/flyer-converted.png');
|
||||
|
||||
console.log('--- [DEBUG] ---: In TIFF test, logging actual AI call arguments:');
|
||||
console.log(
|
||||
JSON.stringify(
|
||||
vi.mocked(mockedAiService.aiService.extractCoreDataFromFlyerImage).mock.calls[0],
|
||||
null,
|
||||
2,
|
||||
),
|
||||
);
|
||||
|
||||
expect(mockedAiService.aiService.extractCoreDataFromFlyerImage).toHaveBeenCalledWith(
|
||||
[{ path: '/tmp/flyer-converted.png', mimetype: 'image/png' }], // masterItems is mocked to []
|
||||
[], // submitterIp is undefined in the mock job
|
||||
undefined, // userProfileAddress is undefined in the mock job
|
||||
undefined, // The job-specific logger
|
||||
expect.any(Object),
|
||||
);
|
||||
|
||||
expect(mockCleanupQueue.add).toHaveBeenCalledWith(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId: 1, paths: ['/tmp/flyer.tiff', '/tmp/flyer-converted.png'] },
|
||||
{ flyerId: 1, paths: ['/tmp/flyer.gif', convertedPath] },
|
||||
expect.any(Object),
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error and not enqueue cleanup if the database service fails', async () => {
|
||||
const job = createMockJob({});
|
||||
const { logger } = await import('./logger.server');
|
||||
const dbError = new Error('Database transaction failed');
|
||||
vi.mocked(createFlyerAndItems).mockRejectedValue(dbError);
|
||||
|
||||
@@ -366,8 +368,11 @@ describe('FlyerProcessingService', () => {
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'UNKNOWN_ERROR',
|
||||
message: 'Database transaction failed',
|
||||
});
|
||||
}); // This was a duplicate, fixed.
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw UnsupportedFileTypeError for an unsupported file type', async () => {
|
||||
@@ -375,25 +380,24 @@ describe('FlyerProcessingService', () => {
|
||||
filePath: '/tmp/document.txt',
|
||||
originalFileName: 'document.txt',
|
||||
});
|
||||
const fileTypeError = new UnsupportedFileTypeError('Unsupported file type: .txt. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.');
|
||||
mockFileHandler.prepareImageInputs.mockRejectedValue(fileTypeError);
|
||||
const { logger } = await import('./logger.server');
|
||||
|
||||
await expect(service.processJob(job)).rejects.toThrow(UnsupportedFileTypeError);
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'UNSUPPORTED_FILE_TYPE',
|
||||
message:
|
||||
'Unsupported file type: .txt. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.',
|
||||
message: 'Unsupported file type: .txt. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.',
|
||||
});
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should log a warning and not enqueue cleanup if the job fails but a flyer ID was somehow generated', async () => {
|
||||
const job = createMockJob({});
|
||||
vi.mocked(createFlyerAndItems).mockRejectedValue(new Error('DB Error'));
|
||||
await expect(service.processJob(job)).rejects.toThrow();
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error and not enqueue cleanup if icon generation fails', async () => {
|
||||
const job = createMockJob({});
|
||||
const { logger } = await import('./logger.server');
|
||||
const iconError = new Error('Icon generation failed.');
|
||||
// The `transform` method calls `generateFlyerIcon`. In `beforeEach`, `transform` is mocked
|
||||
// to always succeed. For this test, we override that mock to simulate a failure
|
||||
@@ -405,235 +409,140 @@ describe('FlyerProcessingService', () => {
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'UNKNOWN_ERROR',
|
||||
message: 'Icon generation failed.',
|
||||
});
|
||||
}); // This was a duplicate, fixed.
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('_prepareImageInputs (private method)', () => {
|
||||
it('should throw UnsupportedFileTypeError for an unsupported file type', async () => {
|
||||
describe('_reportErrorAndThrow (private method)', () => {
|
||||
it('should update progress and throw UnrecoverableError for quota messages', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({
|
||||
filePath: '/tmp/unsupported.doc',
|
||||
originalFileName: 'unsupported.doc',
|
||||
const job = createMockJob({});
|
||||
const quotaError = new Error('RESOURCE_EXHAUSTED');
|
||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||
|
||||
await expect(privateMethod(quotaError, job, logger)).rejects.toThrow(
|
||||
UnrecoverableError,
|
||||
);
|
||||
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'QUOTA_EXCEEDED',
|
||||
message: 'An AI quota has been exceeded. Please try again later.',
|
||||
});
|
||||
const privateMethod = (service as any)._prepareImageInputs;
|
||||
});
|
||||
|
||||
await expect(privateMethod('/tmp/unsupported.doc', job, logger)).rejects.toThrow(
|
||||
UnsupportedFileTypeError,
|
||||
it('should use toErrorPayload for FlyerProcessingError instances', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({});
|
||||
const validationError = new AiDataValidationError(
|
||||
'Validation failed',
|
||||
{ foo: 'bar' },
|
||||
{ raw: 'data' },
|
||||
);
|
||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||
|
||||
await expect(privateMethod(validationError, job, logger)).rejects.toThrow(
|
||||
validationError,
|
||||
);
|
||||
|
||||
// The payload should now come from the error's `toErrorPayload` method
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'AI_VALIDATION_FAILED',
|
||||
message:
|
||||
"The AI couldn't read the flyer's format. Please try a clearer image or a different flyer.",
|
||||
validationErrors: { foo: 'bar' },
|
||||
rawData: { raw: 'data' },
|
||||
});
|
||||
});
|
||||
|
||||
it('should update progress and re-throw standard errors', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({});
|
||||
const genericError = new Error('A standard failure');
|
||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||
|
||||
await expect(privateMethod(genericError, job, logger)).rejects.toThrow(genericError);
|
||||
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({
|
||||
errorCode: 'UNKNOWN_ERROR',
|
||||
message: 'A standard failure', // This was a duplicate, fixed.
|
||||
});
|
||||
});
|
||||
|
||||
it('should wrap and throw non-Error objects', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({});
|
||||
const nonError = 'just a string error';
|
||||
const privateMethod = (service as any)._reportErrorAndThrow;
|
||||
|
||||
await expect(privateMethod(nonError, job, logger)).rejects.toThrow('just a string error');
|
||||
});
|
||||
});
|
||||
|
||||
describe('_convertImageToPng (private method)', () => {
|
||||
it('should throw an error if sharp fails', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const sharpError = new Error('Sharp failed');
|
||||
vi.mocked(mockSharpInstance.toFile).mockRejectedValue(sharpError);
|
||||
const privateMethod = (service as any)._convertImageToPng;
|
||||
describe('processCleanupJob', () => {
|
||||
it('should delete all files successfully', async () => {
|
||||
const job = createMockCleanupJob({ flyerId: 1, paths: ['/tmp/file1', '/tmp/file2'] });
|
||||
mocks.unlink.mockResolvedValue(undefined);
|
||||
|
||||
await expect(privateMethod('/tmp/image.gif', logger)).rejects.toThrow(
|
||||
'Image conversion to PNG failed for image.gif',
|
||||
const result = await service.processCleanupJob(job);
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file1');
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file2');
|
||||
expect(result).toEqual({ status: 'success', deletedCount: 2 });
|
||||
});
|
||||
|
||||
it('should handle ENOENT errors gracefully and still succeed', async () => {
|
||||
const job = createMockCleanupJob({ flyerId: 1, paths: ['/tmp/file1', '/tmp/file2'] });
|
||||
const enoentError: NodeJS.ErrnoException = new Error('File not found');
|
||||
enoentError.code = 'ENOENT';
|
||||
|
||||
mocks.unlink.mockResolvedValueOnce(undefined).mockRejectedValueOnce(enoentError);
|
||||
|
||||
const result = await service.processCleanupJob(job);
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
expect(result).toEqual({ status: 'success', deletedCount: 2 });
|
||||
// Check that the warning was logged
|
||||
const { logger } = await import('./logger.server');
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'File not found during cleanup (already deleted?): /tmp/file2',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an aggregate error if a non-ENOENT error occurs', async () => {
|
||||
const job = createMockCleanupJob({
|
||||
flyerId: 1,
|
||||
paths: ['/tmp/file1', '/tmp/permission-denied'],
|
||||
});
|
||||
const permissionError: NodeJS.ErrnoException = new Error('Permission denied');
|
||||
permissionError.code = 'EACCES';
|
||||
|
||||
mocks.unlink.mockResolvedValueOnce(undefined).mockRejectedValueOnce(permissionError);
|
||||
|
||||
await expect(service.processCleanupJob(job)).rejects.toThrow(
|
||||
'Failed to delete 1 file(s): /tmp/permission-denied',
|
||||
);
|
||||
|
||||
// Check that the error was logged
|
||||
const { logger } = await import('./logger.server');
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: sharpError, filePath: '/tmp/image.gif' },
|
||||
'Failed to convert image to PNG using sharp.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('_extractFlyerDataWithAI (private method)', () => {
|
||||
it('should throw AiDataValidationError if AI response validation fails', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const jobData = createMockJob({}).data;
|
||||
// Mock AI to return data missing a required field ('store_name')
|
||||
vi.mocked(mockedAiService.aiService.extractCoreDataFromFlyerImage).mockResolvedValue({
|
||||
valid_from: '2024-01-01',
|
||||
items: [],
|
||||
} as any);
|
||||
|
||||
await expect((service as any)._extractFlyerDataWithAI([], jobData, logger)).rejects.toThrow(
|
||||
AiDataValidationError,
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('_enqueueCleanup (private method)', () => {
|
||||
it('should enqueue a cleanup job with the correct parameters', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const flyerId = 42;
|
||||
const paths = ['/tmp/file1.jpg', '/tmp/file2.pdf'];
|
||||
|
||||
// Access and call the private method for testing
|
||||
await (
|
||||
service as unknown as {
|
||||
_enqueueCleanup: (flyerId: number, paths: string[], logger: Logger) => Promise<void>;
|
||||
}
|
||||
)._enqueueCleanup(flyerId, paths, logger);
|
||||
|
||||
expect(mockCleanupQueue.add).toHaveBeenCalledWith(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId, paths },
|
||||
{ jobId: `cleanup-flyer-${flyerId}`, removeOnComplete: true },
|
||||
expect.objectContaining({ err: permissionError, path: '/tmp/permission-denied' }),
|
||||
'Failed to delete temporary file.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should not call the queue if the paths array is empty', async () => {
|
||||
it('should skip processing and return "skipped" if paths array is empty', async () => {
|
||||
const job = createMockCleanupJob({ flyerId: 1, paths: [] });
|
||||
const result = await service.processCleanupJob(job);
|
||||
|
||||
expect(mocks.unlink).not.toHaveBeenCalled();
|
||||
expect(result).toEqual({ status: 'skipped', reason: 'no paths' });
|
||||
const { logger } = await import('./logger.server');
|
||||
// Access and call the private method with an empty array
|
||||
await (
|
||||
service as unknown as {
|
||||
_enqueueCleanup: (flyerId: number, paths: string[], logger: Logger) => Promise<void>;
|
||||
}
|
||||
)._enqueueCleanup(123, [], logger);
|
||||
|
||||
expect(mockCleanupQueue.add).not.toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
describe('_saveProcessedFlyerData (private method)', () => {
|
||||
it('should transform data, create flyer in DB, and log activity', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
// Arrange
|
||||
const mockExtractedData = {
|
||||
store_name: 'Test Store',
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Mock St',
|
||||
items: [
|
||||
{
|
||||
item: 'Test Item',
|
||||
price_display: '$1.99',
|
||||
price_in_cents: 199,
|
||||
quantity: 'each',
|
||||
category_name: 'Test Category',
|
||||
master_item_id: 1,
|
||||
},
|
||||
],
|
||||
};
|
||||
const mockImagePaths = [{ path: '/tmp/flyer.jpg', mimetype: 'image/jpeg' }];
|
||||
const mockJobData = {
|
||||
filePath: '/tmp/flyer.jpg',
|
||||
originalFileName: 'flyer.jpg',
|
||||
checksum: 'checksum-123',
|
||||
userId: 'user-abc',
|
||||
};
|
||||
|
||||
// The DB create function is also mocked in beforeEach.
|
||||
// Create a complete mock that satisfies the Flyer type.
|
||||
const mockNewFlyer = createMockFlyer({
|
||||
flyer_id: 1,
|
||||
file_name: 'flyer.jpg',
|
||||
image_url: '/flyer-images/flyer.jpg',
|
||||
icon_url: '/flyer-images/icons/icon-flyer.webp',
|
||||
checksum: 'checksum-123',
|
||||
store_id: 1,
|
||||
item_count: 1,
|
||||
});
|
||||
vi.mocked(createFlyerAndItems).mockResolvedValue({ flyer: mockNewFlyer, items: [] });
|
||||
|
||||
// Act: Access and call the private method for testing
|
||||
const result = await (
|
||||
service as unknown as {
|
||||
_saveProcessedFlyerData: (
|
||||
extractedData: z.infer<typeof AiFlyerDataSchema>,
|
||||
imagePaths: { path: string; mimetype: string }[],
|
||||
jobData: FlyerJobData,
|
||||
logger: Logger,
|
||||
) => Promise<Flyer>;
|
||||
}
|
||||
)._saveProcessedFlyerData(mockExtractedData, mockImagePaths, mockJobData, logger);
|
||||
|
||||
// Assert
|
||||
// 1. Transformer was called correctly
|
||||
expect(FlyerDataTransformer.prototype.transform).toHaveBeenCalledWith(
|
||||
mockExtractedData,
|
||||
mockImagePaths,
|
||||
mockJobData.originalFileName,
|
||||
mockJobData.checksum,
|
||||
mockJobData.userId,
|
||||
logger,
|
||||
);
|
||||
|
||||
// 2. DB function was called with the transformed data
|
||||
// The data comes from the mock defined in `beforeEach`.
|
||||
expect(createFlyerAndItems).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ store_name: 'Mock Store', checksum: 'checksum-123' }),
|
||||
[], // itemsForDb from the mock
|
||||
logger,
|
||||
);
|
||||
|
||||
// 3. Activity was logged with all expected fields
|
||||
expect(mockedDb.adminRepo.logActivity).toHaveBeenCalledWith(
|
||||
{
|
||||
userId: 'user-abc',
|
||||
action: 'flyer_processed' as const,
|
||||
displayText: 'Processed a new flyer for Mock Store.', // This was a duplicate, fixed.
|
||||
details: { flyerId: 1, storeName: 'Mock Store' },
|
||||
},
|
||||
logger,
|
||||
);
|
||||
|
||||
// 4. The method returned the new flyer
|
||||
expect(result).toEqual(mockNewFlyer);
|
||||
});
|
||||
});
|
||||
|
||||
describe('_convertPdfToImages (private method)', () => {
|
||||
it('should call pdftocairo and return sorted image paths on success', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({ filePath: '/tmp/test.pdf' });
|
||||
// Mock readdir to return unsorted Dirent-like objects
|
||||
mocks.readdir.mockResolvedValue([
|
||||
{ name: 'test-10.jpg' },
|
||||
{ name: 'test-1.jpg' },
|
||||
{ name: 'test-2.jpg' },
|
||||
{ name: 'other-file.txt' },
|
||||
] as Dirent[]);
|
||||
|
||||
// Access and call the private method for testing
|
||||
const imagePaths = await (
|
||||
service as unknown as {
|
||||
_convertPdfToImages: (filePath: string, job: Job, logger: Logger) => Promise<string[]>;
|
||||
}
|
||||
)._convertPdfToImages('/tmp/test.pdf', job, logger);
|
||||
|
||||
expect(mocks.execAsync).toHaveBeenCalledWith(
|
||||
'pdftocairo -jpeg -r 150 "/tmp/test.pdf" "/tmp/test"',
|
||||
);
|
||||
expect(job.updateProgress).toHaveBeenCalledWith({ message: 'Converting PDF to images...' });
|
||||
// Verify that the paths are correctly sorted numerically
|
||||
expect(imagePaths).toEqual(['/tmp/test-1.jpg', '/tmp/test-2.jpg', '/tmp/test-10.jpg']);
|
||||
});
|
||||
|
||||
it('should throw PdfConversionError if no images are generated', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({ filePath: '/tmp/empty.pdf' });
|
||||
// Mock readdir to return no matching files
|
||||
mocks.readdir.mockResolvedValue([]);
|
||||
|
||||
await expect(
|
||||
(
|
||||
service as unknown as {
|
||||
_convertPdfToImages: (filePath: string, job: Job, logger: Logger) => Promise<string[]>;
|
||||
}
|
||||
)._convertPdfToImages('/tmp/empty.pdf', job, logger),
|
||||
).rejects.toThrow('PDF conversion resulted in 0 images for file: /tmp/empty.pdf');
|
||||
});
|
||||
|
||||
it('should re-throw an error if the exec command fails', async () => {
|
||||
const { logger } = await import('./logger.server');
|
||||
const job = createMockJob({ filePath: '/tmp/bad.pdf' });
|
||||
const commandError = new Error('pdftocairo not found');
|
||||
mocks.execAsync.mockRejectedValue(commandError);
|
||||
|
||||
await expect(
|
||||
(
|
||||
service as unknown as {
|
||||
_convertPdfToImages: (filePath: string, job: Job, logger: Logger) => Promise<string[]>;
|
||||
}
|
||||
)._convertPdfToImages('/tmp/bad.pdf', job, logger),
|
||||
).rejects.toThrow(commandError);
|
||||
expect(logger.warn).toHaveBeenCalledWith('Job received no paths to clean. Skipping.');
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -1,409 +1,184 @@
|
||||
// src/services/flyerProcessingService.server.ts
|
||||
import type { Job, JobsOptions } from 'bullmq';
|
||||
import sharp from 'sharp';
|
||||
import path from 'path';
|
||||
import type { Dirent } from 'node:fs';
|
||||
import { z } from 'zod';
|
||||
|
||||
import type { AIService } from './aiService.server';
|
||||
import * as db from './db/index.db';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import {
|
||||
PdfConversionError,
|
||||
AiDataValidationError,
|
||||
UnsupportedFileTypeError,
|
||||
} from './processingErrors';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
import type { Job, Queue } from 'bullmq';
|
||||
import { UnrecoverableError } from 'bullmq';
|
||||
import type { Logger } from 'pino';
|
||||
|
||||
// Helper for consistent required string validation (handles missing/null/empty)
|
||||
const requiredString = (message: string) =>
|
||||
z.preprocess((val) => val ?? '', z.string().min(1, message));
|
||||
|
||||
// Define the image formats supported by the AI model
|
||||
const SUPPORTED_IMAGE_EXTENSIONS = ['.jpg', '.jpeg', '.png', '.webp', '.heic', '.heif'];
|
||||
|
||||
// Define image formats that are not directly supported but can be converted to PNG.
|
||||
const CONVERTIBLE_IMAGE_EXTENSIONS = ['.gif', '.tiff', '.svg', '.bmp'];
|
||||
|
||||
// --- Start: Interfaces for Dependency Injection ---
|
||||
|
||||
export interface IFileSystem {
|
||||
readdir(path: string, options: { withFileTypes: true }): Promise<Dirent[]>;
|
||||
unlink(path: string): Promise<void>;
|
||||
}
|
||||
|
||||
export interface ICommandExecutor {
|
||||
(command: string): Promise<{ stdout: string; stderr: string }>;
|
||||
}
|
||||
|
||||
export interface FlyerJobData {
|
||||
filePath: string;
|
||||
originalFileName: string;
|
||||
checksum: string;
|
||||
userId?: string;
|
||||
submitterIp?: string;
|
||||
userProfileAddress?: string;
|
||||
}
|
||||
|
||||
interface CleanupJobData {
|
||||
flyerId: number;
|
||||
// An array of absolute file paths to be deleted. Made optional for manual cleanup triggers.
|
||||
paths?: string[];
|
||||
}
|
||||
import type { FlyerFileHandler, IFileSystem, ICommandExecutor } from './flyerFileHandler.server';
|
||||
import type { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||
import type * as Db from './db/index.db';
|
||||
import type { AdminRepository } from './db/admin.db';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
import type { FlyerJobData, CleanupJobData } from '../types/job-data';
|
||||
import { FlyerProcessingError } from './processingErrors';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
|
||||
/**
|
||||
* Defines the contract for a queue that can have cleanup jobs added to it.
|
||||
* This is used for dependency injection to avoid circular dependencies.
|
||||
*/
|
||||
interface ICleanupQueue {
|
||||
add(name: string, data: CleanupJobData, opts?: JobsOptions): Promise<Job<CleanupJobData>>;
|
||||
}
|
||||
|
||||
// --- Zod Schemas for AI Response Validation (exported for the transformer) ---
|
||||
const ExtractedFlyerItemSchema = z.object({
|
||||
item: z.string().nullable(), // AI might return null or empty, normalize later
|
||||
price_display: z.string().nullable(), // AI might return null or empty, normalize later
|
||||
price_in_cents: z.number().nullable(),
|
||||
quantity: z.string().nullable(), // AI might return null or empty, normalize later
|
||||
category_name: z.string().nullable(), // AI might return null or empty, normalize later
|
||||
master_item_id: z.number().nullish(), // .nullish() allows null or undefined
|
||||
});
|
||||
|
||||
export const AiFlyerDataSchema = z.object({
|
||||
store_name: z.string().nullable(), // AI might return null or empty, normalize later
|
||||
valid_from: z.string().nullable(),
|
||||
valid_to: z.string().nullable(),
|
||||
store_address: z.string().nullable(),
|
||||
items: z.array(ExtractedFlyerItemSchema),
|
||||
});
|
||||
|
||||
/**
|
||||
* This class encapsulates the business logic for processing a flyer from a file.
|
||||
* It handles PDF conversion, AI data extraction, and saving the results to the database.
|
||||
* This service orchestrates the entire flyer processing workflow. It's responsible for
|
||||
* coordinating various sub-services (file handling, AI processing, data transformation,
|
||||
* and database operations) to process a flyer from upload to completion.
|
||||
*/
|
||||
export class FlyerProcessingService {
|
||||
constructor(
|
||||
private ai: AIService,
|
||||
private database: typeof db,
|
||||
private fileHandler: FlyerFileHandler,
|
||||
private aiProcessor: FlyerAiProcessor,
|
||||
// This service only needs the `logActivity` method from the `adminRepo`.
|
||||
// By using `Pick`, we create a more focused and testable dependency.
|
||||
private db: { adminRepo: Pick<AdminRepository, 'logActivity'> },
|
||||
private fs: IFileSystem,
|
||||
private exec: ICommandExecutor,
|
||||
private cleanupQueue: ICleanupQueue,
|
||||
// By depending on `Pick<Queue, 'add'>`, we specify that this service only needs
|
||||
// an object with an `add` method that matches the Queue's `add` method signature.
|
||||
// This decouples the service from the full BullMQ Queue implementation, making it more modular and easier to test.
|
||||
private cleanupQueue: Pick<Queue<CleanupJobData>, 'add'>,
|
||||
private transformer: FlyerDataTransformer,
|
||||
) {}
|
||||
|
||||
/**
|
||||
* Converts a PDF file to a series of JPEG images using an external tool.
|
||||
* @param filePath The path to the PDF file.
|
||||
* @param job The BullMQ job instance for progress updates.
|
||||
* @returns A promise that resolves to an array of paths to the created image files.
|
||||
* Orchestrates the processing of a flyer job.
|
||||
* @param job The BullMQ job containing flyer data.
|
||||
* @returns An object containing the ID of the newly created flyer.
|
||||
*/
|
||||
private async _convertPdfToImages(
|
||||
filePath: string,
|
||||
job: Job<FlyerJobData>,
|
||||
logger: Logger,
|
||||
): Promise<string[]> {
|
||||
logger.info(`Starting PDF conversion for: ${filePath}`);
|
||||
await job.updateProgress({ message: 'Converting PDF to images...' });
|
||||
async processJob(job: Job<FlyerJobData>): Promise<{ flyerId: number }> {
|
||||
// Create a logger instance with job-specific context for better traceability.
|
||||
const logger = globalLogger.child({ jobId: job.id, jobName: job.name, ...job.data });
|
||||
logger.info('Picked up flyer processing job.');
|
||||
|
||||
const outputDir = path.dirname(filePath);
|
||||
const outputFilePrefix = path.join(outputDir, path.basename(filePath, '.pdf'));
|
||||
logger.debug({ outputDir, outputFilePrefix }, `PDF output details`);
|
||||
|
||||
const command = `pdftocairo -jpeg -r 150 "${filePath}" "${outputFilePrefix}"`;
|
||||
logger.info(`Executing PDF conversion command`);
|
||||
logger.debug({ command });
|
||||
const { stdout, stderr } = await this.exec(command);
|
||||
|
||||
if (stdout) logger.debug({ stdout }, `[Worker] pdftocairo stdout for ${filePath}:`);
|
||||
if (stderr) logger.warn({ stderr }, `[Worker] pdftocairo stderr for ${filePath}:`);
|
||||
|
||||
logger.debug(`[Worker] Reading contents of output directory: ${outputDir}`);
|
||||
const filesInDir = await this.fs.readdir(outputDir, { withFileTypes: true });
|
||||
logger.debug(`[Worker] Found ${filesInDir.length} total entries in output directory.`);
|
||||
|
||||
const generatedImages = filesInDir
|
||||
.filter((f) => f.name.startsWith(path.basename(outputFilePrefix)) && f.name.endsWith('.jpg'))
|
||||
.sort((a, b) => a.name.localeCompare(b.name, undefined, { numeric: true }));
|
||||
|
||||
logger.debug(
|
||||
{ imageNames: generatedImages.map((f) => f.name) },
|
||||
`Filtered down to ${generatedImages.length} generated JPGs.`,
|
||||
);
|
||||
|
||||
if (generatedImages.length === 0) {
|
||||
const errorMessage = `PDF conversion resulted in 0 images for file: ${filePath}. The PDF might be blank or corrupt.`;
|
||||
logger.error({ stderr }, `PdfConversionError: ${errorMessage}`);
|
||||
throw new PdfConversionError(errorMessage, stderr);
|
||||
}
|
||||
|
||||
return generatedImages.map((img) => path.join(outputDir, img.name));
|
||||
}
|
||||
|
||||
/**
|
||||
* Converts an image file (e.g., GIF, TIFF) to a PNG format that the AI can process.
|
||||
* @param filePath The path to the source image file.
|
||||
* @param logger A logger instance.
|
||||
* @returns The path to the newly created PNG file.
|
||||
*/
|
||||
private async _convertImageToPng(filePath: string, logger: Logger): Promise<string> {
|
||||
const outputDir = path.dirname(filePath);
|
||||
const originalFileName = path.parse(path.basename(filePath)).name;
|
||||
const newFileName = `${originalFileName}-converted.png`;
|
||||
const outputPath = path.join(outputDir, newFileName);
|
||||
|
||||
logger.info({ from: filePath, to: outputPath }, 'Converting unsupported image format to PNG.');
|
||||
// Keep track of all created file paths for eventual cleanup.
|
||||
const allFilePaths: string[] = [job.data.filePath];
|
||||
|
||||
try {
|
||||
await sharp(filePath).png().toFile(outputPath);
|
||||
return outputPath;
|
||||
} catch (error) {
|
||||
logger.error({ err: error, filePath }, 'Failed to convert image to PNG using sharp.');
|
||||
throw new Error(`Image conversion to PNG failed for ${path.basename(filePath)}.`);
|
||||
}
|
||||
}
|
||||
/**
|
||||
* Prepares the input images for the AI service. If the input is a PDF, it's converted to images.
|
||||
* @param filePath The path to the original uploaded file.
|
||||
* @param job The BullMQ job instance.
|
||||
* @returns An object containing the final image paths for the AI and a list of any newly created image files.
|
||||
*/
|
||||
private async _prepareImageInputs(
|
||||
filePath: string,
|
||||
job: Job<FlyerJobData>,
|
||||
logger: Logger,
|
||||
): Promise<{ imagePaths: { path: string; mimetype: string }[]; createdImagePaths: string[] }> {
|
||||
const fileExt = path.extname(filePath).toLowerCase();
|
||||
|
||||
// Handle PDF conversion separately
|
||||
if (fileExt === '.pdf') {
|
||||
const createdImagePaths = await this._convertPdfToImages(filePath, job, logger);
|
||||
const imagePaths = createdImagePaths.map((p) => ({ path: p, mimetype: 'image/jpeg' }));
|
||||
logger.info(`Converted PDF to ${imagePaths.length} images.`);
|
||||
return { imagePaths, createdImagePaths };
|
||||
// Handle directly supported single-image formats
|
||||
} else if (SUPPORTED_IMAGE_EXTENSIONS.includes(fileExt)) {
|
||||
logger.info(`Processing as a single image file: ${filePath}`);
|
||||
// Normalize .jpg to image/jpeg for consistency
|
||||
const mimetype =
|
||||
fileExt === '.jpg' || fileExt === '.jpeg' ? 'image/jpeg' : `image/${fileExt.slice(1)}`;
|
||||
const imagePaths = [{ path: filePath, mimetype }];
|
||||
return { imagePaths, createdImagePaths: [] };
|
||||
// Handle convertible image formats
|
||||
} else if (CONVERTIBLE_IMAGE_EXTENSIONS.includes(fileExt)) {
|
||||
const createdPngPath = await this._convertImageToPng(filePath, logger);
|
||||
const imagePaths = [{ path: createdPngPath, mimetype: 'image/png' }];
|
||||
// The new PNG is a temporary file that needs to be cleaned up.
|
||||
return { imagePaths, createdImagePaths: [createdPngPath] };
|
||||
} else {
|
||||
// If the file is neither a PDF nor a supported image, throw an error.
|
||||
const errorMessage = `Unsupported file type: ${fileExt}. Supported types are PDF, JPG, PNG, WEBP, HEIC, HEIF, GIF, TIFF, SVG, BMP.`;
|
||||
logger.error({ originalFileName: job.data.originalFileName, fileExt }, errorMessage);
|
||||
throw new UnsupportedFileTypeError(errorMessage);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Calls the AI service to extract structured data from the flyer images.
|
||||
* @param imagePaths An array of paths and mimetypes for the images.
|
||||
* @param jobData The data from the BullMQ job.
|
||||
* @returns A promise that resolves to the validated, structured flyer data.
|
||||
*/
|
||||
private async _extractFlyerDataWithAI(
|
||||
imagePaths: { path: string; mimetype: string }[],
|
||||
jobData: FlyerJobData,
|
||||
logger: Logger,
|
||||
): Promise<z.infer<typeof AiFlyerDataSchema>> {
|
||||
logger.info(`Starting AI data extraction.`);
|
||||
const { submitterIp, userProfileAddress } = jobData;
|
||||
const masterItems = await this.database.personalizationRepo.getAllMasterItems(logger);
|
||||
logger.debug(`Retrieved ${masterItems.length} master items for AI matching.`);
|
||||
|
||||
const extractedData = await this.ai.extractCoreDataFromFlyerImage(
|
||||
imagePaths,
|
||||
masterItems,
|
||||
submitterIp, // Pass the job-specific logger
|
||||
userProfileAddress, // Pass the job-specific logger
|
||||
logger,
|
||||
);
|
||||
|
||||
const validationResult = AiFlyerDataSchema.safeParse(extractedData);
|
||||
if (!validationResult.success) {
|
||||
const errors = validationResult.error.flatten();
|
||||
logger.error({ errors, rawData: extractedData }, 'AI response failed validation.');
|
||||
throw new AiDataValidationError(
|
||||
'AI response validation failed. The returned data structure is incorrect.',
|
||||
errors,
|
||||
extractedData,
|
||||
);
|
||||
}
|
||||
|
||||
logger.info(`AI extracted ${validationResult.data.items.length} items.`);
|
||||
return validationResult.data;
|
||||
}
|
||||
|
||||
/**
|
||||
* Saves the extracted flyer data to the database.
|
||||
* @param extractedData The structured data from the AI.
|
||||
* @param imagePaths The paths to the flyer images.
|
||||
* @param jobData The data from the BullMQ job.
|
||||
* @returns A promise that resolves to the newly created flyer record.
|
||||
*/
|
||||
private async _saveProcessedFlyerData(
|
||||
extractedData: z.infer<typeof AiFlyerDataSchema>,
|
||||
imagePaths: { path: string; mimetype: string }[],
|
||||
jobData: FlyerJobData,
|
||||
logger: Logger,
|
||||
) {
|
||||
logger.info(`Preparing to save extracted data to database.`);
|
||||
|
||||
// Ensure store_name is a non-empty string before passing to the transformer.
|
||||
// This makes the handling of the nullable store_name explicit in this service.
|
||||
const dataForTransformer = { ...extractedData };
|
||||
if (!dataForTransformer.store_name) {
|
||||
logger.warn('AI did not return a store name. Using fallback "Unknown Store (auto)".');
|
||||
dataForTransformer.store_name = 'Unknown Store (auto)';
|
||||
}
|
||||
|
||||
// 1. Transform the AI data into database-ready records.
|
||||
const { flyerData, itemsForDb } = await this.transformer.transform(
|
||||
dataForTransformer,
|
||||
imagePaths,
|
||||
jobData.originalFileName,
|
||||
jobData.checksum,
|
||||
jobData.userId,
|
||||
// Pass the job-specific logger to the transformer
|
||||
logger,
|
||||
);
|
||||
|
||||
// 2. Save the transformed data to the database.
|
||||
const { flyer: newFlyer } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
||||
logger.info({ newFlyerId: newFlyer.flyer_id }, `Successfully saved new flyer.`);
|
||||
|
||||
await this.database.adminRepo.logActivity(
|
||||
{
|
||||
userId: jobData.userId,
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed a new flyer for ${flyerData.store_name}.`,
|
||||
details: { flyerId: newFlyer.flyer_id, storeName: flyerData.store_name },
|
||||
},
|
||||
logger,
|
||||
);
|
||||
|
||||
return newFlyer;
|
||||
}
|
||||
|
||||
/**
|
||||
* Enqueues a job to clean up temporary files associated with a flyer upload.
|
||||
* @param flyerId The ID of the processed flyer.
|
||||
* @param paths An array of file paths to be deleted.
|
||||
*/
|
||||
private async _enqueueCleanup(flyerId: number, paths: string[], logger: Logger): Promise<void> {
|
||||
if (paths.length === 0) return;
|
||||
|
||||
await this.cleanupQueue.add(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId, paths },
|
||||
{
|
||||
jobId: `cleanup-flyer-${flyerId}`,
|
||||
removeOnComplete: true,
|
||||
},
|
||||
);
|
||||
logger.info({ flyerId }, `Enqueued cleanup job.`);
|
||||
}
|
||||
|
||||
async processJob(job: Job<FlyerJobData>) {
|
||||
const { filePath, originalFileName } = job.data;
|
||||
const createdImagePaths: string[] = [];
|
||||
let newFlyerId: number | undefined;
|
||||
|
||||
// Create a job-specific logger instance with context, as per ADR-004
|
||||
const logger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
userId: job.data.userId,
|
||||
checksum: job.data.checksum,
|
||||
originalFileName,
|
||||
});
|
||||
|
||||
logger.info(`Picked up job.`);
|
||||
|
||||
try {
|
||||
await job.updateProgress({ message: 'Starting process...' });
|
||||
const { imagePaths, createdImagePaths: tempImagePaths } = await this._prepareImageInputs(
|
||||
filePath,
|
||||
// Stage 1: Prepare Inputs (e.g., convert PDF to images)
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'in-progress', critical: true, detail: 'Validating and preparing file...' }] });
|
||||
const { imagePaths, createdImagePaths } = await this.fileHandler.prepareImageInputs(
|
||||
job.data.filePath,
|
||||
job,
|
||||
logger,
|
||||
);
|
||||
createdImagePaths.push(...tempImagePaths);
|
||||
allFilePaths.push(...createdImagePaths);
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }] });
|
||||
|
||||
await job.updateProgress({ message: 'Extracting data...' });
|
||||
const extractedData = await this._extractFlyerDataWithAI(imagePaths, job.data, logger);
|
||||
// Stage 2: Extract Data with AI
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'in-progress', critical: true, detail: 'Communicating with AI model...' }] });
|
||||
const aiResult = await this.aiProcessor.extractAndValidateData(imagePaths, job.data, logger);
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }] });
|
||||
|
||||
await job.updateProgress({ message: 'Saving to database...' });
|
||||
const newFlyer = await this._saveProcessedFlyerData(
|
||||
extractedData,
|
||||
// Stage 3: Transform AI Data into DB format
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'in-progress', critical: true }] });
|
||||
const { flyerData, itemsForDb } = await this.transformer.transform(
|
||||
aiResult,
|
||||
imagePaths,
|
||||
job.data,
|
||||
job.data.originalFileName,
|
||||
job.data.checksum,
|
||||
job.data.userId,
|
||||
logger,
|
||||
); // Pass logger
|
||||
);
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'completed', critical: true }] });
|
||||
|
||||
newFlyerId = newFlyer.flyer_id;
|
||||
logger.info({ flyerId: newFlyerId }, `Job processed successfully.`);
|
||||
return { flyerId: newFlyer.flyer_id };
|
||||
} catch (error: unknown) {
|
||||
// Define a structured error payload for job progress updates.
|
||||
// This allows the frontend to provide more specific feedback.
|
||||
let errorPayload = {
|
||||
errorCode: 'UNKNOWN_ERROR',
|
||||
message: 'An unexpected error occurred during processing.',
|
||||
};
|
||||
// Stage 4: Save to Database
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'completed', critical: true }, { name: 'Saving to Database', status: 'in-progress', critical: true }] });
|
||||
const { flyer } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
||||
await job.updateProgress({ stages: [{ name: 'Preparing Inputs', status: 'completed', critical: true, detail: `${imagePaths.length} page(s) ready for AI.` }, { name: 'Extracting Data with AI', status: 'completed', critical: true }, { name: 'Transforming AI Data', status: 'completed', critical: true }, { name: 'Saving to Database', status: 'completed', critical: true }] });
|
||||
|
||||
if (error instanceof UnsupportedFileTypeError) {
|
||||
logger.error({ err: error }, `Unsupported file type error.`);
|
||||
errorPayload = {
|
||||
errorCode: 'UNSUPPORTED_FILE_TYPE',
|
||||
message: error.message, // The message is already user-friendly
|
||||
};
|
||||
} else if (error instanceof PdfConversionError) {
|
||||
logger.error({ err: error, stderr: error.stderr }, `PDF Conversion failed.`);
|
||||
errorPayload = {
|
||||
errorCode: 'PDF_CONVERSION_FAILED',
|
||||
message:
|
||||
'The uploaded PDF could not be processed. It might be blank, corrupt, or password-protected.',
|
||||
};
|
||||
} else if (error instanceof AiDataValidationError) {
|
||||
logger.error(
|
||||
{ err: error, validationErrors: error.validationErrors, rawData: error.rawData },
|
||||
`AI Data Validation failed.`,
|
||||
);
|
||||
errorPayload = {
|
||||
errorCode: 'AI_VALIDATION_FAILED',
|
||||
message:
|
||||
"The AI couldn't read the flyer's format. Please try a clearer image or a different flyer.",
|
||||
};
|
||||
} else if (error instanceof Error) {
|
||||
logger.error(
|
||||
{ err: error, attemptsMade: job.attemptsMade, totalAttempts: job.opts.attempts },
|
||||
`A generic error occurred in job.`,
|
||||
);
|
||||
// For generic errors, we can pass the message along, but still use a code.
|
||||
errorPayload.message = error.message;
|
||||
}
|
||||
// Stage 5: Log Activity
|
||||
await this.db.adminRepo.logActivity(
|
||||
{
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed flyer for ${flyerData.store_name}`,
|
||||
details: { flyer_id: flyer.flyer_id, store_name: flyerData.store_name },
|
||||
userId: job.data.userId,
|
||||
},
|
||||
logger,
|
||||
);
|
||||
|
||||
// Update the job's progress with the structured error payload.
|
||||
await job.updateProgress(errorPayload);
|
||||
// Enqueue a job to clean up the original and any generated files.
|
||||
await this.cleanupQueue.add(
|
||||
'cleanup-flyer-files',
|
||||
{ flyerId: flyer.flyer_id, paths: allFilePaths },
|
||||
{ removeOnComplete: true },
|
||||
);
|
||||
logger.info(`Successfully processed job and enqueued cleanup for flyer ID: ${flyer.flyer_id}`);
|
||||
|
||||
return { flyerId: flyer.flyer_id };
|
||||
} catch (error) {
|
||||
logger.warn('Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.');
|
||||
// This private method handles error reporting and re-throwing.
|
||||
await this._reportErrorAndThrow(error, job, logger);
|
||||
// This line is technically unreachable because the above method always throws,
|
||||
// but it's required to satisfy TypeScript's control flow analysis.
|
||||
throw error;
|
||||
} finally {
|
||||
if (newFlyerId) {
|
||||
const pathsToClean = [filePath, ...createdImagePaths];
|
||||
await this._enqueueCleanup(newFlyerId, pathsToClean, logger);
|
||||
} else {
|
||||
logger.warn(
|
||||
`Job failed. Temporary files will NOT be cleaned up to allow for manual inspection.`,
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Processes a job to clean up temporary files associated with a flyer.
|
||||
* @param job The BullMQ job containing cleanup data.
|
||||
* @returns An object indicating the status of the cleanup operation.
|
||||
*/
|
||||
async processCleanupJob(job: Job<CleanupJobData>): Promise<{ status: string; deletedCount?: number; reason?: string }> {
|
||||
const logger = globalLogger.child({ jobId: job.id, jobName: job.name, ...job.data });
|
||||
logger.info('Picked up file cleanup job.');
|
||||
|
||||
const { paths } = job.data;
|
||||
if (!paths || paths.length === 0) {
|
||||
logger.warn('Job received no paths to clean. Skipping.');
|
||||
return { status: 'skipped', reason: 'no paths' };
|
||||
}
|
||||
|
||||
const results = await Promise.allSettled(
|
||||
paths.map(async (filePath) => {
|
||||
try {
|
||||
await this.fs.unlink(filePath);
|
||||
logger.info(`Successfully deleted temporary file: ${filePath}`);
|
||||
} catch (error) {
|
||||
const nodeError = error as NodeJS.ErrnoException;
|
||||
if (nodeError.code === 'ENOENT') {
|
||||
// This is not a critical error; the file might have been deleted already.
|
||||
logger.warn(`File not found during cleanup (already deleted?): ${filePath}`);
|
||||
} else {
|
||||
logger.error({ err: nodeError, path: filePath }, 'Failed to delete temporary file.');
|
||||
throw error; // Re-throw to mark this specific deletion as failed.
|
||||
}
|
||||
}
|
||||
}),
|
||||
);
|
||||
|
||||
const failedDeletions = results.filter((r) => r.status === 'rejected');
|
||||
if (failedDeletions.length > 0) {
|
||||
const failedPaths = paths.filter((_, i) => results[i].status === 'rejected');
|
||||
throw new Error(`Failed to delete ${failedDeletions.length} file(s): ${failedPaths.join(', ')}`);
|
||||
}
|
||||
|
||||
logger.info(`Successfully deleted all ${paths.length} temporary files.`);
|
||||
return { status: 'success', deletedCount: paths.length };
|
||||
}
|
||||
|
||||
/**
|
||||
* A private helper to normalize errors, update job progress with an error state,
|
||||
* and re-throw the error to be handled by BullMQ.
|
||||
* @param error The error that was caught.
|
||||
* @param job The BullMQ job instance.
|
||||
* @param logger The logger instance.
|
||||
*/
|
||||
private async _reportErrorAndThrow(error: unknown, job: Job, logger: Logger): Promise<never> {
|
||||
const normalizedError = error instanceof Error ? error : new Error(String(error));
|
||||
let errorPayload: { errorCode: string; message: string; [key: string]: any };
|
||||
|
||||
if (normalizedError instanceof FlyerProcessingError) {
|
||||
errorPayload = normalizedError.toErrorPayload();
|
||||
logger.error({ err: normalizedError, ...errorPayload }, `A known processing error occurred: ${normalizedError.name}`);
|
||||
} else {
|
||||
const message = normalizedError.message || 'An unknown error occurred.';
|
||||
errorPayload = { errorCode: 'UNKNOWN_ERROR', message };
|
||||
logger.error({ err: normalizedError }, `An unknown error occurred: ${message}`);
|
||||
}
|
||||
|
||||
// Check for specific error messages that indicate a non-retriable failure, like quota exhaustion.
|
||||
if (errorPayload.message.toLowerCase().includes('quota') || errorPayload.message.toLowerCase().includes('resource_exhausted')) {
|
||||
const unrecoverablePayload = { errorCode: 'QUOTA_EXCEEDED', message: 'An AI quota has been exceeded. Please try again later.' };
|
||||
await job.updateProgress(unrecoverablePayload);
|
||||
throw new UnrecoverableError(unrecoverablePayload.message);
|
||||
}
|
||||
|
||||
await job.updateProgress(errorPayload);
|
||||
throw normalizedError;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3,13 +3,23 @@
|
||||
/**
|
||||
* Base class for all flyer processing errors.
|
||||
* This allows for catching all processing-related errors with a single `catch` block.
|
||||
* Each custom error should define its own `errorCode` and a user-friendly `message`.
|
||||
*/
|
||||
export class FlyerProcessingError extends Error {
|
||||
constructor(message: string) {
|
||||
super(message);
|
||||
public errorCode: string;
|
||||
public userMessage: string;
|
||||
|
||||
constructor(message: string, errorCode: string = 'UNKNOWN_ERROR', userMessage?: string) {
|
||||
super(message); // The 'message' property of Error is for internal/developer use.
|
||||
this.name = this.constructor.name;
|
||||
this.errorCode = errorCode;
|
||||
this.userMessage = userMessage || message; // User-friendly message for UI
|
||||
Object.setPrototypeOf(this, new.target.prototype);
|
||||
}
|
||||
|
||||
toErrorPayload(): { errorCode: string; message: string; [key: string]: any } {
|
||||
return { errorCode: this.errorCode, message: this.userMessage };
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -18,9 +28,17 @@ export class FlyerProcessingError extends Error {
|
||||
export class PdfConversionError extends FlyerProcessingError {
|
||||
public stderr?: string;
|
||||
constructor(message: string, stderr?: string) {
|
||||
super(message);
|
||||
super(
|
||||
message,
|
||||
'PDF_CONVERSION_FAILED',
|
||||
'The uploaded PDF could not be processed. It might be blank, corrupt, or password-protected.',
|
||||
);
|
||||
this.stderr = stderr;
|
||||
}
|
||||
|
||||
toErrorPayload(): { errorCode: string; message: string; [key: string]: any } {
|
||||
return { ...super.toErrorPayload(), stderr: this.stderr };
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -32,16 +50,36 @@ export class AiDataValidationError extends FlyerProcessingError {
|
||||
public validationErrors: object,
|
||||
public rawData: unknown,
|
||||
) {
|
||||
super(message);
|
||||
super(
|
||||
message,
|
||||
'AI_VALIDATION_FAILED',
|
||||
"The AI couldn't read the flyer's format. Please try a clearer image or a different flyer.",
|
||||
);
|
||||
}
|
||||
|
||||
toErrorPayload(): { errorCode: string; message: string; [key: string]: any } {
|
||||
return { ...super.toErrorPayload(), validationErrors: this.validationErrors, rawData: this.rawData };
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Error thrown when an image conversion fails (e.g., using sharp).
|
||||
*/
|
||||
export class ImageConversionError extends FlyerProcessingError {
|
||||
constructor(message: string) {
|
||||
super(
|
||||
message,
|
||||
'IMAGE_CONVERSION_FAILED',
|
||||
'The uploaded image could not be processed. It might be corrupt or in an unsupported format.',
|
||||
);
|
||||
}
|
||||
}
|
||||
/**
|
||||
* Error thrown when all geocoding providers fail to find coordinates for an address.
|
||||
*/
|
||||
export class GeocodingFailedError extends FlyerProcessingError {
|
||||
constructor(message: string) {
|
||||
super(message);
|
||||
super(message, 'GEOCODING_FAILED', 'Failed to geocode the address.');
|
||||
}
|
||||
}
|
||||
|
||||
@@ -50,6 +88,6 @@ export class GeocodingFailedError extends FlyerProcessingError {
|
||||
*/
|
||||
export class UnsupportedFileTypeError extends FlyerProcessingError {
|
||||
constructor(message: string) {
|
||||
super(message);
|
||||
super(message, 'UNSUPPORTED_FILE_TYPE', message); // The message is already user-friendly.
|
||||
}
|
||||
}
|
||||
|
||||
@@ -8,12 +8,17 @@ const mocks = vi.hoisted(() => {
|
||||
const capturedProcessors: Record<string, (job: Job) => Promise<unknown>> = {};
|
||||
|
||||
return {
|
||||
sendEmail: vi.fn(),
|
||||
unlink: vi.fn(),
|
||||
// Service method mocks
|
||||
processFlyerJob: vi.fn(),
|
||||
processCleanupJob: vi.fn(),
|
||||
processEmailJob: vi.fn(),
|
||||
processDailyReportJob: vi.fn(),
|
||||
processWeeklyReportJob: vi.fn(),
|
||||
processTokenCleanupJob: vi.fn(),
|
||||
|
||||
// Test utilities
|
||||
capturedProcessors,
|
||||
deleteExpiredResetTokens: vi.fn(),
|
||||
// Mock the Worker constructor to capture the processor function. It must be a
|
||||
// Mock the Worker constructor to capture the processor function. It must be a`
|
||||
// `function` and not an arrow function so it can be called with `new`.
|
||||
MockWorker: vi.fn(function (name: string, processor: (job: Job) => Promise<unknown>) {
|
||||
if (processor) {
|
||||
@@ -26,23 +31,20 @@ const mocks = vi.hoisted(() => {
|
||||
});
|
||||
|
||||
// --- Mock Modules ---
|
||||
vi.mock('./emailService.server', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('./emailService.server')>();
|
||||
return {
|
||||
...actual,
|
||||
// We only need to mock the specific function being called by the worker.
|
||||
// The rest of the module can retain its original implementation if needed elsewhere.
|
||||
sendEmail: mocks.sendEmail,
|
||||
};
|
||||
});
|
||||
vi.mock('./emailService.server', () => ({
|
||||
processEmailJob: mocks.processEmailJob,
|
||||
}));
|
||||
|
||||
// The workers use an `fsAdapter`. We can mock the underlying `fsPromises`
|
||||
// that the adapter is built from in queueService.server.ts.
|
||||
vi.mock('node:fs/promises', () => ({
|
||||
default: {
|
||||
unlink: mocks.unlink,
|
||||
// Add other fs functions if needed by other tests
|
||||
readdir: vi.fn(),
|
||||
vi.mock('./analyticsService.server', () => ({
|
||||
analyticsService: {
|
||||
processDailyReportJob: mocks.processDailyReportJob,
|
||||
processWeeklyReportJob: mocks.processWeeklyReportJob,
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./userService', () => ({
|
||||
userService: {
|
||||
processTokenCleanupJob: mocks.processTokenCleanupJob,
|
||||
},
|
||||
}));
|
||||
|
||||
@@ -56,28 +58,29 @@ vi.mock('./logger.server', () => ({
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./db/index.db', () => ({
|
||||
userRepo: {
|
||||
deleteExpiredResetTokens: mocks.deleteExpiredResetTokens,
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock bullmq to capture the processor functions passed to the Worker constructor
|
||||
import { logger as mockLogger } from './logger.server';
|
||||
vi.mock('bullmq', () => ({
|
||||
Worker: mocks.MockWorker,
|
||||
// FIX: Use a standard function for the mock constructor to allow `new Queue(...)` to work.
|
||||
Queue: vi.fn(function () {
|
||||
return { add: vi.fn() };
|
||||
}),
|
||||
// Add UnrecoverableError to the mock so it can be used in tests
|
||||
UnrecoverableError: class UnrecoverableError extends Error {},
|
||||
}));
|
||||
|
||||
// Mock flyerProcessingService.server as flyerWorker depends on it
|
||||
vi.mock('./flyerProcessingService.server', () => ({
|
||||
FlyerProcessingService: class {
|
||||
processJob = mocks.processFlyerJob;
|
||||
},
|
||||
}));
|
||||
// Mock flyerProcessingService.server as flyerWorker and cleanupWorker depend on it
|
||||
vi.mock('./flyerProcessingService.server', () => {
|
||||
// Mock the constructor to return an object with the mocked methods
|
||||
return {
|
||||
FlyerProcessingService: vi.fn().mockImplementation(function () {
|
||||
return {
|
||||
processJob: mocks.processFlyerJob,
|
||||
processCleanupJob: mocks.processCleanupJob,
|
||||
};
|
||||
}),
|
||||
};
|
||||
});
|
||||
|
||||
// Mock flyerDataTransformer as it's a dependency of FlyerProcessingService
|
||||
vi.mock('./flyerDataTransformer', () => ({
|
||||
@@ -110,15 +113,16 @@ describe('Queue Workers', () => {
|
||||
let tokenCleanupProcessor: (job: Job) => Promise<unknown>;
|
||||
|
||||
beforeEach(async () => {
|
||||
// Reset default mock implementations for hoisted mocks
|
||||
mocks.processFlyerJob.mockResolvedValue({ flyerId: 123 });
|
||||
mocks.processCleanupJob.mockResolvedValue({ status: 'success' });
|
||||
mocks.processEmailJob.mockResolvedValue(undefined);
|
||||
mocks.processDailyReportJob.mockResolvedValue({ status: 'success' });
|
||||
mocks.processWeeklyReportJob.mockResolvedValue({ status: 'success' });
|
||||
mocks.processTokenCleanupJob.mockResolvedValue({ deletedCount: 5 });
|
||||
|
||||
vi.clearAllMocks();
|
||||
vi.resetModules();
|
||||
|
||||
// Reset default mock implementations for hoisted mocks
|
||||
mocks.sendEmail.mockResolvedValue(undefined);
|
||||
mocks.unlink.mockResolvedValue(undefined);
|
||||
mocks.processFlyerJob.mockResolvedValue({ flyerId: 123 }); // Default success for flyer processing
|
||||
mocks.deleteExpiredResetTokens.mockResolvedValue(5);
|
||||
|
||||
await import('./workers.server');
|
||||
|
||||
flyerProcessor = mocks.capturedProcessors['flyer-processing'];
|
||||
@@ -155,10 +159,24 @@ describe('Queue Workers', () => {
|
||||
|
||||
await expect(flyerProcessor(job)).rejects.toThrow('Flyer processing failed');
|
||||
});
|
||||
|
||||
it('should re-throw UnrecoverableError from the service layer', async () => {
|
||||
const { UnrecoverableError } = await import('bullmq');
|
||||
const job = createMockJob({
|
||||
filePath: '/tmp/fail.pdf',
|
||||
originalFileName: 'fail.pdf',
|
||||
checksum: 'def',
|
||||
});
|
||||
const unrecoverableError = new UnrecoverableError('Quota exceeded');
|
||||
mocks.processFlyerJob.mockRejectedValue(unrecoverableError);
|
||||
|
||||
// The worker should just let this specific error type pass through.
|
||||
await expect(flyerProcessor(job)).rejects.toThrow(unrecoverableError);
|
||||
});
|
||||
});
|
||||
|
||||
describe('emailWorker', () => {
|
||||
it('should call emailService.sendEmail with the job data', async () => {
|
||||
it('should call emailService.processEmailJob with the job', async () => {
|
||||
const jobData = {
|
||||
to: 'test@example.com',
|
||||
subject: 'Test Email',
|
||||
@@ -166,173 +184,84 @@ describe('Queue Workers', () => {
|
||||
text: 'Hello',
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
|
||||
await emailProcessor(job);
|
||||
|
||||
expect(mocks.sendEmail).toHaveBeenCalledTimes(1);
|
||||
// The implementation passes the logger as the second argument
|
||||
expect(mocks.sendEmail).toHaveBeenCalledWith(jobData, expect.anything());
|
||||
expect(mocks.processEmailJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processEmailJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should log and re-throw an error if sendEmail fails with a non-Error object', async () => {
|
||||
const job = createMockJob({ to: 'fail@example.com', subject: 'fail', html: '', text: '' });
|
||||
const emailError = 'SMTP server is down'; // Reject with a string
|
||||
mocks.sendEmail.mockRejectedValue(emailError);
|
||||
|
||||
await expect(emailProcessor(job)).rejects.toThrow(emailError);
|
||||
|
||||
// The worker should wrap the string in an Error object for logging
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: new Error(emailError), jobData: job.data },
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
});
|
||||
|
||||
it('should re-throw an error if sendEmail fails', async () => {
|
||||
it('should re-throw an error if processEmailJob fails', async () => {
|
||||
const job = createMockJob({ to: 'fail@example.com', subject: 'fail', html: '', text: '' });
|
||||
const emailError = new Error('SMTP server is down');
|
||||
mocks.sendEmail.mockRejectedValue(emailError);
|
||||
|
||||
mocks.processEmailJob.mockRejectedValue(emailError);
|
||||
await expect(emailProcessor(job)).rejects.toThrow('SMTP server is down');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: emailError, jobData: job.data },
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('analyticsWorker', () => {
|
||||
it('should complete successfully for a valid report date', async () => {
|
||||
vi.useFakeTimers();
|
||||
it('should call analyticsService.processDailyReportJob with the job', async () => {
|
||||
const job = createMockJob({ reportDate: '2024-01-01' });
|
||||
|
||||
const promise = analyticsProcessor(job);
|
||||
// Advance timers to simulate the 10-second task completing
|
||||
await vi.advanceTimersByTimeAsync(10000);
|
||||
await promise; // Wait for the promise to resolve
|
||||
|
||||
// No error should be thrown
|
||||
expect(true).toBe(true);
|
||||
vi.useRealTimers();
|
||||
await analyticsProcessor(job);
|
||||
expect(mocks.processDailyReportJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processDailyReportJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should throw an error if reportDate is "FAIL"', async () => {
|
||||
it('should re-throw an error if processDailyReportJob fails', async () => {
|
||||
const job = createMockJob({ reportDate: 'FAIL' });
|
||||
|
||||
await expect(analyticsProcessor(job)).rejects.toThrow(
|
||||
'This is a test failure for the analytics job.',
|
||||
);
|
||||
const analyticsError = new Error('Analytics processing failed');
|
||||
mocks.processDailyReportJob.mockRejectedValue(analyticsError);
|
||||
await expect(analyticsProcessor(job)).rejects.toThrow('Analytics processing failed');
|
||||
});
|
||||
});
|
||||
|
||||
describe('cleanupWorker', () => {
|
||||
it('should call unlink for each path provided in the job data', async () => {
|
||||
it('should call flyerProcessingService.processCleanupJob with the job', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/file1.jpg', '/tmp/file2.pdf'],
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
mocks.unlink.mockResolvedValue(undefined);
|
||||
|
||||
await cleanupProcessor(job);
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file1.jpg');
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file2.pdf');
|
||||
expect(mocks.processCleanupJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processCleanupJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should not throw an error if a file is already deleted (ENOENT)', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/existing.jpg', '/tmp/already-deleted.jpg'],
|
||||
};
|
||||
it('should re-throw an error if processCleanupJob fails', async () => {
|
||||
const jobData = { flyerId: 123, paths: ['/tmp/protected-file.jpg'] };
|
||||
const job = createMockJob(jobData);
|
||||
// Use the built-in NodeJS.ErrnoException type for mock system errors.
|
||||
const enoentError: NodeJS.ErrnoException = new Error('File not found');
|
||||
enoentError.code = 'ENOENT';
|
||||
|
||||
// First call succeeds, second call fails with ENOENT
|
||||
mocks.unlink.mockResolvedValueOnce(undefined).mockRejectedValueOnce(enoentError);
|
||||
|
||||
// The processor should complete without throwing
|
||||
await expect(cleanupProcessor(job)).resolves.toBeUndefined();
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
});
|
||||
|
||||
it('should re-throw an error for issues other than ENOENT (e.g., permissions)', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/protected-file.jpg'],
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
// Use the built-in NodeJS.ErrnoException type for mock system errors.
|
||||
const permissionError: NodeJS.ErrnoException = new Error('Permission denied');
|
||||
permissionError.code = 'EACCES';
|
||||
|
||||
mocks.unlink.mockRejectedValue(permissionError);
|
||||
|
||||
const cleanupError = new Error('Permission denied');
|
||||
mocks.processCleanupJob.mockRejectedValue(cleanupError);
|
||||
await expect(cleanupProcessor(job)).rejects.toThrow('Permission denied');
|
||||
|
||||
// Verify the error was logged by the worker's catch block
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: permissionError },
|
||||
expect.stringContaining(
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${job.data.flyerId} failed.`,
|
||||
),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('weeklyAnalyticsWorker', () => {
|
||||
it('should complete successfully for a valid report date', async () => {
|
||||
vi.useFakeTimers();
|
||||
it('should call analyticsService.processWeeklyReportJob with the job', async () => {
|
||||
const job = createMockJob({ reportYear: 2024, reportWeek: 1 });
|
||||
|
||||
const promise = weeklyAnalyticsProcessor(job);
|
||||
// Advance timers to simulate the 30-second task completing
|
||||
await vi.advanceTimersByTimeAsync(30000);
|
||||
await promise; // Wait for the promise to resolve
|
||||
|
||||
// No error should be thrown
|
||||
expect(true).toBe(true);
|
||||
vi.useRealTimers();
|
||||
await weeklyAnalyticsProcessor(job);
|
||||
expect(mocks.processWeeklyReportJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processWeeklyReportJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should re-throw an error if the job fails', async () => {
|
||||
vi.useFakeTimers();
|
||||
it('should re-throw an error if processWeeklyReportJob fails', async () => {
|
||||
const job = createMockJob({ reportYear: 2024, reportWeek: 1 });
|
||||
// Mock the internal logic to throw an error
|
||||
const originalSetTimeout = setTimeout;
|
||||
vi.spyOn(global, 'setTimeout').mockImplementation((callback, ms) => {
|
||||
if (ms === 30000) {
|
||||
// Target the simulated delay
|
||||
throw new Error('Weekly analytics job failed');
|
||||
}
|
||||
return originalSetTimeout(callback, ms);
|
||||
});
|
||||
|
||||
const weeklyError = new Error('Weekly analytics job failed');
|
||||
mocks.processWeeklyReportJob.mockRejectedValue(weeklyError);
|
||||
await expect(weeklyAnalyticsProcessor(job)).rejects.toThrow('Weekly analytics job failed');
|
||||
vi.useRealTimers();
|
||||
vi.restoreAllMocks(); // Restore setTimeout mock
|
||||
});
|
||||
});
|
||||
|
||||
describe('tokenCleanupWorker', () => {
|
||||
it('should call userRepo.deleteExpiredResetTokens and return the count', async () => {
|
||||
it('should call userService.processTokenCleanupJob with the job', async () => {
|
||||
const job = createMockJob({ timestamp: new Date().toISOString() });
|
||||
mocks.deleteExpiredResetTokens.mockResolvedValue(10);
|
||||
|
||||
const result = await tokenCleanupProcessor(job);
|
||||
|
||||
expect(mocks.deleteExpiredResetTokens).toHaveBeenCalledTimes(1);
|
||||
expect(result).toEqual({ deletedCount: 10 });
|
||||
await tokenCleanupProcessor(job);
|
||||
expect(mocks.processTokenCleanupJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processTokenCleanupJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should re-throw an error if the database call fails', async () => {
|
||||
it('should re-throw an error if processTokenCleanupJob fails', async () => {
|
||||
const job = createMockJob({ timestamp: new Date().toISOString() });
|
||||
const dbError = new Error('DB cleanup failed');
|
||||
mocks.deleteExpiredResetTokens.mockRejectedValue(dbError);
|
||||
mocks.processTokenCleanupJob.mockRejectedValue(dbError);
|
||||
await expect(tokenCleanupProcessor(job)).rejects.toThrow(dbError);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -1,33 +1,13 @@
|
||||
import { Queue } from 'bullmq';
|
||||
import { connection } from './redis.server';
|
||||
import type { FlyerJobData } from './flyerProcessingService.server';
|
||||
|
||||
// --- Job Data Interfaces ---
|
||||
|
||||
export interface EmailJobData {
|
||||
to: string;
|
||||
subject: string;
|
||||
text: string;
|
||||
html: string;
|
||||
}
|
||||
|
||||
export interface AnalyticsJobData {
|
||||
reportDate: string; // e.g., '2024-10-26'
|
||||
}
|
||||
|
||||
export interface WeeklyAnalyticsJobData {
|
||||
reportYear: number;
|
||||
reportWeek: number; // ISO week number (1-53)
|
||||
}
|
||||
|
||||
export interface CleanupJobData {
|
||||
flyerId: number;
|
||||
paths?: string[];
|
||||
}
|
||||
|
||||
export interface TokenCleanupJobData {
|
||||
timestamp: string;
|
||||
}
|
||||
import type {
|
||||
FlyerJobData,
|
||||
EmailJobData,
|
||||
AnalyticsJobData,
|
||||
WeeklyAnalyticsJobData,
|
||||
CleanupJobData,
|
||||
TokenCleanupJobData,
|
||||
} from '../types/job-data';
|
||||
|
||||
// --- Queues ---
|
||||
|
||||
|
||||
@@ -1,9 +1,12 @@
|
||||
// src/services/userService.ts
|
||||
import * as db from './db/index.db';
|
||||
import type { Job } from 'bullmq';
|
||||
import type { Logger } from 'pino';
|
||||
import { AddressRepository } from './db/address.db';
|
||||
import { UserRepository } from './db/user.db';
|
||||
import type { Address, UserProfile } from '../types';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
import type { TokenCleanupJobData } from './queues.server';
|
||||
|
||||
/**
|
||||
* Encapsulates user-related business logic that may involve multiple repository calls.
|
||||
@@ -44,6 +47,35 @@ class UserService {
|
||||
return addressId;
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Processes a job to clean up expired password reset tokens from the database.
|
||||
* @param job The BullMQ job object.
|
||||
* @returns An object containing the count of deleted tokens.
|
||||
*/
|
||||
async processTokenCleanupJob(
|
||||
job: Job<TokenCleanupJobData>,
|
||||
): Promise<{ deletedCount: number }> {
|
||||
const logger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
});
|
||||
|
||||
logger.info('Picked up expired token cleanup job.');
|
||||
|
||||
try {
|
||||
const deletedCount = await db.userRepo.deleteExpiredResetTokens(logger);
|
||||
logger.info(`Successfully deleted ${deletedCount} expired tokens.`);
|
||||
return { deletedCount };
|
||||
} catch (error) {
|
||||
const wrappedError = error instanceof Error ? error : new Error(String(error));
|
||||
logger.error(
|
||||
{ err: wrappedError, attemptsMade: job.attemptsMade },
|
||||
'Expired token cleanup job failed.',
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
export const userService = new UserService();
|
||||
|
||||
@@ -6,13 +6,17 @@ import type { Job } from 'bullmq';
|
||||
const mocks = vi.hoisted(() => {
|
||||
// This object will store the processor functions captured from the worker constructors.
|
||||
const capturedProcessors: Record<string, (job: Job) => Promise<unknown>> = {};
|
||||
|
||||
return {
|
||||
sendEmail: vi.fn(),
|
||||
unlink: vi.fn(),
|
||||
// Service method mocks
|
||||
processFlyerJob: vi.fn(),
|
||||
processCleanupJob: vi.fn(),
|
||||
processEmailJob: vi.fn(),
|
||||
processDailyReportJob: vi.fn(),
|
||||
processWeeklyReportJob: vi.fn(),
|
||||
processTokenCleanupJob: vi.fn(),
|
||||
|
||||
// Test utilities
|
||||
capturedProcessors,
|
||||
deleteExpiredResetTokens: vi.fn(),
|
||||
// Mock the Worker constructor to capture the processor function. It must be a
|
||||
// `function` and not an arrow function so it can be called with `new`.
|
||||
MockWorker: vi.fn(function (name: string, processor: (job: Job) => Promise<unknown>) {
|
||||
@@ -26,23 +30,28 @@ const mocks = vi.hoisted(() => {
|
||||
});
|
||||
|
||||
// --- Mock Modules ---
|
||||
vi.mock('./emailService.server', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('./emailService.server')>();
|
||||
return {
|
||||
...actual,
|
||||
// We only need to mock the specific function being called by the worker.
|
||||
// The rest of the module can retain its original implementation if needed elsewhere.
|
||||
sendEmail: mocks.sendEmail,
|
||||
};
|
||||
});
|
||||
vi.mock('./emailService.server', () => ({
|
||||
processEmailJob: mocks.processEmailJob,
|
||||
}));
|
||||
|
||||
vi.mock('./analyticsService.server', () => ({
|
||||
analyticsService: {
|
||||
processDailyReportJob: mocks.processDailyReportJob,
|
||||
processWeeklyReportJob: mocks.processWeeklyReportJob,
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./userService', () => ({
|
||||
userService: {
|
||||
processTokenCleanupJob: mocks.processTokenCleanupJob,
|
||||
},
|
||||
}));
|
||||
|
||||
// The workers use an `fsAdapter`. We can mock the underlying `fsPromises`
|
||||
// that the adapter is built from in queueService.server.ts.
|
||||
vi.mock('node:fs/promises', () => ({
|
||||
default: {
|
||||
unlink: mocks.unlink,
|
||||
// Add other fs functions if needed by other tests
|
||||
readdir: vi.fn(),
|
||||
// unlink is no longer directly called by the worker
|
||||
},
|
||||
}));
|
||||
|
||||
@@ -56,28 +65,29 @@ vi.mock('./logger.server', () => ({
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./db/index.db', () => ({
|
||||
userRepo: {
|
||||
deleteExpiredResetTokens: mocks.deleteExpiredResetTokens,
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock bullmq to capture the processor functions passed to the Worker constructor
|
||||
import { logger as mockLogger } from './logger.server';
|
||||
vi.mock('bullmq', () => ({
|
||||
Worker: mocks.MockWorker,
|
||||
// FIX: Use a standard function for the mock constructor to allow `new Queue(...)` to work.
|
||||
Queue: vi.fn(function () {
|
||||
return { add: vi.fn() };
|
||||
}),
|
||||
// Add UnrecoverableError to the mock so it can be used in tests
|
||||
UnrecoverableError: class UnrecoverableError extends Error {},
|
||||
}));
|
||||
|
||||
// Mock flyerProcessingService.server as flyerWorker depends on it
|
||||
vi.mock('./flyerProcessingService.server', () => ({
|
||||
FlyerProcessingService: class {
|
||||
processJob = mocks.processFlyerJob;
|
||||
},
|
||||
}));
|
||||
vi.mock('./flyerProcessingService.server', () => {
|
||||
// Mock the constructor to return an object with the mocked methods
|
||||
return {
|
||||
FlyerProcessingService: vi.fn().mockImplementation(function () {
|
||||
return {
|
||||
processJob: mocks.processFlyerJob,
|
||||
processCleanupJob: mocks.processCleanupJob,
|
||||
};
|
||||
}),
|
||||
};
|
||||
});
|
||||
|
||||
// Mock flyerDataTransformer as it's a dependency of FlyerProcessingService
|
||||
vi.mock('./flyerDataTransformer', () => ({
|
||||
@@ -112,12 +122,13 @@ describe('Queue Workers', () => {
|
||||
|
||||
beforeEach(async () => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
// Reset default mock implementations for hoisted mocks
|
||||
mocks.sendEmail.mockResolvedValue(undefined);
|
||||
mocks.unlink.mockResolvedValue(undefined);
|
||||
mocks.processFlyerJob.mockResolvedValue({ flyerId: 123 }); // Default success for flyer processing
|
||||
mocks.deleteExpiredResetTokens.mockResolvedValue(5);
|
||||
mocks.processFlyerJob.mockResolvedValue({ flyerId: 123 });
|
||||
mocks.processCleanupJob.mockResolvedValue({ status: 'success' });
|
||||
mocks.processEmailJob.mockResolvedValue(undefined);
|
||||
mocks.processDailyReportJob.mockResolvedValue({ status: 'success' });
|
||||
mocks.processWeeklyReportJob.mockResolvedValue({ status: 'success' });
|
||||
mocks.processTokenCleanupJob.mockResolvedValue({ deletedCount: 5 });
|
||||
|
||||
// Reset modules to re-evaluate the workers.server.ts file with fresh mocks.
|
||||
// This ensures that new worker instances are created and their processors are captured for each test.
|
||||
@@ -162,10 +173,24 @@ describe('Queue Workers', () => {
|
||||
|
||||
await expect(flyerProcessor(job)).rejects.toThrow('Flyer processing failed');
|
||||
});
|
||||
|
||||
it('should re-throw UnrecoverableError from the service layer', async () => {
|
||||
const { UnrecoverableError } = await import('bullmq');
|
||||
const job = createMockJob({
|
||||
filePath: '/tmp/fail.pdf',
|
||||
originalFileName: 'fail.pdf',
|
||||
checksum: 'def',
|
||||
});
|
||||
const unrecoverableError = new UnrecoverableError('Quota exceeded');
|
||||
mocks.processFlyerJob.mockRejectedValue(unrecoverableError);
|
||||
|
||||
// The worker should just let this specific error type pass through.
|
||||
await expect(flyerProcessor(job)).rejects.toThrow(unrecoverableError);
|
||||
});
|
||||
});
|
||||
|
||||
describe('emailWorker', () => {
|
||||
it('should call emailService.sendEmail with the job data', async () => {
|
||||
it('should call emailService.processEmailJob with the job', async () => {
|
||||
const jobData = {
|
||||
to: 'test@example.com',
|
||||
subject: 'Test Email',
|
||||
@@ -173,173 +198,84 @@ describe('Queue Workers', () => {
|
||||
text: 'Hello',
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
|
||||
await emailProcessor(job);
|
||||
|
||||
expect(mocks.sendEmail).toHaveBeenCalledTimes(1);
|
||||
// The implementation passes the logger as the second argument
|
||||
expect(mocks.sendEmail).toHaveBeenCalledWith(jobData, expect.anything());
|
||||
expect(mocks.processEmailJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processEmailJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should log and re-throw an error if sendEmail fails with a non-Error object', async () => {
|
||||
const job = createMockJob({ to: 'fail@example.com', subject: 'fail', html: '', text: '' });
|
||||
const emailError = 'SMTP server is down'; // Reject with a string
|
||||
mocks.sendEmail.mockRejectedValue(emailError);
|
||||
|
||||
await expect(emailProcessor(job)).rejects.toThrow(emailError);
|
||||
|
||||
// The worker should wrap the string in an Error object for logging
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: new Error(emailError), jobData: job.data },
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
});
|
||||
|
||||
it('should re-throw an error if sendEmail fails', async () => {
|
||||
it('should re-throw an error if processEmailJob fails', async () => {
|
||||
const job = createMockJob({ to: 'fail@example.com', subject: 'fail', html: '', text: '' });
|
||||
const emailError = new Error('SMTP server is down');
|
||||
mocks.sendEmail.mockRejectedValue(emailError);
|
||||
|
||||
mocks.processEmailJob.mockRejectedValue(emailError);
|
||||
await expect(emailProcessor(job)).rejects.toThrow('SMTP server is down');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: emailError, jobData: job.data },
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('analyticsWorker', () => {
|
||||
it('should complete successfully for a valid report date', async () => {
|
||||
vi.useFakeTimers();
|
||||
it('should call analyticsService.processDailyReportJob with the job', async () => {
|
||||
const job = createMockJob({ reportDate: '2024-01-01' });
|
||||
|
||||
const promise = analyticsProcessor(job);
|
||||
// Advance timers to simulate the 10-second task completing
|
||||
await vi.advanceTimersByTimeAsync(10000);
|
||||
await promise; // Wait for the promise to resolve
|
||||
|
||||
// No error should be thrown
|
||||
expect(true).toBe(true);
|
||||
vi.useRealTimers();
|
||||
await analyticsProcessor(job);
|
||||
expect(mocks.processDailyReportJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processDailyReportJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should throw an error if reportDate is "FAIL"', async () => {
|
||||
it('should re-throw an error if processDailyReportJob fails', async () => {
|
||||
const job = createMockJob({ reportDate: 'FAIL' });
|
||||
|
||||
await expect(analyticsProcessor(job)).rejects.toThrow(
|
||||
'This is a test failure for the analytics job.',
|
||||
);
|
||||
const analyticsError = new Error('Analytics processing failed');
|
||||
mocks.processDailyReportJob.mockRejectedValue(analyticsError);
|
||||
await expect(analyticsProcessor(job)).rejects.toThrow('Analytics processing failed');
|
||||
});
|
||||
});
|
||||
|
||||
describe('cleanupWorker', () => {
|
||||
it('should call unlink for each path provided in the job data', async () => {
|
||||
it('should call flyerProcessingService.processCleanupJob with the job', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/file1.jpg', '/tmp/file2.pdf'],
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
mocks.unlink.mockResolvedValue(undefined);
|
||||
|
||||
await cleanupProcessor(job);
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file1.jpg');
|
||||
expect(mocks.unlink).toHaveBeenCalledWith('/tmp/file2.pdf');
|
||||
expect(mocks.processCleanupJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processCleanupJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should not throw an error if a file is already deleted (ENOENT)', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/existing.jpg', '/tmp/already-deleted.jpg'],
|
||||
};
|
||||
it('should re-throw an error if processCleanupJob fails', async () => {
|
||||
const jobData = { flyerId: 123, paths: ['/tmp/protected-file.jpg'] };
|
||||
const job = createMockJob(jobData);
|
||||
// Use the built-in NodeJS.ErrnoException type for mock system errors.
|
||||
const enoentError: NodeJS.ErrnoException = new Error('File not found');
|
||||
enoentError.code = 'ENOENT';
|
||||
|
||||
// First call succeeds, second call fails with ENOENT
|
||||
mocks.unlink.mockResolvedValueOnce(undefined).mockRejectedValueOnce(enoentError);
|
||||
|
||||
// The processor should complete without throwing
|
||||
await expect(cleanupProcessor(job)).resolves.toBeUndefined();
|
||||
|
||||
expect(mocks.unlink).toHaveBeenCalledTimes(2);
|
||||
});
|
||||
|
||||
it('should re-throw an error for issues other than ENOENT (e.g., permissions)', async () => {
|
||||
const jobData = {
|
||||
flyerId: 123,
|
||||
paths: ['/tmp/protected-file.jpg'],
|
||||
};
|
||||
const job = createMockJob(jobData);
|
||||
// Use the built-in NodeJS.ErrnoException type for mock system errors.
|
||||
const permissionError: NodeJS.ErrnoException = new Error('Permission denied');
|
||||
permissionError.code = 'EACCES';
|
||||
|
||||
mocks.unlink.mockRejectedValue(permissionError);
|
||||
|
||||
const cleanupError = new Error('Permission denied');
|
||||
mocks.processCleanupJob.mockRejectedValue(cleanupError);
|
||||
await expect(cleanupProcessor(job)).rejects.toThrow('Permission denied');
|
||||
|
||||
// Verify the error was logged by the worker's catch block
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: permissionError },
|
||||
expect.stringContaining(
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${job.data.flyerId} failed.`,
|
||||
),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('weeklyAnalyticsWorker', () => {
|
||||
it('should complete successfully for a valid report date', async () => {
|
||||
vi.useFakeTimers();
|
||||
it('should call analyticsService.processWeeklyReportJob with the job', async () => {
|
||||
const job = createMockJob({ reportYear: 2024, reportWeek: 1 });
|
||||
|
||||
const promise = weeklyAnalyticsProcessor(job);
|
||||
// Advance timers to simulate the 30-second task completing
|
||||
await vi.advanceTimersByTimeAsync(30000);
|
||||
await promise; // Wait for the promise to resolve
|
||||
|
||||
// No error should be thrown
|
||||
expect(true).toBe(true);
|
||||
vi.useRealTimers();
|
||||
await weeklyAnalyticsProcessor(job);
|
||||
expect(mocks.processWeeklyReportJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processWeeklyReportJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should re-throw an error if the job fails', async () => {
|
||||
vi.useFakeTimers();
|
||||
it('should re-throw an error if processWeeklyReportJob fails', async () => {
|
||||
const job = createMockJob({ reportYear: 2024, reportWeek: 1 });
|
||||
// Mock the internal logic to throw an error
|
||||
const originalSetTimeout = setTimeout;
|
||||
vi.spyOn(global, 'setTimeout').mockImplementation((callback, ms) => {
|
||||
if (ms === 30000) {
|
||||
// Target the simulated delay
|
||||
throw new Error('Weekly analytics job failed');
|
||||
}
|
||||
return originalSetTimeout(callback, ms);
|
||||
});
|
||||
|
||||
const weeklyError = new Error('Weekly analytics job failed');
|
||||
mocks.processWeeklyReportJob.mockRejectedValue(weeklyError);
|
||||
await expect(weeklyAnalyticsProcessor(job)).rejects.toThrow('Weekly analytics job failed');
|
||||
vi.useRealTimers();
|
||||
vi.restoreAllMocks(); // Restore setTimeout mock
|
||||
});
|
||||
});
|
||||
|
||||
describe('tokenCleanupWorker', () => {
|
||||
it('should call userRepo.deleteExpiredResetTokens and return the count', async () => {
|
||||
it('should call userService.processTokenCleanupJob with the job', async () => {
|
||||
const job = createMockJob({ timestamp: new Date().toISOString() });
|
||||
mocks.deleteExpiredResetTokens.mockResolvedValue(10);
|
||||
|
||||
const result = await tokenCleanupProcessor(job);
|
||||
|
||||
expect(mocks.deleteExpiredResetTokens).toHaveBeenCalledTimes(1);
|
||||
expect(result).toEqual({ deletedCount: 10 });
|
||||
await tokenCleanupProcessor(job);
|
||||
expect(mocks.processTokenCleanupJob).toHaveBeenCalledTimes(1);
|
||||
expect(mocks.processTokenCleanupJob).toHaveBeenCalledWith(job);
|
||||
});
|
||||
|
||||
it('should re-throw an error if the database call fails', async () => {
|
||||
it('should re-throw an error if processTokenCleanupJob fails', async () => {
|
||||
const job = createMockJob({ timestamp: new Date().toISOString() });
|
||||
const dbError = new Error('DB cleanup failed');
|
||||
mocks.deleteExpiredResetTokens.mockRejectedValue(dbError);
|
||||
mocks.processTokenCleanupJob.mockRejectedValue(dbError);
|
||||
await expect(tokenCleanupProcessor(job)).rejects.toThrow(dbError);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -6,27 +6,30 @@ import { promisify } from 'util';
|
||||
import { logger } from './logger.server';
|
||||
import { connection } from './redis.server';
|
||||
import { aiService } from './aiService.server';
|
||||
import { analyticsService } from './analyticsService.server';
|
||||
import { userService } from './userService';
|
||||
import * as emailService from './emailService.server';
|
||||
import * as db from './db/index.db';
|
||||
import {
|
||||
FlyerProcessingService,
|
||||
type FlyerJobData,
|
||||
type IFileSystem,
|
||||
} from './flyerProcessingService.server';
|
||||
import { FlyerProcessingService } from './flyerProcessingService.server';
|
||||
import { FlyerAiProcessor } from './flyerAiProcessor.server';
|
||||
import { FlyerDataTransformer } from './flyerDataTransformer';
|
||||
import {
|
||||
cleanupQueue,
|
||||
flyerQueue,
|
||||
emailQueue,
|
||||
analyticsQueue,
|
||||
weeklyAnalyticsQueue,
|
||||
cleanupQueue,
|
||||
tokenCleanupQueue,
|
||||
type EmailJobData,
|
||||
type AnalyticsJobData,
|
||||
type CleanupJobData,
|
||||
type WeeklyAnalyticsJobData,
|
||||
type TokenCleanupJobData,
|
||||
} from './queues.server';
|
||||
import type {
|
||||
FlyerJobData,
|
||||
EmailJobData,
|
||||
AnalyticsJobData,
|
||||
WeeklyAnalyticsJobData,
|
||||
CleanupJobData,
|
||||
TokenCleanupJobData,
|
||||
} from '../types/job-data';
|
||||
import { FlyerFileHandler, type IFileSystem } from './flyerFileHandler.server';
|
||||
|
||||
const execAsync = promisify(exec);
|
||||
|
||||
@@ -38,10 +41,10 @@ const fsAdapter: IFileSystem = {
|
||||
};
|
||||
|
||||
const flyerProcessingService = new FlyerProcessingService(
|
||||
aiService,
|
||||
new FlyerFileHandler(fsAdapter, execAsync),
|
||||
new FlyerAiProcessor(aiService, db.personalizationRepo),
|
||||
db,
|
||||
fsAdapter,
|
||||
execAsync,
|
||||
cleanupQueue,
|
||||
new FlyerDataTransformer(),
|
||||
);
|
||||
@@ -50,6 +53,25 @@ const normalizeError = (error: unknown): Error => {
|
||||
return error instanceof Error ? error : new Error(String(error));
|
||||
};
|
||||
|
||||
/**
|
||||
* Creates a higher-order function to wrap worker processors with common logic.
|
||||
* This includes error normalization to ensure that any thrown value is an Error instance,
|
||||
* which is a best practice for BullMQ workers.
|
||||
* @param processor The core logic for the worker.
|
||||
* @returns An async function that takes a job and executes the processor.
|
||||
*/
|
||||
const createWorkerProcessor = <T>(processor: (job: Job<T>) => Promise<any>) => {
|
||||
return async (job: Job<T>) => {
|
||||
try {
|
||||
return await processor(job);
|
||||
} catch (error: unknown) {
|
||||
// The service layer now handles detailed logging. This block just ensures
|
||||
// any unexpected errors are normalized before BullMQ handles them.
|
||||
throw normalizeError(error);
|
||||
}
|
||||
};
|
||||
};
|
||||
|
||||
const attachWorkerEventListeners = (worker: Worker) => {
|
||||
worker.on('completed', (job: Job, returnValue: unknown) => {
|
||||
logger.info({ returnValue }, `[${worker.name}] Job ${job.id} completed successfully.`);
|
||||
@@ -65,26 +87,7 @@ const attachWorkerEventListeners = (worker: Worker) => {
|
||||
|
||||
export const flyerWorker = new Worker<FlyerJobData>(
|
||||
'flyer-processing',
|
||||
async (job) => {
|
||||
try {
|
||||
return await flyerProcessingService.processJob(job);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
const errorMessage = wrappedError.message || '';
|
||||
if (
|
||||
errorMessage.includes('quota') ||
|
||||
errorMessage.includes('429') ||
|
||||
errorMessage.includes('RESOURCE_EXHAUSTED')
|
||||
) {
|
||||
logger.error(
|
||||
{ err: wrappedError, jobId: job.id },
|
||||
'[FlyerWorker] Unrecoverable quota error detected. Failing job immediately.',
|
||||
);
|
||||
throw new UnrecoverableError(errorMessage);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
createWorkerProcessor((job) => flyerProcessingService.processJob(job)),
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.WORKER_CONCURRENCY || '1', 10),
|
||||
@@ -93,24 +96,7 @@ export const flyerWorker = new Worker<FlyerJobData>(
|
||||
|
||||
export const emailWorker = new Worker<EmailJobData>(
|
||||
'email-sending',
|
||||
async (job: Job<EmailJobData>) => {
|
||||
const { to, subject } = job.data;
|
||||
const jobLogger = logger.child({ jobId: job.id, jobName: job.name });
|
||||
jobLogger.info({ to, subject }, `[EmailWorker] Sending email for job ${job.id}`);
|
||||
try {
|
||||
await emailService.sendEmail(job.data, jobLogger);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error(
|
||||
{
|
||||
err: wrappedError,
|
||||
jobData: job.data,
|
||||
},
|
||||
`[EmailWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
createWorkerProcessor((job) => emailService.processEmailJob(job)),
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.EMAIL_WORKER_CONCURRENCY || '10', 10),
|
||||
@@ -119,23 +105,7 @@ export const emailWorker = new Worker<EmailJobData>(
|
||||
|
||||
export const analyticsWorker = new Worker<AnalyticsJobData>(
|
||||
'analytics-reporting',
|
||||
async (job: Job<AnalyticsJobData>) => {
|
||||
const { reportDate } = job.data;
|
||||
logger.info({ reportDate }, `[AnalyticsWorker] Starting report generation for job ${job.id}`);
|
||||
try {
|
||||
if (reportDate === 'FAIL') {
|
||||
throw new Error('This is a test failure for the analytics job.');
|
||||
}
|
||||
await new Promise((resolve) => setTimeout(resolve, 10000));
|
||||
logger.info(`[AnalyticsWorker] Successfully generated report for ${reportDate}.`);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error({ err: wrappedError, jobData: job.data },
|
||||
`[AnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
createWorkerProcessor((job) => analyticsService.processDailyReportJob(job)),
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.ANALYTICS_WORKER_CONCURRENCY || '1', 10),
|
||||
@@ -144,51 +114,7 @@ export const analyticsWorker = new Worker<AnalyticsJobData>(
|
||||
|
||||
export const cleanupWorker = new Worker<CleanupJobData>(
|
||||
'file-cleanup',
|
||||
async (job: Job<CleanupJobData>) => {
|
||||
const { flyerId, paths } = job.data;
|
||||
logger.info(
|
||||
{ paths },
|
||||
`[CleanupWorker] Starting file cleanup for job ${job.id} (Flyer ID: ${flyerId})`,
|
||||
);
|
||||
|
||||
try {
|
||||
if (!paths || paths.length === 0) {
|
||||
logger.warn(
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${flyerId} received no paths to clean. Skipping.`,
|
||||
);
|
||||
return;
|
||||
}
|
||||
|
||||
for (const filePath of paths) {
|
||||
try {
|
||||
await fsAdapter.unlink(filePath);
|
||||
logger.info(`[CleanupWorker] Deleted temporary file: ${filePath}`);
|
||||
} catch (unlinkError: unknown) {
|
||||
if (
|
||||
unlinkError instanceof Error &&
|
||||
'code' in unlinkError &&
|
||||
(unlinkError as any).code === 'ENOENT'
|
||||
) {
|
||||
logger.warn(
|
||||
`[CleanupWorker] File not found during cleanup (already deleted?): ${filePath}`,
|
||||
);
|
||||
} else {
|
||||
throw unlinkError;
|
||||
}
|
||||
}
|
||||
}
|
||||
logger.info(
|
||||
`[CleanupWorker] Successfully cleaned up ${paths.length} file(s) for flyer ${flyerId}.`,
|
||||
);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error(
|
||||
{ err: wrappedError },
|
||||
`[CleanupWorker] Job ${job.id} for flyer ${flyerId} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
createWorkerProcessor((job) => flyerProcessingService.processCleanupJob(job)),
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.CLEANUP_WORKER_CONCURRENCY || '10', 10),
|
||||
@@ -197,26 +123,7 @@ export const cleanupWorker = new Worker<CleanupJobData>(
|
||||
|
||||
export const weeklyAnalyticsWorker = new Worker<WeeklyAnalyticsJobData>(
|
||||
'weekly-analytics-reporting',
|
||||
async (job: Job<WeeklyAnalyticsJobData>) => {
|
||||
const { reportYear, reportWeek } = job.data;
|
||||
logger.info(
|
||||
{ reportYear, reportWeek },
|
||||
`[WeeklyAnalyticsWorker] Starting weekly report generation for job ${job.id}`,
|
||||
);
|
||||
try {
|
||||
await new Promise((resolve) => setTimeout(resolve, 30000));
|
||||
logger.info(
|
||||
`[WeeklyAnalyticsWorker] Successfully generated weekly report for week ${reportWeek}, ${reportYear}.`,
|
||||
);
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
logger.error(
|
||||
{ err: wrappedError, jobData: job.data },
|
||||
`[WeeklyAnalyticsWorker] Job ${job.id} failed. Attempt ${job.attemptsMade}/${job.opts.attempts}.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
createWorkerProcessor((job) => analyticsService.processWeeklyReportJob(job)),
|
||||
{
|
||||
connection,
|
||||
concurrency: parseInt(process.env.WEEKLY_ANALYTICS_WORKER_CONCURRENCY || '1', 10),
|
||||
@@ -225,19 +132,7 @@ export const weeklyAnalyticsWorker = new Worker<WeeklyAnalyticsJobData>(
|
||||
|
||||
export const tokenCleanupWorker = new Worker<TokenCleanupJobData>(
|
||||
'token-cleanup',
|
||||
async (job: Job<TokenCleanupJobData>) => {
|
||||
const jobLogger = logger.child({ jobId: job.id, jobName: job.name });
|
||||
jobLogger.info('[TokenCleanupWorker] Starting cleanup of expired password reset tokens.');
|
||||
try {
|
||||
const deletedCount = await db.userRepo.deleteExpiredResetTokens(jobLogger);
|
||||
jobLogger.info(`[TokenCleanupWorker] Successfully deleted ${deletedCount} expired tokens.`);
|
||||
return { deletedCount };
|
||||
} catch (error: unknown) {
|
||||
const wrappedError = normalizeError(error);
|
||||
jobLogger.error({ err: wrappedError }, `[TokenCleanupWorker] Job ${job.id} failed.`);
|
||||
throw wrappedError;
|
||||
}
|
||||
},
|
||||
createWorkerProcessor((job) => userService.processTokenCleanupJob(job)),
|
||||
{
|
||||
connection,
|
||||
concurrency: 1,
|
||||
|
||||
@@ -39,6 +39,7 @@ import {
|
||||
ShoppingTripItem,
|
||||
Receipt,
|
||||
ReceiptItem,
|
||||
SearchQuery,
|
||||
ProcessingStage,
|
||||
UserAlert,
|
||||
UserSubmittedPrice,
|
||||
@@ -199,6 +200,7 @@ export const createMockFlyer = (
|
||||
valid_from: new Date().toISOString().split('T')[0],
|
||||
valid_to: new Date(Date.now() + 7 * 24 * 60 * 60 * 1000).toISOString().split('T')[0], // 7 days from now
|
||||
store_address: '123 Main St, Anytown, USA',
|
||||
status: 'processed',
|
||||
item_count: 50,
|
||||
uploaded_by: null,
|
||||
store,
|
||||
@@ -1451,3 +1453,66 @@ export const createMockAppliance = (overrides: Partial<Appliance> = {}): Applian
|
||||
...overrides,
|
||||
};
|
||||
};
|
||||
|
||||
// src/tests/utils/mockFactories.ts
|
||||
|
||||
// ... existing factories
|
||||
|
||||
export const createMockShoppingListItemPayload = (overrides: Partial<{ masterItemId: number; customItemName: string }> = {}): { masterItemId?: number; customItemName?: string } => ({
|
||||
customItemName: 'Mock Item',
|
||||
...overrides,
|
||||
});
|
||||
|
||||
export const createMockRecipeCommentPayload = (overrides: Partial<{ content: string; parentCommentId: number }> = {}): { content: string; parentCommentId?: number } => ({
|
||||
content: 'This is a mock comment.',
|
||||
...overrides,
|
||||
});
|
||||
|
||||
export const createMockProfileUpdatePayload = (overrides: Partial<Profile> = {}): Partial<Profile> => ({
|
||||
full_name: 'Mock User',
|
||||
...overrides,
|
||||
});
|
||||
|
||||
export const createMockAddressPayload = (overrides: Partial<Address> = {}): Partial<Address> => ({
|
||||
address_line_1: '123 Mock St',
|
||||
city: 'Mockville',
|
||||
province_state: 'MS',
|
||||
postal_code: '12345',
|
||||
country: 'Mockland',
|
||||
...overrides,
|
||||
});
|
||||
|
||||
export const createMockSearchQueryPayload = (overrides: Partial<Omit<SearchQuery, 'search_query_id' | 'id' | 'created_at' | 'user_id'>> = {}): Omit<SearchQuery, 'search_query_id' | 'id' | 'created_at' | 'user_id'> => ({
|
||||
query_text: 'mock search',
|
||||
result_count: 5,
|
||||
was_successful: true,
|
||||
...overrides,
|
||||
});
|
||||
|
||||
export const createMockWatchedItemPayload = (overrides: Partial<{ itemName: string; category: string }> = {}): { itemName: string; category: string } => ({
|
||||
itemName: 'Mock Watched Item',
|
||||
category: 'Pantry',
|
||||
...overrides,
|
||||
});
|
||||
|
||||
export const createMockRegisterUserPayload = (
|
||||
overrides: Partial<{
|
||||
email: string;
|
||||
password: string;
|
||||
full_name: string;
|
||||
avatar_url: string | undefined;
|
||||
}> = {},
|
||||
) => ({
|
||||
email: 'mock@example.com',
|
||||
password: 'password123',
|
||||
full_name: 'Mock User',
|
||||
avatar_url: undefined,
|
||||
...overrides,
|
||||
});
|
||||
|
||||
export const createMockLoginPayload = (overrides: Partial<{ email: string; password: string; rememberMe: boolean }> = {}) => ({
|
||||
email: 'mock@example.com',
|
||||
password: 'password123',
|
||||
rememberMe: false,
|
||||
...overrides,
|
||||
});
|
||||
|
||||
@@ -8,6 +8,8 @@ export interface Store {
|
||||
created_by?: string | null;
|
||||
}
|
||||
|
||||
export type FlyerStatus = 'processed' | 'needs_review' | 'archived';
|
||||
|
||||
export interface Flyer {
|
||||
flyer_id: number;
|
||||
created_at: string;
|
||||
@@ -20,6 +22,7 @@ export interface Flyer {
|
||||
valid_from?: string | null;
|
||||
valid_to?: string | null;
|
||||
store_address?: string | null;
|
||||
status: FlyerStatus;
|
||||
item_count: number;
|
||||
uploaded_by?: string | null; // UUID of the user who uploaded it, can be null for anonymous uploads
|
||||
store?: Store;
|
||||
@@ -38,6 +41,7 @@ export interface FlyerInsert {
|
||||
valid_from: string | null;
|
||||
valid_to: string | null;
|
||||
store_address: string | null;
|
||||
status: FlyerStatus;
|
||||
item_count: number;
|
||||
uploaded_by?: string | null;
|
||||
}
|
||||
|
||||
54
src/types/job-data.ts
Normal file
54
src/types/job-data.ts
Normal file
@@ -0,0 +1,54 @@
|
||||
// src/types/job-data.ts
|
||||
|
||||
/**
|
||||
* Defines the data structure for a flyer processing job.
|
||||
* This is the information passed to the worker when a new flyer is uploaded.
|
||||
*/
|
||||
export interface FlyerJobData {
|
||||
filePath: string;
|
||||
originalFileName: string;
|
||||
checksum: string;
|
||||
userId?: string;
|
||||
submitterIp?: string;
|
||||
userProfileAddress?: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for an email sending job.
|
||||
*/
|
||||
export interface EmailJobData {
|
||||
to: string;
|
||||
subject: string;
|
||||
text: string;
|
||||
html: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for a daily analytics reporting job.
|
||||
*/
|
||||
export interface AnalyticsJobData {
|
||||
reportDate: string; // e.g., '2024-10-26'
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for a weekly analytics reporting job.
|
||||
*/
|
||||
export interface WeeklyAnalyticsJobData {
|
||||
reportYear: number;
|
||||
reportWeek: number; // ISO week number (1-53)
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for a file cleanup job, which runs after a flyer is successfully processed.
|
||||
*/
|
||||
export interface CleanupJobData {
|
||||
flyerId: number;
|
||||
paths?: string[];
|
||||
}
|
||||
|
||||
/**
|
||||
* Defines the data structure for the job that cleans up expired password reset tokens.
|
||||
*/
|
||||
export interface TokenCleanupJobData {
|
||||
timestamp: string;
|
||||
}
|
||||
Reference in New Issue
Block a user