Compare commits
120 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
e4d70a9b37 | ||
| c30f1c4162 | |||
|
|
44062a9f5b | ||
| 17fac8cf86 | |||
|
|
9fa8553486 | ||
|
|
f5b0b3b543 | ||
| e3ed5c7e63 | |||
|
|
ae0040e092 | ||
| 1f3f99d430 | |||
|
|
7be72f1758 | ||
| 0967c7a33d | |||
| 1f1c0fa6f3 | |||
|
|
728b1a20d3 | ||
| f248f7cbd0 | |||
|
|
0ad9bb16c2 | ||
| 510787bc5b | |||
|
|
9f696e7676 | ||
|
|
a77105316f | ||
| cadacb63f5 | |||
|
|
62592f707e | ||
| 023e48d99a | |||
|
|
99efca0371 | ||
| 1448950b81 | |||
|
|
a811fdac63 | ||
| 1201fe4d3c | |||
|
|
ba9228c9cb | ||
| b392b82c25 | |||
|
|
87825d13d6 | ||
| 21a6a796cf | |||
|
|
ecd0a73bc8 | ||
|
|
39d61dc7ad | ||
|
|
43491359d9 | ||
| 5ed2cea7e9 | |||
|
|
cbb16a8d52 | ||
| 70e94a6ce0 | |||
|
|
b61a00003a | ||
| 52dba6f890 | |||
| 4242678aab | |||
|
|
b2e086d5ba | ||
| 07a9787570 | |||
|
|
4bf5dc3d58 | ||
| be3d269928 | |||
|
|
80a53fae94 | ||
| e15d2b6c2f | |||
|
|
7a52bf499e | ||
| 2489ec8d2d | |||
|
|
4a4f349805 | ||
| 517a268307 | |||
|
|
a94b2a97b1 | ||
| 542cdfbb82 | |||
|
|
262062f468 | ||
| 0a14193371 | |||
|
|
7f665f5117 | ||
| 2782a8fb3b | |||
|
|
c182ef6d30 | ||
| fdb3b76cbd | |||
|
|
01e7c843cb | ||
| a0dbefbfa0 | |||
|
|
ab3fc318a0 | ||
| e658b35e43 | |||
|
|
67e106162a | ||
| b7f3182fd6 | |||
|
|
ac60072d88 | ||
| 9390f38bf6 | |||
|
|
236d5518c9 | ||
| fd52a79a72 | |||
|
|
f72819e343 | ||
| 1af8be3f15 | |||
|
|
28d03f4e21 | ||
| 2e72ee81dd | |||
|
|
ba67ace190 | ||
|
|
50782c30e5 | ||
| 4a2ff8afc5 | |||
|
|
7a1c14ce89 | ||
| 6fafc3d089 | |||
|
|
4316866bce | ||
| 356c1a1894 | |||
|
|
2a310648ca | ||
| 8592633c22 | |||
|
|
0a9cdb8709 | ||
| 0d21e098f8 | |||
| b6799ed167 | |||
|
|
be5bda169e | ||
| 4ede403356 | |||
| 5d31605b80 | |||
| ddd4ad024e | |||
|
|
4e927f48bd | ||
| af5644d17a | |||
|
|
016c0a883a | ||
| c6a5f889b4 | |||
|
|
c895ecdb28 | ||
| 05e3f8a61c | |||
|
|
f79a2abc65 | ||
| a726c270bb | |||
|
|
8a4965c45b | ||
| 93497bf7c7 | |||
|
|
20584af729 | ||
| be9f452656 | |||
| ef4b8e58fe | |||
|
|
a42f7d7007 | ||
| 768d02b9ed | |||
|
|
c4742959e4 | ||
| 97c54c0c5c | |||
| 7cc50907d1 | |||
|
|
b4199f7c48 | ||
| dda36f7bc5 | |||
| 27810bbb36 | |||
|
|
7a1421d5c2 | ||
| 1b52478f97 | |||
| fe8b000737 | |||
|
|
d2babbe3b0 | ||
|
|
684d81db2a | ||
| 59ffa65562 | |||
| 0c0dd852ac | |||
|
|
cde766872e | ||
| 604b543c12 | |||
| fd67fe2941 | |||
|
|
582035b60e | ||
| 44e7670a89 | |||
| 2abfb3ed6e |
@@ -138,6 +138,10 @@ jobs:
|
||||
cd /var/www/flyer-crawler.projectium.com
|
||||
npm install --omit=dev
|
||||
|
||||
# --- Cleanup Errored Processes ---
|
||||
echo "Cleaning up errored or stopped PM2 processes..."
|
||||
node -e "const exec = require('child_process').execSync; try { const list = JSON.parse(exec('pm2 jlist').toString()); list.forEach(p => { if (p.pm2_env.status === 'errored' || p.pm2_env.status === 'stopped') { console.log('Deleting ' + p.pm2_env.status + ' process: ' + p.name + ' (' + p.pm2_env.pm_id + ')'); try { exec('pm2 delete ' + p.pm2_env.pm_id); } catch(e) { console.error('Failed to delete ' + p.pm2_env.pm_id); } } }); } catch (e) { console.error('Error cleaning up processes:', e); }"
|
||||
|
||||
# --- Version Check Logic ---
|
||||
# Get the version from the newly deployed package.json
|
||||
NEW_VERSION=$(node -p "require('./package.json').version")
|
||||
@@ -154,7 +158,7 @@ jobs:
|
||||
else
|
||||
echo "Version mismatch (Running: $RUNNING_VERSION -> Deployed: $NEW_VERSION) or app not running. Reloading PM2..."
|
||||
fi
|
||||
pm2 startOrReload ecosystem.config.cjs --env production && pm2 save
|
||||
pm2 startOrReload ecosystem.config.cjs --env production --update-env && pm2 save
|
||||
echo "Production backend server reloaded successfully."
|
||||
else
|
||||
echo "Version $NEW_VERSION is already running. Skipping PM2 reload."
|
||||
@@ -181,7 +185,17 @@ jobs:
|
||||
- name: Show PM2 Environment for Production
|
||||
run: |
|
||||
echo "--- Displaying recent PM2 logs for flyer-crawler-api ---"
|
||||
sleep 5
|
||||
pm2 describe flyer-crawler-api || echo "Could not find production pm2 process."
|
||||
pm2 logs flyer-crawler-api --lines 20 --nostream || echo "Could not find production pm2 process."
|
||||
pm2 env flyer-crawler-api || echo "Could not find production pm2 process."
|
||||
sleep 5 # Wait a few seconds for the app to start and log its output.
|
||||
|
||||
# Resolve the PM2 ID dynamically to ensure we target the correct process
|
||||
PM2_ID=$(pm2 jlist | node -e "try { const list = JSON.parse(require('fs').readFileSync(0, 'utf-8')); const app = list.find(p => p.name === 'flyer-crawler-api'); console.log(app ? app.pm2_env.pm_id : ''); } catch(e) { console.log(''); }")
|
||||
|
||||
if [ -n "$PM2_ID" ]; then
|
||||
echo "Found process ID: $PM2_ID"
|
||||
pm2 describe "$PM2_ID" || echo "Failed to describe process $PM2_ID"
|
||||
pm2 logs "$PM2_ID" --lines 20 --nostream || echo "Failed to get logs for $PM2_ID"
|
||||
pm2 env "$PM2_ID" || echo "Failed to get env for $PM2_ID"
|
||||
else
|
||||
echo "Could not find process 'flyer-crawler-api' in pm2 list."
|
||||
pm2 list # Fallback to listing everything to help debug
|
||||
fi
|
||||
|
||||
@@ -90,10 +90,11 @@ jobs:
|
||||
# integration test suite can launch its own, fresh server instance.
|
||||
# '|| true' ensures the workflow doesn't fail if the process isn't running.
|
||||
run: |
|
||||
pm2 stop flyer-crawler-api-test || true
|
||||
pm2 stop flyer-crawler-worker-test || true
|
||||
pm2 delete flyer-crawler-api-test || true
|
||||
pm2 delete flyer-crawler-worker-test || true
|
||||
echo "--- Stopping and deleting all test processes ---"
|
||||
# Use a script to parse pm2's JSON output and delete any process whose name ends with '-test'.
|
||||
# This is safer than 'pm2 delete all' and more robust than naming each process individually.
|
||||
# It prevents the accumulation of duplicate processes from previous test runs.
|
||||
node -e "const exec = require('child_process').execSync; try { const list = JSON.parse(exec('pm2 jlist').toString()); list.forEach(p => { if (p.name && p.name.endsWith('-test')) { console.log('Deleting test process: ' + p.name + ' (' + p.pm2_env.pm_id + ')'); try { exec('pm2 delete ' + p.pm2_env.pm_id); } catch(e) { console.error('Failed to delete ' + p.pm2_env.pm_id, e.message); } } }); console.log('✅ Test process cleanup complete.'); } catch (e) { if (e.stdout.toString().includes('No process found')) { console.log('No PM2 processes running, cleanup not needed.'); } else { console.error('Error cleaning up test processes:', e.message); } }" || true
|
||||
|
||||
- name: Run All Tests and Generate Merged Coverage Report
|
||||
# This single step runs both unit and integration tests, then merges their
|
||||
@@ -126,7 +127,7 @@ jobs:
|
||||
|
||||
# --- Increase Node.js memory limit to prevent heap out of memory errors ---
|
||||
# This is crucial for memory-intensive tasks like running tests and coverage.
|
||||
NODE_OPTIONS: '--max-old-space-size=8192'
|
||||
NODE_OPTIONS: '--max-old-space-size=8192 --trace-warnings --unhandled-rejections=strict'
|
||||
|
||||
run: |
|
||||
# Fail-fast check to ensure secrets are configured in Gitea for testing.
|
||||
@@ -142,15 +143,48 @@ jobs:
|
||||
# The `|| true` ensures the workflow continues even if tests fail, allowing coverage to run.
|
||||
echo "--- Running Unit Tests ---"
|
||||
# npm run test:unit -- --coverage --reporter=verbose --includeTaskLocation --testTimeout=10000 --silent=passed-only || true
|
||||
npm run test:unit -- --coverage --coverage.exclude='**/*.test.ts' --coverage.exclude='**/tests/**' --coverage.exclude='**/mocks/**' --reporter=verbose --includeTaskLocation --testTimeout=10000 --silent=passed-only --no-file-parallelism || true
|
||||
npm run test:unit -- --coverage \
|
||||
--coverage.exclude='**/*.test.ts' \
|
||||
--coverage.exclude='**/tests/**' \
|
||||
--coverage.exclude='**/mocks/**' \
|
||||
--coverage.exclude='src/components/icons/**' \
|
||||
--coverage.exclude='src/db/**' \
|
||||
--coverage.exclude='src/lib/**' \
|
||||
--coverage.exclude='src/types/**' \
|
||||
--coverage.exclude='**/index.tsx' \
|
||||
--coverage.exclude='**/vite-env.d.ts' \
|
||||
--coverage.exclude='**/vitest.setup.ts' \
|
||||
--reporter=verbose --includeTaskLocation --testTimeout=10000 --silent=passed-only --no-file-parallelism || true
|
||||
|
||||
echo "--- Running Integration Tests ---"
|
||||
npm run test:integration -- --coverage --coverage.exclude='**/*.test.ts' --coverage.exclude='**/tests/**' --coverage.exclude='**/mocks/**' --reporter=verbose --includeTaskLocation --testTimeout=10000 --silent=passed-only || true
|
||||
npm run test:integration -- --coverage \
|
||||
--coverage.exclude='**/*.test.ts' \
|
||||
--coverage.exclude='**/tests/**' \
|
||||
--coverage.exclude='**/mocks/**' \
|
||||
--coverage.exclude='src/components/icons/**' \
|
||||
--coverage.exclude='src/db/**' \
|
||||
--coverage.exclude='src/lib/**' \
|
||||
--coverage.exclude='src/types/**' \
|
||||
--coverage.exclude='**/index.tsx' \
|
||||
--coverage.exclude='**/vite-env.d.ts' \
|
||||
--coverage.exclude='**/vitest.setup.ts' \
|
||||
--reporter=verbose --includeTaskLocation --testTimeout=10000 --silent=passed-only || true
|
||||
|
||||
echo "--- Running E2E Tests ---"
|
||||
# Run E2E tests using the dedicated E2E config which inherits from integration config.
|
||||
# We still pass --coverage to enable it, but directory and timeout are now in the config.
|
||||
npx vitest run --config vitest.config.e2e.ts --coverage --coverage.exclude='**/*.test.ts' --coverage.exclude='**/tests/**' --coverage.exclude='**/mocks/**' --reporter=verbose --no-file-parallelism || true
|
||||
npx vitest run --config vitest.config.e2e.ts --coverage \
|
||||
--coverage.exclude='**/*.test.ts' \
|
||||
--coverage.exclude='**/tests/**' \
|
||||
--coverage.exclude='**/mocks/**' \
|
||||
--coverage.exclude='src/components/icons/**' \
|
||||
--coverage.exclude='src/db/**' \
|
||||
--coverage.exclude='src/lib/**' \
|
||||
--coverage.exclude='src/types/**' \
|
||||
--coverage.exclude='**/index.tsx' \
|
||||
--coverage.exclude='**/vite-env.d.ts' \
|
||||
--coverage.exclude='**/vitest.setup.ts' \
|
||||
--reporter=verbose --no-file-parallelism || true
|
||||
|
||||
# Re-enable secret masking for subsequent steps.
|
||||
echo "::secret-masking::"
|
||||
@@ -221,7 +255,10 @@ jobs:
|
||||
--temp-dir "$NYC_SOURCE_DIR" \
|
||||
--exclude "**/*.test.ts" \
|
||||
--exclude "**/tests/**" \
|
||||
--exclude "**/mocks/**"
|
||||
--exclude "**/mocks/**" \
|
||||
--exclude "**/index.tsx" \
|
||||
--exclude "**/vite-env.d.ts" \
|
||||
--exclude "**/vitest.setup.ts"
|
||||
|
||||
# Re-enable secret masking for subsequent steps.
|
||||
echo "::secret-masking::"
|
||||
@@ -234,16 +271,6 @@ jobs:
|
||||
if: always() # This step runs even if the previous test or coverage steps failed.
|
||||
run: echo "Skipping test artifact cleanup on runner; this is handled on the server."
|
||||
|
||||
- name: Deploy Coverage Report to Public URL
|
||||
if: always()
|
||||
run: |
|
||||
TARGET_DIR="/var/www/flyer-crawler-test.projectium.com/coverage"
|
||||
echo "Deploying HTML coverage report to $TARGET_DIR..."
|
||||
mkdir -p "$TARGET_DIR"
|
||||
rm -rf "$TARGET_DIR"/*
|
||||
cp -r .coverage/* "$TARGET_DIR/"
|
||||
echo "✅ Coverage report deployed to https://flyer-crawler-test.projectium.com/coverage"
|
||||
|
||||
- name: Archive Code Coverage Report
|
||||
# This action saves the generated HTML coverage report as a downloadable artifact.
|
||||
uses: actions/upload-artifact@v3
|
||||
@@ -333,6 +360,17 @@ jobs:
|
||||
rsync -avz dist/ "$APP_PATH"
|
||||
echo "Application deployment complete."
|
||||
|
||||
- name: Deploy Coverage Report to Public URL
|
||||
if: always()
|
||||
run: |
|
||||
TARGET_DIR="/var/www/flyer-crawler-test.projectium.com/coverage"
|
||||
echo "Deploying HTML coverage report to $TARGET_DIR..."
|
||||
mkdir -p "$TARGET_DIR"
|
||||
rm -rf "$TARGET_DIR"/*
|
||||
# The merged nyc report is generated in the .coverage directory. We copy its contents.
|
||||
cp -r .coverage/* "$TARGET_DIR/"
|
||||
echo "✅ Coverage report deployed to https://flyer-crawler-test.projectium.com/coverage"
|
||||
|
||||
- name: Install Backend Dependencies and Restart Test Server
|
||||
env:
|
||||
# --- Test Secrets Injection ---
|
||||
@@ -351,7 +389,7 @@ jobs:
|
||||
|
||||
# Application Secrets
|
||||
FRONTEND_URL: 'https://flyer-crawler-test.projectium.com'
|
||||
JWT_SECRET: ${{ secrets.JWT_SECRET_TEST }}
|
||||
JWT_SECRET: ${{ secrets.JWT_SECRET }}
|
||||
GEMINI_API_KEY: ${{ secrets.VITE_GOOGLE_GENAI_API_KEY_TEST }}
|
||||
GOOGLE_MAPS_API_KEY: ${{ secrets.GOOGLE_MAPS_API_KEY }}
|
||||
|
||||
@@ -365,18 +403,30 @@ jobs:
|
||||
|
||||
run: |
|
||||
# Fail-fast check to ensure secrets are configured in Gitea.
|
||||
if [ -z "$DB_HOST" ] || [ -z "$DB_USER" ] || [ -z "$DB_PASSWORD" ] || [ -z "$DB_NAME" ]; then
|
||||
echo "ERROR: One or more test database secrets (DB_HOST, DB_USER, DB_PASSWORD, DB_DATABASE_TEST) are not set in Gitea repository settings."
|
||||
MISSING_SECRETS=""
|
||||
if [ -z "$DB_HOST" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_HOST"; fi
|
||||
if [ -z "$DB_USER" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_USER"; fi
|
||||
if [ -z "$DB_PASSWORD" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_PASSWORD"; fi
|
||||
if [ -z "$DB_NAME" ]; then MISSING_SECRETS="${MISSING_SECRETS} DB_NAME"; fi
|
||||
if [ -z "$JWT_SECRET" ]; then MISSING_SECRETS="${MISSING_SECRETS} JWT_SECRET"; fi
|
||||
|
||||
if [ ! -z "$MISSING_SECRETS" ]; then
|
||||
echo "ERROR: The following required secrets are missing in Gitea:${MISSING_SECRETS}"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
echo "Installing production dependencies and restarting test server..."
|
||||
cd /var/www/flyer-crawler-test.projectium.com
|
||||
npm install --omit=dev
|
||||
|
||||
# --- Cleanup Errored Processes ---
|
||||
echo "Cleaning up errored or stopped PM2 processes..."
|
||||
node -e "const exec = require('child_process').execSync; try { const list = JSON.parse(exec('pm2 jlist').toString()); list.forEach(p => { if (p.pm2_env.status === 'errored' || p.pm2_env.status === 'stopped') { console.log('Deleting ' + p.pm2_env.status + ' process: ' + p.name + ' (' + p.pm2_env.pm_id + ')'); try { exec('pm2 delete ' + p.pm2_env.pm_id); } catch(e) { console.error('Failed to delete ' + p.pm2_env.pm_id); } } }); } catch (e) { console.error('Error cleaning up processes:', e); }"
|
||||
|
||||
# Use `startOrReload` with the ecosystem file. This is the standard, idempotent way to deploy.
|
||||
# It will START the process if it's not running, or RELOAD it if it is.
|
||||
# We also add `&& pm2 save` to persist the process list across server reboots.
|
||||
pm2 startOrReload ecosystem.config.cjs --env test && pm2 save
|
||||
pm2 startOrReload ecosystem.config.cjs --env test --update-env && pm2 save
|
||||
echo "Test backend server reloaded successfully."
|
||||
|
||||
# After a successful deployment, update the schema hash in the database.
|
||||
@@ -411,7 +461,17 @@ jobs:
|
||||
run: |
|
||||
echo "--- Displaying recent PM2 logs for flyer-crawler-api-test ---"
|
||||
# After a reload, the server restarts. We'll show the last 20 lines of the log to see the startup messages.
|
||||
sleep 5 # Wait a few seconds for the app to start and log its output.
|
||||
pm2 describe flyer-crawler-api-test || echo "Could not find test pm2 process."
|
||||
pm2 logs flyer-crawler-api-test --lines 20 --nostream || echo "Could not find test pm2 process."
|
||||
pm2 env flyer-crawler-api-test || echo "Could not find test pm2 process."
|
||||
sleep 5
|
||||
|
||||
# Resolve the PM2 ID dynamically to ensure we target the correct process
|
||||
PM2_ID=$(pm2 jlist | node -e "try { const list = JSON.parse(require('fs').readFileSync(0, 'utf-8')); const app = list.find(p => p.name === 'flyer-crawler-api-test'); console.log(app ? app.pm2_env.pm_id : ''); } catch(e) { console.log(''); }")
|
||||
|
||||
if [ -n "$PM2_ID" ]; then
|
||||
echo "Found process ID: $PM2_ID"
|
||||
pm2 describe "$PM2_ID" || echo "Failed to describe process $PM2_ID"
|
||||
pm2 logs "$PM2_ID" --lines 20 --nostream || echo "Failed to get logs for $PM2_ID"
|
||||
pm2 env "$PM2_ID" || echo "Failed to get env for $PM2_ID"
|
||||
else
|
||||
echo "Could not find process 'flyer-crawler-api-test' in pm2 list."
|
||||
pm2 list # Fallback to listing everything to help debug
|
||||
fi
|
||||
|
||||
@@ -137,6 +137,10 @@ jobs:
|
||||
cd /var/www/flyer-crawler.projectium.com
|
||||
npm install --omit=dev
|
||||
|
||||
# --- Cleanup Errored Processes ---
|
||||
echo "Cleaning up errored or stopped PM2 processes..."
|
||||
node -e "const exec = require('child_process').execSync; try { const list = JSON.parse(exec('pm2 jlist').toString()); list.forEach(p => { if (p.pm2_env.status === 'errored' || p.pm2_env.status === 'stopped') { console.log('Deleting ' + p.pm2_env.status + ' process: ' + p.name + ' (' + p.pm2_env.pm_id + ')'); try { exec('pm2 delete ' + p.pm2_env.pm_id); } catch(e) { console.error('Failed to delete ' + p.pm2_env.pm_id); } } }); } catch (e) { console.error('Error cleaning up processes:', e); }"
|
||||
|
||||
# --- Version Check Logic ---
|
||||
# Get the version from the newly deployed package.json
|
||||
NEW_VERSION=$(node -p "require('./package.json').version")
|
||||
@@ -153,7 +157,7 @@ jobs:
|
||||
else
|
||||
echo "Version mismatch (Running: $RUNNING_VERSION -> Deployed: $NEW_VERSION) or app not running. Reloading PM2..."
|
||||
fi
|
||||
pm2 startOrReload ecosystem.config.cjs --env production && pm2 save
|
||||
pm2 startOrReload ecosystem.config.cjs --env production --update-env && pm2 save
|
||||
echo "Production backend server reloaded successfully."
|
||||
else
|
||||
echo "Version $NEW_VERSION is already running. Skipping PM2 reload."
|
||||
|
||||
@@ -3,22 +3,39 @@
|
||||
// It allows us to define all the settings for our application in one place.
|
||||
// The .cjs extension is required because the project's package.json has "type": "module".
|
||||
|
||||
// --- Environment Variable Validation ---
|
||||
const requiredSecrets = ['DB_HOST', 'JWT_SECRET', 'GEMINI_API_KEY'];
|
||||
const missingSecrets = requiredSecrets.filter(key => !process.env[key]);
|
||||
|
||||
if (missingSecrets.length > 0) {
|
||||
console.warn('\n[ecosystem.config.cjs] ⚠️ WARNING: The following environment variables are MISSING in the shell:');
|
||||
missingSecrets.forEach(key => console.warn(` - ${key}`));
|
||||
console.warn('[ecosystem.config.cjs] The application may crash if these are required for startup.\n');
|
||||
process.exit(1); // Fail fast so PM2 doesn't attempt to start a broken app
|
||||
} else {
|
||||
console.log('[ecosystem.config.cjs] ✅ Critical environment variables are present.');
|
||||
}
|
||||
|
||||
module.exports = {
|
||||
apps: [
|
||||
{
|
||||
// --- API Server ---
|
||||
// The name is now dynamically set based on the environment.
|
||||
// This is a common pattern but requires you to call pm2 with the correct name.
|
||||
// The deploy script handles this by using 'flyer-crawler-api' for prod and 'flyer-crawler-api-test' for test.
|
||||
name: 'flyer-crawler-api',
|
||||
// Note: The process names below are referenced in .gitea/workflows/ for status checks.
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'server.ts', // tsx will execute this file
|
||||
args: 'server.ts',
|
||||
max_memory_restart: '500M',
|
||||
|
||||
// Restart Logic
|
||||
max_restarts: 40,
|
||||
exp_backoff_restart_delay: 100,
|
||||
min_uptime: '10s',
|
||||
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production', // Set the Node.js environment to production
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-api',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -38,10 +55,9 @@ module.exports = {
|
||||
},
|
||||
// Test Environment Settings
|
||||
env_test: {
|
||||
NODE_ENV: 'test', // Set to 'test' to match the environment purpose and disable pino-pretty
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-api-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -65,7 +81,6 @@ module.exports = {
|
||||
name: 'flyer-crawler-api-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -88,13 +103,19 @@ module.exports = {
|
||||
// --- General Worker ---
|
||||
name: 'flyer-crawler-worker',
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'src/worker.ts', // tsx will execute this file
|
||||
args: 'src/services/worker.ts',
|
||||
max_memory_restart: '1G',
|
||||
|
||||
// Restart Logic
|
||||
max_restarts: 40,
|
||||
exp_backoff_restart_delay: 100,
|
||||
min_uptime: '10s',
|
||||
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-worker',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -117,7 +138,6 @@ module.exports = {
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-worker-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -141,7 +161,6 @@ module.exports = {
|
||||
name: 'flyer-crawler-worker-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -164,13 +183,19 @@ module.exports = {
|
||||
// --- Analytics Worker ---
|
||||
name: 'flyer-crawler-analytics-worker',
|
||||
script: './node_modules/.bin/tsx',
|
||||
args: 'src/worker.ts', // tsx will execute this file
|
||||
args: 'src/services/worker.ts',
|
||||
max_memory_restart: '1G',
|
||||
|
||||
// Restart Logic
|
||||
max_restarts: 40,
|
||||
exp_backoff_restart_delay: 100,
|
||||
min_uptime: '10s',
|
||||
|
||||
// Production Environment Settings
|
||||
env_production: {
|
||||
NODE_ENV: 'production',
|
||||
name: 'flyer-crawler-analytics-worker',
|
||||
cwd: '/var/www/flyer-crawler.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -193,7 +218,6 @@ module.exports = {
|
||||
NODE_ENV: 'test',
|
||||
name: 'flyer-crawler-analytics-worker-test',
|
||||
cwd: '/var/www/flyer-crawler-test.projectium.com',
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
@@ -217,7 +241,6 @@ module.exports = {
|
||||
name: 'flyer-crawler-analytics-worker-dev',
|
||||
watch: true,
|
||||
ignore_watch: ['node_modules', 'logs', '*.log', 'flyer-images', '.git'],
|
||||
// Inherit secrets from the deployment environment
|
||||
DB_HOST: process.env.DB_HOST,
|
||||
DB_USER: process.env.DB_USER,
|
||||
DB_PASSWORD: process.env.DB_PASSWORD,
|
||||
|
||||
118
notes-to-ai4.txt
Normal file
118
notes-to-ai4.txt
Normal file
@@ -0,0 +1,118 @@
|
||||
RULES:
|
||||
1) if you do not have a file that you need, stop, and request it immediately.
|
||||
2) never remove logging or comments
|
||||
3) you cannot ever use 'any' or 'unknown' to solve possible typescript issues
|
||||
4) when creating new files, output there entire path in your explanation, to make it easier to know where to save those new files and directories to
|
||||
5) add comments when you can, as that will help ensure ideas persist into the app
|
||||
6) Your knowledge of package version, like nodejs, is always old, like a year or more old - ask me for the best version to use, as your knowledge is incomplete
|
||||
7) Stop making predictions and/or guessing at solutions. Focus on adding logging and debugging to issues that are not solved right away.
|
||||
8) Do not make obsequious statements - we're here to do a job, not get patted on the shoulder for insignificant achievements.
|
||||
9) Provide me with the npm command to execute rather than wanting to edit the package.json file. That is not the correct way to handle a package update.
|
||||
10) Provide the code solution in DIFF format for brevity.
|
||||
11) Always add logging and debugging to prove a solution works, and only remove logging and debugging when it is not possible to adapt it to the new conditions.
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
UPC SCANNING !
|
||||
|
||||
EXPIRY DATES !
|
||||
|
||||
|
||||
|
||||
|
||||
are there any other unit test files that should be refactored :
|
||||
|
||||
Extract Shared Utility Functions Several utility functions are duplicated across multiple route files.
|
||||
|
||||
Update Route Files to Use New Utilities route files will be updated to import these new utility functions
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
Review the `getJobStatus` API client function and add more robust error handling for non-JSON responses.
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
Review the `src/services/apiClient.test.ts` file. It seems to have a lot of manual data creation for API payloads. Can we refactor it to use factories where appropriate, or create new factories for API payloads if needed?
|
||||
|
||||
|
||||
|
||||
The `useApi` hook has complex state management. Can you suggest a way to simplify it, perhaps by using a reducer or a state machine?
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
Are there any opportunities to improve the mock factories themselves, for example, by adding more realistic default data?
|
||||
|
||||
Now that the user creation is centralized, can we add more complex test user scenarios to the helper, like a user with a pre-existing shopping list?
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
How can I add a new "Recipes Created" stat to the AdminStatsPage and its tests?
|
||||
|
||||
NO CODE just outline the steps
|
||||
|
||||
|
||||
Gemini
|
||||
Outlining the Implementation Path
|
||||
Gemini 2.5 generated this response due to high traffic on Gemini 3
|
||||
|
||||
Of course! Here is a step-by-step outline of how to add a new "Recipes Created" statistic to your AdminStatsPage.
|
||||
|
||||
Backend Steps
|
||||
|
||||
|
||||
Update Database Query:
|
||||
|
||||
In the database repository file responsible for fetching admin statistics (e.g., admin.db.ts), you'll need to add a query that performs a COUNT(*) on your recipes table.
|
||||
|
||||
|
||||
Update Service Layer:
|
||||
|
||||
The service function that aggregates the statistics (which is called by your API controller) needs to be modified to execute the new recipe count query.
|
||||
The result of this query should be added to the statistics object that is returned.
|
||||
|
||||
|
||||
Update API Endpoint:
|
||||
|
||||
The controller for the /api/admin/stats endpoint must be updated to include the new recipeCount field in the JSON object it sends to the frontend.
|
||||
Frontend Steps
|
||||
|
||||
|
||||
Update AppStats Type:
|
||||
|
||||
In src/services/apiClient.ts, where the AppStats type is defined, add the new property: recipeCount: number;.
|
||||
Update Mock Factory:
|
||||
|
||||
In src/tests/utils/mockFactories.ts, modify the createMockAppStats function. Add recipeCount to the defaultStats object with a default value (e.g., recipeCount: 500). This ensures any test that uses the factory will not break due to the missing property.
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
Update AdminStatsPage Component (AdminStatsPage.tsx):
|
||||
|
||||
Import a suitable icon for recipes, like BookOpenIcon, at the top of the file.
|
||||
Inside the main div where the other StatCard components are rendered, add a new one for the recipe count:
|
||||
jsx
|
||||
<StatCard title="Total Recipes" value={stats.recipeCount.toLocaleString()} icon={<BookOpenIcon className="w-6 h-6" />} />
|
||||
Update AdminStatsPage Tests (AdminStatsPage.test.tsx):
|
||||
|
||||
In the test 'should display stats cards when data is fetched successfully', update the mockStats object to include a value for recipeCount. Add a new expect assertion to verify that the "Total Recipes" title and its corresponding value are rendered.
|
||||
In the test 'should format large numbers with commas for readability', add a large number for recipeCount in the mock data and assert that it is rendered with commas.
|
||||
In the test 'should correctly display zero values for all stats', add recipeCount: 0 to the mock data and increase the expected count of "0" elements found in the document from 5 to 6.
|
||||
52
package-lock.json
generated
52
package-lock.json
generated
@@ -1,22 +1,24 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.1.16",
|
||||
"version": "0.6.2",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "flyer-crawler",
|
||||
"version": "0.1.16",
|
||||
"version": "0.6.2",
|
||||
"dependencies": {
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
"@google/genai": "^1.30.0",
|
||||
"@tanstack/react-query": "^5.90.12",
|
||||
"@types/connect-timeout": "^1.9.0",
|
||||
"bcrypt": "^5.1.1",
|
||||
"bullmq": "^5.65.1",
|
||||
"connect-timeout": "^1.9.1",
|
||||
"cookie-parser": "^1.4.7",
|
||||
"date-fns": "^4.1.0",
|
||||
"exif-parser": "^0.1.12",
|
||||
"express": "^5.1.0",
|
||||
"express-list-endpoints": "^7.1.1",
|
||||
"express-rate-limit": "^8.2.1",
|
||||
@@ -34,6 +36,7 @@
|
||||
"passport-local": "^1.0.0",
|
||||
"pdfjs-dist": "^5.4.394",
|
||||
"pg": "^8.16.3",
|
||||
"piexifjs": "^1.0.6",
|
||||
"pino": "^10.1.0",
|
||||
"react": "^19.2.0",
|
||||
"react-dom": "^19.2.0",
|
||||
@@ -65,6 +68,7 @@
|
||||
"@types/passport-jwt": "^4.0.1",
|
||||
"@types/passport-local": "^1.0.38",
|
||||
"@types/pg": "^8.15.6",
|
||||
"@types/piexifjs": "^1.0.0",
|
||||
"@types/pino": "^7.0.4",
|
||||
"@types/react": "^19.2.7",
|
||||
"@types/react-dom": "^19.2.3",
|
||||
@@ -4882,6 +4886,32 @@
|
||||
"dev": true,
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/@tanstack/query-core": {
|
||||
"version": "5.90.12",
|
||||
"resolved": "https://registry.npmjs.org/@tanstack/query-core/-/query-core-5.90.12.tgz",
|
||||
"integrity": "sha512-T1/8t5DhV/SisWjDnaiU2drl6ySvsHj1bHBCWNXd+/T+Hh1cf6JodyEYMd5sgwm+b/mETT4EV3H+zCVczCU5hg==",
|
||||
"license": "MIT",
|
||||
"funding": {
|
||||
"type": "github",
|
||||
"url": "https://github.com/sponsors/tannerlinsley"
|
||||
}
|
||||
},
|
||||
"node_modules/@tanstack/react-query": {
|
||||
"version": "5.90.12",
|
||||
"resolved": "https://registry.npmjs.org/@tanstack/react-query/-/react-query-5.90.12.tgz",
|
||||
"integrity": "sha512-graRZspg7EoEaw0a8faiUASCyJrqjKPdqJ9EwuDRUF9mEYJ1YPczI9H+/agJ0mOJkPCJDk0lsz5QTrLZ/jQ2rg==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"@tanstack/query-core": "5.90.12"
|
||||
},
|
||||
"funding": {
|
||||
"type": "github",
|
||||
"url": "https://github.com/sponsors/tannerlinsley"
|
||||
},
|
||||
"peerDependencies": {
|
||||
"react": "^18 || ^19"
|
||||
}
|
||||
},
|
||||
"node_modules/@testcontainers/postgresql": {
|
||||
"version": "11.10.0",
|
||||
"resolved": "https://registry.npmjs.org/@testcontainers/postgresql/-/postgresql-11.10.0.tgz",
|
||||
@@ -5408,6 +5438,13 @@
|
||||
"pg-types": "^2.2.0"
|
||||
}
|
||||
},
|
||||
"node_modules/@types/piexifjs": {
|
||||
"version": "1.0.0",
|
||||
"resolved": "https://registry.npmjs.org/@types/piexifjs/-/piexifjs-1.0.0.tgz",
|
||||
"integrity": "sha512-PPiGeCkmkZQgYjvqtjD3kp4OkbCox2vEFVuK4DaLVOIazJLAXk+/ujbizkIPH5CN4AnN9Clo5ckzUlaj3+SzCA==",
|
||||
"dev": true,
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/@types/pino": {
|
||||
"version": "7.0.4",
|
||||
"resolved": "https://registry.npmjs.org/@types/pino/-/pino-7.0.4.tgz",
|
||||
@@ -8938,6 +8975,11 @@
|
||||
"bare-events": "^2.7.0"
|
||||
}
|
||||
},
|
||||
"node_modules/exif-parser": {
|
||||
"version": "0.1.12",
|
||||
"resolved": "https://registry.npmjs.org/exif-parser/-/exif-parser-0.1.12.tgz",
|
||||
"integrity": "sha512-c2bQfLNbMzLPmzQuOr8fy0csy84WmwnER81W88DzTp9CYNPJ6yzOj2EZAh9pywYpqHnshVLHQJ8WzldAyfY+Iw=="
|
||||
},
|
||||
"node_modules/expect-type": {
|
||||
"version": "1.3.0",
|
||||
"resolved": "https://registry.npmjs.org/expect-type/-/expect-type-1.3.0.tgz",
|
||||
@@ -13336,6 +13378,12 @@
|
||||
"url": "https://github.com/sponsors/jonschlinkert"
|
||||
}
|
||||
},
|
||||
"node_modules/piexifjs": {
|
||||
"version": "1.0.6",
|
||||
"resolved": "https://registry.npmjs.org/piexifjs/-/piexifjs-1.0.6.tgz",
|
||||
"integrity": "sha512-0wVyH0cKohzBQ5Gi2V1BuxYpxWfxF3cSqfFXfPIpl5tl9XLS5z4ogqhUCD20AbHi0h9aJkqXNJnkVev6gwh2ag==",
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/pino": {
|
||||
"version": "10.1.0",
|
||||
"resolved": "https://registry.npmjs.org/pino/-/pino-10.1.0.tgz",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "flyer-crawler",
|
||||
"private": true,
|
||||
"version": "0.1.16",
|
||||
"version": "0.6.2",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "concurrently \"npm:start:dev\" \"vite\"",
|
||||
@@ -30,12 +30,14 @@
|
||||
"@bull-board/api": "^6.14.2",
|
||||
"@bull-board/express": "^6.14.2",
|
||||
"@google/genai": "^1.30.0",
|
||||
"@tanstack/react-query": "^5.90.12",
|
||||
"@types/connect-timeout": "^1.9.0",
|
||||
"bcrypt": "^5.1.1",
|
||||
"bullmq": "^5.65.1",
|
||||
"connect-timeout": "^1.9.1",
|
||||
"cookie-parser": "^1.4.7",
|
||||
"date-fns": "^4.1.0",
|
||||
"exif-parser": "^0.1.12",
|
||||
"express": "^5.1.0",
|
||||
"express-list-endpoints": "^7.1.1",
|
||||
"express-rate-limit": "^8.2.1",
|
||||
@@ -53,6 +55,7 @@
|
||||
"passport-local": "^1.0.0",
|
||||
"pdfjs-dist": "^5.4.394",
|
||||
"pg": "^8.16.3",
|
||||
"piexifjs": "^1.0.6",
|
||||
"pino": "^10.1.0",
|
||||
"react": "^19.2.0",
|
||||
"react-dom": "^19.2.0",
|
||||
@@ -84,6 +87,7 @@
|
||||
"@types/passport-jwt": "^4.0.1",
|
||||
"@types/passport-local": "^1.0.38",
|
||||
"@types/pg": "^8.15.6",
|
||||
"@types/piexifjs": "^1.0.0",
|
||||
"@types/pino": "^7.0.4",
|
||||
"@types/react": "^19.2.7",
|
||||
"@types/react-dom": "^19.2.3",
|
||||
|
||||
@@ -115,6 +115,7 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
||||
valid_from DATE,
|
||||
valid_to DATE,
|
||||
store_address TEXT,
|
||||
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
||||
item_count INTEGER DEFAULT 0 NOT NULL,
|
||||
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
@@ -130,11 +131,13 @@ COMMENT ON COLUMN public.flyers.store_id IS 'Foreign key linking this flyer to a
|
||||
COMMENT ON COLUMN public.flyers.valid_from IS 'The start date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.valid_to IS 'The end date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.store_address IS 'The physical store address if it was successfully extracted from the flyer image.';
|
||||
COMMENT ON COLUMN public.flyers.status IS 'The processing status of the flyer, e.g., if it needs manual review.';
|
||||
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
||||
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
||||
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||
-- 7. The 'master_grocery_items' table. This is the master dictionary.
|
||||
CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
||||
master_grocery_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
|
||||
@@ -131,6 +131,7 @@ CREATE TABLE IF NOT EXISTS public.flyers (
|
||||
valid_from DATE,
|
||||
valid_to DATE,
|
||||
store_address TEXT,
|
||||
status TEXT DEFAULT 'processed' NOT NULL CHECK (status IN ('processed', 'needs_review', 'archived')),
|
||||
item_count INTEGER DEFAULT 0 NOT NULL,
|
||||
uploaded_by UUID REFERENCES public.users(user_id) ON DELETE SET NULL,
|
||||
created_at TIMESTAMPTZ DEFAULT now() NOT NULL,
|
||||
@@ -146,11 +147,13 @@ COMMENT ON COLUMN public.flyers.store_id IS 'Foreign key linking this flyer to a
|
||||
COMMENT ON COLUMN public.flyers.valid_from IS 'The start date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.valid_to IS 'The end date of the sale period for this flyer, extracted by the AI.';
|
||||
COMMENT ON COLUMN public.flyers.store_address IS 'The physical store address if it was successfully extracted from the flyer image.';
|
||||
COMMENT ON COLUMN public.flyers.status IS 'The processing status of the flyer, e.g., if it needs manual review.';
|
||||
COMMENT ON COLUMN public.flyers.item_count IS 'A cached count of the number of items in this flyer, maintained by a trigger.';
|
||||
COMMENT ON COLUMN public.flyers.uploaded_by IS 'The user who uploaded the flyer. Can be null for anonymous or system uploads.';
|
||||
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_created_at ON public.flyers (created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_valid_to_file_name ON public.flyers (valid_to DESC, file_name ASC);
|
||||
CREATE INDEX IF NOT EXISTS idx_flyers_status ON public.flyers(status);
|
||||
-- 7. The 'master_grocery_items' table. This is the master dictionary.
|
||||
CREATE TABLE IF NOT EXISTS public.master_grocery_items (
|
||||
master_grocery_item_id BIGINT PRIMARY KEY GENERATED ALWAYS AS IDENTITY,
|
||||
|
||||
360
src/App.test.tsx
360
src/App.test.tsx
@@ -20,6 +20,7 @@ import {
|
||||
mockUseUserData,
|
||||
mockUseFlyerItems,
|
||||
} from './tests/setup/mockHooks';
|
||||
import { useAppInitialization } from './hooks/useAppInitialization';
|
||||
|
||||
// Mock top-level components rendered by App's routes
|
||||
|
||||
@@ -52,6 +53,9 @@ vi.mock('./hooks/useFlyerItems', async () => {
|
||||
return { useFlyerItems: hooks.mockUseFlyerItems };
|
||||
});
|
||||
|
||||
vi.mock('./hooks/useAppInitialization');
|
||||
const mockedUseAppInitialization = vi.mocked(useAppInitialization);
|
||||
|
||||
vi.mock('./hooks/useAuth', async () => {
|
||||
const hooks = await import('./tests/setup/mockHooks');
|
||||
return { useAuth: hooks.mockUseAuth };
|
||||
@@ -122,7 +126,23 @@ vi.mock('./layouts/MainLayout', async () => {
|
||||
return { MainLayout: MockMainLayout };
|
||||
});
|
||||
|
||||
const mockedAiApiClient = vi.mocked(aiApiClient); // Mock aiApiClient
|
||||
vi.mock('./components/AppGuard', async () => {
|
||||
// We need to use the real useModal hook inside our mock AppGuard
|
||||
const { useModal } = await vi.importActual<typeof import('./hooks/useModal')>('./hooks/useModal');
|
||||
return {
|
||||
AppGuard: ({ children }: { children: React.ReactNode }) => {
|
||||
const { isModalOpen } = useModal();
|
||||
return (
|
||||
<div data-testid="app-guard-mock">
|
||||
{children}
|
||||
{isModalOpen('whatsNew') && <div data-testid="whats-new-modal-mock" />}
|
||||
</div>
|
||||
);
|
||||
},
|
||||
};
|
||||
});
|
||||
|
||||
const mockedAiApiClient = vi.mocked(aiApiClient);
|
||||
const mockedApiClient = vi.mocked(apiClient);
|
||||
|
||||
const mockFlyers: Flyer[] = [
|
||||
@@ -131,33 +151,6 @@ const mockFlyers: Flyer[] = [
|
||||
];
|
||||
|
||||
describe('App Component', () => {
|
||||
// Mock localStorage
|
||||
let storage: { [key: string]: string } = {};
|
||||
const localStorageMock = {
|
||||
getItem: vi.fn((key: string) => storage[key] || null),
|
||||
setItem: vi.fn((key: string, value: string) => {
|
||||
storage[key] = value;
|
||||
}),
|
||||
removeItem: vi.fn((key: string) => {
|
||||
delete storage[key];
|
||||
}),
|
||||
clear: vi.fn(() => {
|
||||
storage = {};
|
||||
}),
|
||||
};
|
||||
|
||||
// Mock matchMedia
|
||||
const matchMediaMock = vi.fn().mockImplementation((query) => ({
|
||||
matches: false, // Default to light mode
|
||||
media: query,
|
||||
onchange: null,
|
||||
addListener: vi.fn(), // deprecated
|
||||
removeListener: vi.fn(), // deprecated
|
||||
addEventListener: vi.fn(),
|
||||
removeEventListener: vi.fn(),
|
||||
dispatchEvent: vi.fn(),
|
||||
}));
|
||||
|
||||
beforeEach(() => {
|
||||
console.log('[TEST DEBUG] beforeEach: Clearing mocks and setting up defaults');
|
||||
vi.clearAllMocks();
|
||||
@@ -205,11 +198,9 @@ describe('App Component', () => {
|
||||
mockUseFlyerItems.mockReturnValue({
|
||||
flyerItems: [],
|
||||
isLoading: false,
|
||||
error: null,
|
||||
});
|
||||
// Clear local storage to prevent state from leaking between tests.
|
||||
localStorage.clear();
|
||||
Object.defineProperty(window, 'localStorage', { value: localStorageMock, configurable: true });
|
||||
Object.defineProperty(window, 'matchMedia', { value: matchMediaMock, configurable: true });
|
||||
mockedUseAppInitialization.mockReturnValue({ isDarkMode: false, unitSystem: 'imperial' });
|
||||
|
||||
// Default mocks for API calls
|
||||
// Use mockImplementation to create a new Response object for each call,
|
||||
@@ -261,6 +252,7 @@ describe('App Component', () => {
|
||||
|
||||
it('should render the main layout and header', async () => {
|
||||
// Simulate the auth hook finishing its initial check
|
||||
mockedUseAppInitialization.mockReturnValue({ isDarkMode: false, unitSystem: 'imperial' });
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: null,
|
||||
authStatus: 'SIGNED_OUT',
|
||||
@@ -272,6 +264,7 @@ describe('App Component', () => {
|
||||
|
||||
renderApp();
|
||||
await waitFor(() => {
|
||||
expect(screen.getByTestId('app-guard-mock')).toBeInTheDocument();
|
||||
expect(screen.getByTestId('header-mock')).toBeInTheDocument();
|
||||
// Check that the main layout and home page are rendered for the root path
|
||||
expect(screen.getByTestId('main-layout-mock')).toBeInTheDocument();
|
||||
@@ -364,193 +357,6 @@ describe('App Component', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('Theme and Unit System Synchronization', () => {
|
||||
it('should set dark mode based on user profile preferences', async () => {
|
||||
console.log(
|
||||
'[TEST DEBUG] Test Start: should set dark mode based on user profile preferences',
|
||||
);
|
||||
const profileWithDarkMode: UserProfile = createMockUserProfile({
|
||||
user: createMockUser({ user_id: 'user-1', email: 'dark@mode.com' }),
|
||||
role: 'user',
|
||||
points: 0,
|
||||
preferences: { darkMode: true },
|
||||
});
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: profileWithDarkMode,
|
||||
authStatus: 'AUTHENTICATED',
|
||||
isLoading: false,
|
||||
login: vi.fn(),
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
|
||||
console.log('[TEST DEBUG] Rendering App');
|
||||
renderApp();
|
||||
// The useEffect that sets the theme is asynchronous. We must wait for the update.
|
||||
await waitFor(() => {
|
||||
console.log(
|
||||
'[TEST DEBUG] Checking for dark class. Current classes:',
|
||||
document.documentElement.className,
|
||||
);
|
||||
expect(document.documentElement).toHaveClass('dark');
|
||||
});
|
||||
});
|
||||
|
||||
it('should set light mode based on user profile preferences', async () => {
|
||||
const profileWithLightMode: UserProfile = createMockUserProfile({
|
||||
user: createMockUser({ user_id: 'user-1', email: 'light@mode.com' }),
|
||||
role: 'user',
|
||||
points: 0,
|
||||
preferences: { darkMode: false },
|
||||
});
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: profileWithLightMode,
|
||||
authStatus: 'AUTHENTICATED',
|
||||
isLoading: false,
|
||||
login: vi.fn(),
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
renderApp();
|
||||
await waitFor(() => {
|
||||
expect(document.documentElement).not.toHaveClass('dark');
|
||||
});
|
||||
});
|
||||
|
||||
it('should set dark mode based on localStorage if profile has no preference', async () => {
|
||||
localStorageMock.setItem('darkMode', 'true');
|
||||
renderApp();
|
||||
await waitFor(() => {
|
||||
expect(document.documentElement).toHaveClass('dark');
|
||||
});
|
||||
});
|
||||
|
||||
it('should set dark mode based on system preference if no other setting exists', async () => {
|
||||
matchMediaMock.mockImplementationOnce((query) => ({ matches: true, media: query }));
|
||||
renderApp();
|
||||
await waitFor(() => {
|
||||
expect(document.documentElement).toHaveClass('dark');
|
||||
});
|
||||
});
|
||||
|
||||
it('should set unit system based on user profile preferences', async () => {
|
||||
const profileWithMetric: UserProfile = createMockUserProfile({
|
||||
user: createMockUser({ user_id: 'user-1', email: 'metric@user.com' }),
|
||||
role: 'user',
|
||||
points: 0,
|
||||
preferences: { unitSystem: 'metric' },
|
||||
});
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: profileWithMetric,
|
||||
authStatus: 'AUTHENTICATED',
|
||||
isLoading: false,
|
||||
login: vi.fn(),
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
|
||||
renderApp();
|
||||
// The unit system is passed as a prop to Header, which is mocked.
|
||||
// We can't directly see the result in the DOM easily, so we trust the state is set.
|
||||
// A more integrated test would be needed to verify the Header receives the prop.
|
||||
// For now, this test ensures the useEffect logic runs without crashing.
|
||||
await waitFor(() => {
|
||||
expect(screen.getByTestId('header-mock')).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('OAuth Token Handling', () => {
|
||||
it('should call login when a googleAuthToken is in the URL', async () => {
|
||||
console.log(
|
||||
'[TEST DEBUG] Test Start: should call login when a googleAuthToken is in the URL',
|
||||
);
|
||||
const mockLogin = vi.fn().mockResolvedValue(undefined);
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: null,
|
||||
authStatus: 'SIGNED_OUT',
|
||||
isLoading: false,
|
||||
login: mockLogin,
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
|
||||
console.log('[TEST DEBUG] Rendering App with googleAuthToken');
|
||||
renderApp(['/?googleAuthToken=test-google-token']);
|
||||
|
||||
await waitFor(() => {
|
||||
console.log('[TEST DEBUG] Checking mockLogin calls:', mockLogin.mock.calls);
|
||||
expect(mockLogin).toHaveBeenCalledWith('test-google-token');
|
||||
});
|
||||
});
|
||||
|
||||
it('should call login when a githubAuthToken is in the URL', async () => {
|
||||
console.log(
|
||||
'[TEST DEBUG] Test Start: should call login when a githubAuthToken is in the URL',
|
||||
);
|
||||
const mockLogin = vi.fn().mockResolvedValue(undefined);
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: null,
|
||||
authStatus: 'SIGNED_OUT',
|
||||
isLoading: false,
|
||||
login: mockLogin,
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
|
||||
console.log('[TEST DEBUG] Rendering App with githubAuthToken');
|
||||
renderApp(['/?githubAuthToken=test-github-token']);
|
||||
|
||||
await waitFor(() => {
|
||||
console.log('[TEST DEBUG] Checking mockLogin calls:', mockLogin.mock.calls);
|
||||
expect(mockLogin).toHaveBeenCalledWith('test-github-token');
|
||||
});
|
||||
});
|
||||
|
||||
it('should log an error if login with a GitHub token fails', async () => {
|
||||
console.log(
|
||||
'[TEST DEBUG] Test Start: should log an error if login with a GitHub token fails',
|
||||
);
|
||||
const mockLogin = vi.fn().mockRejectedValue(new Error('GitHub login failed'));
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: null,
|
||||
authStatus: 'SIGNED_OUT',
|
||||
isLoading: false,
|
||||
login: mockLogin,
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
|
||||
console.log('[TEST DEBUG] Rendering App with githubAuthToken');
|
||||
renderApp(['/?githubAuthToken=bad-token']);
|
||||
|
||||
await waitFor(() => {
|
||||
console.log('[TEST DEBUG] Checking mockLogin calls:', mockLogin.mock.calls);
|
||||
expect(mockLogin).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
it('should log an error if login with a token fails', async () => {
|
||||
console.log('[TEST DEBUG] Test Start: should log an error if login with a token fails');
|
||||
const mockLogin = vi.fn().mockRejectedValue(new Error('Token login failed'));
|
||||
mockUseAuth.mockReturnValue({
|
||||
userProfile: null,
|
||||
authStatus: 'SIGNED_OUT',
|
||||
isLoading: false,
|
||||
login: mockLogin,
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
|
||||
console.log('[TEST DEBUG] Rendering App with googleAuthToken');
|
||||
renderApp(['/?googleAuthToken=bad-token']);
|
||||
await waitFor(() => {
|
||||
console.log('[TEST DEBUG] Checking mockLogin calls:', mockLogin.mock.calls);
|
||||
expect(mockLogin).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('Flyer Selection from URL', () => {
|
||||
it('should select a flyer when flyerId is present in the URL', async () => {
|
||||
renderApp(['/flyers/2']);
|
||||
@@ -583,23 +389,9 @@ describe('App Component', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('Version and "What\'s New" Modal', () => {
|
||||
it('should show the "What\'s New" modal if the app version is new', async () => {
|
||||
// Mock the config module for this specific test
|
||||
vi.mock('./config', () => ({
|
||||
default: {
|
||||
app: { version: '20250101-1200:abc1234:1.0.1', commitMessage: 'New feature!', commitUrl: '#' },
|
||||
google: { mapsEmbedApiKey: 'mock-key' },
|
||||
},
|
||||
}));
|
||||
localStorageMock.setItem('lastSeenVersion', '20250101-1200:abc1234:1.0.0');
|
||||
renderApp();
|
||||
await expect(screen.findByTestId('whats-new-modal-mock')).resolves.toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Modal Interactions', () => {
|
||||
it('should open and close the ProfileManager modal', async () => {
|
||||
console.log('[TEST DEBUG] Test Start: should open and close the ProfileManager modal');
|
||||
renderApp();
|
||||
expect(screen.queryByTestId('profile-manager-mock')).not.toBeInTheDocument();
|
||||
|
||||
@@ -607,11 +399,13 @@ describe('App Component', () => {
|
||||
fireEvent.click(screen.getByText('Open Profile'));
|
||||
expect(await screen.findByTestId('profile-manager-mock')).toBeInTheDocument();
|
||||
|
||||
console.log('[TEST DEBUG] ProfileManager modal opened. Now closing...');
|
||||
// Close modal
|
||||
fireEvent.click(screen.getByText('Close Profile'));
|
||||
await waitFor(() => {
|
||||
expect(screen.queryByTestId('profile-manager-mock')).not.toBeInTheDocument();
|
||||
});
|
||||
console.log('[TEST DEBUG] ProfileManager modal closed.');
|
||||
});
|
||||
|
||||
it('should open and close the VoiceAssistant modal for authenticated users', async () => {
|
||||
@@ -636,7 +430,7 @@ describe('App Component', () => {
|
||||
fireEvent.click(screen.getByText('Open Voice Assistant'));
|
||||
|
||||
console.log('[TEST DEBUG] Waiting for voice-assistant-mock');
|
||||
expect(await screen.findByTestId('voice-assistant-mock')).toBeInTheDocument();
|
||||
expect(await screen.findByTestId('voice-assistant-mock', {}, { timeout: 3000 })).toBeInTheDocument();
|
||||
|
||||
// Close modal
|
||||
fireEvent.click(screen.getByText('Close Voice Assistant'));
|
||||
@@ -735,64 +529,6 @@ describe('App Component', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe("Version Display and What's New", () => {
|
||||
beforeEach(() => {
|
||||
// Also mock the config module to reflect this change
|
||||
vi.mock('./config', () => ({
|
||||
default: {
|
||||
app: {
|
||||
version: '20250101-1200:abc1234:2.0.0',
|
||||
commitMessage: 'A new version!',
|
||||
commitUrl: 'http://example.com/commit/2.0.0',
|
||||
},
|
||||
google: { mapsEmbedApiKey: 'mock-key' },
|
||||
},
|
||||
}));
|
||||
});
|
||||
|
||||
it('should display the version number and commit link', () => {
|
||||
renderApp();
|
||||
const versionLink = screen.getByText(`Version: 20250101-1200:abc1234:2.0.0`);
|
||||
expect(versionLink).toBeInTheDocument();
|
||||
expect(versionLink).toHaveAttribute('href', 'http://example.com/commit/2.0.0');
|
||||
});
|
||||
|
||||
it('should open the "What\'s New" modal when the question mark icon is clicked', async () => {
|
||||
// Pre-set the localStorage to prevent the modal from opening automatically
|
||||
localStorageMock.setItem('lastSeenVersion', '20250101-1200:abc1234:2.0.0');
|
||||
|
||||
renderApp();
|
||||
expect(screen.queryByTestId('whats-new-modal-mock')).not.toBeInTheDocument();
|
||||
|
||||
const openButton = await screen.findByTitle("Show what's new in this version");
|
||||
fireEvent.click(openButton);
|
||||
|
||||
expect(await screen.findByTestId('whats-new-modal-mock')).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Dynamic Toaster Styles', () => {
|
||||
it('should render the correct CSS variables for toast styling in light mode', async () => {
|
||||
renderApp();
|
||||
await waitFor(() => {
|
||||
const styleTag = document.querySelector('style');
|
||||
expect(styleTag).not.toBeNull();
|
||||
expect(styleTag!.innerHTML).toContain('--toast-bg: #FFFFFF');
|
||||
expect(styleTag!.innerHTML).toContain('--toast-color: #1F2937');
|
||||
});
|
||||
});
|
||||
|
||||
it('should render the correct CSS variables for toast styling in dark mode', async () => {
|
||||
localStorageMock.setItem('darkMode', 'true');
|
||||
renderApp();
|
||||
await waitFor(() => {
|
||||
const styleTag = document.querySelector('style');
|
||||
expect(styleTag).not.toBeNull();
|
||||
expect(styleTag!.innerHTML).toContain('--toast-bg: #4B5563');
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('Profile and Login Handlers', () => {
|
||||
it('should call updateProfile when handleProfileUpdate is triggered', async () => {
|
||||
console.log(
|
||||
@@ -841,12 +577,19 @@ describe('App Component', () => {
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
// Mock the login function to simulate a successful login. Signature: (token, profile)
|
||||
const mockLoginSuccess = vi.fn(async (_token: string, _profile?: UserProfile) => {
|
||||
// Simulate fetching profile after login
|
||||
const profileResponse = await mockedApiClient.getAuthenticatedUserProfile();
|
||||
const userProfileData: UserProfile = await profileResponse.json();
|
||||
mockUseAuth.mockReturnValue({ ...mockUseAuth(), userProfile: userProfileData, authStatus: 'AUTHENTICATED' });
|
||||
});
|
||||
|
||||
console.log('[TEST DEBUG] Rendering App');
|
||||
renderApp();
|
||||
console.log('[TEST DEBUG] Opening Profile');
|
||||
fireEvent.click(screen.getByText('Open Profile'));
|
||||
const loginButton = await screen.findByText('Login');
|
||||
const loginButton = await screen.findByRole('button', { name: 'Login' });
|
||||
console.log('[TEST DEBUG] Clicking Login');
|
||||
fireEvent.click(loginButton);
|
||||
|
||||
@@ -857,4 +600,33 @@ describe('App Component', () => {
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe("Version Display and What's New", () => {
|
||||
beforeEach(() => {
|
||||
vi.mock('./config', () => ({
|
||||
default: {
|
||||
app: {
|
||||
version: '2.0.0',
|
||||
commitMessage: 'A new version!',
|
||||
commitUrl: 'http://example.com/commit/2.0.0',
|
||||
},
|
||||
},
|
||||
}));
|
||||
});
|
||||
|
||||
it('should display the version number and commit link', () => {
|
||||
renderApp();
|
||||
const versionLink = screen.getByText(`Version: 2.0.0`);
|
||||
expect(versionLink).toBeInTheDocument();
|
||||
expect(versionLink).toHaveAttribute('href', 'http://example.com/commit/2.0.0');
|
||||
});
|
||||
|
||||
it('should open the "What\'s New" modal when the question mark icon is clicked', async () => {
|
||||
renderApp();
|
||||
const openButton = await screen.findByTitle("Show what's new in this version");
|
||||
fireEvent.click(openButton);
|
||||
// The mock AppGuard now renders the modal when it's open
|
||||
expect(await screen.findByTestId('whats-new-modal-mock')).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
132
src/App.tsx
132
src/App.tsx
@@ -1,9 +1,9 @@
|
||||
// src/App.tsx
|
||||
import React, { useState, useCallback, useEffect } from 'react';
|
||||
import { Routes, Route, useParams, useLocation, useNavigate } from 'react-router-dom';
|
||||
import { Toaster } from 'react-hot-toast';
|
||||
import { Routes, Route, useParams } from 'react-router-dom';
|
||||
import { QueryClient, QueryClientProvider } from '@tanstack/react-query';
|
||||
import * as pdfjsLib from 'pdfjs-dist';
|
||||
import { Footer } from './components/Footer'; // Assuming this is where your Footer component will live
|
||||
import { Footer } from './components/Footer';
|
||||
import { Header } from './components/Header';
|
||||
import { logger } from './services/logger.client';
|
||||
import type { Flyer, Profile, UserProfile } from './types';
|
||||
@@ -13,18 +13,20 @@ import { AdminPage } from './pages/admin/AdminPage';
|
||||
import { AdminRoute } from './components/AdminRoute';
|
||||
import { CorrectionsPage } from './pages/admin/CorrectionsPage';
|
||||
import { AdminStatsPage } from './pages/admin/AdminStatsPage';
|
||||
import { FlyerReviewPage } from './pages/admin/FlyerReviewPage';
|
||||
import { ResetPasswordPage } from './pages/ResetPasswordPage';
|
||||
import { VoiceLabPage } from './pages/VoiceLabPage';
|
||||
import { WhatsNewModal } from './components/WhatsNewModal';
|
||||
import { FlyerCorrectionTool } from './components/FlyerCorrectionTool';
|
||||
import { QuestionMarkCircleIcon } from './components/icons/QuestionMarkCircleIcon';
|
||||
import { useAuth } from './hooks/useAuth';
|
||||
import { useFlyers } from './hooks/useFlyers'; // Assuming useFlyers fetches all flyers
|
||||
import { useFlyerItems } from './hooks/useFlyerItems'; // Import the new hook for flyer items
|
||||
import { useFlyers } from './hooks/useFlyers';
|
||||
import { useFlyerItems } from './hooks/useFlyerItems';
|
||||
import { useModal } from './hooks/useModal';
|
||||
import { MainLayout } from './layouts/MainLayout';
|
||||
import config from './config';
|
||||
import { HomePage } from './pages/HomePage';
|
||||
import { AppGuard } from './components/AppGuard';
|
||||
import { useAppInitialization } from './hooks/useAppInitialization';
|
||||
|
||||
// pdf.js worker configuration
|
||||
// This is crucial for allowing pdf.js to process PDFs in a separate thread, preventing the UI from freezing.
|
||||
@@ -35,15 +37,20 @@ pdfjsLib.GlobalWorkerOptions.workerSrc = new URL(
|
||||
import.meta.url,
|
||||
).toString();
|
||||
|
||||
// Create a client
|
||||
const queryClient = new QueryClient();
|
||||
|
||||
function App() {
|
||||
const { userProfile, authStatus, login, logout, updateProfile } = useAuth();
|
||||
const { flyers } = useFlyers();
|
||||
const [selectedFlyer, setSelectedFlyer] = useState<Flyer | null>(null);
|
||||
const { openModal, closeModal, isModalOpen } = useModal();
|
||||
const location = useLocation();
|
||||
const navigate = useNavigate();
|
||||
const params = useParams<{ flyerId?: string }>();
|
||||
|
||||
// This hook now handles initialization effects (OAuth, version check, theme)
|
||||
// and returns the theme/unit state needed by other components.
|
||||
const { isDarkMode, unitSystem } = useAppInitialization();
|
||||
|
||||
// Debugging: Log renders to identify infinite loops
|
||||
useEffect(() => {
|
||||
if (process.env.NODE_ENV === 'test') {
|
||||
@@ -53,14 +60,11 @@ function App() {
|
||||
paramsFlyerId: params?.flyerId, // This was a duplicate, fixed.
|
||||
authStatus,
|
||||
profileId: userProfile?.user.user_id,
|
||||
locationSearch: location.search,
|
||||
});
|
||||
}
|
||||
});
|
||||
|
||||
const [isDarkMode, setIsDarkMode] = useState(false);
|
||||
const { flyerItems } = useFlyerItems(selectedFlyer);
|
||||
const [unitSystem, setUnitSystem] = useState<'metric' | 'imperial'>('imperial');
|
||||
|
||||
// Define modal handlers with useCallback at the top level to avoid Rules of Hooks violations
|
||||
const handleOpenProfile = useCallback(() => openModal('profile'), [openModal]);
|
||||
@@ -105,37 +109,6 @@ function App() {
|
||||
|
||||
// --- State Synchronization and Error Handling ---
|
||||
|
||||
// Effect to set initial theme based on user profile, local storage, or system preference
|
||||
useEffect(() => {
|
||||
if (process.env.NODE_ENV === 'test')
|
||||
console.log('[App] Effect: Theme Update', { profileId: userProfile?.user.user_id });
|
||||
if (userProfile && userProfile.preferences?.darkMode !== undefined) {
|
||||
// Preference from DB
|
||||
const dbDarkMode = userProfile.preferences.darkMode;
|
||||
setIsDarkMode(dbDarkMode);
|
||||
document.documentElement.classList.toggle('dark', dbDarkMode);
|
||||
} else {
|
||||
// Fallback to local storage or system preference
|
||||
const savedMode = localStorage.getItem('darkMode');
|
||||
const prefersDark = window.matchMedia('(prefers-color-scheme: dark)').matches;
|
||||
const initialDarkMode = savedMode !== null ? savedMode === 'true' : prefersDark;
|
||||
setIsDarkMode(initialDarkMode);
|
||||
document.documentElement.classList.toggle('dark', initialDarkMode);
|
||||
}
|
||||
}, [userProfile?.preferences?.darkMode, userProfile?.user.user_id]);
|
||||
|
||||
// Effect to set initial unit system based on user profile or local storage
|
||||
useEffect(() => {
|
||||
if (userProfile && userProfile.preferences?.unitSystem) {
|
||||
setUnitSystem(userProfile.preferences.unitSystem);
|
||||
} else {
|
||||
const savedSystem = localStorage.getItem('unitSystem') as 'metric' | 'imperial' | null;
|
||||
if (savedSystem) {
|
||||
setUnitSystem(savedSystem);
|
||||
}
|
||||
}
|
||||
}, [userProfile?.preferences?.unitSystem, userProfile?.user.user_id]);
|
||||
|
||||
// This is the login handler that will be passed to the ProfileManager component.
|
||||
const handleLoginSuccess = useCallback(
|
||||
async (userProfile: UserProfile, token: string, _rememberMe: boolean) => {
|
||||
@@ -153,36 +126,6 @@ function App() {
|
||||
[login],
|
||||
);
|
||||
|
||||
// Effect to handle the token from Google OAuth redirect
|
||||
useEffect(() => {
|
||||
const urlParams = new URLSearchParams(location.search);
|
||||
const googleToken = urlParams.get('googleAuthToken');
|
||||
|
||||
if (googleToken) {
|
||||
logger.info('Received Google Auth token from URL. Authenticating...');
|
||||
// The login flow is now handled by the useAuth hook. We just need to trigger it.
|
||||
// We pass only the token; the AuthProvider will fetch the user profile.
|
||||
login(googleToken).catch((err) =>
|
||||
logger.error('Failed to log in with Google token', { error: err }),
|
||||
);
|
||||
// Clean the token from the URL
|
||||
navigate(location.pathname, { replace: true });
|
||||
}
|
||||
|
||||
const githubToken = urlParams.get('githubAuthToken');
|
||||
if (githubToken) {
|
||||
logger.info('Received GitHub Auth token from URL. Authenticating...');
|
||||
login(githubToken).catch((err) => {
|
||||
logger.error('Failed to log in with GitHub token', { error: err });
|
||||
// Optionally, redirect to a page with an error message
|
||||
// navigate('/login?error=github_auth_failed');
|
||||
});
|
||||
|
||||
// Clean the token from the URL
|
||||
navigate(location.pathname, { replace: true });
|
||||
}
|
||||
}, [login, location.search, navigate, location.pathname]);
|
||||
|
||||
const handleFlyerSelect = useCallback(async (flyer: Flyer) => {
|
||||
setSelectedFlyer(flyer);
|
||||
}, []);
|
||||
@@ -210,31 +153,10 @@ function App() {
|
||||
// Read the application version injected at build time.
|
||||
// This will only be available in the production build, not during local development.
|
||||
const appVersion = config.app.version;
|
||||
const commitMessage = config.app.commitMessage;
|
||||
useEffect(() => {
|
||||
if (appVersion) {
|
||||
logger.info(`Application version: ${appVersion}`);
|
||||
const lastSeenVersion = localStorage.getItem('lastSeenVersion');
|
||||
// If the current version is new, show the "What's New" modal.
|
||||
if (appVersion !== lastSeenVersion) {
|
||||
openModal('whatsNew');
|
||||
localStorage.setItem('lastSeenVersion', appVersion);
|
||||
}
|
||||
}
|
||||
}, [appVersion]);
|
||||
|
||||
return (
|
||||
<div className="bg-gray-100 dark:bg-gray-950 min-h-screen font-sans text-gray-800 dark:text-gray-200">
|
||||
{/* Toaster component for displaying notifications. It's placed at the top level. */}
|
||||
<Toaster position="top-center" reverseOrder={false} />
|
||||
{/* Add CSS variables for toast theming based on dark mode */}
|
||||
<style>{`
|
||||
:root {
|
||||
--toast-bg: ${isDarkMode ? '#4B5563' : '#FFFFFF'};
|
||||
--toast-color: ${isDarkMode ? '#F9FAFB' : '#1F2937'};
|
||||
}
|
||||
`}</style>
|
||||
|
||||
// AppGuard now handles the main page wrapper, theme styles, and "What's New" modal
|
||||
<AppGuard>
|
||||
<Header
|
||||
isDarkMode={isDarkMode}
|
||||
unitSystem={unitSystem}
|
||||
@@ -261,15 +183,6 @@ function App() {
|
||||
/>
|
||||
)}
|
||||
|
||||
{appVersion && commitMessage && (
|
||||
<WhatsNewModal
|
||||
isOpen={isModalOpen('whatsNew')}
|
||||
onClose={handleCloseWhatsNew}
|
||||
version={appVersion}
|
||||
commitMessage={commitMessage}
|
||||
/>
|
||||
)}
|
||||
|
||||
{selectedFlyer && (
|
||||
<FlyerCorrectionTool
|
||||
isOpen={isModalOpen('correctionTool')}
|
||||
@@ -316,6 +229,7 @@ function App() {
|
||||
<Route path="/admin" element={<AdminPage />} />
|
||||
<Route path="/admin/corrections" element={<CorrectionsPage />} />
|
||||
<Route path="/admin/stats" element={<AdminStatsPage />} />
|
||||
<Route path="/admin/flyer-review" element={<FlyerReviewPage />} />
|
||||
<Route path="/admin/voice-lab" element={<VoiceLabPage />} />
|
||||
</Route>
|
||||
<Route path="/reset-password/:token" element={<ResetPasswordPage />} />
|
||||
@@ -341,8 +255,14 @@ function App() {
|
||||
)}
|
||||
|
||||
<Footer />
|
||||
</div>
|
||||
</AppGuard>
|
||||
);
|
||||
}
|
||||
|
||||
export default App;
|
||||
const WrappedApp = () => (
|
||||
<QueryClientProvider client={queryClient}>
|
||||
<App />
|
||||
</QueryClientProvider>
|
||||
);
|
||||
|
||||
export default WrappedApp;
|
||||
|
||||
@@ -5,7 +5,7 @@ import { describe, it, expect, vi } from 'vitest';
|
||||
import { AnonymousUserBanner } from './AnonymousUserBanner';
|
||||
|
||||
// Mock the icon to ensure it is rendered correctly
|
||||
vi.mock('../../../components/icons/InformationCircleIcon', () => ({
|
||||
vi.mock('./icons/InformationCircleIcon', () => ({
|
||||
InformationCircleIcon: (props: React.SVGProps<SVGSVGElement>) => (
|
||||
<svg data-testid="info-icon" {...props} />
|
||||
),
|
||||
@@ -1,6 +1,6 @@
|
||||
// src/pages/admin/components/AnonymousUserBanner.tsx
|
||||
// src/components/AnonymousUserBanner.tsx
|
||||
import React from 'react';
|
||||
import { InformationCircleIcon } from '../../../components/icons/InformationCircleIcon';
|
||||
import { InformationCircleIcon } from './icons/InformationCircleIcon';
|
||||
|
||||
interface AnonymousUserBannerProps {
|
||||
/**
|
||||
93
src/components/AppGuard.test.tsx
Normal file
93
src/components/AppGuard.test.tsx
Normal file
@@ -0,0 +1,93 @@
|
||||
// src/components/AppGuard.test.tsx
|
||||
import React from 'react';
|
||||
import { render, screen, waitFor } from '@testing-library/react';
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { AppGuard } from './AppGuard';
|
||||
import { useAppInitialization } from '../hooks/useAppInitialization';
|
||||
import { useModal } from '../hooks/useModal';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('../hooks/useAppInitialization');
|
||||
vi.mock('../hooks/useModal');
|
||||
vi.mock('./WhatsNewModal', () => ({
|
||||
WhatsNewModal: ({ isOpen }: { isOpen: boolean }) =>
|
||||
isOpen ? <div data-testid="whats-new-modal-mock" /> : null,
|
||||
}));
|
||||
vi.mock('../config', () => ({
|
||||
default: {
|
||||
app: { version: '1.0.0', commitMessage: 'Test commit' },
|
||||
},
|
||||
}));
|
||||
|
||||
const mockedUseAppInitialization = vi.mocked(useAppInitialization);
|
||||
const mockedUseModal = vi.mocked(useModal);
|
||||
|
||||
describe('AppGuard', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Default mocks
|
||||
mockedUseAppInitialization.mockReturnValue({
|
||||
isDarkMode: false,
|
||||
unitSystem: 'imperial',
|
||||
});
|
||||
mockedUseModal.mockReturnValue({
|
||||
isModalOpen: vi.fn().mockReturnValue(false),
|
||||
openModal: vi.fn(),
|
||||
closeModal: vi.fn(),
|
||||
});
|
||||
});
|
||||
|
||||
it('should render children', () => {
|
||||
render(
|
||||
<AppGuard>
|
||||
<div>Child Content</div>
|
||||
</AppGuard>,
|
||||
);
|
||||
expect(screen.getByText('Child Content')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should render WhatsNewModal when it is open', () => {
|
||||
mockedUseModal.mockReturnValue({
|
||||
...mockedUseModal(),
|
||||
isModalOpen: (modalId) => modalId === 'whatsNew',
|
||||
});
|
||||
render(
|
||||
<AppGuard>
|
||||
<div>Child</div>
|
||||
</AppGuard>,
|
||||
);
|
||||
expect(screen.getByTestId('whats-new-modal-mock')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should set dark mode styles for toaster', async () => {
|
||||
mockedUseAppInitialization.mockReturnValue({
|
||||
isDarkMode: true,
|
||||
unitSystem: 'imperial',
|
||||
});
|
||||
render(
|
||||
<AppGuard>
|
||||
<div>Child</div>
|
||||
</AppGuard>,
|
||||
);
|
||||
await waitFor(() => {
|
||||
const styleTag = document.querySelector('style');
|
||||
expect(styleTag).not.toBeNull();
|
||||
expect(styleTag!.innerHTML).toContain('--toast-bg: #4B5563');
|
||||
expect(styleTag!.innerHTML).toContain('--toast-color: #F9FAFB');
|
||||
});
|
||||
});
|
||||
|
||||
it('should set light mode styles for toaster', async () => {
|
||||
render(
|
||||
<AppGuard>
|
||||
<div>Child</div>
|
||||
</AppGuard>,
|
||||
);
|
||||
await waitFor(() => {
|
||||
const styleTag = document.querySelector('style');
|
||||
expect(styleTag).not.toBeNull();
|
||||
expect(styleTag!.innerHTML).toContain('--toast-bg: #FFFFFF');
|
||||
expect(styleTag!.innerHTML).toContain('--toast-color: #1F2937');
|
||||
});
|
||||
});
|
||||
});
|
||||
47
src/components/AppGuard.tsx
Normal file
47
src/components/AppGuard.tsx
Normal file
@@ -0,0 +1,47 @@
|
||||
// src/components/AppGuard.tsx
|
||||
import React, { useCallback } from 'react';
|
||||
import { Toaster } from 'react-hot-toast';
|
||||
import { useAppInitialization } from '../hooks/useAppInitialization';
|
||||
import { useModal } from '../hooks/useModal';
|
||||
import { WhatsNewModal } from './WhatsNewModal';
|
||||
import config from '../config';
|
||||
|
||||
interface AppGuardProps {
|
||||
children: React.ReactNode;
|
||||
}
|
||||
|
||||
export const AppGuard: React.FC<AppGuardProps> = ({ children }) => {
|
||||
// This hook handles OAuth tokens, version checks, and returns theme state.
|
||||
const { isDarkMode } = useAppInitialization();
|
||||
const { isModalOpen, closeModal } = useModal();
|
||||
|
||||
const handleCloseWhatsNew = useCallback(() => closeModal('whatsNew'), [closeModal]);
|
||||
|
||||
const appVersion = config.app.version;
|
||||
const commitMessage = config.app.commitMessage;
|
||||
|
||||
return (
|
||||
<div className="bg-gray-100 dark:bg-gray-950 min-h-screen font-sans text-gray-800 dark:text-gray-200">
|
||||
{/* Toaster component for displaying notifications. It's placed at the top level. */}
|
||||
<Toaster position="top-center" reverseOrder={false} />
|
||||
{/* Add CSS variables for toast theming based on dark mode */}
|
||||
<style>{`
|
||||
:root {
|
||||
--toast-bg: ${isDarkMode ? '#4B5563' : '#FFFFFF'};
|
||||
--toast-color: ${isDarkMode ? '#F9FAFB' : '#1F2937'};
|
||||
}
|
||||
`}</style>
|
||||
|
||||
{appVersion && commitMessage && (
|
||||
<WhatsNewModal
|
||||
isOpen={isModalOpen('whatsNew')}
|
||||
onClose={handleCloseWhatsNew}
|
||||
version={appVersion}
|
||||
commitMessage={commitMessage}
|
||||
/>
|
||||
)}
|
||||
|
||||
{children}
|
||||
</div>
|
||||
);
|
||||
};
|
||||
@@ -1,7 +1,7 @@
|
||||
// src/pages/admin/components/PasswordInput.tsx
|
||||
// src/components/PasswordInput.tsx
|
||||
import React, { useState } from 'react';
|
||||
import { EyeIcon } from '../../../components/icons/EyeIcon';
|
||||
import { EyeSlashIcon } from '../../../components/icons/EyeSlashIcon';
|
||||
import { EyeIcon } from './icons/EyeIcon';
|
||||
import { EyeSlashIcon } from './icons/EyeSlashIcon';
|
||||
import { PasswordStrengthIndicator } from './PasswordStrengthIndicator';
|
||||
|
||||
/**
|
||||
@@ -1,4 +1,5 @@
|
||||
// src/pages/admin/components/PasswordStrengthIndicator.tsx
|
||||
// src/components/PasswordStrengthIndicator.tsx
|
||||
import React from 'react';
|
||||
import zxcvbn from 'zxcvbn';
|
||||
|
||||
18
src/components/icons/DocumentMagnifyingGlassIcon.tsx
Normal file
18
src/components/icons/DocumentMagnifyingGlassIcon.tsx
Normal file
@@ -0,0 +1,18 @@
|
||||
import React from 'react';
|
||||
|
||||
export const DocumentMagnifyingGlassIcon: React.FC<React.SVGProps<SVGSVGElement>> = (props) => (
|
||||
<svg
|
||||
xmlns="http://www.w3.org/2000/svg"
|
||||
fill="none"
|
||||
viewBox="0 0 24 24"
|
||||
strokeWidth={1.5}
|
||||
stroke="currentColor"
|
||||
{...props}
|
||||
>
|
||||
<path
|
||||
strokeLinecap="round"
|
||||
strokeLinejoin="round"
|
||||
d="M19.5 14.25v-2.625a3.375 3.375 0 0 0-3.375-3.375h-1.5A1.125 1.125 0 0 1 13.5 7.125v-1.5a3.375 3.375 0 0 0-3.375-3.375H8.25m5.231 13.481L15 17.25m-4.5 4.5L6.75 21.75m0 0L2.25 17.25m4.5 4.5v-4.5m13.5-3V9A2.25 2.25 0 0 0 16.5 6.75h-9A2.25 2.25 0 0 0 5.25 9v9.75m14.25-10.5a2.25 2.25 0 0 0-2.25-2.25H5.25a2.25 2.25 0 0 0-2.25 2.25v10.5a2.25 2.25 0 0 0 2.25 225h5.25"
|
||||
/>
|
||||
</svg>
|
||||
);
|
||||
@@ -38,8 +38,26 @@ vi.mock('recharts', () => ({
|
||||
),
|
||||
CartesianGrid: () => <div data-testid="cartesian-grid" />,
|
||||
XAxis: () => <div data-testid="x-axis" />,
|
||||
YAxis: () => <div data-testid="y-axis" />,
|
||||
Tooltip: () => <div data-testid="tooltip" />,
|
||||
YAxis: ({ tickFormatter, domain }: any) => {
|
||||
// Execute functions for coverage
|
||||
if (typeof tickFormatter === 'function') {
|
||||
tickFormatter(1000);
|
||||
}
|
||||
if (Array.isArray(domain)) {
|
||||
domain.forEach((d) => {
|
||||
if (typeof d === 'function') d(100);
|
||||
});
|
||||
}
|
||||
return <div data-testid="y-axis" />;
|
||||
},
|
||||
Tooltip: ({ formatter }: any) => {
|
||||
// Execute formatter for coverage
|
||||
if (typeof formatter === 'function') {
|
||||
formatter(1000);
|
||||
formatter(undefined);
|
||||
}
|
||||
return <div data-testid="tooltip" />;
|
||||
},
|
||||
Legend: () => <div data-testid="legend" />,
|
||||
// Fix: Use dataKey if name is not explicitly provided, as the component relies on dataKey
|
||||
Line: ({ name, dataKey }: { name?: string; dataKey?: string }) => (
|
||||
@@ -301,4 +319,66 @@ describe('PriceHistoryChart', () => {
|
||||
expect(chartData).toHaveLength(2);
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle malformed data points and unmatched items gracefully', async () => {
|
||||
const malformedData: any[] = [
|
||||
{ master_item_id: null, summary_date: '2024-10-01', avg_price_in_cents: 100 }, // Missing ID
|
||||
{ master_item_id: 1, summary_date: null, avg_price_in_cents: 100 }, // Missing date
|
||||
{ master_item_id: 1, summary_date: '2024-10-01', avg_price_in_cents: null }, // Missing price
|
||||
{ master_item_id: 999, summary_date: '2024-10-01', avg_price_in_cents: 100 }, // ID not in watchlist
|
||||
];
|
||||
vi.mocked(apiClient.fetchHistoricalPriceData).mockResolvedValue(
|
||||
new Response(JSON.stringify(malformedData)),
|
||||
);
|
||||
render(<PriceHistoryChart />);
|
||||
|
||||
await waitFor(() => {
|
||||
// Should show "Not enough historical data" because all points are invalid or filtered
|
||||
expect(
|
||||
screen.getByText(
|
||||
'Not enough historical data for your watched items. Process more flyers to build a trend.',
|
||||
),
|
||||
).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
|
||||
it('should ignore higher prices for the same day', async () => {
|
||||
const dataWithHigherPrice: HistoricalPriceDataPoint[] = [
|
||||
createMockHistoricalPriceDataPoint({
|
||||
master_item_id: 1,
|
||||
summary_date: '2024-10-01',
|
||||
avg_price_in_cents: 100,
|
||||
}),
|
||||
createMockHistoricalPriceDataPoint({
|
||||
master_item_id: 1,
|
||||
summary_date: '2024-10-01',
|
||||
avg_price_in_cents: 150, // Higher price should be ignored
|
||||
}),
|
||||
createMockHistoricalPriceDataPoint({
|
||||
master_item_id: 1,
|
||||
summary_date: '2024-10-08',
|
||||
avg_price_in_cents: 100,
|
||||
}),
|
||||
];
|
||||
vi.mocked(apiClient.fetchHistoricalPriceData).mockResolvedValue(
|
||||
new Response(JSON.stringify(dataWithHigherPrice)),
|
||||
);
|
||||
render(<PriceHistoryChart />);
|
||||
|
||||
await waitFor(() => {
|
||||
const chart = screen.getByTestId('line-chart');
|
||||
const chartData = JSON.parse(chart.getAttribute('data-chartdata')!);
|
||||
const dataPoint = chartData.find((d: any) => d.date === 'Oct 1');
|
||||
expect(dataPoint['Organic Bananas']).toBe(100);
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle non-Error objects thrown during fetch', async () => {
|
||||
vi.mocked(apiClient.fetchHistoricalPriceData).mockRejectedValue('String Error');
|
||||
render(<PriceHistoryChart />);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.getByText('Failed to load price history.')).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -406,6 +406,74 @@ describe('ExtractedDataTable', () => {
|
||||
render(<ExtractedDataTable {...defaultProps} items={singleCategoryItems} />);
|
||||
expect(screen.queryByLabelText('Filter by category')).not.toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should allow switching filter back to All Categories', () => {
|
||||
render(<ExtractedDataTable {...defaultProps} />);
|
||||
const categoryFilter = screen.getByLabelText('Filter by category');
|
||||
|
||||
// Filter to Dairy
|
||||
fireEvent.change(categoryFilter, { target: { value: 'Dairy' } });
|
||||
expect(screen.queryByText('Gala Apples')).not.toBeInTheDocument();
|
||||
expect(screen.getByText('2% Milk')).toBeInTheDocument();
|
||||
|
||||
// Filter back to All
|
||||
fireEvent.change(categoryFilter, { target: { value: 'all' } });
|
||||
expect(screen.getByText('Gala Apples')).toBeInTheDocument();
|
||||
expect(screen.getByText('2% Milk')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should sort items alphabetically within watched and unwatched groups', () => {
|
||||
const items = [
|
||||
createMockFlyerItem({
|
||||
flyer_item_id: 1,
|
||||
item: 'Yam',
|
||||
master_item_id: 3,
|
||||
category_name: 'Produce',
|
||||
}), // Unwatched
|
||||
createMockFlyerItem({
|
||||
flyer_item_id: 2,
|
||||
item: 'Zebra',
|
||||
master_item_id: 1,
|
||||
category_name: 'Produce',
|
||||
}), // Watched
|
||||
createMockFlyerItem({
|
||||
flyer_item_id: 3,
|
||||
item: 'Banana',
|
||||
master_item_id: 4,
|
||||
category_name: 'Produce',
|
||||
}), // Unwatched
|
||||
createMockFlyerItem({
|
||||
flyer_item_id: 4,
|
||||
item: 'Apple',
|
||||
master_item_id: 2,
|
||||
category_name: 'Produce',
|
||||
}), // Watched
|
||||
];
|
||||
|
||||
vi.mocked(useUserData).mockReturnValue({
|
||||
watchedItems: [
|
||||
createMockMasterGroceryItem({ master_grocery_item_id: 1, name: 'Zebra' }),
|
||||
createMockMasterGroceryItem({ master_grocery_item_id: 2, name: 'Apple' }),
|
||||
],
|
||||
shoppingLists: [],
|
||||
setWatchedItems: vi.fn(),
|
||||
setShoppingLists: vi.fn(),
|
||||
isLoading: false,
|
||||
error: null,
|
||||
});
|
||||
|
||||
render(<ExtractedDataTable {...defaultProps} items={items} />);
|
||||
|
||||
const rows = screen.getAllByRole('row');
|
||||
// Extract item names based on the bold/semibold classes used for names
|
||||
const itemNames = rows.map((row) => {
|
||||
const nameEl = row.querySelector('.font-bold, .font-semibold');
|
||||
return nameEl?.textContent;
|
||||
});
|
||||
|
||||
// Expected: Watched items first (Apple, Zebra), then Unwatched (Banana, Yam)
|
||||
expect(itemNames).toEqual(['Apple', 'Zebra', 'Banana', 'Yam']);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Data Edge Cases', () => {
|
||||
@@ -460,5 +528,46 @@ describe('ExtractedDataTable', () => {
|
||||
// Check for the unit suffix, which might be in a separate element or part of the string
|
||||
expect(within(chickenItemRow).getAllByText(/\/kg/i).length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
it('should handle activeListId pointing to a non-existent list', () => {
|
||||
vi.mocked(useShoppingLists).mockReturnValue({
|
||||
activeListId: 999, // Non-existent
|
||||
shoppingLists: mockShoppingLists,
|
||||
addItemToList: mockAddItemToList,
|
||||
setActiveListId: vi.fn(),
|
||||
createList: vi.fn(),
|
||||
deleteList: vi.fn(),
|
||||
updateItemInList: vi.fn(),
|
||||
removeItemFromList: vi.fn(),
|
||||
isCreatingList: false,
|
||||
isDeletingList: false,
|
||||
isAddingItem: false,
|
||||
isUpdatingItem: false,
|
||||
isRemovingItem: false,
|
||||
error: null,
|
||||
});
|
||||
|
||||
render(<ExtractedDataTable {...defaultProps} />);
|
||||
|
||||
// Should behave as if item is not in list (Add button enabled)
|
||||
const appleItemRow = screen.getByText('Gala Apples').closest('tr')!;
|
||||
const addToListButton = within(appleItemRow).getByTitle('Add Apples to list');
|
||||
expect(addToListButton).toBeInTheDocument();
|
||||
expect(addToListButton).not.toBeDisabled();
|
||||
});
|
||||
|
||||
it('should display numeric quantity in parentheses if available', () => {
|
||||
const itemWithQtyNum = createMockFlyerItem({
|
||||
flyer_item_id: 999,
|
||||
item: 'Bulk Rice',
|
||||
quantity: 'Bag',
|
||||
quantity_num: 5,
|
||||
unit_price: { value: 10, unit: 'kg' },
|
||||
category_name: 'Pantry',
|
||||
flyer_id: 1,
|
||||
});
|
||||
render(<ExtractedDataTable {...defaultProps} items={[itemWithQtyNum]} />);
|
||||
expect(screen.getByText('(5)')).toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
// src/features/flyer/FlyerList.test.tsx
|
||||
import React from 'react';
|
||||
import { render, screen, fireEvent, waitFor } from '@testing-library/react';
|
||||
import { describe, it, expect, vi, beforeEach, type Mocked } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterEach, type Mocked } from 'vitest';
|
||||
import { FlyerList } from './FlyerList';
|
||||
import { formatShortDate } from './dateUtils';
|
||||
import type { Flyer, UserProfile } from '../../types';
|
||||
@@ -257,6 +257,73 @@ describe('FlyerList', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('Expiration Status Logic', () => {
|
||||
beforeEach(() => {
|
||||
vi.useFakeTimers();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
vi.useRealTimers();
|
||||
});
|
||||
|
||||
it('should show "Expired" for past dates', () => {
|
||||
// Flyer 1 valid_to is 2023-10-11
|
||||
vi.setSystemTime(new Date('2023-10-12T12:00:00Z'));
|
||||
render(
|
||||
<FlyerList
|
||||
flyers={[mockFlyers[0]]}
|
||||
onFlyerSelect={mockOnFlyerSelect}
|
||||
selectedFlyerId={null}
|
||||
profile={mockProfile}
|
||||
/>,
|
||||
);
|
||||
expect(screen.getByText('• Expired')).toBeInTheDocument();
|
||||
expect(screen.getByText('• Expired')).toHaveClass('text-red-500');
|
||||
});
|
||||
|
||||
it('should show "Expires today" when valid_to is today', () => {
|
||||
vi.setSystemTime(new Date('2023-10-11T12:00:00Z'));
|
||||
render(
|
||||
<FlyerList
|
||||
flyers={[mockFlyers[0]]}
|
||||
onFlyerSelect={mockOnFlyerSelect}
|
||||
selectedFlyerId={null}
|
||||
profile={mockProfile}
|
||||
/>,
|
||||
);
|
||||
expect(screen.getByText('• Expires today')).toBeInTheDocument();
|
||||
expect(screen.getByText('• Expires today')).toHaveClass('text-orange-500');
|
||||
});
|
||||
|
||||
it('should show "Expires in X days" (orange) for <= 3 days', () => {
|
||||
vi.setSystemTime(new Date('2023-10-09T12:00:00Z')); // 2 days left
|
||||
render(
|
||||
<FlyerList
|
||||
flyers={[mockFlyers[0]]}
|
||||
onFlyerSelect={mockOnFlyerSelect}
|
||||
selectedFlyerId={null}
|
||||
profile={mockProfile}
|
||||
/>,
|
||||
);
|
||||
expect(screen.getByText('• Expires in 2 days')).toBeInTheDocument();
|
||||
expect(screen.getByText('• Expires in 2 days')).toHaveClass('text-orange-500');
|
||||
});
|
||||
|
||||
it('should show "Expires in X days" (green) for > 3 days', () => {
|
||||
vi.setSystemTime(new Date('2023-10-05T12:00:00Z')); // 6 days left
|
||||
render(
|
||||
<FlyerList
|
||||
flyers={[mockFlyers[0]]}
|
||||
onFlyerSelect={mockOnFlyerSelect}
|
||||
selectedFlyerId={null}
|
||||
profile={mockProfile}
|
||||
/>,
|
||||
);
|
||||
expect(screen.getByText('• Expires in 6 days')).toBeInTheDocument();
|
||||
expect(screen.getByText('• Expires in 6 days')).toHaveClass('text-green-600');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Admin Functionality', () => {
|
||||
const adminProfile: UserProfile = createMockUserProfile({
|
||||
user: { user_id: 'admin-1', email: 'admin@example.com' },
|
||||
|
||||
@@ -6,14 +6,24 @@ import { FlyerUploader } from './FlyerUploader';
|
||||
import * as aiApiClientModule from '../../services/aiApiClient';
|
||||
import * as checksumModule from '../../utils/checksum';
|
||||
import { useNavigate, MemoryRouter } from 'react-router-dom';
|
||||
import { QueryClient, QueryClientProvider, onlineManager } from '@tanstack/react-query';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('../../services/aiApiClient');
|
||||
vi.mock('../../services/aiApiClient', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('../../services/aiApiClient')>();
|
||||
return {
|
||||
...actual,
|
||||
uploadAndProcessFlyer: vi.fn(),
|
||||
getJobStatus: vi.fn(),
|
||||
};
|
||||
});
|
||||
vi.mock('../../services/logger.client', () => ({
|
||||
// Keep the original logger.info/error but also spy on it for test assertions if needed
|
||||
logger: {
|
||||
info: vi.fn((...args) => console.log('[LOGGER.INFO]', ...args)),
|
||||
error: vi.fn((...args) => console.error('[LOGGER.ERROR]', ...args)),
|
||||
warn: vi.fn((...args) => console.warn('[LOGGER.WARN]', ...args)),
|
||||
debug: vi.fn((...args) => console.debug('[LOGGER.DEBUG]', ...args)),
|
||||
},
|
||||
}));
|
||||
vi.mock('../../utils/checksum', () => ({
|
||||
@@ -39,10 +49,19 @@ const mockedChecksumModule = checksumModule as unknown as {
|
||||
|
||||
const renderComponent = (onProcessingComplete = vi.fn()) => {
|
||||
console.log('--- [TEST LOG] ---: Rendering component inside MemoryRouter.');
|
||||
const queryClient = new QueryClient({
|
||||
defaultOptions: {
|
||||
queries: {
|
||||
retry: false,
|
||||
},
|
||||
},
|
||||
});
|
||||
return render(
|
||||
<MemoryRouter>
|
||||
<FlyerUploader onProcessingComplete={onProcessingComplete} />
|
||||
</MemoryRouter>,
|
||||
<QueryClientProvider client={queryClient}>
|
||||
<MemoryRouter>
|
||||
<FlyerUploader onProcessingComplete={onProcessingComplete} />
|
||||
</MemoryRouter>
|
||||
</QueryClientProvider>,
|
||||
);
|
||||
};
|
||||
|
||||
@@ -50,10 +69,11 @@ describe('FlyerUploader', () => {
|
||||
const navigateSpy = vi.fn();
|
||||
|
||||
beforeEach(() => {
|
||||
// Disable react-query's online manager to prevent it from interfering with fake timers
|
||||
onlineManager.setEventListener((setOnline) => {
|
||||
return () => {};
|
||||
});
|
||||
console.log(`\n--- [TEST LOG] ---: Starting test: "${expect.getState().currentTestName}"`);
|
||||
// Use the 'modern' implementation of fake timers to handle promise microtasks correctly.
|
||||
vi.useFakeTimers({ toFake: ['setTimeout'], shouldAdvanceTime: true });
|
||||
console.log('--- [TEST LOG] ---: MODERN fake timers enabled.');
|
||||
vi.resetAllMocks(); // Resets mock implementations AND call history.
|
||||
console.log('--- [TEST LOG] ---: Mocks reset.');
|
||||
mockedChecksumModule.generateFileChecksum.mockResolvedValue('mock-checksum');
|
||||
@@ -61,7 +81,6 @@ describe('FlyerUploader', () => {
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
vi.useRealTimers();
|
||||
console.log(`--- [TEST LOG] ---: Finished test: "${expect.getState().currentTestName}"\n`);
|
||||
});
|
||||
|
||||
@@ -104,21 +123,18 @@ describe('FlyerUploader', () => {
|
||||
expect(mockedAiApiClient.getJobStatus).toHaveBeenCalledTimes(1);
|
||||
console.log('--- [TEST LOG] ---: 7. Mocks verified. Advancing timers now...');
|
||||
|
||||
await act(async () => {
|
||||
console.log('--- [TEST LOG] ---: 8a. vi.advanceTimersByTime(3000) starting...');
|
||||
vi.advanceTimersByTime(3000);
|
||||
console.log('--- [TEST LOG] ---: 8b. vi.advanceTimersByTime(3000) complete.');
|
||||
});
|
||||
// With real timers, we now wait for the polling interval to elapse.
|
||||
console.log(
|
||||
`--- [TEST LOG] ---: 9. Act block finished. Now checking if getJobStatus was called again.`,
|
||||
);
|
||||
|
||||
try {
|
||||
// The polling interval is 3s, so we wait for a bit longer.
|
||||
await waitFor(() => {
|
||||
const calls = mockedAiApiClient.getJobStatus.mock.calls.length;
|
||||
console.log(`--- [TEST LOG] ---: 10. waitFor check: getJobStatus calls = ${calls}`);
|
||||
expect(mockedAiApiClient.getJobStatus).toHaveBeenCalledTimes(2);
|
||||
});
|
||||
}, { timeout: 4000 });
|
||||
console.log('--- [TEST LOG] ---: 11. SUCCESS: Second poll confirmed.');
|
||||
} catch (error) {
|
||||
console.error('--- [TEST LOG] ---: 11. ERROR: waitFor for second poll timed out.');
|
||||
@@ -181,24 +197,21 @@ describe('FlyerUploader', () => {
|
||||
expect(mockedAiApiClient.getJobStatus).toHaveBeenCalledTimes(1);
|
||||
console.log('--- [TEST LOG] ---: 5. First poll confirmed. Now AWAITING timer advancement.');
|
||||
|
||||
await act(async () => {
|
||||
console.log(`--- [TEST LOG] ---: 6. Advancing timers by 4000ms for the second poll...`);
|
||||
vi.advanceTimersByTime(4000);
|
||||
});
|
||||
console.log(`--- [TEST LOG] ---: 7. Timers advanced. Now AWAITING completion message.`);
|
||||
|
||||
try {
|
||||
console.log(
|
||||
'--- [TEST LOG] ---: 8a. waitFor check: Waiting for completion text and job status count.',
|
||||
);
|
||||
// Wait for the second poll to occur and the UI to update.
|
||||
await waitFor(() => {
|
||||
console.log(
|
||||
`--- [TEST LOG] ---: 8b. waitFor interval: calls=${mockedAiApiClient.getJobStatus.mock.calls.length}`,
|
||||
`--- [TEST LOG] ---: 8b. waitFor interval: calls=${
|
||||
mockedAiApiClient.getJobStatus.mock.calls.length
|
||||
}`,
|
||||
);
|
||||
expect(
|
||||
screen.getByText('Processing complete! Redirecting to flyer 42...'),
|
||||
).toBeInTheDocument();
|
||||
});
|
||||
}, { timeout: 4000 });
|
||||
console.log('--- [TEST LOG] ---: 9. SUCCESS: Completion message found.');
|
||||
} catch (error) {
|
||||
console.error('--- [TEST LOG] ---: 9. ERROR: waitFor for completion message timed out.');
|
||||
@@ -208,12 +221,9 @@ describe('FlyerUploader', () => {
|
||||
}
|
||||
expect(mockedAiApiClient.getJobStatus).toHaveBeenCalledTimes(2);
|
||||
|
||||
await act(async () => {
|
||||
console.log(`--- [TEST LOG] ---: 10. Advancing timers by 2000ms for redirect...`);
|
||||
vi.advanceTimersByTime(2000);
|
||||
});
|
||||
// Wait for the redirect timer (1.5s in component) to fire.
|
||||
await act(() => new Promise((r) => setTimeout(r, 2000)));
|
||||
console.log(`--- [TEST LOG] ---: 11. Timers advanced. Now asserting navigation.`);
|
||||
|
||||
expect(onProcessingComplete).toHaveBeenCalled();
|
||||
expect(navigateSpy).toHaveBeenCalledWith('/flyers/42');
|
||||
console.log('--- [TEST LOG] ---: 12. Callback and navigation confirmed.');
|
||||
@@ -222,10 +232,10 @@ describe('FlyerUploader', () => {
|
||||
it('should handle a failed job', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for a failed job.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-fail' });
|
||||
mockedAiApiClient.getJobStatus.mockResolvedValue({
|
||||
state: 'failed',
|
||||
failedReason: 'AI model exploded',
|
||||
});
|
||||
// The getJobStatus function throws a specific error when the job fails,
|
||||
// which is then caught by react-query and placed in the `error` state.
|
||||
const jobFailedError = new aiApiClientModule.JobFailedError('AI model exploded', 'UNKNOWN_ERROR');
|
||||
mockedAiApiClient.getJobStatus.mockRejectedValue(jobFailedError);
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering and uploading.');
|
||||
renderComponent();
|
||||
@@ -238,7 +248,8 @@ describe('FlyerUploader', () => {
|
||||
|
||||
try {
|
||||
console.log('--- [TEST LOG] ---: 4. AWAITING failure message...');
|
||||
expect(await screen.findByText(/Processing failed: AI model exploded/i)).toBeInTheDocument();
|
||||
// The UI should now display the error from the `pollError` state, which includes the "Polling failed" prefix.
|
||||
expect(await screen.findByText(/Polling failed: AI model exploded/i)).toBeInTheDocument();
|
||||
console.log('--- [TEST LOG] ---: 5. SUCCESS: Failure message found.');
|
||||
} catch (error) {
|
||||
console.error('--- [TEST LOG] ---: 5. ERROR: findByText for failure message timed out.');
|
||||
@@ -252,14 +263,18 @@ describe('FlyerUploader', () => {
|
||||
});
|
||||
|
||||
it('should clear the polling timeout when a job fails', async () => {
|
||||
const clearTimeoutSpy = vi.spyOn(global, 'clearTimeout');
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mocks for failed job timeout clearance.');
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: 'job-fail-timeout' });
|
||||
|
||||
// We need at least one 'active' response to establish a timeout loop so we have something to clear
|
||||
mockedAiApiClient.getJobStatus
|
||||
.mockResolvedValueOnce({ state: 'active', progress: { message: 'Working...' } })
|
||||
.mockResolvedValueOnce({ state: 'failed', failedReason: 'Fatal Error' });
|
||||
// The second call should be a rejection, as this is how getJobStatus signals a failure.
|
||||
mockedAiApiClient.getJobStatus.mockImplementation(async () => {
|
||||
// Fail on the second call by checking the number of times the mock has been invoked.
|
||||
if (mockedAiApiClient.getJobStatus.mock.calls.length > 1) {
|
||||
throw new aiApiClientModule.JobFailedError('Fatal Error', 'UNKNOWN_ERROR');
|
||||
}
|
||||
return { state: 'active', progress: { message: 'Working...' } } as aiApiClientModule.JobStatus;
|
||||
});
|
||||
|
||||
renderComponent();
|
||||
const file = new File(['content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
@@ -270,25 +285,8 @@ describe('FlyerUploader', () => {
|
||||
// Wait for the first poll to complete and UI to update to "Working..."
|
||||
await screen.findByText('Working...');
|
||||
|
||||
// Advance time to trigger the second poll
|
||||
await act(async () => {
|
||||
vi.advanceTimersByTime(3000);
|
||||
});
|
||||
|
||||
// Wait for the failure UI
|
||||
await screen.findByText(/Processing failed: Fatal Error/i);
|
||||
|
||||
// Verify clearTimeout was called
|
||||
expect(clearTimeoutSpy).toHaveBeenCalled();
|
||||
|
||||
// Verify no further polling occurs
|
||||
const callsBefore = mockedAiApiClient.getJobStatus.mock.calls.length;
|
||||
await act(async () => {
|
||||
vi.advanceTimersByTime(10000);
|
||||
});
|
||||
expect(mockedAiApiClient.getJobStatus).toHaveBeenCalledTimes(callsBefore);
|
||||
|
||||
clearTimeoutSpy.mockRestore();
|
||||
await waitFor(() => expect(screen.getByText(/Polling failed: Fatal Error/i)).toBeInTheDocument(), { timeout: 4000 });
|
||||
});
|
||||
|
||||
it('should clear the polling timeout when the component unmounts', async () => {
|
||||
@@ -322,10 +320,11 @@ describe('FlyerUploader', () => {
|
||||
|
||||
it('should handle a duplicate flyer error (409)', async () => {
|
||||
console.log('--- [TEST LOG] ---: 1. Setting up mock for 409 duplicate error.');
|
||||
// The API client now throws a structured error for non-2xx responses.
|
||||
// The API client throws a structured error, which useFlyerUploader now parses
|
||||
// to set both the errorMessage and the duplicateFlyerId.
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockRejectedValue({
|
||||
status: 409,
|
||||
body: { flyerId: 99, message: 'Duplicate' },
|
||||
body: { flyerId: 99, message: 'This flyer has already been processed.' },
|
||||
});
|
||||
|
||||
console.log('--- [TEST LOG] ---: 2. Rendering and uploading.');
|
||||
@@ -339,9 +338,10 @@ describe('FlyerUploader', () => {
|
||||
|
||||
try {
|
||||
console.log('--- [TEST LOG] ---: 4. AWAITING duplicate flyer message...');
|
||||
expect(
|
||||
await screen.findByText('This flyer has already been processed. You can view it here:'),
|
||||
).toBeInTheDocument();
|
||||
// With the fix, the duplicate error message and the link are combined into a single paragraph.
|
||||
// We now look for this combined message.
|
||||
const errorMessage = await screen.findByText(/This flyer has already been processed. You can view it here:/i);
|
||||
expect(errorMessage).toBeInTheDocument();
|
||||
console.log('--- [TEST LOG] ---: 5. SUCCESS: Duplicate message found.');
|
||||
} catch (error) {
|
||||
console.error('--- [TEST LOG] ---: 5. ERROR: findByText for duplicate message timed out.');
|
||||
@@ -453,7 +453,7 @@ describe('FlyerUploader', () => {
|
||||
fireEvent.change(input, { target: { files: [file] } });
|
||||
|
||||
console.log('--- [TEST LOG] ---: 3. Awaiting error message.');
|
||||
expect(await screen.findByText(/Polling Network Error/i)).toBeInTheDocument();
|
||||
expect(await screen.findByText(/Polling failed: Polling Network Error/i)).toBeInTheDocument();
|
||||
console.log('--- [TEST LOG] ---: 4. Assertions passed.');
|
||||
});
|
||||
|
||||
@@ -495,7 +495,9 @@ describe('FlyerUploader', () => {
|
||||
fireEvent.change(input, { target: { files: [file] } });
|
||||
|
||||
console.log('--- [TEST LOG] ---: 3. Awaiting error message.');
|
||||
expect(await screen.findByText(/Failed to parse JSON response from server/i)).toBeInTheDocument();
|
||||
expect(
|
||||
await screen.findByText(/Polling failed: Failed to parse JSON response from server/i),
|
||||
).toBeInTheDocument();
|
||||
console.log('--- [TEST LOG] ---: 4. Assertions passed.');
|
||||
});
|
||||
|
||||
|
||||
@@ -1,208 +1,68 @@
|
||||
// src/features/flyer/FlyerUploader.tsx
|
||||
import React, { useState, useEffect, useRef, useCallback } from 'react';
|
||||
import React, { useEffect, useCallback } from 'react';
|
||||
import { useNavigate, Link } from 'react-router-dom';
|
||||
import { uploadAndProcessFlyer, getJobStatus } from '../../services/aiApiClient';
|
||||
import { generateFileChecksum } from '../../utils/checksum';
|
||||
import { logger } from '../../services/logger.client';
|
||||
import { ProcessingStatus } from './ProcessingStatus';
|
||||
import type { ProcessingStage } from '../../types';
|
||||
import { useDragAndDrop } from '../../hooks/useDragAndDrop';
|
||||
|
||||
type ProcessingState = 'idle' | 'uploading' | 'polling' | 'completed' | 'error';
|
||||
import { useFlyerUploader } from '../../hooks/useFlyerUploader';
|
||||
|
||||
interface FlyerUploaderProps {
|
||||
onProcessingComplete: () => void;
|
||||
}
|
||||
|
||||
export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComplete }) => {
|
||||
const [processingState, setProcessingState] = useState<ProcessingState>('idle');
|
||||
const [statusMessage, setStatusMessage] = useState<string | null>(null);
|
||||
const [jobId, setJobId] = useState<string | null>(null);
|
||||
const [errorMessage, setErrorMessage] = useState<string | null>(null);
|
||||
const [duplicateFlyerId, setDuplicateFlyerId] = useState<number | null>(null);
|
||||
const navigate = useNavigate();
|
||||
|
||||
const pollingTimeoutRef = useRef<number | null>(null);
|
||||
|
||||
const [processingStages, setProcessingStages] = useState<ProcessingStage[]>([]);
|
||||
const [estimatedTime, setEstimatedTime] = useState(0);
|
||||
const [currentFile, setCurrentFile] = useState<string | null>(null);
|
||||
|
||||
// DEBUG: Log component mount and unmount
|
||||
useEffect(() => {
|
||||
console.debug('[DEBUG] FlyerUploader: Component did mount.');
|
||||
return () => {
|
||||
console.debug('[DEBUG] FlyerUploader: Component will unmount.');
|
||||
};
|
||||
}, []);
|
||||
|
||||
// DEBUG: Log state changes
|
||||
useEffect(() => {
|
||||
console.debug(`[DEBUG] FlyerUploader: processingState changed to -> ${processingState}`);
|
||||
}, [processingState]);
|
||||
const {
|
||||
processingState,
|
||||
statusMessage,
|
||||
errorMessage,
|
||||
duplicateFlyerId,
|
||||
processingStages,
|
||||
estimatedTime,
|
||||
currentFile,
|
||||
flyerId,
|
||||
upload,
|
||||
resetUploaderState,
|
||||
} = useFlyerUploader();
|
||||
|
||||
useEffect(() => {
|
||||
if (statusMessage) logger.info(`FlyerUploader Status: ${statusMessage}`);
|
||||
}, [statusMessage]);
|
||||
|
||||
useEffect(() => {
|
||||
console.debug(`[DEBUG] Polling Effect Triggered: state=${processingState}, jobId=${jobId}`);
|
||||
if (processingState !== 'polling' || !jobId) {
|
||||
if (pollingTimeoutRef.current) {
|
||||
console.debug(
|
||||
`[DEBUG] Polling Effect: Clearing timeout ID ${pollingTimeoutRef.current} because state is not 'polling' or no jobId exists.`,
|
||||
);
|
||||
clearTimeout(pollingTimeoutRef.current);
|
||||
}
|
||||
return;
|
||||
if (errorMessage) {
|
||||
logger.error(`[FlyerUploader] Error encountered: ${errorMessage}`, { duplicateFlyerId });
|
||||
}
|
||||
}, [errorMessage, duplicateFlyerId]);
|
||||
|
||||
const pollStatus = async () => {
|
||||
console.debug(`[DEBUG] pollStatus(): Polling for jobId: ${jobId}`);
|
||||
try {
|
||||
const job = await getJobStatus(jobId); // Now returns parsed JSON directly
|
||||
console.debug('[DEBUG] pollStatus(): Job status received:', job); // The rest of the logic remains the same
|
||||
|
||||
if (job.progress) {
|
||||
setProcessingStages(job.progress.stages || []);
|
||||
setEstimatedTime(job.progress.estimatedTimeRemaining || 0);
|
||||
setStatusMessage(job.progress.message || null);
|
||||
}
|
||||
|
||||
switch (job.state) {
|
||||
case 'completed':
|
||||
console.debug('[DEBUG] pollStatus(): Job state is "completed".');
|
||||
const flyerId = job.returnValue?.flyerId;
|
||||
if (flyerId) {
|
||||
setStatusMessage(`Processing complete! Redirecting to flyer ${flyerId}...`);
|
||||
setProcessingState('completed');
|
||||
onProcessingComplete();
|
||||
console.debug('[DEBUG] pollStatus(): Setting 1500ms timeout for redirect.');
|
||||
setTimeout(() => {
|
||||
console.debug(`[DEBUG] pollStatus(): Redirecting to /flyers/${flyerId}`);
|
||||
navigate(`/flyers/${flyerId}`);
|
||||
}, 1500);
|
||||
} else {
|
||||
throw new Error('Job completed but did not return a flyer ID.');
|
||||
}
|
||||
break;
|
||||
|
||||
case 'failed':
|
||||
console.debug(
|
||||
`[DEBUG] pollStatus(): Job state is "failed". Reason: ${job.failedReason}`,
|
||||
);
|
||||
// Explicitly clear any pending timeout to stop the polling loop immediately.
|
||||
if (pollingTimeoutRef.current) {
|
||||
clearTimeout(pollingTimeoutRef.current);
|
||||
}
|
||||
setErrorMessage(`Processing failed: ${job.failedReason || 'Unknown error'}`);
|
||||
// Clear any stale "in-progress" messages to avoid user confusion.
|
||||
setStatusMessage(null);
|
||||
setProcessingState('error');
|
||||
break;
|
||||
|
||||
case 'active':
|
||||
case 'waiting':
|
||||
default:
|
||||
console.debug(
|
||||
`[DEBUG] pollStatus(): Job state is "${job.state}". Setting timeout for next poll (3000ms).`,
|
||||
);
|
||||
pollingTimeoutRef.current = window.setTimeout(pollStatus, 3000);
|
||||
console.debug(`[DEBUG] pollStatus(): Timeout ID ${pollingTimeoutRef.current} set.`);
|
||||
break;
|
||||
}
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error during polling:');
|
||||
setErrorMessage(
|
||||
error instanceof Error ? error.message : 'An unexpected error occurred during polling.',
|
||||
);
|
||||
setProcessingState('error');
|
||||
}
|
||||
};
|
||||
|
||||
pollStatus();
|
||||
|
||||
return () => {
|
||||
if (pollingTimeoutRef.current) {
|
||||
console.debug(
|
||||
`[DEBUG] Polling Effect Cleanup: Clearing timeout ID ${pollingTimeoutRef.current}`,
|
||||
);
|
||||
clearTimeout(pollingTimeoutRef.current);
|
||||
pollingTimeoutRef.current = null;
|
||||
} else {
|
||||
console.debug('[DEBUG] Polling Effect Cleanup: No active timeout to clear.');
|
||||
}
|
||||
};
|
||||
}, [processingState, jobId, onProcessingComplete, navigate]);
|
||||
|
||||
const processFile = useCallback(async (file: File) => {
|
||||
console.debug('[DEBUG] processFile(): Starting file processing for', file.name);
|
||||
setProcessingState('uploading');
|
||||
setErrorMessage(null);
|
||||
setDuplicateFlyerId(null);
|
||||
setCurrentFile(file.name);
|
||||
|
||||
try {
|
||||
console.debug('[DEBUG] processFile(): Generating file checksum.');
|
||||
const checksum = await generateFileChecksum(file);
|
||||
setStatusMessage('Uploading file...');
|
||||
console.debug(
|
||||
`[DEBUG] processFile(): Checksum generated: ${checksum}. Calling uploadAndProcessFlyer.`,
|
||||
);
|
||||
|
||||
// The API client now returns parsed JSON on success or throws a structured error on failure.
|
||||
const { jobId: newJobId } = await uploadAndProcessFlyer(file, checksum);
|
||||
console.debug(`[DEBUG] processFile(): Upload successful. Received jobId: ${newJobId}`);
|
||||
setJobId(newJobId);
|
||||
setProcessingState('polling');
|
||||
} catch (error: any) {
|
||||
// Handle the structured error thrown by the API client.
|
||||
logger.error({ error }, 'An error occurred during file upload:');
|
||||
// Handle 409 Conflict for duplicate flyers
|
||||
if (error?.status === 409 && error.body?.flyerId) {
|
||||
setErrorMessage(`This flyer has already been processed. You can view it here:`);
|
||||
setDuplicateFlyerId(error.body.flyerId);
|
||||
} else {
|
||||
// Handle other errors (e.g., validation, server errors)
|
||||
const message =
|
||||
error?.body?.message || error?.message || 'An unexpected error occurred during upload.';
|
||||
setErrorMessage(message);
|
||||
}
|
||||
setProcessingState('error');
|
||||
// Handle completion and navigation
|
||||
useEffect(() => {
|
||||
if (processingState === 'completed' && flyerId) {
|
||||
onProcessingComplete();
|
||||
// Small delay to show the "Complete" state before redirecting
|
||||
const timer = setTimeout(() => {
|
||||
navigate(`/flyers/${flyerId}`);
|
||||
}, 1500);
|
||||
return () => clearTimeout(timer);
|
||||
}
|
||||
}, []);
|
||||
}, [processingState, flyerId, onProcessingComplete, navigate]);
|
||||
|
||||
const handleFileChange = (event: React.ChangeEvent<HTMLInputElement>) => {
|
||||
console.debug('[DEBUG] handleFileChange(): File input changed.');
|
||||
const file = event.target.files?.[0];
|
||||
if (file) {
|
||||
processFile(file);
|
||||
upload(file);
|
||||
}
|
||||
event.target.value = '';
|
||||
};
|
||||
|
||||
const resetUploaderState = useCallback(() => {
|
||||
console.debug(
|
||||
`[DEBUG] resetUploaderState(): User triggered reset. Previous jobId was: ${jobId}`,
|
||||
);
|
||||
setProcessingState('idle');
|
||||
setJobId(null);
|
||||
setErrorMessage(null);
|
||||
setDuplicateFlyerId(null);
|
||||
setCurrentFile(null);
|
||||
setProcessingStages([]);
|
||||
setEstimatedTime(0);
|
||||
logger.info('Uploader state has been reset. Previous job ID was:', jobId);
|
||||
}, [jobId]);
|
||||
|
||||
const onFilesDropped = useCallback(
|
||||
(files: FileList) => {
|
||||
console.debug('[DEBUG] onFilesDropped(): Files were dropped.');
|
||||
if (files && files.length > 0) {
|
||||
processFile(files[0]);
|
||||
upload(files[0]);
|
||||
}
|
||||
},
|
||||
[processFile],
|
||||
[upload],
|
||||
);
|
||||
|
||||
const isProcessing = processingState === 'uploading' || processingState === 'polling';
|
||||
@@ -216,11 +76,6 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
? 'bg-brand-light/50 dark:bg-brand-dark/20'
|
||||
: 'bg-gray-50/50 dark:bg-gray-800/20';
|
||||
|
||||
// If processing, show the detailed status component. Otherwise, show the uploader.
|
||||
console.debug(
|
||||
`[DEBUG] FlyerUploader: Rendering. State=${processingState}, Msg=${statusMessage}, Err=${!!errorMessage}`,
|
||||
);
|
||||
|
||||
if (isProcessing || processingState === 'completed' || processingState === 'error') {
|
||||
return (
|
||||
<div className="max-w-4xl mx-auto">
|
||||
@@ -230,22 +85,30 @@ export const FlyerUploader: React.FC<FlyerUploaderProps> = ({ onProcessingComple
|
||||
currentFile={currentFile}
|
||||
/>
|
||||
<div className="mt-4 text-center">
|
||||
{/* Display the current status message to the user and the test runner */}
|
||||
{statusMessage && (
|
||||
{/* Display status message if not completed (completed has its own redirect logic) */}
|
||||
{statusMessage && processingState !== 'completed' && (
|
||||
<p className="text-gray-600 dark:text-gray-400 mt-2 italic animate-pulse">
|
||||
{statusMessage}
|
||||
</p>
|
||||
)}
|
||||
|
||||
{processingState === 'completed' && (
|
||||
<p className="text-green-600 dark:text-green-400 mt-2 font-bold">
|
||||
Processing complete! Redirecting to flyer {flyerId}...
|
||||
</p>
|
||||
)}
|
||||
|
||||
{errorMessage && (
|
||||
<div className="text-red-600 dark:text-red-400 font-semibold p-4 bg-red-100 dark:bg-red-900/30 rounded-md">
|
||||
<p>{errorMessage}</p>
|
||||
{duplicateFlyerId && (
|
||||
{duplicateFlyerId ? (
|
||||
<p>
|
||||
<Link to={`/flyers/${duplicateFlyerId}`} className="text-blue-500 underline">
|
||||
{errorMessage} You can view it here:{' '}
|
||||
<Link to={`/flyers/${duplicateFlyerId}`} className="text-blue-500 underline" data-discover="true">
|
||||
Flyer #{duplicateFlyerId}
|
||||
</Link>
|
||||
</p>
|
||||
) : (
|
||||
<p>{errorMessage}</p>
|
||||
)}
|
||||
</div>
|
||||
)}
|
||||
|
||||
@@ -236,6 +236,24 @@ describe('ShoppingListComponent (in shopping feature)', () => {
|
||||
alertSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should show a generic alert if reading aloud fails with a non-Error object', async () => {
|
||||
const alertSpy = vi.spyOn(window, 'alert').mockImplementation(() => {});
|
||||
vi.spyOn(aiApiClient, 'generateSpeechFromText').mockRejectedValue('A string error');
|
||||
|
||||
render(<ShoppingListComponent {...defaultProps} />);
|
||||
const readAloudButton = screen.getByTitle(/read list aloud/i);
|
||||
|
||||
fireEvent.click(readAloudButton);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(alertSpy).toHaveBeenCalledWith(
|
||||
'Could not read list aloud: An unknown error occurred while generating audio.',
|
||||
);
|
||||
});
|
||||
|
||||
alertSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle interactions with purchased items', () => {
|
||||
render(<ShoppingListComponent {...defaultProps} />);
|
||||
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
// src/features/shopping/ShoppingList.tsx
|
||||
import React, { useState, useMemo, useCallback, useEffect } from 'react';
|
||||
import React, { useState, useMemo, useCallback } from 'react';
|
||||
import type { ShoppingList, ShoppingListItem, User } from '../../types';
|
||||
import { UserIcon } from '../../components/icons/UserIcon';
|
||||
import { ListBulletIcon } from '../../components/icons/ListBulletIcon';
|
||||
@@ -56,28 +56,6 @@ export const ShoppingListComponent: React.FC<ShoppingListComponentProps> = ({
|
||||
return { neededItems, purchasedItems };
|
||||
}, [activeList]);
|
||||
|
||||
useEffect(() => {
|
||||
if (activeList) {
|
||||
console.log('ShoppingList Debug: Active List:', activeList.name);
|
||||
console.log(
|
||||
'ShoppingList Debug: Needed Items:',
|
||||
neededItems.map((i) => ({
|
||||
id: i.shopping_list_item_id,
|
||||
name: i.custom_item_name || i.master_item?.name,
|
||||
raw: i,
|
||||
})),
|
||||
);
|
||||
console.log(
|
||||
'ShoppingList Debug: Purchased Items:',
|
||||
purchasedItems.map((i) => ({
|
||||
id: i.shopping_list_item_id,
|
||||
name: i.custom_item_name || i.master_item?.name,
|
||||
raw: i,
|
||||
})),
|
||||
);
|
||||
}
|
||||
}, [activeList, neededItems, purchasedItems]);
|
||||
|
||||
const handleCreateList = async () => {
|
||||
const name = prompt('Enter a name for your new shopping list:');
|
||||
if (name && name.trim()) {
|
||||
|
||||
@@ -164,6 +164,15 @@ describe('WatchedItemsList (in shopping feature)', () => {
|
||||
expect(itemsDesc[1]).toHaveTextContent('Eggs');
|
||||
expect(itemsDesc[2]).toHaveTextContent('Bread');
|
||||
expect(itemsDesc[3]).toHaveTextContent('Apples');
|
||||
|
||||
// Click again to sort ascending
|
||||
fireEvent.click(sortButton);
|
||||
|
||||
const itemsAscAgain = screen.getAllByRole('listitem');
|
||||
expect(itemsAscAgain[0]).toHaveTextContent('Apples');
|
||||
expect(itemsAscAgain[1]).toHaveTextContent('Bread');
|
||||
expect(itemsAscAgain[2]).toHaveTextContent('Eggs');
|
||||
expect(itemsAscAgain[3]).toHaveTextContent('Milk');
|
||||
});
|
||||
|
||||
it('should call onAddItemToList when plus icon is clicked', () => {
|
||||
@@ -222,6 +231,18 @@ describe('WatchedItemsList (in shopping feature)', () => {
|
||||
fireEvent.change(nameInput, { target: { value: 'Grapes' } });
|
||||
expect(addButton).toBeDisabled();
|
||||
});
|
||||
|
||||
it('should not submit if form is submitted with invalid data', () => {
|
||||
render(<WatchedItemsList {...defaultProps} />);
|
||||
const nameInput = screen.getByPlaceholderText(/add item/i);
|
||||
const form = nameInput.closest('form')!;
|
||||
const categorySelect = screen.getByDisplayValue('Select a category');
|
||||
fireEvent.change(categorySelect, { target: { value: 'Dairy & Eggs' } });
|
||||
|
||||
fireEvent.change(nameInput, { target: { value: ' ' } });
|
||||
fireEvent.submit(form);
|
||||
expect(mockOnAddItem).not.toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Error Handling', () => {
|
||||
|
||||
@@ -26,8 +26,17 @@ export function useApi<T, TArgs extends unknown[]>(
|
||||
const [isRefetching, setIsRefetching] = useState<boolean>(false);
|
||||
const [error, setError] = useState<Error | null>(null);
|
||||
const hasBeenExecuted = useRef(false);
|
||||
const lastErrorMessageRef = useRef<string | null>(null);
|
||||
const abortControllerRef = useRef<AbortController>(new AbortController());
|
||||
|
||||
// Use a ref to track the latest apiFunction. This allows us to keep `execute` stable
|
||||
// even if `apiFunction` is recreated on every render (common with inline arrow functions).
|
||||
const apiFunctionRef = useRef(apiFunction);
|
||||
|
||||
useEffect(() => {
|
||||
apiFunctionRef.current = apiFunction;
|
||||
}, [apiFunction]);
|
||||
|
||||
// This effect ensures that when the component using the hook unmounts,
|
||||
// any in-flight request is cancelled.
|
||||
useEffect(() => {
|
||||
@@ -52,12 +61,13 @@ export function useApi<T, TArgs extends unknown[]>(
|
||||
async (...args: TArgs): Promise<T | null> => {
|
||||
setLoading(true);
|
||||
setError(null);
|
||||
lastErrorMessageRef.current = null;
|
||||
if (hasBeenExecuted.current) {
|
||||
setIsRefetching(true);
|
||||
}
|
||||
|
||||
try {
|
||||
const response = await apiFunction(...args, abortControllerRef.current.signal);
|
||||
const response = await apiFunctionRef.current(...args, abortControllerRef.current.signal);
|
||||
|
||||
if (!response.ok) {
|
||||
// Attempt to parse a JSON error response. This is aligned with ADR-003,
|
||||
@@ -96,7 +106,17 @@ export function useApi<T, TArgs extends unknown[]>(
|
||||
}
|
||||
return result;
|
||||
} catch (e) {
|
||||
const err = e instanceof Error ? e : new Error('An unknown error occurred.');
|
||||
let err: Error;
|
||||
if (e instanceof Error) {
|
||||
err = e;
|
||||
} else if (typeof e === 'object' && e !== null && 'status' in e) {
|
||||
// Handle structured errors (e.g. { status: 409, body: { ... } })
|
||||
const structuredError = e as { status: number; body?: { message?: string } };
|
||||
const message = structuredError.body?.message || `Request failed with status ${structuredError.status}`;
|
||||
err = new Error(message);
|
||||
} else {
|
||||
err = new Error('An unknown error occurred.');
|
||||
}
|
||||
// If the error is an AbortError, it's an intentional cancellation, so we don't set an error state.
|
||||
if (err.name === 'AbortError') {
|
||||
logger.info('API request was cancelled.', { functionName: apiFunction.name });
|
||||
@@ -106,7 +126,13 @@ export function useApi<T, TArgs extends unknown[]>(
|
||||
error: err.message,
|
||||
functionName: apiFunction.name,
|
||||
});
|
||||
setError(err);
|
||||
// Only set a new error object if the message is different from the last one.
|
||||
// This prevents creating new object references for the same error (e.g. repeated timeouts)
|
||||
// and helps break infinite loops in components that depend on the `error` object.
|
||||
if (err.message !== lastErrorMessageRef.current) {
|
||||
setError(err);
|
||||
lastErrorMessageRef.current = err.message;
|
||||
}
|
||||
notifyError(err.message); // Optionally notify the user automatically.
|
||||
return null; // Return null on failure.
|
||||
} finally {
|
||||
@@ -114,7 +140,7 @@ export function useApi<T, TArgs extends unknown[]>(
|
||||
setIsRefetching(false);
|
||||
}
|
||||
},
|
||||
[apiFunction],
|
||||
[], // execute is now stable because it uses apiFunctionRef
|
||||
); // abortControllerRef is stable
|
||||
|
||||
return { execute, loading, isRefetching, error, data, reset };
|
||||
|
||||
174
src/hooks/useAppInitialization.test.tsx
Normal file
174
src/hooks/useAppInitialization.test.tsx
Normal file
@@ -0,0 +1,174 @@
|
||||
// src/hooks/useAppInitialization.test.tsx
|
||||
import { renderHook, waitFor } from '@testing-library/react';
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { MemoryRouter, useNavigate } from 'react-router-dom';
|
||||
import { useAppInitialization } from './useAppInitialization';
|
||||
import { useAuth } from './useAuth';
|
||||
import { useModal } from './useModal';
|
||||
import { createMockUserProfile } from '../tests/utils/mockFactories';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('./useAuth');
|
||||
vi.mock('./useModal');
|
||||
vi.mock('react-router-dom', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('react-router-dom')>();
|
||||
return {
|
||||
...actual,
|
||||
useNavigate: vi.fn(),
|
||||
};
|
||||
});
|
||||
vi.mock('../services/logger.client');
|
||||
vi.mock('../config', () => ({
|
||||
default: {
|
||||
app: { version: '1.0.1' },
|
||||
},
|
||||
}));
|
||||
|
||||
const mockedUseAuth = vi.mocked(useAuth);
|
||||
const mockedUseModal = vi.mocked(useModal);
|
||||
const mockedUseNavigate = vi.mocked(useNavigate);
|
||||
|
||||
const mockLogin = vi.fn().mockResolvedValue(undefined);
|
||||
const mockNavigate = vi.fn();
|
||||
const mockOpenModal = vi.fn();
|
||||
|
||||
// Wrapper with MemoryRouter is needed because the hook uses useLocation and useNavigate
|
||||
const wrapper = ({
|
||||
children,
|
||||
initialEntries = ['/'],
|
||||
}: {
|
||||
children: React.ReactNode;
|
||||
initialEntries?: string[];
|
||||
}) => <MemoryRouter initialEntries={initialEntries}>{children}</MemoryRouter>;
|
||||
|
||||
describe('useAppInitialization Hook', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
mockedUseNavigate.mockReturnValue(mockNavigate);
|
||||
mockedUseAuth.mockReturnValue({
|
||||
userProfile: null,
|
||||
login: mockLogin,
|
||||
authStatus: 'SIGNED_OUT',
|
||||
isLoading: false,
|
||||
logout: vi.fn(),
|
||||
updateProfile: vi.fn(),
|
||||
});
|
||||
mockedUseModal.mockReturnValue({
|
||||
openModal: mockOpenModal,
|
||||
closeModal: vi.fn(),
|
||||
isModalOpen: vi.fn(),
|
||||
});
|
||||
// Mock localStorage
|
||||
Object.defineProperty(window, 'localStorage', {
|
||||
value: {
|
||||
getItem: vi.fn().mockReturnValue(null),
|
||||
setItem: vi.fn(),
|
||||
removeItem: vi.fn(),
|
||||
clear: vi.fn(),
|
||||
},
|
||||
writable: true,
|
||||
});
|
||||
// Mock matchMedia
|
||||
Object.defineProperty(window, 'matchMedia', {
|
||||
value: vi.fn().mockImplementation((query) => ({
|
||||
matches: false, // default to light mode
|
||||
})),
|
||||
writable: true,
|
||||
configurable: true,
|
||||
});
|
||||
});
|
||||
|
||||
it('should call login when googleAuthToken is in URL', async () => {
|
||||
renderHook(() => useAppInitialization(), {
|
||||
wrapper: (props) => wrapper({ ...props, initialEntries: ['/?googleAuthToken=test-token'] }),
|
||||
});
|
||||
await waitFor(() => {
|
||||
expect(mockLogin).toHaveBeenCalledWith('test-token');
|
||||
});
|
||||
});
|
||||
|
||||
it('should call login when githubAuthToken is in URL', async () => {
|
||||
renderHook(() => useAppInitialization(), {
|
||||
wrapper: (props) => wrapper({ ...props, initialEntries: ['/?githubAuthToken=test-token'] }),
|
||||
});
|
||||
await waitFor(() => {
|
||||
expect(mockLogin).toHaveBeenCalledWith('test-token');
|
||||
});
|
||||
});
|
||||
|
||||
it('should call navigate to clean the URL after processing a token', async () => {
|
||||
renderHook(() => useAppInitialization(), {
|
||||
wrapper: (props) => wrapper({ ...props, initialEntries: ['/some/path?googleAuthToken=test-token'] }),
|
||||
});
|
||||
await waitFor(() => {
|
||||
expect(mockLogin).toHaveBeenCalledWith('test-token');
|
||||
});
|
||||
expect(mockNavigate).toHaveBeenCalledWith('/some/path', { replace: true });
|
||||
});
|
||||
|
||||
it("should open \"What's New\" modal if version is new", () => {
|
||||
vi.spyOn(window.localStorage, 'getItem').mockReturnValue('1.0.0');
|
||||
renderHook(() => useAppInitialization(), { wrapper });
|
||||
expect(mockOpenModal).toHaveBeenCalledWith('whatsNew');
|
||||
expect(window.localStorage.setItem).toHaveBeenCalledWith('lastSeenVersion', '1.0.1');
|
||||
});
|
||||
|
||||
it("should not open \"What's New\" modal if version is the same", () => {
|
||||
vi.spyOn(window.localStorage, 'getItem').mockReturnValue('1.0.1');
|
||||
renderHook(() => useAppInitialization(), { wrapper });
|
||||
expect(mockOpenModal).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should set dark mode from user profile', async () => {
|
||||
mockedUseAuth.mockReturnValue({
|
||||
...mockedUseAuth(),
|
||||
userProfile: createMockUserProfile({ preferences: { darkMode: true } }),
|
||||
});
|
||||
const { result } = renderHook(() => useAppInitialization(), { wrapper });
|
||||
await waitFor(() => {
|
||||
expect(result.current.isDarkMode).toBe(true);
|
||||
expect(document.documentElement.classList.contains('dark')).toBe(true);
|
||||
});
|
||||
});
|
||||
|
||||
it('should set dark mode from localStorage', async () => {
|
||||
vi.spyOn(window.localStorage, 'getItem').mockImplementation((key) =>
|
||||
key === 'darkMode' ? 'true' : null,
|
||||
);
|
||||
const { result } = renderHook(() => useAppInitialization(), { wrapper });
|
||||
await waitFor(() => {
|
||||
expect(result.current.isDarkMode).toBe(true);
|
||||
expect(document.documentElement.classList.contains('dark')).toBe(true);
|
||||
});
|
||||
});
|
||||
|
||||
it('should set dark mode from system preference', async () => {
|
||||
vi.spyOn(window, 'matchMedia').mockReturnValue({ matches: true } as any);
|
||||
const { result } = renderHook(() => useAppInitialization(), { wrapper });
|
||||
await waitFor(() => {
|
||||
expect(result.current.isDarkMode).toBe(true);
|
||||
expect(document.documentElement.classList.contains('dark')).toBe(true);
|
||||
});
|
||||
});
|
||||
|
||||
it('should set unit system from user profile', async () => {
|
||||
mockedUseAuth.mockReturnValue({
|
||||
...mockedUseAuth(),
|
||||
userProfile: createMockUserProfile({ preferences: { unitSystem: 'metric' } }),
|
||||
});
|
||||
const { result } = renderHook(() => useAppInitialization(), { wrapper });
|
||||
await waitFor(() => {
|
||||
expect(result.current.unitSystem).toBe('metric');
|
||||
});
|
||||
});
|
||||
|
||||
it('should set unit system from localStorage', async () => {
|
||||
vi.spyOn(window.localStorage, 'getItem').mockImplementation((key) =>
|
||||
key === 'unitSystem' ? 'metric' : null,
|
||||
);
|
||||
const { result } = renderHook(() => useAppInitialization(), { wrapper });
|
||||
await waitFor(() => {
|
||||
expect(result.current.unitSystem).toBe('metric');
|
||||
});
|
||||
});
|
||||
});
|
||||
88
src/hooks/useAppInitialization.ts
Normal file
88
src/hooks/useAppInitialization.ts
Normal file
@@ -0,0 +1,88 @@
|
||||
// src/hooks/useAppInitialization.ts
|
||||
import { useState, useEffect } from 'react';
|
||||
import { useLocation, useNavigate } from 'react-router-dom';
|
||||
import { useAuth } from './useAuth';
|
||||
import { useModal } from './useModal';
|
||||
import { logger } from '../services/logger.client';
|
||||
import config from '../config';
|
||||
|
||||
export const useAppInitialization = () => {
|
||||
const { userProfile, login } = useAuth();
|
||||
const { openModal } = useModal();
|
||||
const location = useLocation();
|
||||
const navigate = useNavigate();
|
||||
|
||||
const [isDarkMode, setIsDarkMode] = useState(false);
|
||||
const [unitSystem, setUnitSystem] = useState<'metric' | 'imperial'>('imperial');
|
||||
|
||||
// Effect to handle the token from Google/GitHub OAuth redirect
|
||||
useEffect(() => {
|
||||
const urlParams = new URLSearchParams(location.search);
|
||||
const googleToken = urlParams.get('googleAuthToken');
|
||||
|
||||
if (googleToken) {
|
||||
logger.info('Received Google Auth token from URL. Authenticating...');
|
||||
login(googleToken).catch((err) =>
|
||||
logger.error('Failed to log in with Google token', { error: err }),
|
||||
);
|
||||
navigate(location.pathname, { replace: true });
|
||||
}
|
||||
|
||||
const githubToken = urlParams.get('githubAuthToken');
|
||||
if (githubToken) {
|
||||
logger.info('Received GitHub Auth token from URL. Authenticating...');
|
||||
login(githubToken).catch((err) => {
|
||||
logger.error('Failed to log in with GitHub token', { error: err });
|
||||
});
|
||||
navigate(location.pathname, { replace: true });
|
||||
}
|
||||
}, [login, location.search, navigate, location.pathname]);
|
||||
|
||||
// Effect to handle "What's New" modal
|
||||
useEffect(() => {
|
||||
const appVersion = config.app.version;
|
||||
if (appVersion) {
|
||||
logger.info(`Application version: ${appVersion}`);
|
||||
const lastSeenVersion = localStorage.getItem('lastSeenVersion');
|
||||
if (appVersion !== lastSeenVersion) {
|
||||
openModal('whatsNew');
|
||||
localStorage.setItem('lastSeenVersion', appVersion);
|
||||
}
|
||||
}
|
||||
}, [openModal]);
|
||||
|
||||
// Effect to set initial theme based on user profile, local storage, or system preference
|
||||
useEffect(() => {
|
||||
let darkModeValue: boolean;
|
||||
if (userProfile && userProfile.preferences?.darkMode !== undefined) {
|
||||
// Preference from DB
|
||||
darkModeValue = userProfile.preferences.darkMode;
|
||||
} else {
|
||||
// Fallback to local storage or system preference
|
||||
const savedMode = localStorage.getItem('darkMode');
|
||||
const prefersDark = window.matchMedia('(prefers-color-scheme: dark)').matches;
|
||||
darkModeValue = savedMode !== null ? savedMode === 'true' : prefersDark;
|
||||
}
|
||||
setIsDarkMode(darkModeValue);
|
||||
document.documentElement.classList.toggle('dark', darkModeValue);
|
||||
// Also save to local storage if coming from profile, to persist on logout
|
||||
if (userProfile && userProfile.preferences?.darkMode !== undefined) {
|
||||
localStorage.setItem('darkMode', String(userProfile.preferences.darkMode));
|
||||
}
|
||||
}, [userProfile]);
|
||||
|
||||
// Effect to set initial unit system based on user profile or local storage
|
||||
useEffect(() => {
|
||||
if (userProfile && userProfile.preferences?.unitSystem) {
|
||||
setUnitSystem(userProfile.preferences.unitSystem);
|
||||
localStorage.setItem('unitSystem', userProfile.preferences.unitSystem);
|
||||
} else {
|
||||
const savedSystem = localStorage.getItem('unitSystem') as 'metric' | 'imperial' | null;
|
||||
if (savedSystem) {
|
||||
setUnitSystem(savedSystem);
|
||||
}
|
||||
}
|
||||
}, [userProfile?.preferences?.unitSystem, userProfile?.user.user_id]);
|
||||
|
||||
return { isDarkMode, unitSystem };
|
||||
};
|
||||
@@ -6,24 +6,28 @@ import { useAuth } from './useAuth';
|
||||
import { AuthProvider } from '../providers/AuthProvider';
|
||||
import * as apiClient from '../services/apiClient';
|
||||
import type { UserProfile } from '../types';
|
||||
import * as tokenStorage from '../services/tokenStorage';
|
||||
import { createMockUserProfile } from '../tests/utils/mockFactories';
|
||||
import { logger } from '../services/logger.client';
|
||||
|
||||
// Mock the dependencies
|
||||
vi.mock('../services/apiClient', () => ({
|
||||
// Mock other functions if needed
|
||||
getAuthenticatedUserProfile: vi.fn(),
|
||||
}));
|
||||
vi.mock('../services/tokenStorage');
|
||||
|
||||
// Mock the logger to see auth provider logs during test execution
|
||||
// Mock the logger to spy on its methods
|
||||
vi.mock('../services/logger.client', () => ({
|
||||
logger: {
|
||||
info: vi.fn((...args) => console.log('[AUTH-INFO]', ...args)),
|
||||
warn: vi.fn((...args) => console.warn('[AUTH-WARN]', ...args)),
|
||||
error: vi.fn((...args) => console.error('[AUTH-ERROR]', ...args)),
|
||||
info: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
error: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
const mockedApiClient = vi.mocked(apiClient);
|
||||
const mockedTokenStorage = vi.mocked(tokenStorage);
|
||||
|
||||
const mockProfile: UserProfile = createMockUserProfile({
|
||||
full_name: 'Test User',
|
||||
@@ -36,26 +40,9 @@ const mockProfile: UserProfile = createMockUserProfile({
|
||||
const wrapper = ({ children }: { children: ReactNode }) => <AuthProvider>{children}</AuthProvider>;
|
||||
|
||||
describe('useAuth Hook and AuthProvider', () => {
|
||||
// Mock localStorage
|
||||
let storage: { [key: string]: string } = {};
|
||||
const localStorageMock = {
|
||||
getItem: vi.fn((key: string) => storage[key] || null),
|
||||
setItem: vi.fn((key: string, value: string) => {
|
||||
storage[key] = value;
|
||||
}),
|
||||
removeItem: vi.fn((key: string) => {
|
||||
delete storage[key];
|
||||
}),
|
||||
clear: vi.fn(() => {
|
||||
storage = {};
|
||||
}),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset mocks and storage before each test
|
||||
vi.clearAllMocks();
|
||||
storage = {};
|
||||
Object.defineProperty(window, 'localStorage', { value: localStorageMock, configurable: true });
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
@@ -85,7 +72,8 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
});
|
||||
|
||||
describe('Initial Auth Check (useEffect)', () => {
|
||||
it('sets state to SIGNED_OUT if no token is found', async () => {
|
||||
it('sets state to SIGNED_OUT if no token is found in storage', async () => {
|
||||
mockedTokenStorage.getToken.mockReturnValue(null);
|
||||
const { result } = renderHook(() => useAuth(), { wrapper });
|
||||
|
||||
await waitFor(() => {
|
||||
@@ -97,7 +85,7 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
});
|
||||
|
||||
it('sets state to AUTHENTICATED if a valid token is found', async () => {
|
||||
localStorageMock.setItem('authToken', 'valid-token');
|
||||
mockedTokenStorage.getToken.mockReturnValue('valid-token');
|
||||
mockedApiClient.getAuthenticatedUserProfile.mockResolvedValue({
|
||||
ok: true,
|
||||
status: 200,
|
||||
@@ -121,7 +109,7 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
});
|
||||
|
||||
it('sets state to SIGNED_OUT and removes token if validation fails', async () => {
|
||||
localStorageMock.setItem('authToken', 'invalid-token');
|
||||
mockedTokenStorage.getToken.mockReturnValue('invalid-token');
|
||||
mockedApiClient.getAuthenticatedUserProfile.mockRejectedValue(new Error('Invalid token'));
|
||||
|
||||
const { result } = renderHook(() => useAuth(), { wrapper });
|
||||
@@ -132,13 +120,40 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
|
||||
expect(result.current.authStatus).toBe('SIGNED_OUT');
|
||||
expect(result.current.userProfile).toBeNull();
|
||||
expect(localStorageMock.removeItem).toHaveBeenCalledWith('authToken');
|
||||
expect(mockedTokenStorage.removeToken).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
it('sets state to SIGNED_OUT and removes token if profile fetch returns null after token validation', async () => {
|
||||
mockedTokenStorage.getToken.mockReturnValue('valid-token');
|
||||
// Mock getAuthenticatedUserProfile to return a 200 OK response with a null body
|
||||
mockedApiClient.getAuthenticatedUserProfile.mockResolvedValue({
|
||||
ok: true,
|
||||
status: 200,
|
||||
json: () => Promise.resolve(null), // Simulate API returning no profile data
|
||||
} as unknown as Response);
|
||||
|
||||
const { result } = renderHook(() => useAuth(), { wrapper });
|
||||
|
||||
await waitFor(() => {
|
||||
expect(result.current.isLoading).toBe(false);
|
||||
});
|
||||
|
||||
expect(result.current.authStatus).toBe('SIGNED_OUT');
|
||||
expect(result.current.userProfile).toBeNull();
|
||||
expect(mockedTokenStorage.removeToken).toHaveBeenCalled();
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
'[AuthProvider-Effect] Token was present but validation returned no profile. Signing out.',
|
||||
);
|
||||
});
|
||||
|
||||
describe('login function', () => {
|
||||
// This was the failing test
|
||||
it('sets token, fetches profile, and updates state on successful login', async () => {
|
||||
// --- FIX ---
|
||||
// Explicitly mock that no token exists initially to prevent state leakage from other tests.
|
||||
mockedTokenStorage.getToken.mockReturnValue(null);
|
||||
|
||||
// --- FIX ---
|
||||
// The mock for `getAuthenticatedUserProfile` must resolve to a `Response`-like object,
|
||||
// as this is the return type of the actual function. The `useApi` hook then
|
||||
@@ -172,7 +187,7 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
console.log('[TEST-DEBUG] State immediately after login `act` call:', result.current);
|
||||
|
||||
// 3. Assertions
|
||||
expect(localStorageMock.setItem).toHaveBeenCalledWith('authToken', 'new-valid-token');
|
||||
expect(mockedTokenStorage.setToken).toHaveBeenCalledWith('new-valid-token');
|
||||
|
||||
// 4. We must wait for the state update inside the hook to propagate
|
||||
await waitFor(() => {
|
||||
@@ -202,16 +217,44 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
});
|
||||
|
||||
// Should trigger the logout flow
|
||||
expect(localStorageMock.removeItem).toHaveBeenCalledWith('authToken');
|
||||
expect(mockedTokenStorage.removeToken).toHaveBeenCalled();
|
||||
expect(result.current.authStatus).toBe('SIGNED_OUT'); // This was a duplicate, fixed.
|
||||
expect(result.current.userProfile).toBeNull();
|
||||
});
|
||||
|
||||
it('logs out and throws an error if profile fetch returns null after login (no profileData)', async () => {
|
||||
// Simulate successful token setting, but subsequent profile fetch returns null
|
||||
mockedApiClient.getAuthenticatedUserProfile.mockResolvedValue({
|
||||
ok: true,
|
||||
status: 200,
|
||||
json: () => Promise.resolve(null), // Simulate API returning no profile data
|
||||
} as unknown as Response);
|
||||
|
||||
const { result } = renderHook(() => useAuth(), { wrapper });
|
||||
await waitFor(() => expect(result.current.isLoading).toBe(false));
|
||||
|
||||
// Call login without profileData, forcing a profile fetch
|
||||
await act(async () => {
|
||||
await expect(result.current.login('new-token-no-profile-data')).rejects.toThrow(
|
||||
'Login succeeded, but failed to fetch your data: Received null or undefined profile from API.',
|
||||
);
|
||||
});
|
||||
|
||||
// Should trigger the logout flow
|
||||
expect(mockedTokenStorage.removeToken).toHaveBeenCalled();
|
||||
expect(result.current.authStatus).toBe('SIGNED_OUT');
|
||||
expect(result.current.userProfile).toBeNull();
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.any(String), // The error message
|
||||
expect.objectContaining({ error: 'Received null or undefined profile from API.' }),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('logout function', () => {
|
||||
it('removes token and resets auth state', async () => {
|
||||
// Start in a logged-in state
|
||||
localStorageMock.setItem('authToken', 'valid-token');
|
||||
// Start in a logged-in state by mocking the token storage
|
||||
mockedTokenStorage.getToken.mockReturnValue('valid-token');
|
||||
mockedApiClient.getAuthenticatedUserProfile.mockResolvedValue({
|
||||
ok: true,
|
||||
status: 200,
|
||||
@@ -227,16 +270,15 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
result.current.logout();
|
||||
});
|
||||
|
||||
expect(localStorageMock.removeItem).toHaveBeenCalledWith('authToken');
|
||||
expect(mockedTokenStorage.removeToken).toHaveBeenCalled();
|
||||
expect(result.current.authStatus).toBe('SIGNED_OUT');
|
||||
expect(result.current.userProfile).toBeNull();
|
||||
});
|
||||
});
|
||||
|
||||
describe('updateProfile function', () => {
|
||||
it('merges new data into the existing profile state', async () => {
|
||||
// Start in a logged-in state
|
||||
localStorageMock.setItem('authToken', 'valid-token');
|
||||
it('merges new data into the existing profile state', async () => { // Start in a logged-in state
|
||||
mockedTokenStorage.getToken.mockReturnValue('valid-token');
|
||||
mockedApiClient.getAuthenticatedUserProfile.mockResolvedValue({
|
||||
ok: true,
|
||||
status: 200,
|
||||
@@ -264,6 +306,10 @@ describe('useAuth Hook and AuthProvider', () => {
|
||||
});
|
||||
|
||||
it('should not update profile if user is not authenticated', async () => {
|
||||
// --- FIX ---
|
||||
// Explicitly mock that no token exists initially to prevent state leakage from other tests.
|
||||
mockedTokenStorage.getToken.mockReturnValue(null);
|
||||
|
||||
const { result } = renderHook(() => useAuth(), { wrapper });
|
||||
|
||||
// Wait for initial check to complete
|
||||
|
||||
136
src/hooks/useFlyerUploader.test.tsx
Normal file
136
src/hooks/useFlyerUploader.test.tsx
Normal file
@@ -0,0 +1,136 @@
|
||||
import { renderHook, act, waitFor } from '@testing-library/react';
|
||||
import { QueryClient, QueryClientProvider } from '@tanstack/react-query';
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { useFlyerUploader } from './useFlyerUploader';
|
||||
import * as aiApiClient from '../services/aiApiClient';
|
||||
import * as checksumUtil from '../utils/checksum';
|
||||
|
||||
// Import the actual error class because the module is mocked
|
||||
const { JobFailedError } = await vi.importActual<typeof import('../services/aiApiClient')>(
|
||||
'../services/aiApiClient',
|
||||
);
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('../services/aiApiClient');
|
||||
vi.mock('../utils/checksum');
|
||||
vi.mock('../services/logger.client', () => ({
|
||||
logger: {
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
const mockedAiApiClient = vi.mocked(aiApiClient);
|
||||
const mockedChecksumUtil = vi.mocked(checksumUtil);
|
||||
|
||||
// Helper to wrap the hook with QueryClientProvider, which is required by react-query
|
||||
const createWrapper = () => {
|
||||
const queryClient = new QueryClient({
|
||||
defaultOptions: {
|
||||
queries: {
|
||||
retry: false, // Disable retries for tests for predictable behavior
|
||||
},
|
||||
},
|
||||
});
|
||||
return ({ children }: { children: React.ReactNode }) => (
|
||||
<QueryClientProvider client={queryClient}>{children}</QueryClientProvider>
|
||||
);
|
||||
};
|
||||
|
||||
describe('useFlyerUploader Hook with React Query', () => {
|
||||
beforeEach(() => {
|
||||
vi.resetAllMocks();
|
||||
mockedChecksumUtil.generateFileChecksum.mockResolvedValue('mock-checksum');
|
||||
});
|
||||
|
||||
it('should handle a successful upload and polling flow', async () => {
|
||||
// Arrange
|
||||
const mockJobId = 'job-123';
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: mockJobId });
|
||||
mockedAiApiClient.getJobStatus
|
||||
.mockResolvedValueOnce({
|
||||
// First poll: active
|
||||
id: mockJobId,
|
||||
state: 'active',
|
||||
progress: { message: 'Processing...' },
|
||||
returnValue: null,
|
||||
failedReason: null,
|
||||
} as aiApiClient.JobStatus)
|
||||
.mockResolvedValueOnce({
|
||||
// Second poll: completed
|
||||
id: mockJobId,
|
||||
state: 'completed',
|
||||
progress: { message: 'Complete!' },
|
||||
returnValue: { flyerId: 777 },
|
||||
failedReason: null,
|
||||
} as aiApiClient.JobStatus);
|
||||
|
||||
const { result } = renderHook(() => useFlyerUploader(), { wrapper: createWrapper() });
|
||||
const mockFile = new File([''], 'flyer.pdf');
|
||||
|
||||
// Act
|
||||
await act(async () => {
|
||||
result.current.upload(mockFile);
|
||||
});
|
||||
|
||||
// Assert initial upload state
|
||||
await waitFor(() => expect(result.current.processingState).toBe('polling'));
|
||||
expect(result.current.jobId).toBe(mockJobId);
|
||||
|
||||
// Assert polling state
|
||||
await waitFor(() => expect(result.current.statusMessage).toBe('Processing...'));
|
||||
|
||||
// Assert completed state
|
||||
await waitFor(() => expect(result.current.processingState).toBe('completed'), { timeout: 5000 });
|
||||
expect(result.current.flyerId).toBe(777);
|
||||
});
|
||||
|
||||
it('should handle an upload failure', async () => {
|
||||
// Arrange
|
||||
const uploadError = {
|
||||
status: 409,
|
||||
body: { message: 'Duplicate flyer detected.', flyerId: 99 },
|
||||
};
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockRejectedValue(uploadError);
|
||||
|
||||
const { result } = renderHook(() => useFlyerUploader(), { wrapper: createWrapper() });
|
||||
const mockFile = new File([''], 'flyer.pdf');
|
||||
|
||||
// Act
|
||||
await act(async () => {
|
||||
result.current.upload(mockFile);
|
||||
});
|
||||
|
||||
// Assert error state
|
||||
await waitFor(() => expect(result.current.processingState).toBe('error'));
|
||||
expect(result.current.errorMessage).toBe('Duplicate flyer detected.');
|
||||
expect(result.current.duplicateFlyerId).toBe(99);
|
||||
expect(result.current.jobId).toBeNull();
|
||||
});
|
||||
|
||||
it('should handle a job failure during polling', async () => {
|
||||
// Arrange
|
||||
const mockJobId = 'job-456';
|
||||
mockedAiApiClient.uploadAndProcessFlyer.mockResolvedValue({ jobId: mockJobId });
|
||||
|
||||
// Mock getJobStatus to throw a JobFailedError
|
||||
mockedAiApiClient.getJobStatus.mockRejectedValue(
|
||||
new JobFailedError('AI validation failed.', 'AI_VALIDATION_FAILED'),
|
||||
);
|
||||
|
||||
const { result } = renderHook(() => useFlyerUploader(), { wrapper: createWrapper() });
|
||||
const mockFile = new File([''], 'flyer.pdf');
|
||||
|
||||
// Act
|
||||
await act(async () => {
|
||||
result.current.upload(mockFile);
|
||||
});
|
||||
|
||||
// Assert error state after polling fails
|
||||
await waitFor(() => expect(result.current.processingState).toBe('error'));
|
||||
expect(result.current.errorMessage).toBe('Polling failed: AI validation failed.');
|
||||
expect(result.current.flyerId).toBeNull();
|
||||
});
|
||||
});
|
||||
169
src/hooks/useFlyerUploader.ts
Normal file
169
src/hooks/useFlyerUploader.ts
Normal file
@@ -0,0 +1,169 @@
|
||||
// src/hooks/useFlyerUploader.ts
|
||||
// src/hooks/useFlyerUploader.ts
|
||||
import { useState, useCallback, useMemo } from 'react';
|
||||
import { useQuery, useMutation, useQueryClient } from '@tanstack/react-query';
|
||||
import {
|
||||
uploadAndProcessFlyer,
|
||||
getJobStatus,
|
||||
type JobStatus,
|
||||
JobFailedError,
|
||||
} from '../services/aiApiClient';
|
||||
import { logger } from '../services/logger.client';
|
||||
import { generateFileChecksum } from '../utils/checksum';
|
||||
import type { ProcessingStage } from '../types';
|
||||
|
||||
export type ProcessingState = 'idle' | 'uploading' | 'polling' | 'completed' | 'error';
|
||||
|
||||
// Define a type for the structured error thrown by the API client
|
||||
interface ApiError {
|
||||
status: number;
|
||||
body: {
|
||||
message: string;
|
||||
flyerId?: number;
|
||||
};
|
||||
}
|
||||
|
||||
// Type guard to check if an error is a structured API error
|
||||
function isApiError(error: unknown): error is ApiError {
|
||||
return (
|
||||
typeof error === 'object' &&
|
||||
error !== null &&
|
||||
'status' in error &&
|
||||
typeof (error as { status: unknown }).status === 'number' &&
|
||||
'body' in error &&
|
||||
typeof (error as { body: unknown }).body === 'object' &&
|
||||
(error as { body: unknown }).body !== null &&
|
||||
'message' in ((error as { body: unknown }).body as object)
|
||||
);
|
||||
}
|
||||
export const useFlyerUploader = () => {
|
||||
const queryClient = useQueryClient();
|
||||
const [jobId, setJobId] = useState<string | null>(null);
|
||||
const [currentFile, setCurrentFile] = useState<string | null>(null);
|
||||
|
||||
// Mutation for the initial file upload
|
||||
const uploadMutation = useMutation({
|
||||
mutationFn: async (file: File) => {
|
||||
setCurrentFile(file.name);
|
||||
const checksum = await generateFileChecksum(file);
|
||||
return uploadAndProcessFlyer(file, checksum);
|
||||
},
|
||||
onSuccess: (data) => {
|
||||
// When upload is successful, we get a jobId and can start polling.
|
||||
setJobId(data.jobId);
|
||||
},
|
||||
// onError is handled automatically by react-query and exposed in `uploadMutation.error`
|
||||
});
|
||||
|
||||
// Query for polling the job status
|
||||
const { data: jobStatus, error: pollError } = useQuery({
|
||||
queryKey: ['jobStatus', jobId],
|
||||
queryFn: () => {
|
||||
if (!jobId) throw new Error('No job ID to poll');
|
||||
return getJobStatus(jobId);
|
||||
},
|
||||
// Only run this query if there is a jobId
|
||||
enabled: !!jobId,
|
||||
// Polling logic: react-query handles the interval
|
||||
refetchInterval: (query) => {
|
||||
const data = query.state.data as JobStatus | undefined;
|
||||
// Stop polling if the job is completed or has failed
|
||||
if (data?.state === 'completed' || data?.state === 'failed') {
|
||||
return false;
|
||||
}
|
||||
// Also stop polling if the query itself has errored (e.g. network error, or JobFailedError thrown from getJobStatus)
|
||||
if (query.state.status === 'error') {
|
||||
logger.warn('[useFlyerUploader] Polling stopped due to query error state.');
|
||||
return false;
|
||||
}
|
||||
// Otherwise, poll every 3 seconds
|
||||
return 3000;
|
||||
},
|
||||
refetchOnWindowFocus: false, // No need to refetch on focus, interval is enough
|
||||
// If a poll fails (e.g., network error), don't retry automatically.
|
||||
// The user can see the error and choose to retry manually if we build that feature.
|
||||
retry: false,
|
||||
});
|
||||
|
||||
const upload = useCallback(
|
||||
(file: File) => {
|
||||
// Reset previous state before a new upload
|
||||
setJobId(null);
|
||||
setCurrentFile(null);
|
||||
queryClient.removeQueries({ queryKey: ['jobStatus'] });
|
||||
uploadMutation.mutate(file);
|
||||
},
|
||||
[uploadMutation, queryClient],
|
||||
);
|
||||
|
||||
const resetUploaderState = useCallback(() => {
|
||||
setJobId(null);
|
||||
setCurrentFile(null);
|
||||
uploadMutation.reset();
|
||||
queryClient.removeQueries({ queryKey: ['jobStatus'] });
|
||||
}, [uploadMutation, queryClient]);
|
||||
|
||||
// Consolidate state derivation for the UI from the react-query hooks using useMemo.
|
||||
// This improves performance by memoizing the derived state and makes the logic easier to follow.
|
||||
const { processingState, errorMessage, duplicateFlyerId, flyerId, statusMessage } = useMemo(() => {
|
||||
const state: ProcessingState = (() => {
|
||||
if (uploadMutation.isPending) return 'uploading';
|
||||
if (jobStatus && (jobStatus.state === 'active' || jobStatus.state === 'waiting'))
|
||||
return 'polling';
|
||||
if (jobStatus?.state === 'completed') {
|
||||
if (!jobStatus.returnValue?.flyerId) return 'error';
|
||||
return 'completed';
|
||||
}
|
||||
if (uploadMutation.isError || jobStatus?.state === 'failed' || pollError) return 'error';
|
||||
return 'idle';
|
||||
})();
|
||||
|
||||
let msg: string | null = null;
|
||||
let dupId: number | null = null;
|
||||
|
||||
if (state === 'error') {
|
||||
if (uploadMutation.isError) {
|
||||
const uploadError = uploadMutation.error;
|
||||
if (isApiError(uploadError)) {
|
||||
msg = uploadError.body.message;
|
||||
// Specifically handle 409 Conflict for duplicate flyers
|
||||
if (uploadError.status === 409) {
|
||||
dupId = uploadError.body.flyerId ?? null;
|
||||
}
|
||||
} else if (uploadError instanceof Error) {
|
||||
msg = uploadError.message;
|
||||
} else {
|
||||
msg = 'An unknown upload error occurred.';
|
||||
}
|
||||
} else if (pollError) {
|
||||
msg = `Polling failed: ${pollError.message}`;
|
||||
} else if (jobStatus?.state === 'failed') {
|
||||
msg = `Processing failed: ${jobStatus.progress?.message || jobStatus.failedReason || 'Unknown reason'}`;
|
||||
} else if (jobStatus?.state === 'completed' && !jobStatus.returnValue?.flyerId) {
|
||||
msg = 'Job completed but did not return a flyer ID.';
|
||||
}
|
||||
}
|
||||
|
||||
return {
|
||||
processingState: state,
|
||||
errorMessage: msg,
|
||||
duplicateFlyerId: dupId,
|
||||
flyerId: jobStatus?.state === 'completed' ? jobStatus.returnValue?.flyerId ?? null : null,
|
||||
statusMessage: uploadMutation.isPending ? 'Uploading file...' : jobStatus?.progress?.message,
|
||||
};
|
||||
}, [uploadMutation, jobStatus, pollError]);
|
||||
|
||||
return {
|
||||
processingState,
|
||||
statusMessage: uploadMutation.isPending ? 'Uploading file...' : jobStatus?.progress?.message,
|
||||
errorMessage,
|
||||
duplicateFlyerId,
|
||||
processingStages: jobStatus?.progress?.stages || [],
|
||||
estimatedTime: jobStatus?.progress?.estimatedTimeRemaining || 0,
|
||||
currentFile,
|
||||
flyerId,
|
||||
upload,
|
||||
resetUploaderState,
|
||||
jobId,
|
||||
};
|
||||
};
|
||||
@@ -47,6 +47,7 @@ export function useInfiniteQuery<T>(
|
||||
|
||||
// Use a ref to store the cursor for the next page.
|
||||
const nextCursorRef = useRef<number | string | null | undefined>(initialCursor);
|
||||
const lastErrorMessageRef = useRef<string | null>(null);
|
||||
|
||||
const fetchPage = useCallback(
|
||||
async (cursor?: number | string | null) => {
|
||||
@@ -59,6 +60,7 @@ export function useInfiniteQuery<T>(
|
||||
setIsFetchingNextPage(true);
|
||||
}
|
||||
setError(null);
|
||||
lastErrorMessageRef.current = null;
|
||||
|
||||
try {
|
||||
const response = await apiFunction(cursor);
|
||||
@@ -99,7 +101,10 @@ export function useInfiniteQuery<T>(
|
||||
error: err.message,
|
||||
functionName: apiFunction.name,
|
||||
});
|
||||
setError(err);
|
||||
if (err.message !== lastErrorMessageRef.current) {
|
||||
setError(err);
|
||||
lastErrorMessageRef.current = err.message;
|
||||
}
|
||||
notifyError(err.message);
|
||||
} finally {
|
||||
setIsLoading(false);
|
||||
@@ -125,6 +130,7 @@ export function useInfiniteQuery<T>(
|
||||
// Function to be called by the UI to refetch the entire query from the beginning.
|
||||
const refetch = useCallback(() => {
|
||||
setIsRefetching(true);
|
||||
lastErrorMessageRef.current = null;
|
||||
setData([]);
|
||||
fetchPage(initialCursor);
|
||||
}, [fetchPage, initialCursor]);
|
||||
|
||||
@@ -495,6 +495,22 @@ describe('useShoppingLists Hook', () => {
|
||||
expect(currentLists[0].items).toHaveLength(1); // Length should remain 1
|
||||
console.log(' LOG: SUCCESS! Duplicate was not added and API was not called.');
|
||||
});
|
||||
|
||||
it('should log an error and not call the API if the listId does not exist', async () => {
|
||||
const consoleErrorSpy = vi.spyOn(console, 'error').mockImplementation(() => {});
|
||||
const { result } = renderHook(() => useShoppingLists());
|
||||
|
||||
await act(async () => {
|
||||
// Call with a non-existent list ID (mock lists have IDs 1 and 2)
|
||||
await result.current.addItemToList(999, { customItemName: 'Wont be added' });
|
||||
});
|
||||
|
||||
// The API should not have been called because the list was not found.
|
||||
expect(mockAddItemApi).not.toHaveBeenCalled();
|
||||
expect(consoleErrorSpy).toHaveBeenCalledWith('useShoppingLists: List with ID 999 not found.');
|
||||
|
||||
consoleErrorSpy.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
describe('updateItemInList', () => {
|
||||
@@ -656,24 +672,14 @@ describe('useShoppingLists Hook', () => {
|
||||
},
|
||||
{
|
||||
name: 'updateItemInList',
|
||||
action: (hook: any) => {
|
||||
act(() => {
|
||||
hook.setActiveListId(1);
|
||||
});
|
||||
return hook.updateItemInList(101, { is_purchased: true });
|
||||
},
|
||||
action: (hook: any) => hook.updateItemInList(101, { is_purchased: true }),
|
||||
apiMock: mockUpdateItemApi,
|
||||
mockIndex: 3,
|
||||
errorMessage: 'Update failed',
|
||||
},
|
||||
{
|
||||
name: 'removeItemFromList',
|
||||
action: (hook: any) => {
|
||||
act(() => {
|
||||
hook.setActiveListId(1);
|
||||
});
|
||||
return hook.removeItemFromList(101);
|
||||
},
|
||||
action: (hook: any) => hook.removeItemFromList(101),
|
||||
apiMock: mockRemoveItemApi,
|
||||
mockIndex: 4,
|
||||
errorMessage: 'Removal failed',
|
||||
@@ -681,6 +687,17 @@ describe('useShoppingLists Hook', () => {
|
||||
])(
|
||||
'should set an error for $name if the API call fails',
|
||||
async ({ action, apiMock, mockIndex, errorMessage }) => {
|
||||
// Setup a default list so activeListId is set automatically
|
||||
const mockList = createMockShoppingList({ shopping_list_id: 1, name: 'List 1' });
|
||||
mockedUseUserData.mockReturnValue({
|
||||
shoppingLists: [mockList],
|
||||
setShoppingLists: mockSetShoppingLists,
|
||||
watchedItems: [],
|
||||
setWatchedItems: vi.fn(),
|
||||
isLoading: false,
|
||||
error: null,
|
||||
});
|
||||
|
||||
const apiMocksWithError = [...defaultApiMocks];
|
||||
apiMocksWithError[mockIndex] = {
|
||||
...apiMocksWithError[mockIndex],
|
||||
@@ -689,11 +706,25 @@ describe('useShoppingLists Hook', () => {
|
||||
setupApiMocks(apiMocksWithError);
|
||||
apiMock.mockRejectedValue(new Error(errorMessage));
|
||||
|
||||
// Spy on console.error to ensure the catch block is executed for logging
|
||||
const consoleErrorSpy = vi.spyOn(console, 'error').mockImplementation(() => {});
|
||||
|
||||
const { result } = renderHook(() => useShoppingLists());
|
||||
|
||||
// Wait for the effect to set the active list ID
|
||||
await waitFor(() => expect(result.current.activeListId).toBe(1));
|
||||
|
||||
await act(async () => {
|
||||
await action(result.current);
|
||||
});
|
||||
await waitFor(() => expect(result.current.error).toBe(errorMessage));
|
||||
|
||||
await waitFor(() => {
|
||||
expect(result.current.error).toBe(errorMessage);
|
||||
// Verify that our custom logging within the catch block was called
|
||||
expect(consoleErrorSpy).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
consoleErrorSpy.mockRestore();
|
||||
},
|
||||
);
|
||||
});
|
||||
|
||||
@@ -79,7 +79,7 @@ vi.mock('../pages/admin/ActivityLog', async () => {
|
||||
),
|
||||
};
|
||||
});
|
||||
vi.mock('../pages/admin/components/AnonymousUserBanner', () => ({
|
||||
vi.mock('../components/AnonymousUserBanner', () => ({
|
||||
AnonymousUserBanner: () => <div data-testid="anonymous-banner" />,
|
||||
}));
|
||||
vi.mock('../components/ErrorDisplay', () => ({
|
||||
|
||||
@@ -16,7 +16,7 @@ import { PriceChart } from '../features/charts/PriceChart';
|
||||
import { PriceHistoryChart } from '../features/charts/PriceHistoryChart';
|
||||
import Leaderboard from '../components/Leaderboard';
|
||||
import { ActivityLog, ActivityLogClickHandler } from '../pages/admin/ActivityLog';
|
||||
import { AnonymousUserBanner } from '../pages/admin/components/AnonymousUserBanner';
|
||||
import { AnonymousUserBanner } from '../components/AnonymousUserBanner';
|
||||
import { ErrorDisplay } from '../components/ErrorDisplay';
|
||||
|
||||
export interface MainLayoutProps {
|
||||
|
||||
@@ -113,13 +113,14 @@ describe('errorHandler Middleware', () => {
|
||||
expect(response.body.message).toBe('A generic server error occurred.');
|
||||
expect(response.body.stack).toBeDefined();
|
||||
expect(response.body.errorId).toEqual(expect.any(String));
|
||||
console.log('[DEBUG] errorHandler.test.ts: Received 500 error response with ID:', response.body.errorId);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
errorId: expect.any(String),
|
||||
req: expect.objectContaining({ method: 'GET', url: '/generic-error' }),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
expect(consoleErrorSpy).toHaveBeenCalledWith(
|
||||
expect.stringMatching(/--- \[TEST\] UNHANDLED ERROR \(ID: \w+\) ---/),
|
||||
@@ -226,7 +227,7 @@ describe('errorHandler Middleware', () => {
|
||||
errorId: expect.any(String),
|
||||
req: expect.objectContaining({ method: 'GET', url: '/db-error-500' }),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
expect(consoleErrorSpy).toHaveBeenCalledWith(
|
||||
expect.stringMatching(/--- \[TEST\] UNHANDLED ERROR \(ID: \w+\) ---/),
|
||||
|
||||
269
src/middleware/multer.middleware.test.ts
Normal file
269
src/middleware/multer.middleware.test.ts
Normal file
@@ -0,0 +1,269 @@
|
||||
// src/middleware/multer.middleware.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, afterEach, Mock } from 'vitest';
|
||||
import multer from 'multer';
|
||||
import type { Request, Response, NextFunction } from 'express';
|
||||
import { createUploadMiddleware, handleMulterError } from './multer.middleware';
|
||||
import { createMockUserProfile } from '../tests/utils/mockFactories';
|
||||
import { ValidationError } from '../services/db/errors.db';
|
||||
|
||||
// 1. Hoist the mocks so they can be referenced inside vi.mock factories.
|
||||
const mocks = vi.hoisted(() => ({
|
||||
mkdir: vi.fn(),
|
||||
logger: {
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// 2. Mock node:fs/promises.
|
||||
// We mock the default export because that's how it's imported in the source file.
|
||||
vi.mock('node:fs/promises', () => ({
|
||||
default: {
|
||||
mkdir: mocks.mkdir,
|
||||
},
|
||||
}));
|
||||
|
||||
// 3. Mock the logger service.
|
||||
vi.mock('../services/logger.server', () => ({
|
||||
logger: mocks.logger,
|
||||
}));
|
||||
|
||||
// 4. Mock multer to prevent it from doing anything during import.
|
||||
vi.mock('multer', () => {
|
||||
const diskStorage = vi.fn((options) => options);
|
||||
// A more realistic mock for MulterError that maps error codes to messages,
|
||||
// similar to how the actual multer library works.
|
||||
class MulterError extends Error {
|
||||
code: string;
|
||||
field?: string;
|
||||
|
||||
constructor(code: string, field?: string) {
|
||||
const messages: { [key: string]: string } = {
|
||||
LIMIT_FILE_SIZE: 'File too large',
|
||||
LIMIT_UNEXPECTED_FILE: 'Unexpected file',
|
||||
// Add other codes as needed for tests
|
||||
};
|
||||
const message = messages[code] || code;
|
||||
super(message);
|
||||
this.code = code;
|
||||
this.name = 'MulterError';
|
||||
if (field) {
|
||||
this.field = field;
|
||||
}
|
||||
}
|
||||
}
|
||||
const multer = vi.fn(() => ({
|
||||
single: vi.fn().mockImplementation(() => (req: any, res: any, next: any) => next()),
|
||||
array: vi.fn().mockImplementation(() => (req: any, res: any, next: any) => next()),
|
||||
}));
|
||||
(multer as any).diskStorage = diskStorage;
|
||||
(multer as any).MulterError = MulterError;
|
||||
return {
|
||||
default: multer,
|
||||
diskStorage,
|
||||
MulterError,
|
||||
};
|
||||
});
|
||||
|
||||
describe('Multer Middleware Directory Creation', () => {
|
||||
beforeEach(() => {
|
||||
// Critical: Reset modules to ensure the top-level IIFE runs again for each test.
|
||||
vi.resetModules();
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
it('should attempt to create directories on module load and log success', async () => {
|
||||
// Arrange
|
||||
mocks.mkdir.mockResolvedValue(undefined);
|
||||
|
||||
// Act: Dynamic import triggers the top-level code execution
|
||||
await import('./multer.middleware');
|
||||
|
||||
// Assert
|
||||
// It should try to create both the flyer storage and avatar storage paths
|
||||
expect(mocks.mkdir).toHaveBeenCalledTimes(2);
|
||||
expect(mocks.mkdir).toHaveBeenCalledWith(expect.any(String), { recursive: true });
|
||||
expect(mocks.logger.info).toHaveBeenCalledWith('Ensured multer storage directories exist.');
|
||||
expect(mocks.logger.error).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should log an error if directory creation fails', async () => {
|
||||
// Arrange
|
||||
const error = new Error('Permission denied');
|
||||
mocks.mkdir.mockRejectedValue(error);
|
||||
|
||||
// Act
|
||||
await import('./multer.middleware');
|
||||
|
||||
// Assert
|
||||
expect(mocks.mkdir).toHaveBeenCalled();
|
||||
expect(mocks.logger.error).toHaveBeenCalledWith(
|
||||
{ error },
|
||||
'Failed to create multer storage directories on startup.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('createUploadMiddleware', () => {
|
||||
const mockFile = { originalname: 'test.png' } as Express.Multer.File;
|
||||
const mockUser = createMockUserProfile({ user: { user_id: 'user-123', email: 'test@user.com' } });
|
||||
let originalNodeEnv: string | undefined;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
originalNodeEnv = process.env.NODE_ENV;
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
process.env.NODE_ENV = originalNodeEnv;
|
||||
});
|
||||
|
||||
describe('Avatar Storage', () => {
|
||||
it('should generate a unique filename for an authenticated user', () => {
|
||||
process.env.NODE_ENV = 'production';
|
||||
createUploadMiddleware({ storageType: 'avatar' });
|
||||
const storageOptions = vi.mocked(multer.diskStorage).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
const mockReq = { user: mockUser } as unknown as Request;
|
||||
|
||||
storageOptions.filename!(mockReq, mockFile, cb);
|
||||
|
||||
expect(cb).toHaveBeenCalledWith(null, expect.stringContaining('user-123-'));
|
||||
expect(cb).toHaveBeenCalledWith(null, expect.stringContaining('.png'));
|
||||
});
|
||||
|
||||
it('should call the callback with an error for an unauthenticated user', () => {
|
||||
// This test covers line 37
|
||||
createUploadMiddleware({ storageType: 'avatar' });
|
||||
const storageOptions = vi.mocked(multer.diskStorage).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
const mockReq = {} as Request; // No user on request
|
||||
|
||||
storageOptions.filename!(mockReq, mockFile, cb);
|
||||
|
||||
expect(cb).toHaveBeenCalledWith(
|
||||
new Error('User not authenticated for avatar upload'),
|
||||
expect.any(String),
|
||||
);
|
||||
});
|
||||
|
||||
it('should use a predictable filename in test environment', () => {
|
||||
process.env.NODE_ENV = 'test';
|
||||
createUploadMiddleware({ storageType: 'avatar' });
|
||||
const storageOptions = vi.mocked(multer.diskStorage).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
const mockReq = { user: mockUser } as unknown as Request;
|
||||
|
||||
storageOptions.filename!(mockReq, mockFile, cb);
|
||||
|
||||
expect(cb).toHaveBeenCalledWith(null, 'test-avatar.png');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Flyer Storage', () => {
|
||||
it('should generate a unique, sanitized filename in production environment', () => {
|
||||
process.env.NODE_ENV = 'production';
|
||||
const mockFlyerFile = {
|
||||
fieldname: 'flyerFile',
|
||||
originalname: 'My Flyer (Special!).pdf',
|
||||
} as Express.Multer.File;
|
||||
createUploadMiddleware({ storageType: 'flyer' });
|
||||
const storageOptions = vi.mocked(multer.diskStorage).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
const mockReq = {} as Request;
|
||||
|
||||
storageOptions.filename!(mockReq, mockFlyerFile, cb);
|
||||
|
||||
expect(cb).toHaveBeenCalledWith(
|
||||
null,
|
||||
expect.stringMatching(/^flyerFile-\d+-\d+-my-flyer-special\.pdf$/i),
|
||||
);
|
||||
});
|
||||
|
||||
it('should generate a predictable filename in test environment', () => {
|
||||
// This test covers lines 43-46
|
||||
process.env.NODE_ENV = 'test';
|
||||
const mockFlyerFile = {
|
||||
fieldname: 'flyerFile',
|
||||
originalname: 'test-flyer.jpg',
|
||||
} as Express.Multer.File;
|
||||
createUploadMiddleware({ storageType: 'flyer' });
|
||||
const storageOptions = vi.mocked(multer.diskStorage).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
const mockReq = {} as Request;
|
||||
|
||||
storageOptions.filename!(mockReq, mockFlyerFile, cb);
|
||||
|
||||
expect(cb).toHaveBeenCalledWith(null, 'flyerFile-test-flyer-image.jpg');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Image File Filter', () => {
|
||||
it('should accept files with an image mimetype', () => {
|
||||
createUploadMiddleware({ storageType: 'flyer', fileFilter: 'image' });
|
||||
const multerOptions = vi.mocked(multer).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
const mockImageFile = { mimetype: 'image/png' } as Express.Multer.File;
|
||||
|
||||
multerOptions!.fileFilter!({} as Request, mockImageFile, cb);
|
||||
|
||||
expect(cb).toHaveBeenCalledWith(null, true);
|
||||
});
|
||||
|
||||
it('should reject files without an image mimetype', () => {
|
||||
createUploadMiddleware({ storageType: 'flyer', fileFilter: 'image' });
|
||||
const multerOptions = vi.mocked(multer).mock.calls[0][0];
|
||||
const cb = vi.fn();
|
||||
const mockTextFile = { mimetype: 'text/plain' } as Express.Multer.File;
|
||||
|
||||
multerOptions!.fileFilter!({} as Request, { ...mockTextFile, fieldname: 'test' }, cb);
|
||||
|
||||
const error = (cb as Mock).mock.calls[0][0];
|
||||
expect(error).toBeInstanceOf(ValidationError);
|
||||
expect(error.validationErrors[0].message).toBe('Only image files are allowed!');
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('handleMulterError Middleware', () => {
|
||||
let mockRequest: Partial<Request>;
|
||||
let mockResponse: Partial<Response>;
|
||||
let mockNext: NextFunction;
|
||||
|
||||
beforeEach(() => {
|
||||
mockRequest = {};
|
||||
mockResponse = {
|
||||
status: vi.fn().mockReturnThis(),
|
||||
json: vi.fn(),
|
||||
};
|
||||
mockNext = vi.fn();
|
||||
});
|
||||
|
||||
it('should handle a MulterError (e.g., file too large)', () => {
|
||||
const err = new multer.MulterError('LIMIT_FILE_SIZE');
|
||||
handleMulterError(err, mockRequest as Request, mockResponse as Response, mockNext);
|
||||
expect(mockResponse.status).toHaveBeenCalledWith(400);
|
||||
expect(mockResponse.json).toHaveBeenCalledWith({
|
||||
message: 'File upload error: File too large',
|
||||
});
|
||||
expect(mockNext).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should pass on a ValidationError to the next handler', () => {
|
||||
const err = new ValidationError([], 'Only image files are allowed!');
|
||||
handleMulterError(err, mockRequest as Request, mockResponse as Response, mockNext);
|
||||
// It should now pass the error to the global error handler
|
||||
expect(mockNext).toHaveBeenCalledWith(err);
|
||||
expect(mockResponse.status).not.toHaveBeenCalled();
|
||||
expect(mockResponse.json).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should pass on non-multer errors to the next error handler', () => {
|
||||
const err = new Error('A generic error');
|
||||
handleMulterError(err, mockRequest as Request, mockResponse as Response, mockNext);
|
||||
expect(mockNext).toHaveBeenCalledWith(err);
|
||||
expect(mockResponse.status).not.toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
122
src/middleware/multer.middleware.ts
Normal file
122
src/middleware/multer.middleware.ts
Normal file
@@ -0,0 +1,122 @@
|
||||
// src/middleware/multer.middleware.ts
|
||||
import multer from 'multer';
|
||||
import path from 'path';
|
||||
import fs from 'node:fs/promises';
|
||||
import { Request, Response, NextFunction } from 'express';
|
||||
import { UserProfile } from '../types';
|
||||
import { sanitizeFilename } from '../utils/stringUtils';
|
||||
import { ValidationError } from '../services/db/errors.db';
|
||||
import { logger } from '../services/logger.server';
|
||||
|
||||
export const flyerStoragePath =
|
||||
process.env.STORAGE_PATH || '/var/www/flyer-crawler.projectium.com/flyer-images';
|
||||
export const avatarStoragePath = path.join(process.cwd(), 'public', 'uploads', 'avatars');
|
||||
|
||||
// Ensure directories exist at startup
|
||||
(async () => {
|
||||
try {
|
||||
await fs.mkdir(flyerStoragePath, { recursive: true });
|
||||
await fs.mkdir(avatarStoragePath, { recursive: true });
|
||||
logger.info('Ensured multer storage directories exist.');
|
||||
} catch (error) {
|
||||
const err = error instanceof Error ? error : new Error(String(error));
|
||||
logger.error({ error: err }, 'Failed to create multer storage directories on startup.');
|
||||
}
|
||||
})();
|
||||
|
||||
type StorageType = 'flyer' | 'avatar';
|
||||
|
||||
const getStorageConfig = (type: StorageType) => {
|
||||
switch (type) {
|
||||
case 'avatar':
|
||||
return multer.diskStorage({
|
||||
destination: (req, file, cb) => cb(null, avatarStoragePath),
|
||||
filename: (req, file, cb) => {
|
||||
const user = req.user as UserProfile | undefined;
|
||||
if (!user) {
|
||||
// This should ideally not happen if auth middleware runs first.
|
||||
return cb(new Error('User not authenticated for avatar upload'), '');
|
||||
}
|
||||
if (process.env.NODE_ENV === 'test') {
|
||||
// Use a predictable filename for test avatars for easy cleanup.
|
||||
return cb(null, `test-avatar${path.extname(file.originalname) || '.png'}`);
|
||||
}
|
||||
const uniqueSuffix = `${user.user.user_id}-${Date.now()}${path.extname(
|
||||
file.originalname,
|
||||
)}`;
|
||||
cb(null, uniqueSuffix);
|
||||
},
|
||||
});
|
||||
case 'flyer':
|
||||
default:
|
||||
return multer.diskStorage({
|
||||
destination: (req, file, cb) => cb(null, flyerStoragePath),
|
||||
filename: (req, file, cb) => {
|
||||
if (process.env.NODE_ENV === 'test') {
|
||||
// Use a predictable filename for test flyers for easy cleanup.
|
||||
const ext = path.extname(file.originalname);
|
||||
return cb(null, `${file.fieldname}-test-flyer-image${ext || '.jpg'}`);
|
||||
}
|
||||
const uniqueSuffix = `${Date.now()}-${Math.round(Math.random() * 1e9)}`;
|
||||
const sanitizedOriginalName = sanitizeFilename(file.originalname);
|
||||
cb(null, `${file.fieldname}-${uniqueSuffix}-${sanitizedOriginalName}`);
|
||||
},
|
||||
});
|
||||
}
|
||||
};
|
||||
|
||||
const imageFileFilter = (req: Request, file: Express.Multer.File, cb: multer.FileFilterCallback) => {
|
||||
if (file.mimetype.startsWith('image/')) {
|
||||
cb(null, true);
|
||||
} else {
|
||||
// Reject the file with a specific error that can be caught by a middleware.
|
||||
const validationIssue = { path: ['file', file.fieldname], message: 'Only image files are allowed!' };
|
||||
const err = new ValidationError([validationIssue], 'Only image files are allowed!');
|
||||
cb(err as Error); // Cast to Error to satisfy multer's type, though ValidationError extends Error.
|
||||
}
|
||||
};
|
||||
|
||||
interface MulterOptions {
|
||||
storageType: StorageType;
|
||||
fileSize?: number;
|
||||
fileFilter?: 'image';
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a configured multer instance for file uploads.
|
||||
* @param options - Configuration for storage type, file size, and file filter.
|
||||
* @returns A multer instance.
|
||||
*/
|
||||
export const createUploadMiddleware = (options: MulterOptions) => {
|
||||
const multerOptions: multer.Options = {
|
||||
storage: getStorageConfig(options.storageType),
|
||||
};
|
||||
|
||||
if (options.fileSize) {
|
||||
multerOptions.limits = { fileSize: options.fileSize };
|
||||
}
|
||||
|
||||
if (options.fileFilter === 'image') {
|
||||
multerOptions.fileFilter = imageFileFilter;
|
||||
}
|
||||
|
||||
return multer(multerOptions);
|
||||
};
|
||||
|
||||
/**
|
||||
* A general error handler for multer. Place this after all routes using multer in your router file.
|
||||
* It catches errors from `fileFilter` and other multer issues (e.g., file size limits).
|
||||
*/
|
||||
export const handleMulterError = (
|
||||
err: Error,
|
||||
req: Request,
|
||||
res: Response,
|
||||
next: NextFunction,
|
||||
) => {
|
||||
if (err instanceof multer.MulterError) {
|
||||
// A Multer error occurred when uploading (e.g., file too large).
|
||||
return res.status(400).json({ message: `File upload error: ${err.message}` });
|
||||
}
|
||||
// If it's not a multer error, pass it on.
|
||||
next(err);
|
||||
};
|
||||
@@ -4,7 +4,7 @@ import { useParams, useNavigate, Link } from 'react-router-dom';
|
||||
import * as apiClient from '../services/apiClient';
|
||||
import { logger } from '../services/logger.client';
|
||||
import { LoadingSpinner } from '../components/LoadingSpinner';
|
||||
import { PasswordInput } from './admin/components/PasswordInput';
|
||||
import { PasswordInput } from '../components/PasswordInput';
|
||||
|
||||
export const ResetPasswordPage: React.FC = () => {
|
||||
const { token } = useParams<{ token: string }>();
|
||||
|
||||
@@ -4,6 +4,7 @@ import { SystemCheck } from './components/SystemCheck';
|
||||
import { Link } from 'react-router-dom';
|
||||
import { ShieldExclamationIcon } from '../../components/icons/ShieldExclamationIcon';
|
||||
import { ChartBarIcon } from '../../components/icons/ChartBarIcon';
|
||||
import { DocumentMagnifyingGlassIcon } from '../../components/icons/DocumentMagnifyingGlassIcon';
|
||||
|
||||
export const AdminPage: React.FC = () => {
|
||||
// The onReady prop for SystemCheck is present to allow for future UI changes,
|
||||
@@ -39,6 +40,13 @@ export const AdminPage: React.FC = () => {
|
||||
<ChartBarIcon className="w-6 h-6 mr-3 text-brand-primary" />
|
||||
<span className="font-semibold">View Statistics</span>
|
||||
</Link>
|
||||
<Link
|
||||
to="/admin/flyer-review"
|
||||
className="flex items-center p-3 rounded-lg hover:bg-gray-100 dark:hover:bg-gray-700/50 transition-colors"
|
||||
>
|
||||
<DocumentMagnifyingGlassIcon className="w-6 h-6 mr-3 text-brand-primary" />
|
||||
<span className="font-semibold">Flyer Review Queue</span>
|
||||
</Link>
|
||||
</div>
|
||||
</div>
|
||||
<SystemCheck />
|
||||
|
||||
179
src/pages/admin/FlyerReviewPage.test.tsx
Normal file
179
src/pages/admin/FlyerReviewPage.test.tsx
Normal file
@@ -0,0 +1,179 @@
|
||||
// src/pages/admin/FlyerReviewPage.test.tsx
|
||||
import { render, screen, waitFor, within } from '@testing-library/react';
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { FlyerReviewPage } from './FlyerReviewPage';
|
||||
import { MemoryRouter } from 'react-router-dom';
|
||||
import * as apiClient from '../../services/apiClient';
|
||||
import { logger } from '../../services/logger.client';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('../../services/apiClient', () => ({
|
||||
getFlyersForReview: vi.fn(),
|
||||
}));
|
||||
|
||||
vi.mock('../../services/logger.client', () => ({
|
||||
logger: {
|
||||
error: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock LoadingSpinner to simplify DOM and avoid potential issues
|
||||
vi.mock('../../components/LoadingSpinner', () => ({
|
||||
LoadingSpinner: () => <div data-testid="loading-spinner">Loading...</div>,
|
||||
}));
|
||||
|
||||
describe('FlyerReviewPage', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
it('renders loading spinner initially', () => {
|
||||
// Mock a promise that doesn't resolve immediately to check loading state
|
||||
vi.mocked(apiClient.getFlyersForReview).mockReturnValue(new Promise(() => {}));
|
||||
|
||||
render(
|
||||
<MemoryRouter>
|
||||
<FlyerReviewPage />
|
||||
</MemoryRouter>
|
||||
);
|
||||
|
||||
expect(screen.getByRole('status', { name: /loading flyers for review/i })).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('renders empty state when no flyers are returned', async () => {
|
||||
vi.mocked(apiClient.getFlyersForReview).mockResolvedValue({
|
||||
ok: true,
|
||||
json: async () => [],
|
||||
} as Response);
|
||||
|
||||
render(
|
||||
<MemoryRouter>
|
||||
<FlyerReviewPage />
|
||||
</MemoryRouter>
|
||||
);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.queryByRole('status')).not.toBeInTheDocument();
|
||||
});
|
||||
|
||||
expect(screen.getByText(/the review queue is empty/i)).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('renders a list of flyers when API returns data', async () => {
|
||||
const mockFlyers = [
|
||||
{
|
||||
flyer_id: 1,
|
||||
file_name: 'flyer1.jpg',
|
||||
created_at: '2023-01-01T00:00:00Z',
|
||||
store: { name: 'Store A' },
|
||||
icon_url: 'icon1.jpg',
|
||||
},
|
||||
{
|
||||
flyer_id: 2,
|
||||
file_name: 'flyer2.jpg',
|
||||
created_at: '2023-01-02T00:00:00Z',
|
||||
store: { name: 'Store B' },
|
||||
icon_url: 'icon2.jpg',
|
||||
},
|
||||
{
|
||||
flyer_id: 3,
|
||||
file_name: 'flyer3.jpg',
|
||||
created_at: '2023-01-03T00:00:00Z',
|
||||
store: null,
|
||||
icon_url: null,
|
||||
},
|
||||
];
|
||||
|
||||
vi.mocked(apiClient.getFlyersForReview).mockResolvedValue({
|
||||
ok: true,
|
||||
json: async () => mockFlyers,
|
||||
} as Response);
|
||||
|
||||
render(
|
||||
<MemoryRouter>
|
||||
<FlyerReviewPage />
|
||||
</MemoryRouter>
|
||||
);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.queryByRole('status')).not.toBeInTheDocument();
|
||||
});
|
||||
|
||||
expect(screen.getByText('Store A')).toBeInTheDocument();
|
||||
expect(screen.getByText('flyer1.jpg')).toBeInTheDocument();
|
||||
expect(screen.getByText('Store B')).toBeInTheDocument();
|
||||
expect(screen.getByText('flyer2.jpg')).toBeInTheDocument();
|
||||
|
||||
// Test fallback for null store and icon_url
|
||||
expect(screen.getByText('Unknown Store')).toBeInTheDocument();
|
||||
expect(screen.getByText('flyer3.jpg')).toBeInTheDocument();
|
||||
const unknownStoreItem = screen.getByText('Unknown Store').closest('li');
|
||||
const unknownStoreImage = within(unknownStoreItem!).getByRole('img');
|
||||
expect(unknownStoreImage).not.toHaveAttribute('src');
|
||||
expect(unknownStoreImage).not.toHaveAttribute('alt');
|
||||
});
|
||||
|
||||
it('renders error message when API response is not ok', async () => {
|
||||
vi.mocked(apiClient.getFlyersForReview).mockResolvedValue({
|
||||
ok: false,
|
||||
json: async () => ({ message: 'Server error' }),
|
||||
} as Response);
|
||||
|
||||
render(
|
||||
<MemoryRouter>
|
||||
<FlyerReviewPage />
|
||||
</MemoryRouter>
|
||||
);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.queryByRole('status')).not.toBeInTheDocument();
|
||||
});
|
||||
|
||||
expect(screen.getByText('Server error')).toBeInTheDocument();
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ err: expect.any(Error) }),
|
||||
'Failed to fetch flyers for review'
|
||||
);
|
||||
});
|
||||
|
||||
it('renders error message when API throws an error', async () => {
|
||||
const networkError = new Error('Network error');
|
||||
vi.mocked(apiClient.getFlyersForReview).mockRejectedValue(networkError);
|
||||
|
||||
render(
|
||||
<MemoryRouter>
|
||||
<FlyerReviewPage />
|
||||
</MemoryRouter>
|
||||
);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.queryByRole('status')).not.toBeInTheDocument();
|
||||
});
|
||||
|
||||
expect(screen.getByText('Network error')).toBeInTheDocument();
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: networkError },
|
||||
'Failed to fetch flyers for review'
|
||||
);
|
||||
});
|
||||
|
||||
it('renders a generic error for non-Error rejections', async () => {
|
||||
const nonErrorRejection = { message: 'This is not an Error object' };
|
||||
vi.mocked(apiClient.getFlyersForReview).mockRejectedValue(nonErrorRejection);
|
||||
|
||||
render(
|
||||
<MemoryRouter>
|
||||
<FlyerReviewPage />
|
||||
</MemoryRouter>,
|
||||
);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(screen.getByText('An unknown error occurred while fetching data.')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: nonErrorRejection },
|
||||
'Failed to fetch flyers for review',
|
||||
);
|
||||
});
|
||||
});
|
||||
93
src/pages/admin/FlyerReviewPage.tsx
Normal file
93
src/pages/admin/FlyerReviewPage.tsx
Normal file
@@ -0,0 +1,93 @@
|
||||
// src/pages/admin/FlyerReviewPage.tsx
|
||||
import React, { useEffect, useState } from 'react';
|
||||
import { Link } from 'react-router-dom';
|
||||
import { getFlyersForReview } from '../../services/apiClient';
|
||||
import { logger } from '../../services/logger.client';
|
||||
import type { Flyer } from '../../types';
|
||||
import { LoadingSpinner } from '../../components/LoadingSpinner';
|
||||
import { format } from 'date-fns';
|
||||
|
||||
export const FlyerReviewPage: React.FC = () => {
|
||||
const [flyers, setFlyers] = useState<Flyer[]>([]);
|
||||
const [isLoading, setIsLoading] = useState(true);
|
||||
const [error, setError] = useState<string | null>(null);
|
||||
|
||||
useEffect(() => {
|
||||
const fetchFlyers = async () => {
|
||||
setIsLoading(true);
|
||||
setError(null);
|
||||
try {
|
||||
const response = await getFlyersForReview();
|
||||
if (!response.ok) {
|
||||
throw new Error((await response.json()).message || 'Failed to fetch flyers for review.');
|
||||
}
|
||||
setFlyers(await response.json());
|
||||
} catch (err) {
|
||||
const errorMessage =
|
||||
err instanceof Error ? err.message : 'An unknown error occurred while fetching data.';
|
||||
logger.error({ err }, 'Failed to fetch flyers for review');
|
||||
setError(errorMessage);
|
||||
} finally {
|
||||
setIsLoading(false);
|
||||
}
|
||||
};
|
||||
|
||||
fetchFlyers();
|
||||
}, []);
|
||||
|
||||
return (
|
||||
<div className="max-w-7xl mx-auto py-8 px-4">
|
||||
<div className="mb-8">
|
||||
<Link to="/admin" className="text-brand-primary hover:underline">
|
||||
← Back to Admin Dashboard
|
||||
</Link>
|
||||
<h1 className="text-3xl font-bold text-gray-800 dark:text-white mt-2">
|
||||
Flyer Review Queue
|
||||
</h1>
|
||||
<p className="text-gray-500 dark:text-gray-400">
|
||||
Review flyers that were processed with low confidence by the AI.
|
||||
</p>
|
||||
</div>
|
||||
|
||||
{isLoading && (
|
||||
<div
|
||||
role="status"
|
||||
aria-label="Loading flyers for review"
|
||||
className="flex justify-center items-center h-64"
|
||||
>
|
||||
<LoadingSpinner />
|
||||
</div>
|
||||
)}
|
||||
{error && (
|
||||
<div className="text-red-500 bg-red-100 dark:bg-red-900/20 p-4 rounded-lg">{error}</div>
|
||||
)}
|
||||
|
||||
{!isLoading && !error && (
|
||||
<div className="bg-white dark:bg-gray-800 rounded-lg border border-gray-200 dark:border-gray-700 overflow-hidden">
|
||||
<ul className="divide-y divide-gray-200 dark:divide-gray-700">
|
||||
{flyers.length === 0 ? (
|
||||
<li className="p-6 text-center text-gray-500">
|
||||
The review queue is empty. Great job!
|
||||
</li>
|
||||
) : (
|
||||
flyers.map((flyer) => (
|
||||
<li key={flyer.flyer_id} className="p-4 hover:bg-gray-50 dark:hover:bg-gray-700/50">
|
||||
<Link to={`/flyers/${flyer.flyer_id}`} className="flex items-center space-x-4">
|
||||
<img src={flyer.icon_url || undefined} alt={flyer.store?.name} className="w-12 h-12 rounded-md object-cover" />
|
||||
<div className="flex-1">
|
||||
<p className="font-semibold text-gray-800 dark:text-white">{flyer.store?.name || 'Unknown Store'}</p>
|
||||
<p className="text-sm text-gray-500 dark:text-gray-400">{flyer.file_name}</p>
|
||||
</div>
|
||||
<div className="text-right text-sm text-gray-500 dark:text-gray-400">
|
||||
<p>Uploaded: {format(new Date(flyer.created_at), 'MMM d, yyyy')}</p>
|
||||
</div>
|
||||
</Link>
|
||||
</li>
|
||||
))
|
||||
)}
|
||||
</ul>
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
);
|
||||
};
|
||||
@@ -1,6 +1,6 @@
|
||||
// src/pages/admin/components/AuthView.test.tsx
|
||||
import React from 'react';
|
||||
import { render, screen, fireEvent, waitFor } from '@testing-library/react';
|
||||
import { render, screen, fireEvent, waitFor, act } from '@testing-library/react';
|
||||
import { describe, it, expect, vi, beforeEach, type Mock } from 'vitest';
|
||||
import { AuthView } from './AuthView';
|
||||
import * as apiClient from '../../../services/apiClient';
|
||||
@@ -12,6 +12,11 @@ const mockedApiClient = vi.mocked(apiClient, true);
|
||||
const mockOnClose = vi.fn();
|
||||
const mockOnLoginSuccess = vi.fn();
|
||||
|
||||
vi.mock('../../../components/PasswordInput', () => ({
|
||||
// Mock the moved component
|
||||
PasswordInput: (props: any) => <input {...props} data-testid="password-input" />,
|
||||
}));
|
||||
|
||||
const defaultProps = {
|
||||
onClose: mockOnClose,
|
||||
onLoginSuccess: mockOnLoginSuccess,
|
||||
@@ -353,4 +358,27 @@ describe('AuthView', () => {
|
||||
expect(screen.queryByText('Send Reset Link')).not.toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
|
||||
it('should show loading state during registration submission', async () => {
|
||||
// Mock a promise that doesn't resolve immediately
|
||||
(mockedApiClient.registerUser as Mock).mockReturnValue(new Promise(() => {}));
|
||||
render(<AuthView {...defaultProps} />);
|
||||
|
||||
// Switch to registration view
|
||||
fireEvent.click(screen.getByRole('button', { name: /don't have an account\? register/i }));
|
||||
|
||||
fireEvent.change(screen.getByLabelText(/email address/i), {
|
||||
target: { value: 'test@example.com' },
|
||||
});
|
||||
fireEvent.change(screen.getByTestId('password-input'), { target: { value: 'password' } });
|
||||
fireEvent.submit(screen.getByTestId('auth-form'));
|
||||
|
||||
await waitFor(() => {
|
||||
const submitButton = screen.getByTestId('auth-form').querySelector('button[type="submit"]');
|
||||
expect(submitButton).toBeInTheDocument();
|
||||
expect(submitButton).toBeDisabled();
|
||||
// Verify the text 'Register' is gone from any button
|
||||
expect(screen.queryByRole('button', { name: 'Register' })).not.toBeInTheDocument();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -7,7 +7,7 @@ import { notifySuccess } from '../../../services/notificationService';
|
||||
import { LoadingSpinner } from '../../../components/LoadingSpinner';
|
||||
import { GoogleIcon } from '../../../components/icons/GoogleIcon';
|
||||
import { GithubIcon } from '../../../components/icons/GithubIcon';
|
||||
import { PasswordInput } from './PasswordInput';
|
||||
import { PasswordInput } from '../../../components/PasswordInput';
|
||||
|
||||
interface AuthResponse {
|
||||
userprofile: UserProfile;
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
// src/pages/admin/components/ProfileManager.test.tsx
|
||||
import React from 'react';
|
||||
import { render, screen, fireEvent, waitFor, cleanup, act } from '@testing-library/react';
|
||||
import { describe, it, expect, vi, beforeEach, afterEach, type Mock } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterEach, type Mock, test } from 'vitest';
|
||||
import { ProfileManager } from './ProfileManager';
|
||||
import * as apiClient from '../../../services/apiClient';
|
||||
import { notifySuccess, notifyError } from '../../../services/notificationService';
|
||||
@@ -16,6 +16,11 @@ import {
|
||||
// Unmock the component to test the real implementation
|
||||
vi.unmock('./ProfileManager');
|
||||
|
||||
vi.mock('../../../components/PasswordInput', () => ({
|
||||
// Mock the moved component
|
||||
PasswordInput: (props: any) => <input {...props} data-testid="password-input" />,
|
||||
}));
|
||||
|
||||
const mockedApiClient = vi.mocked(apiClient, true);
|
||||
|
||||
vi.mock('../../../services/notificationService');
|
||||
@@ -242,6 +247,17 @@ describe('ProfileManager', () => {
|
||||
expect(screen.queryByRole('heading', { name: /^sign in$/i })).not.toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should close the modal when clicking the backdrop', async () => {
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
// The backdrop is the element with role="dialog"
|
||||
const backdrop = screen.getByRole('dialog');
|
||||
fireEvent.click(backdrop);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(mockOnClose).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
it('should reset state when the modal is closed and reopened', async () => {
|
||||
const { rerender } = render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
await waitFor(() => expect(screen.getByLabelText(/full name/i)).toHaveValue('Test User'));
|
||||
@@ -308,6 +324,41 @@ describe('ProfileManager', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle partial success when saving profile and address', async () => {
|
||||
const loggerSpy = vi.spyOn(logger.logger, 'warn');
|
||||
// Mock profile update to succeed
|
||||
mockedApiClient.updateUserProfile.mockResolvedValue(
|
||||
new Response(JSON.stringify({ ...authenticatedProfile, full_name: 'New Name' })),
|
||||
);
|
||||
// Mock address update to fail (useApi will return null)
|
||||
mockedApiClient.updateUserAddress.mockRejectedValue(new Error('Address update failed'));
|
||||
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
await waitFor(() => expect(screen.getByLabelText(/city/i)).toHaveValue(mockAddress.city));
|
||||
|
||||
// Change both profile and address data
|
||||
fireEvent.change(screen.getByLabelText(/full name/i), { target: { value: 'New Name' } });
|
||||
fireEvent.change(screen.getByLabelText(/city/i), { target: { value: 'NewCity' } });
|
||||
|
||||
fireEvent.click(screen.getByRole('button', { name: /save profile/i }));
|
||||
|
||||
await waitFor(() => {
|
||||
// The useApi hook for the failed call will show its own error
|
||||
expect(notifyError).toHaveBeenCalledWith('Address update failed');
|
||||
// The profile update should still go through
|
||||
expect(mockOnProfileUpdate).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ full_name: 'New Name' }),
|
||||
);
|
||||
// The specific warning for partial failure should be logged
|
||||
expect(loggerSpy).toHaveBeenCalledWith(
|
||||
'[handleProfileSave] One or more operations failed. The useApi hook should have shown an error. The modal will remain open.',
|
||||
);
|
||||
// The modal should remain open and no global success message shown
|
||||
expect(mockOnClose).not.toHaveBeenCalled();
|
||||
expect(notifySuccess).not.toHaveBeenCalledWith('Profile updated successfully!');
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle unexpected critical error during profile save', async () => {
|
||||
const loggerSpy = vi.spyOn(logger.logger, 'error');
|
||||
mockedApiClient.updateUserProfile.mockRejectedValue(new Error('Catastrophic failure'));
|
||||
@@ -324,6 +375,31 @@ describe('ProfileManager', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle unexpected Promise.allSettled rejection during save', async () => {
|
||||
const allSettledSpy = vi
|
||||
.spyOn(Promise, 'allSettled')
|
||||
.mockRejectedValueOnce(new Error('AllSettled failed'));
|
||||
const loggerSpy = vi.spyOn(logger.logger, 'error');
|
||||
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
await waitFor(() => expect(screen.getByLabelText(/city/i)).toHaveValue(mockAddress.city));
|
||||
|
||||
fireEvent.change(screen.getByLabelText(/full name/i), { target: { value: 'New Name' } });
|
||||
fireEvent.click(screen.getByRole('button', { name: /save profile/i }));
|
||||
|
||||
await waitFor(() => {
|
||||
expect(loggerSpy).toHaveBeenCalledWith(
|
||||
{ err: new Error('AllSettled failed') },
|
||||
"[CRITICAL] An unexpected error was caught directly in handleProfileSave's catch block.",
|
||||
);
|
||||
expect(notifyError).toHaveBeenCalledWith(
|
||||
'An unexpected critical error occurred: AllSettled failed',
|
||||
);
|
||||
});
|
||||
|
||||
allSettledSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should show map view when address has coordinates', async () => {
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
await waitFor(() => {
|
||||
@@ -365,51 +441,52 @@ describe('ProfileManager', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should automatically geocode address after user stops typing', async () => {
|
||||
it('should automatically geocode address after user stops typing (using fake timers)', async () => {
|
||||
// Use fake timers for the entire test to control the debounce.
|
||||
vi.useFakeTimers();
|
||||
const addressWithoutCoords = { ...mockAddress, latitude: undefined, longitude: undefined };
|
||||
mockedApiClient.getUserAddress.mockResolvedValue(
|
||||
new Response(JSON.stringify(addressWithoutCoords)),
|
||||
);
|
||||
|
||||
console.log('[TEST LOG] Rendering for automatic geocode test (Real Timers + Wait)');
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
|
||||
console.log('[TEST LOG] Waiting for initial address load...');
|
||||
await waitFor(() => expect(screen.getByLabelText(/city/i)).toHaveValue('Anytown'));
|
||||
|
||||
console.log('[TEST LOG] Initial address loaded. Changing city...');
|
||||
// Wait for initial async address load to complete by flushing promises.
|
||||
await act(async () => {
|
||||
await vi.runAllTimersAsync();
|
||||
});
|
||||
expect(screen.getByLabelText(/city/i)).toHaveValue('Anytown');
|
||||
|
||||
// Change address, geocode should not be called immediately
|
||||
fireEvent.change(screen.getByLabelText(/city/i), { target: { value: 'NewCity' } });
|
||||
expect(mockedApiClient.geocodeAddress).not.toHaveBeenCalled();
|
||||
|
||||
console.log('[TEST LOG] Waiting 1600ms for debounce...');
|
||||
// Wait for debounce (1500ms) + buffer using real timers to avoid freeze
|
||||
// Advance timers to fire the debounce and resolve the subsequent geocode promise.
|
||||
await act(async () => {
|
||||
await new Promise((resolve) => setTimeout(resolve, 1600));
|
||||
await vi.runAllTimersAsync();
|
||||
});
|
||||
console.log('[TEST LOG] Wait complete. Checking results.');
|
||||
|
||||
await waitFor(() => {
|
||||
expect(mockedApiClient.geocodeAddress).toHaveBeenCalledWith(
|
||||
expect.stringContaining('NewCity'),
|
||||
expect.anything(),
|
||||
);
|
||||
expect(toast.success).toHaveBeenCalledWith('Address geocoded successfully!');
|
||||
});
|
||||
// Now check the final result.
|
||||
expect(mockedApiClient.geocodeAddress).toHaveBeenCalledWith(
|
||||
expect.stringContaining('NewCity'),
|
||||
expect.anything(),
|
||||
);
|
||||
expect(toast.success).toHaveBeenCalledWith('Address geocoded successfully!');
|
||||
});
|
||||
|
||||
it('should not geocode if address already has coordinates', async () => {
|
||||
console.log('[TEST LOG] Rendering for no-geocode test (Real Timers + Wait)');
|
||||
it('should not geocode if address already has coordinates (using fake timers)', async () => {
|
||||
// Use real timers for the initial async render and data fetch
|
||||
vi.useRealTimers();
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
console.log('[TEST LOG] Waiting for initial address load...');
|
||||
await waitFor(() => expect(screen.getByLabelText(/city/i)).toHaveValue('Anytown'));
|
||||
|
||||
console.log(
|
||||
'[TEST LOG] Initial address loaded. Waiting 1600ms to ensure no geocode triggers...',
|
||||
);
|
||||
await act(async () => {
|
||||
await new Promise((resolve) => setTimeout(resolve, 1600));
|
||||
// Switch to fake timers to control the debounce check
|
||||
vi.useFakeTimers();
|
||||
|
||||
// Advance timers past the debounce threshold. Nothing should happen.
|
||||
act(() => {
|
||||
vi.advanceTimersByTime(1600);
|
||||
});
|
||||
console.log('[TEST LOG] Wait complete. Verifying no geocode call.');
|
||||
|
||||
@@ -434,6 +511,29 @@ describe('ProfileManager', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should switch between all tabs correctly', async () => {
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
|
||||
// Initial state: Profile tab
|
||||
expect(screen.getByLabelText('Profile Form')).toBeInTheDocument();
|
||||
|
||||
// Switch to Security
|
||||
fireEvent.click(screen.getByRole('button', { name: /security/i }));
|
||||
expect(await screen.findByLabelText('New Password')).toBeInTheDocument();
|
||||
|
||||
// Switch to Data & Privacy
|
||||
fireEvent.click(screen.getByRole('button', { name: /data & privacy/i }));
|
||||
expect(await screen.findByRole('heading', { name: /export your data/i })).toBeInTheDocument();
|
||||
|
||||
// Switch to Preferences
|
||||
fireEvent.click(screen.getByRole('button', { name: /preferences/i }));
|
||||
expect(await screen.findByRole('heading', { name: /theme/i })).toBeInTheDocument();
|
||||
|
||||
// Switch back to Profile
|
||||
fireEvent.click(screen.getByRole('button', { name: /^profile$/i }));
|
||||
expect(await screen.findByLabelText('Profile Form')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('should show an error if password is too short', async () => {
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
fireEvent.click(screen.getByRole('button', { name: /security/i }));
|
||||
@@ -442,7 +542,7 @@ describe('ProfileManager', () => {
|
||||
fireEvent.change(screen.getByLabelText('Confirm New Password'), {
|
||||
target: { value: 'short' },
|
||||
});
|
||||
fireEvent.submit(screen.getByTestId('update-password-form'));
|
||||
fireEvent.submit(screen.getByTestId('update-password-form'), {});
|
||||
|
||||
await waitFor(() => {
|
||||
expect(notifyError).toHaveBeenCalledWith('Password must be at least 6 characters long.');
|
||||
@@ -456,7 +556,7 @@ describe('ProfileManager', () => {
|
||||
fireEvent.click(screen.getByRole('button', { name: /data & privacy/i }));
|
||||
fireEvent.click(screen.getByRole('button', { name: /delete my account/i }));
|
||||
|
||||
fireEvent.change(screen.getByPlaceholderText(/enter your password/i), {
|
||||
fireEvent.change(screen.getByTestId('password-input'), {
|
||||
target: { value: 'password' },
|
||||
});
|
||||
fireEvent.submit(screen.getByTestId('delete-account-form'));
|
||||
@@ -593,7 +693,7 @@ describe('ProfileManager', () => {
|
||||
fireEvent.change(screen.getByLabelText('Confirm New Password'), {
|
||||
target: { value: 'newpassword123' },
|
||||
});
|
||||
fireEvent.submit(screen.getByTestId('update-password-form'));
|
||||
fireEvent.submit(screen.getByTestId('update-password-form'), {});
|
||||
|
||||
await waitFor(() => {
|
||||
expect(mockedApiClient.updateUserPassword).toHaveBeenCalledWith(
|
||||
@@ -614,7 +714,7 @@ describe('ProfileManager', () => {
|
||||
fireEvent.change(screen.getByLabelText('Confirm New Password'), {
|
||||
target: { value: 'mismatch' },
|
||||
});
|
||||
fireEvent.submit(screen.getByTestId('update-password-form'));
|
||||
fireEvent.submit(screen.getByTestId('update-password-form'), {});
|
||||
|
||||
await waitFor(() => {
|
||||
expect(notifyError).toHaveBeenCalledWith('Passwords do not match.');
|
||||
@@ -641,9 +741,10 @@ describe('ProfileManager', () => {
|
||||
});
|
||||
|
||||
it('should handle account deletion flow', async () => {
|
||||
// Use spy instead of fake timers to avoid blocking waitFor during async API calls
|
||||
const setTimeoutSpy = vi.spyOn(window, 'setTimeout');
|
||||
const { unmount } = render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
// Use fake timers to control the setTimeout call for the entire test.
|
||||
vi.useFakeTimers();
|
||||
|
||||
render(<ProfileManager {...defaultAuthenticatedProps} />);
|
||||
|
||||
fireEvent.click(screen.getByRole('button', { name: /data & privacy/i }));
|
||||
|
||||
@@ -654,39 +755,28 @@ describe('ProfileManager', () => {
|
||||
).toBeInTheDocument();
|
||||
|
||||
// Fill password and submit to open modal
|
||||
fireEvent.change(screen.getByPlaceholderText(/enter your password/i), {
|
||||
fireEvent.change(screen.getByTestId('password-input'), {
|
||||
target: { value: 'correctpassword' },
|
||||
});
|
||||
fireEvent.submit(screen.getByTestId('delete-account-form'));
|
||||
|
||||
// Confirm in the modal
|
||||
const confirmButton = await screen.findByRole('button', { name: /yes, delete my account/i });
|
||||
// Use getByRole since the modal appears synchronously after the form submit.
|
||||
const confirmButton = screen.getByRole('button', { name: /yes, delete my account/i });
|
||||
fireEvent.click(confirmButton);
|
||||
|
||||
await waitFor(() => {
|
||||
expect(mockedApiClient.deleteUserAccount).toHaveBeenCalledWith(
|
||||
'correctpassword',
|
||||
expect.objectContaining({ signal: expect.anything() }),
|
||||
);
|
||||
expect(notifySuccess).toHaveBeenCalledWith(
|
||||
'Account deleted successfully. You will be logged out shortly.',
|
||||
);
|
||||
});
|
||||
|
||||
// Verify setTimeout was called with 3000ms
|
||||
const deletionTimeoutCall = setTimeoutSpy.mock.calls.find((call) => call[1] === 3000);
|
||||
expect(deletionTimeoutCall).toBeDefined();
|
||||
|
||||
// Manually trigger the callback to verify cleanup
|
||||
act(() => {
|
||||
if (deletionTimeoutCall) (deletionTimeoutCall[0] as Function)();
|
||||
// The async deleteAccount call is now pending. We need to flush promises
|
||||
// and then advance the timers to run the subsequent setTimeout.
|
||||
// `runAllTimersAsync` will resolve pending promises and run timers recursively.
|
||||
await act(async () => {
|
||||
await vi.runAllTimersAsync();
|
||||
});
|
||||
|
||||
// Now that all timers and promises have been flushed, we can check the final state.
|
||||
expect(mockedApiClient.deleteUserAccount).toHaveBeenCalled();
|
||||
expect(notifySuccess).toHaveBeenCalled();
|
||||
expect(mockOnClose).toHaveBeenCalled();
|
||||
expect(mockOnSignOut).toHaveBeenCalled();
|
||||
|
||||
unmount();
|
||||
setTimeoutSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should allow toggling dark mode', async () => {
|
||||
|
||||
@@ -9,8 +9,8 @@ import { LoadingSpinner } from '../../../components/LoadingSpinner';
|
||||
import { XMarkIcon } from '../../../components/icons/XMarkIcon';
|
||||
import { GoogleIcon } from '../../../components/icons/GoogleIcon';
|
||||
import { GithubIcon } from '../../../components/icons/GithubIcon';
|
||||
import { ConfirmationModal } from '../../../components/ConfirmationModal';
|
||||
import { PasswordInput } from './PasswordInput';
|
||||
import { ConfirmationModal } from '../../../components/ConfirmationModal'; // This path is correct
|
||||
import { PasswordInput } from '../../../components/PasswordInput';
|
||||
import { MapView } from '../../../components/MapView';
|
||||
import type { AuthStatus } from '../../../hooks/useAuth';
|
||||
import { AuthView } from './AuthView';
|
||||
|
||||
55
src/providers/ApiProvider.test.tsx
Normal file
55
src/providers/ApiProvider.test.tsx
Normal file
@@ -0,0 +1,55 @@
|
||||
// src/providers/ApiProvider.test.tsx
|
||||
import React, { useContext } from 'react';
|
||||
import { render, screen } from '@testing-library/react';
|
||||
import { describe, it, expect, vi } from 'vitest';
|
||||
import { ApiProvider } from './ApiProvider';
|
||||
import { ApiContext } from '../contexts/ApiContext';
|
||||
import * as apiClient from '../services/apiClient';
|
||||
|
||||
// Mock the apiClient module.
|
||||
// Since ApiProvider and ApiContext import * as apiClient, mocking it ensures
|
||||
// we control the reference identity and can verify it's being passed correctly.
|
||||
vi.mock('../services/apiClient', () => ({
|
||||
fetchFlyers: vi.fn(),
|
||||
fetchMasterItems: vi.fn(),
|
||||
// Add other mocked methods as needed for the shape to be valid-ish
|
||||
}));
|
||||
|
||||
describe('ApiProvider & ApiContext', () => {
|
||||
const TestConsumer = () => {
|
||||
const contextValue = useContext(ApiContext);
|
||||
// We check if the context value is strictly equal to the imported module
|
||||
return (
|
||||
<div>
|
||||
<span data-testid="value-check">
|
||||
{contextValue === apiClient ? 'Matches apiClient' : 'Does not match'}
|
||||
</span>
|
||||
</div>
|
||||
);
|
||||
};
|
||||
|
||||
it('renders children correctly', () => {
|
||||
render(
|
||||
<ApiProvider>
|
||||
<div data-testid="child">Child Content</div>
|
||||
</ApiProvider>
|
||||
);
|
||||
expect(screen.getByTestId('child')).toBeInTheDocument();
|
||||
expect(screen.getByText('Child Content')).toBeInTheDocument();
|
||||
});
|
||||
|
||||
it('provides the apiClient module via context', () => {
|
||||
render(
|
||||
<ApiProvider>
|
||||
<TestConsumer />
|
||||
</ApiProvider>
|
||||
);
|
||||
expect(screen.getByTestId('value-check')).toHaveTextContent('Matches apiClient');
|
||||
});
|
||||
|
||||
it('ApiContext has apiClient as the default value (when no provider is present)', () => {
|
||||
// This verifies the logic in ApiContext.tsx: createContext(apiClient)
|
||||
render(<TestConsumer />);
|
||||
expect(screen.getByTestId('value-check')).toHaveTextContent('Matches apiClient');
|
||||
});
|
||||
});
|
||||
@@ -4,6 +4,7 @@ import { AuthContext, AuthContextType } from '../contexts/AuthContext';
|
||||
import type { UserProfile } from '../types';
|
||||
import * as apiClient from '../services/apiClient';
|
||||
import { useApi } from '../hooks/useApi';
|
||||
import { getToken, setToken, removeToken } from '../services/tokenStorage';
|
||||
import { logger } from '../services/logger.client';
|
||||
|
||||
export const AuthProvider: React.FC<{ children: ReactNode }> = ({ children }) => {
|
||||
@@ -27,7 +28,7 @@ export const AuthProvider: React.FC<{ children: ReactNode }> = ({ children }) =>
|
||||
logger.info('[AuthProvider-Effect] Starting initial authentication check.');
|
||||
|
||||
const checkAuthToken = async () => {
|
||||
const token = localStorage.getItem('authToken');
|
||||
const token = getToken();
|
||||
if (token) {
|
||||
logger.info('[AuthProvider-Effect] Found auth token. Validating...');
|
||||
try {
|
||||
@@ -41,7 +42,7 @@ export const AuthProvider: React.FC<{ children: ReactNode }> = ({ children }) =>
|
||||
logger.warn(
|
||||
'[AuthProvider-Effect] Token was present but validation returned no profile. Signing out.',
|
||||
);
|
||||
localStorage.removeItem('authToken');
|
||||
removeToken();
|
||||
setUserProfile(null);
|
||||
setAuthStatus('SIGNED_OUT');
|
||||
}
|
||||
@@ -49,7 +50,7 @@ export const AuthProvider: React.FC<{ children: ReactNode }> = ({ children }) =>
|
||||
// This catch block is now primarily for unexpected errors, as useApi handles API errors.
|
||||
logger.warn('Auth token validation failed. Clearing token.', { error: e });
|
||||
if (isMounted) {
|
||||
localStorage.removeItem('authToken');
|
||||
removeToken();
|
||||
setUserProfile(null);
|
||||
setAuthStatus('SIGNED_OUT');
|
||||
}
|
||||
@@ -79,7 +80,7 @@ export const AuthProvider: React.FC<{ children: ReactNode }> = ({ children }) =>
|
||||
|
||||
const logout = useCallback(() => {
|
||||
logger.info('[AuthProvider-Logout] Clearing user data and auth token.');
|
||||
localStorage.removeItem('authToken');
|
||||
removeToken();
|
||||
setUserProfile(null);
|
||||
setAuthStatus('SIGNED_OUT');
|
||||
}, []);
|
||||
@@ -87,7 +88,7 @@ export const AuthProvider: React.FC<{ children: ReactNode }> = ({ children }) =>
|
||||
const login = useCallback(
|
||||
async (token: string, profileData?: UserProfile) => {
|
||||
logger.info(`[AuthProvider-Login] Attempting login.`);
|
||||
localStorage.setItem('authToken', token);
|
||||
setToken(token);
|
||||
|
||||
if (profileData) {
|
||||
// If profile is provided (e.g., from credential login), use it directly.
|
||||
|
||||
@@ -1,18 +1,22 @@
|
||||
// src/routes/admin.content.routes.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterAll } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import type { Request, Response, NextFunction } from 'express';
|
||||
import path from 'path';
|
||||
import {
|
||||
createMockUserProfile,
|
||||
createMockSuggestedCorrection,
|
||||
createMockBrand,
|
||||
createMockRecipe,
|
||||
createMockFlyer,
|
||||
createMockRecipeComment,
|
||||
createMockUnmatchedFlyerItem,
|
||||
} from '../tests/utils/mockFactories';
|
||||
import type { SuggestedCorrection, Brand, UserProfile, UnmatchedFlyerItem } from '../types';
|
||||
import { NotFoundError } from '../services/db/errors.db'; // This can stay, it's a type/class not a module with side effects.
|
||||
import fs from 'node:fs/promises';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
import { cleanupFiles } from '../tests/utils/cleanupFiles';
|
||||
|
||||
// Mock the file upload middleware to allow testing the controller's internal check
|
||||
vi.mock('../middleware/fileUpload.middleware', () => ({
|
||||
@@ -37,9 +41,11 @@ const { mockedDb } = vi.hoisted(() => {
|
||||
rejectCorrection: vi.fn(),
|
||||
updateSuggestedCorrection: vi.fn(),
|
||||
getUnmatchedFlyerItems: vi.fn(),
|
||||
getFlyersForReview: vi.fn(), // Added for flyer review tests
|
||||
updateRecipeStatus: vi.fn(),
|
||||
updateRecipeCommentStatus: vi.fn(),
|
||||
updateBrandLogo: vi.fn(),
|
||||
getApplicationStats: vi.fn(),
|
||||
},
|
||||
flyerRepo: {
|
||||
getAllBrands: vi.fn(),
|
||||
@@ -72,10 +78,12 @@ vi.mock('node:fs/promises', () => ({
|
||||
// Named exports
|
||||
writeFile: vi.fn().mockResolvedValue(undefined),
|
||||
unlink: vi.fn().mockResolvedValue(undefined),
|
||||
mkdir: vi.fn().mockResolvedValue(undefined),
|
||||
// FIX: Add default export to handle `import fs from ...` syntax.
|
||||
default: {
|
||||
writeFile: vi.fn().mockResolvedValue(undefined),
|
||||
unlink: vi.fn().mockResolvedValue(undefined),
|
||||
mkdir: vi.fn().mockResolvedValue(undefined),
|
||||
},
|
||||
}));
|
||||
vi.mock('../services/backgroundJobService');
|
||||
@@ -134,6 +142,26 @@ describe('Admin Content Management Routes (/api/admin)', () => {
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
afterAll(async () => {
|
||||
// Safeguard to clean up any logo files created during tests.
|
||||
const uploadDir = path.resolve(__dirname, '../../../flyer-images');
|
||||
try {
|
||||
const allFiles = await fs.readdir(uploadDir);
|
||||
// Files are named like 'logoImage-timestamp-original.ext'
|
||||
const testFiles = allFiles
|
||||
.filter((f) => f.startsWith('logoImage-'))
|
||||
.map((f) => path.join(uploadDir, f));
|
||||
|
||||
if (testFiles.length > 0) {
|
||||
await cleanupFiles(testFiles);
|
||||
}
|
||||
} catch (error) {
|
||||
if (error instanceof Error && (error as NodeJS.ErrnoException).code !== 'ENOENT') {
|
||||
console.error('Error during admin content test file cleanup:', error);
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
describe('Corrections Routes', () => {
|
||||
it('GET /corrections should return corrections data', async () => {
|
||||
const mockCorrections: SuggestedCorrection[] = [
|
||||
@@ -224,6 +252,39 @@ describe('Admin Content Management Routes (/api/admin)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('Flyer Review Routes', () => {
|
||||
it('GET /review/flyers should return flyers for review', async () => {
|
||||
const mockFlyers = [
|
||||
createMockFlyer({ flyer_id: 1, status: 'needs_review' }),
|
||||
createMockFlyer({ flyer_id: 2, status: 'needs_review' }),
|
||||
];
|
||||
vi.mocked(mockedDb.adminRepo.getFlyersForReview).mockResolvedValue(mockFlyers);
|
||||
const response = await supertest(app).get('/api/admin/review/flyers');
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body).toEqual(mockFlyers);
|
||||
expect(vi.mocked(mockedDb.adminRepo.getFlyersForReview)).toHaveBeenCalledWith(
|
||||
expect.anything(),
|
||||
);
|
||||
});
|
||||
|
||||
it('GET /review/flyers should return 500 on DB error', async () => {
|
||||
vi.mocked(mockedDb.adminRepo.getFlyersForReview).mockRejectedValue(new Error('DB Error'));
|
||||
const response = await supertest(app).get('/api/admin/review/flyers');
|
||||
expect(response.status).toBe(500);
|
||||
expect(response.body.message).toBe('DB Error');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Stats Routes', () => {
|
||||
// This test covers the error path for GET /stats
|
||||
it('GET /stats should return 500 on DB error', async () => {
|
||||
vi.mocked(mockedDb.adminRepo.getApplicationStats).mockRejectedValue(new Error('DB Error'));
|
||||
const response = await supertest(app).get('/api/admin/stats');
|
||||
expect(response.status).toBe(500);
|
||||
expect(response.body.message).toBe('DB Error');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Brand Routes', () => {
|
||||
it('GET /brands should return a list of all brands', async () => {
|
||||
const mockBrands: Brand[] = [createMockBrand({ brand_id: 1, name: 'Brand A' })];
|
||||
@@ -243,7 +304,7 @@ describe('Admin Content Management Routes (/api/admin)', () => {
|
||||
expect(response.body.message).toBe('Brand logo updated successfully.');
|
||||
expect(vi.mocked(mockedDb.adminRepo.updateBrandLogo)).toHaveBeenCalledWith(
|
||||
brandId,
|
||||
expect.stringContaining('/assets/'),
|
||||
expect.stringContaining('/flyer-images/'),
|
||||
expect.anything(),
|
||||
);
|
||||
});
|
||||
@@ -265,6 +326,32 @@ describe('Admin Content Management Routes (/api/admin)', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should clean up the uploaded file if updating the brand logo fails', async () => {
|
||||
const brandId = 55;
|
||||
const dbError = new Error('DB Connection Failed');
|
||||
vi.mocked(mockedDb.adminRepo.updateBrandLogo).mockRejectedValue(dbError);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post(`/api/admin/brands/${brandId}/logo`)
|
||||
.attach('logoImage', Buffer.from('dummy-logo-content'), 'test-logo.png');
|
||||
|
||||
expect(response.status).toBe(500);
|
||||
// Verify that the cleanup function was called via the mocked fs module
|
||||
expect(fs.unlink).toHaveBeenCalledTimes(1);
|
||||
// The filename is predictable because of the multer config in admin.routes.ts
|
||||
expect(fs.unlink).toHaveBeenCalledWith(expect.stringContaining('logoImage-'));
|
||||
});
|
||||
|
||||
it('POST /brands/:id/logo should return 400 if a non-image file is uploaded', async () => {
|
||||
const brandId = 55;
|
||||
const response = await supertest(app)
|
||||
.post(`/api/admin/brands/${brandId}/logo`)
|
||||
.attach('logoImage', Buffer.from('this is not an image'), 'document.txt');
|
||||
expect(response.status).toBe(400);
|
||||
// This message comes from the handleMulterError middleware for the imageFileFilter
|
||||
expect(response.body.message).toBe('Only image files are allowed!');
|
||||
});
|
||||
|
||||
it('POST /brands/:id/logo should return 400 for an invalid brand ID', async () => {
|
||||
const response = await supertest(app)
|
||||
.post('/api/admin/brands/abc/logo')
|
||||
|
||||
@@ -11,6 +11,8 @@ import { createTestApp } from '../tests/utils/createTestApp';
|
||||
vi.mock('../services/backgroundJobService', () => ({
|
||||
backgroundJobService: {
|
||||
runDailyDealCheck: vi.fn(),
|
||||
triggerAnalyticsReport: vi.fn(),
|
||||
triggerWeeklyAnalyticsReport: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
@@ -142,22 +144,17 @@ describe('Admin Job Trigger Routes (/api/admin/trigger)', () => {
|
||||
|
||||
describe('POST /trigger/analytics-report', () => {
|
||||
it('should trigger the analytics report job and return 202 Accepted', async () => {
|
||||
const mockJob = { id: 'manual-report-job-123' } as Job;
|
||||
vi.mocked(analyticsQueue.add).mockResolvedValue(mockJob);
|
||||
vi.mocked(backgroundJobService.triggerAnalyticsReport).mockResolvedValue('manual-report-job-123');
|
||||
|
||||
const response = await supertest(app).post('/api/admin/trigger/analytics-report');
|
||||
|
||||
expect(response.status).toBe(202);
|
||||
expect(response.body.message).toContain('Analytics report generation job has been enqueued');
|
||||
expect(analyticsQueue.add).toHaveBeenCalledWith(
|
||||
'generate-daily-report',
|
||||
expect.objectContaining({ reportDate: expect.any(String) }),
|
||||
expect.any(Object),
|
||||
);
|
||||
expect(backgroundJobService.triggerAnalyticsReport).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should return 500 if enqueuing the analytics job fails', async () => {
|
||||
vi.mocked(analyticsQueue.add).mockRejectedValue(new Error('Queue error'));
|
||||
vi.mocked(backgroundJobService.triggerAnalyticsReport).mockRejectedValue(new Error('Queue error'));
|
||||
const response = await supertest(app).post('/api/admin/trigger/analytics-report');
|
||||
expect(response.status).toBe(500);
|
||||
});
|
||||
@@ -165,22 +162,17 @@ describe('Admin Job Trigger Routes (/api/admin/trigger)', () => {
|
||||
|
||||
describe('POST /trigger/weekly-analytics', () => {
|
||||
it('should trigger the weekly analytics job and return 202 Accepted', async () => {
|
||||
const mockJob = { id: 'manual-weekly-report-job-123' } as Job;
|
||||
vi.mocked(weeklyAnalyticsQueue.add).mockResolvedValue(mockJob);
|
||||
vi.mocked(backgroundJobService.triggerWeeklyAnalyticsReport).mockResolvedValue('manual-weekly-report-job-123');
|
||||
|
||||
const response = await supertest(app).post('/api/admin/trigger/weekly-analytics');
|
||||
|
||||
expect(response.status).toBe(202);
|
||||
expect(response.body.message).toContain('Successfully enqueued weekly analytics job');
|
||||
expect(weeklyAnalyticsQueue.add).toHaveBeenCalledWith(
|
||||
'generate-weekly-report',
|
||||
expect.objectContaining({ reportYear: expect.any(Number), reportWeek: expect.any(Number) }),
|
||||
expect.any(Object),
|
||||
);
|
||||
expect(backgroundJobService.triggerWeeklyAnalyticsReport).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should return 500 if enqueuing the weekly analytics job fails', async () => {
|
||||
vi.mocked(weeklyAnalyticsQueue.add).mockRejectedValue(new Error('Queue error'));
|
||||
vi.mocked(backgroundJobService.triggerWeeklyAnalyticsReport).mockRejectedValue(new Error('Queue error'));
|
||||
const response = await supertest(app).post('/api/admin/trigger/weekly-analytics');
|
||||
expect(response.status).toBe(500);
|
||||
});
|
||||
@@ -242,15 +234,17 @@ describe('Admin Job Trigger Routes (/api/admin/trigger)', () => {
|
||||
expect(response.status).toBe(400);
|
||||
});
|
||||
|
||||
it('should return 404 if the queue name is valid but not in the retry map', async () => {
|
||||
const queueName = 'weekly-analytics-reporting'; // This is in the Zod enum but not the queueMap
|
||||
it('should return 404 if the job ID is not found in the weekly-analytics-reporting queue', async () => {
|
||||
const queueName = 'weekly-analytics-reporting';
|
||||
const jobId = 'some-job-id';
|
||||
|
||||
// Ensure getJob returns undefined (not found)
|
||||
vi.mocked(weeklyAnalyticsQueue.getJob).mockResolvedValue(undefined);
|
||||
|
||||
const response = await supertest(app).post(`/api/admin/jobs/${queueName}/${jobId}/retry`);
|
||||
|
||||
// The route throws a NotFoundError, which the error handler should convert to a 404.
|
||||
expect(response.status).toBe(404);
|
||||
expect(response.body.message).toBe(`Queue 'weekly-analytics-reporting' not found.`);
|
||||
expect(response.body.message).toBe(`Job with ID '${jobId}' not found in queue '${queueName}'.`);
|
||||
});
|
||||
|
||||
it('should return 404 if the job ID is not found in the queue', async () => {
|
||||
|
||||
@@ -9,6 +9,10 @@ import * as db from '../services/db/index.db';
|
||||
import type { UserProfile } from '../types';
|
||||
import { geocodingService } from '../services/geocodingService.server';
|
||||
import { requireFileUpload } from '../middleware/fileUpload.middleware'; // This was a duplicate, fixed.
|
||||
import {
|
||||
createUploadMiddleware,
|
||||
handleMulterError,
|
||||
} from '../middleware/multer.middleware';
|
||||
import { NotFoundError, ValidationError } from '../services/db/errors.db';
|
||||
import { validateRequest } from '../middleware/validation.middleware';
|
||||
|
||||
@@ -16,35 +20,25 @@ import { validateRequest } from '../middleware/validation.middleware';
|
||||
import { createBullBoard } from '@bull-board/api';
|
||||
import { BullMQAdapter } from '@bull-board/api/bullMQAdapter';
|
||||
import { ExpressAdapter } from '@bull-board/express';
|
||||
|
||||
import type { Queue } from 'bullmq';
|
||||
import { backgroundJobService } from '../services/backgroundJobService';
|
||||
import {
|
||||
flyerQueue,
|
||||
emailQueue,
|
||||
analyticsQueue,
|
||||
cleanupQueue,
|
||||
weeklyAnalyticsQueue,
|
||||
} from '../services/queueService.server'; // Import your queues
|
||||
import {
|
||||
analyticsWorker,
|
||||
cleanupWorker,
|
||||
emailWorker,
|
||||
flyerWorker,
|
||||
weeklyAnalyticsWorker,
|
||||
} from '../services/workers.server';
|
||||
import { flyerQueue, emailQueue, analyticsQueue, cleanupQueue, weeklyAnalyticsQueue } from '../services/queueService.server';
|
||||
import { getSimpleWeekAndYear } from '../utils/dateUtils';
|
||||
import {
|
||||
requiredString,
|
||||
numericIdParam,
|
||||
uuidParamSchema,
|
||||
optionalNumeric,
|
||||
optionalString,
|
||||
} from '../utils/zodUtils';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { logger } from '../services/logger.server'; // This was a duplicate, fixed.
|
||||
import { monitoringService } from '../services/monitoringService.server';
|
||||
import { userService } from '../services/userService';
|
||||
import { cleanupUploadedFile } from '../utils/fileUtils';
|
||||
import { brandService } from '../services/brandService';
|
||||
|
||||
const updateCorrectionSchema = numericIdParam('id').extend({
|
||||
body: z.object({
|
||||
suggested_value: requiredString('A new suggested_value is required.'),
|
||||
suggested_value: z.string().trim().min(1, 'A new suggested_value is required.'),
|
||||
}),
|
||||
});
|
||||
|
||||
@@ -82,25 +76,19 @@ const jobRetrySchema = z.object({
|
||||
'file-cleanup',
|
||||
'weekly-analytics-reporting',
|
||||
]),
|
||||
jobId: requiredString('A valid Job ID is required.'),
|
||||
jobId: z.string().trim().min(1, 'A valid Job ID is required.'),
|
||||
}),
|
||||
});
|
||||
|
||||
const emptySchema = z.object({});
|
||||
|
||||
const router = Router();
|
||||
|
||||
// --- Multer Configuration for File Uploads ---
|
||||
const storagePath =
|
||||
process.env.STORAGE_PATH || '/var/www/flyer-crawler.projectium.com/flyer-images';
|
||||
const storage = multer.diskStorage({
|
||||
destination: function (req, file, cb) {
|
||||
cb(null, storagePath);
|
||||
},
|
||||
filename: function (req, file, cb) {
|
||||
const uniqueSuffix = Date.now() + '-' + Math.round(Math.random() * 1e9);
|
||||
cb(null, file.fieldname + '-' + uniqueSuffix + '-' + file.originalname);
|
||||
},
|
||||
const brandLogoUpload = createUploadMiddleware({
|
||||
storageType: 'flyer', // Using flyer storage path is acceptable for brand logos.
|
||||
fileSize: 2 * 1024 * 1024, // 2MB limit for logos
|
||||
fileFilter: 'image',
|
||||
});
|
||||
const upload = multer({ storage: storage });
|
||||
|
||||
// --- Bull Board (Job Queue UI) Setup ---
|
||||
const serverAdapter = new ExpressAdapter();
|
||||
@@ -132,7 +120,7 @@ router.use(passport.authenticate('jwt', { session: false }), isAdmin);
|
||||
|
||||
// --- Admin Routes ---
|
||||
|
||||
router.get('/corrections', async (req, res, next: NextFunction) => {
|
||||
router.get('/corrections', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const corrections = await db.adminRepo.getSuggestedCorrections(req.log);
|
||||
res.json(corrections);
|
||||
@@ -142,7 +130,19 @@ router.get('/corrections', async (req, res, next: NextFunction) => {
|
||||
}
|
||||
});
|
||||
|
||||
router.get('/brands', async (req, res, next: NextFunction) => {
|
||||
router.get('/review/flyers', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
req.log.debug('Fetching flyers for review via adminRepo');
|
||||
const flyers = await db.adminRepo.getFlyersForReview(req.log);
|
||||
req.log.info({ count: Array.isArray(flyers) ? flyers.length : 'unknown' }, 'Successfully fetched flyers for review');
|
||||
res.json(flyers);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching flyers for review');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
|
||||
router.get('/brands', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const brands = await db.flyerRepo.getAllBrands(req.log);
|
||||
res.json(brands);
|
||||
@@ -152,7 +152,7 @@ router.get('/brands', async (req, res, next: NextFunction) => {
|
||||
}
|
||||
});
|
||||
|
||||
router.get('/stats', async (req, res, next: NextFunction) => {
|
||||
router.get('/stats', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const stats = await db.adminRepo.getApplicationStats(req.log);
|
||||
res.json(stats);
|
||||
@@ -162,7 +162,7 @@ router.get('/stats', async (req, res, next: NextFunction) => {
|
||||
}
|
||||
});
|
||||
|
||||
router.get('/stats/daily', async (req, res, next: NextFunction) => {
|
||||
router.get('/stats/daily', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const dailyStats = await db.adminRepo.getDailyStatsForLast30Days(req.log);
|
||||
res.json(dailyStats);
|
||||
@@ -243,10 +243,9 @@ router.put(
|
||||
router.post(
|
||||
'/brands/:id/logo',
|
||||
validateRequest(numericIdParam('id')),
|
||||
upload.single('logoImage'),
|
||||
brandLogoUpload.single('logoImage'),
|
||||
requireFileUpload('logoImage'),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
// Apply ADR-003 pattern for type safety
|
||||
const { params } = req as unknown as z.infer<ReturnType<typeof numericIdParam>>;
|
||||
try {
|
||||
// Although requireFileUpload middleware should ensure the file exists,
|
||||
@@ -254,19 +253,22 @@ router.post(
|
||||
if (!req.file) {
|
||||
throw new ValidationError([], 'Logo image file is missing.');
|
||||
}
|
||||
const logoUrl = `/assets/${req.file.filename}`;
|
||||
await db.adminRepo.updateBrandLogo(params.id, logoUrl, req.log);
|
||||
|
||||
const logoUrl = await brandService.updateBrandLogo(params.id, req.file, req.log);
|
||||
|
||||
logger.info({ brandId: params.id, logoUrl }, `Brand logo updated for brand ID: ${params.id}`);
|
||||
res.status(200).json({ message: 'Brand logo updated successfully.', logoUrl });
|
||||
} catch (error) {
|
||||
// If an error occurs after the file has been uploaded (e.g., DB error),
|
||||
// we must clean up the orphaned file from the disk.
|
||||
await cleanupUploadedFile(req.file);
|
||||
logger.error({ error }, 'Error updating brand logo');
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
);
|
||||
|
||||
router.get('/unmatched-items', async (req, res, next: NextFunction) => {
|
||||
router.get('/unmatched-items', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const items = await db.adminRepo.getUnmatchedFlyerItems(req.log);
|
||||
res.json(items);
|
||||
@@ -336,7 +338,7 @@ router.put(
|
||||
},
|
||||
);
|
||||
|
||||
router.get('/users', async (req, res, next: NextFunction) => {
|
||||
router.get('/users', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const users = await db.adminRepo.getAllUsers(req.log);
|
||||
res.json(users);
|
||||
@@ -351,14 +353,11 @@ router.get(
|
||||
validateRequest(activityLogSchema),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
// Apply ADR-003 pattern for type safety.
|
||||
// We explicitly coerce query params here because the validation middleware might not
|
||||
// replace req.query with the coerced values in all environments.
|
||||
const query = req.query as unknown as { limit?: string; offset?: string };
|
||||
const limit = query.limit ? Number(query.limit) : 50;
|
||||
const offset = query.offset ? Number(query.offset) : 0;
|
||||
// We parse the query here to apply Zod's coercions (string to number) and defaults.
|
||||
const { limit, offset } = activityLogSchema.shape.query.parse(req.query);
|
||||
|
||||
try {
|
||||
const logs = await db.adminRepo.getActivityLog(limit, offset, req.log);
|
||||
const logs = await db.adminRepo.getActivityLog(limit!, offset!, req.log);
|
||||
res.json(logs);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching activity log');
|
||||
@@ -407,10 +406,7 @@ router.delete(
|
||||
// Apply ADR-003 pattern for type safety
|
||||
const { params } = req as unknown as z.infer<ReturnType<typeof uuidParamSchema>>;
|
||||
try {
|
||||
if (userProfile.user.user_id === params.id) {
|
||||
throw new ValidationError([], 'Admins cannot delete their own account.');
|
||||
}
|
||||
await db.userRepo.deleteUserById(params.id, req.log);
|
||||
await userService.deleteUserAsAdmin(userProfile.user.user_id, params.id, req.log);
|
||||
res.status(204).send();
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error deleting user');
|
||||
@@ -425,6 +421,7 @@ router.delete(
|
||||
*/
|
||||
router.post(
|
||||
'/trigger/daily-deal-check',
|
||||
validateRequest(emptySchema),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
const userProfile = req.user as UserProfile;
|
||||
logger.info(
|
||||
@@ -452,6 +449,7 @@ router.post(
|
||||
*/
|
||||
router.post(
|
||||
'/trigger/analytics-report',
|
||||
validateRequest(emptySchema),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
const userProfile = req.user as UserProfile;
|
||||
logger.info(
|
||||
@@ -459,14 +457,9 @@ router.post(
|
||||
);
|
||||
|
||||
try {
|
||||
const reportDate = new Date().toISOString().split('T')[0]; // YYYY-MM-DD
|
||||
// Use a unique job ID for manual triggers to distinguish them from scheduled jobs.
|
||||
const jobId = `manual-report-${reportDate}-${Date.now()}`;
|
||||
|
||||
const job = await analyticsQueue.add('generate-daily-report', { reportDate }, { jobId });
|
||||
|
||||
const jobId = await backgroundJobService.triggerAnalyticsReport();
|
||||
res.status(202).json({
|
||||
message: `Analytics report generation job has been enqueued successfully. Job ID: ${job.id}`,
|
||||
message: `Analytics report generation job has been enqueued successfully. Job ID: ${jobId}`,
|
||||
});
|
||||
} catch (error) {
|
||||
logger.error({ error }, '[Admin] Failed to enqueue analytics report job.');
|
||||
@@ -507,7 +500,10 @@ router.post(
|
||||
* POST /api/admin/trigger/failing-job - Enqueue a test job designed to fail.
|
||||
* This is for testing the retry mechanism and Bull Board UI.
|
||||
*/
|
||||
router.post('/trigger/failing-job', async (req: Request, res: Response, next: NextFunction) => {
|
||||
router.post(
|
||||
'/trigger/failing-job',
|
||||
validateRequest(emptySchema),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
const userProfile = req.user as UserProfile;
|
||||
logger.info(
|
||||
`[Admin] Manual trigger for a failing job received from user: ${userProfile.user.user_id}`,
|
||||
@@ -523,7 +519,8 @@ router.post('/trigger/failing-job', async (req: Request, res: Response, next: Ne
|
||||
logger.error({ error }, 'Error enqueuing failing job');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
}
|
||||
);
|
||||
|
||||
/**
|
||||
* POST /api/admin/system/clear-geocode-cache - Clears the Redis cache for geocoded addresses.
|
||||
@@ -531,6 +528,7 @@ router.post('/trigger/failing-job', async (req: Request, res: Response, next: Ne
|
||||
*/
|
||||
router.post(
|
||||
'/system/clear-geocode-cache',
|
||||
validateRequest(emptySchema),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
const userProfile = req.user as UserProfile;
|
||||
logger.info(
|
||||
@@ -553,44 +551,23 @@ router.post(
|
||||
* GET /api/admin/workers/status - Get the current running status of all BullMQ workers.
|
||||
* This is useful for a system health dashboard to see if any workers have crashed.
|
||||
*/
|
||||
router.get('/workers/status', async (req: Request, res: Response) => {
|
||||
const workers = [flyerWorker, emailWorker, analyticsWorker, cleanupWorker, weeklyAnalyticsWorker];
|
||||
|
||||
const workerStatuses = await Promise.all(
|
||||
workers.map(async (worker) => {
|
||||
return {
|
||||
name: worker.name,
|
||||
isRunning: worker.isRunning(),
|
||||
};
|
||||
}),
|
||||
);
|
||||
|
||||
res.json(workerStatuses);
|
||||
router.get('/workers/status', validateRequest(emptySchema), async (req: Request, res: Response, next: NextFunction) => {
|
||||
try {
|
||||
const workerStatuses = await monitoringService.getWorkerStatuses();
|
||||
res.json(workerStatuses);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching worker statuses');
|
||||
next(error);
|
||||
}
|
||||
});
|
||||
|
||||
/**
|
||||
* GET /api/admin/queues/status - Get job counts for all BullMQ queues.
|
||||
* This is useful for monitoring the health and backlog of background jobs.
|
||||
*/
|
||||
router.get('/queues/status', async (req: Request, res: Response, next: NextFunction) => {
|
||||
router.get('/queues/status', validateRequest(emptySchema), async (req: Request, res: Response, next: NextFunction) => {
|
||||
try {
|
||||
const queues = [flyerQueue, emailQueue, analyticsQueue, cleanupQueue, weeklyAnalyticsQueue];
|
||||
|
||||
const queueStatuses = await Promise.all(
|
||||
queues.map(async (queue) => {
|
||||
return {
|
||||
name: queue.name,
|
||||
counts: await queue.getJobCounts(
|
||||
'waiting',
|
||||
'active',
|
||||
'completed',
|
||||
'failed',
|
||||
'delayed',
|
||||
'paused',
|
||||
),
|
||||
};
|
||||
}),
|
||||
);
|
||||
const queueStatuses = await monitoringService.getQueueStatuses();
|
||||
res.json(queueStatuses);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching queue statuses');
|
||||
@@ -610,35 +587,11 @@ router.post(
|
||||
params: { queueName, jobId },
|
||||
} = req as unknown as z.infer<typeof jobRetrySchema>;
|
||||
|
||||
const queueMap: { [key: string]: Queue } = {
|
||||
'flyer-processing': flyerQueue,
|
||||
'email-sending': emailQueue,
|
||||
'analytics-reporting': analyticsQueue,
|
||||
'file-cleanup': cleanupQueue,
|
||||
};
|
||||
|
||||
const queue = queueMap[queueName];
|
||||
|
||||
if (!queue) {
|
||||
// Throw a NotFoundError to be handled by the central error handler.
|
||||
throw new NotFoundError(`Queue '${queueName}' not found.`);
|
||||
}
|
||||
|
||||
try {
|
||||
const job = await queue.getJob(jobId);
|
||||
if (!job)
|
||||
throw new NotFoundError(`Job with ID '${jobId}' not found in queue '${queueName}'.`);
|
||||
|
||||
const jobState = await job.getState();
|
||||
if (jobState !== 'failed')
|
||||
throw new ValidationError(
|
||||
[],
|
||||
`Job is not in a 'failed' state. Current state: ${jobState}.`,
|
||||
); // This was a duplicate, fixed.
|
||||
|
||||
await job.retry();
|
||||
logger.info(
|
||||
`[Admin] User ${userProfile.user.user_id} manually retried job ${jobId} in queue ${queueName}.`,
|
||||
await monitoringService.retryFailedJob(
|
||||
queueName,
|
||||
jobId,
|
||||
userProfile.user.user_id,
|
||||
);
|
||||
res.status(200).json({ message: `Job ${jobId} has been successfully marked for retry.` });
|
||||
} catch (error) {
|
||||
@@ -653,6 +606,7 @@ router.post(
|
||||
*/
|
||||
router.post(
|
||||
'/trigger/weekly-analytics',
|
||||
validateRequest(emptySchema),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
const userProfile = req.user as UserProfile; // This was a duplicate, fixed.
|
||||
logger.info(
|
||||
@@ -660,19 +614,10 @@ router.post(
|
||||
);
|
||||
|
||||
try {
|
||||
const { year: reportYear, week: reportWeek } = getSimpleWeekAndYear();
|
||||
const { weeklyAnalyticsQueue } = await import('../services/queueService.server');
|
||||
const job = await weeklyAnalyticsQueue.add(
|
||||
'generate-weekly-report',
|
||||
{ reportYear, reportWeek },
|
||||
{
|
||||
jobId: `manual-weekly-report-${reportYear}-${reportWeek}-${Date.now()}`, // Add timestamp to avoid ID conflict
|
||||
},
|
||||
);
|
||||
|
||||
const jobId = await backgroundJobService.triggerWeeklyAnalyticsReport();
|
||||
res
|
||||
.status(202)
|
||||
.json({ message: 'Successfully enqueued weekly analytics job.', jobId: job.id });
|
||||
.json({ message: 'Successfully enqueued weekly analytics job.', jobId });
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error enqueuing weekly analytics job');
|
||||
next(error);
|
||||
@@ -680,4 +625,8 @@ router.post(
|
||||
},
|
||||
);
|
||||
|
||||
/* Catches errors from multer (e.g., file size, file filter) */
|
||||
router.use(handleMulterError);
|
||||
|
||||
|
||||
export default router;
|
||||
|
||||
@@ -4,7 +4,7 @@ import supertest from 'supertest';
|
||||
import type { Request, Response, NextFunction } from 'express';
|
||||
import { createMockUserProfile, createMockAdminUserView } from '../tests/utils/mockFactories';
|
||||
import type { UserProfile, Profile } from '../types';
|
||||
import { NotFoundError } from '../services/db/errors.db';
|
||||
import { NotFoundError, ValidationError } from '../services/db/errors.db';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
|
||||
vi.mock('../services/db/index.db', () => ({
|
||||
@@ -22,6 +22,12 @@ vi.mock('../services/db/index.db', () => ({
|
||||
notificationRepo: {},
|
||||
}));
|
||||
|
||||
vi.mock('../services/userService', () => ({
|
||||
userService: {
|
||||
deleteUserAsAdmin: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// Mock other dependencies that are not directly tested but are part of the adminRouter setup
|
||||
vi.mock('../services/db/flyer.db');
|
||||
vi.mock('../services/db/recipe.db');
|
||||
@@ -53,6 +59,7 @@ import adminRouter from './admin.routes';
|
||||
|
||||
// Import the mocked repos to control them in tests
|
||||
import { adminRepo, userRepo } from '../services/db/index.db';
|
||||
import { userService } from '../services/userService';
|
||||
|
||||
// Mock the passport middleware
|
||||
vi.mock('./passport.routes', () => ({
|
||||
@@ -191,22 +198,27 @@ describe('Admin User Management Routes (/api/admin/users)', () => {
|
||||
it('should successfully delete a user', async () => {
|
||||
const targetId = '123e4567-e89b-12d3-a456-426614174999';
|
||||
vi.mocked(userRepo.deleteUserById).mockResolvedValue(undefined);
|
||||
vi.mocked(userService.deleteUserAsAdmin).mockResolvedValue(undefined);
|
||||
const response = await supertest(app).delete(`/api/admin/users/${targetId}`);
|
||||
expect(response.status).toBe(204);
|
||||
expect(userRepo.deleteUserById).toHaveBeenCalledWith(targetId, expect.any(Object));
|
||||
expect(userService.deleteUserAsAdmin).toHaveBeenCalledWith(adminId, targetId, expect.any(Object));
|
||||
});
|
||||
|
||||
it('should prevent an admin from deleting their own account', async () => {
|
||||
const validationError = new ValidationError([], 'Admins cannot delete their own account.');
|
||||
vi.mocked(userService.deleteUserAsAdmin).mockRejectedValue(validationError);
|
||||
const response = await supertest(app).delete(`/api/admin/users/${adminId}`);
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.message).toMatch(/Admins cannot delete their own account/);
|
||||
expect(userRepo.deleteUserById).not.toHaveBeenCalled();
|
||||
expect(userService.deleteUserAsAdmin).toHaveBeenCalledWith(adminId, adminId, expect.any(Object));
|
||||
});
|
||||
|
||||
it('should return 500 on a generic database error', async () => {
|
||||
const targetId = '123e4567-e89b-12d3-a456-426614174999';
|
||||
const dbError = new Error('DB Error');
|
||||
vi.mocked(userRepo.deleteUserById).mockRejectedValue(dbError);
|
||||
vi.mocked(userService.deleteUserAsAdmin).mockRejectedValue(dbError);
|
||||
const response = await supertest(app).delete(`/api/admin/users/${targetId}`);
|
||||
expect(response.status).toBe(500);
|
||||
});
|
||||
|
||||
@@ -13,14 +13,21 @@ import {
|
||||
import * as aiService from '../services/aiService.server';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
import { mockLogger } from '../tests/utils/mockLogger';
|
||||
import { ValidationError } from '../services/db/errors.db';
|
||||
|
||||
// Mock the AI service methods to avoid making real AI calls
|
||||
vi.mock('../services/aiService.server', () => ({
|
||||
aiService: {
|
||||
extractTextFromImageArea: vi.fn(),
|
||||
planTripWithMaps: vi.fn(), // Added this missing mock
|
||||
},
|
||||
}));
|
||||
vi.mock('../services/aiService.server', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof import('../services/aiService.server')>();
|
||||
return {
|
||||
...actual,
|
||||
aiService: {
|
||||
extractTextFromImageArea: vi.fn(),
|
||||
planTripWithMaps: vi.fn(),
|
||||
enqueueFlyerProcessing: vi.fn(),
|
||||
processLegacyFlyerUpload: vi.fn(),
|
||||
},
|
||||
};
|
||||
});
|
||||
|
||||
const { mockedDb } = vi.hoisted(() => ({
|
||||
mockedDb: {
|
||||
@@ -30,6 +37,9 @@ const { mockedDb } = vi.hoisted(() => ({
|
||||
adminRepo: {
|
||||
logActivity: vi.fn(),
|
||||
},
|
||||
personalizationRepo: {
|
||||
getAllMasterItems: vi.fn(),
|
||||
},
|
||||
// This function is a standalone export, not part of a repo
|
||||
createFlyerAndItems: vi.fn(),
|
||||
},
|
||||
@@ -40,6 +50,7 @@ vi.mock('../services/db/flyer.db', () => ({ createFlyerAndItems: mockedDb.create
|
||||
vi.mock('../services/db/index.db', () => ({
|
||||
flyerRepo: mockedDb.flyerRepo,
|
||||
adminRepo: mockedDb.adminRepo,
|
||||
personalizationRepo: mockedDb.personalizationRepo,
|
||||
}));
|
||||
|
||||
// Mock the queue service
|
||||
@@ -83,36 +94,6 @@ describe('AI Routes (/api/ai)', () => {
|
||||
});
|
||||
const app = createTestApp({ router: aiRouter, basePath: '/api/ai' });
|
||||
|
||||
describe('Module-level error handling', () => {
|
||||
it('should log an error if storage path creation fails', async () => {
|
||||
// Arrange
|
||||
const mkdirError = new Error('EACCES: permission denied');
|
||||
vi.resetModules(); // Reset modules to re-run top-level code
|
||||
vi.doMock('node:fs', () => {
|
||||
const mockFs = {
|
||||
...fs,
|
||||
mkdirSync: vi.fn().mockImplementation(() => {
|
||||
throw mkdirError;
|
||||
}),
|
||||
};
|
||||
return { ...mockFs, default: mockFs };
|
||||
});
|
||||
const { logger } = await import('../services/logger.server');
|
||||
|
||||
// Act: Dynamically import the router to trigger the mkdirSync call
|
||||
await import('./ai.routes');
|
||||
|
||||
// Assert
|
||||
const storagePath =
|
||||
process.env.STORAGE_PATH || '/var/www/flyer-crawler.projectium.com/flyer-images';
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ error: 'EACCES: permission denied' },
|
||||
`Failed to create storage path (${storagePath}). File uploads may fail.`,
|
||||
);
|
||||
vi.doUnmock('node:fs'); // Cleanup
|
||||
});
|
||||
});
|
||||
|
||||
// New test to cover the router.use diagnostic middleware's catch block and errMsg branches
|
||||
describe('Diagnostic Middleware Error Handling', () => {
|
||||
it('should log an error if logger.debug throws an object with a message property', async () => {
|
||||
@@ -166,26 +147,27 @@ describe('AI Routes (/api/ai)', () => {
|
||||
|
||||
describe('POST /upload-and-process', () => {
|
||||
const imagePath = path.resolve(__dirname, '../tests/assets/test-flyer-image.jpg');
|
||||
// A valid SHA-256 checksum is 64 hex characters.
|
||||
const validChecksum = 'a'.repeat(64);
|
||||
|
||||
it('should enqueue a job and return 202 on success', async () => {
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job-123' } as unknown as Job);
|
||||
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockResolvedValue({ id: 'job-123' } as unknown as Job);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', 'new-checksum')
|
||||
.field('checksum', validChecksum)
|
||||
.attach('flyerFile', imagePath);
|
||||
|
||||
expect(response.status).toBe(202);
|
||||
expect(response.body.message).toBe('Flyer accepted for processing.');
|
||||
expect(response.body.jobId).toBe('job-123');
|
||||
expect(flyerQueue.add).toHaveBeenCalledWith('process-flyer', expect.any(Object));
|
||||
expect(aiService.aiService.enqueueFlyerProcessing).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should return 400 if no file is provided', async () => {
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', 'some-checksum');
|
||||
.field('checksum', validChecksum);
|
||||
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.message).toBe('A flyer file (PDF or image) is required.');
|
||||
@@ -202,13 +184,12 @@ describe('AI Routes (/api/ai)', () => {
|
||||
});
|
||||
|
||||
it('should return 409 if flyer checksum already exists', async () => {
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(
|
||||
createMockFlyer({ flyer_id: 99 }),
|
||||
);
|
||||
const duplicateError = new aiService.DuplicateFlyerError('This flyer has already been processed.', 99);
|
||||
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockRejectedValue(duplicateError);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', 'duplicate-checksum')
|
||||
.field('checksum', validChecksum)
|
||||
.attach('flyerFile', imagePath);
|
||||
|
||||
expect(response.status).toBe(409);
|
||||
@@ -216,12 +197,11 @@ describe('AI Routes (/api/ai)', () => {
|
||||
});
|
||||
|
||||
it('should return 500 if enqueuing the job fails', async () => {
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(flyerQueue.add).mockRejectedValueOnce(new Error('Redis connection failed'));
|
||||
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockRejectedValueOnce(new Error('Redis connection failed'));
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', 'new-checksum')
|
||||
.field('checksum', validChecksum)
|
||||
.attach('flyerFile', imagePath);
|
||||
|
||||
expect(response.status).toBe(500);
|
||||
@@ -239,19 +219,20 @@ describe('AI Routes (/api/ai)', () => {
|
||||
basePath: '/api/ai',
|
||||
authenticatedUser: mockUser,
|
||||
});
|
||||
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job-456' } as unknown as Job);
|
||||
|
||||
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockResolvedValue({ id: 'job-456' } as unknown as Job);
|
||||
|
||||
// Act
|
||||
await supertest(authenticatedApp)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', 'auth-checksum')
|
||||
.field('checksum', validChecksum)
|
||||
.attach('flyerFile', imagePath);
|
||||
|
||||
// Assert
|
||||
expect(flyerQueue.add).toHaveBeenCalled();
|
||||
expect(vi.mocked(flyerQueue.add).mock.calls[0][1].userId).toBe('auth-user-1');
|
||||
expect(aiService.aiService.enqueueFlyerProcessing).toHaveBeenCalled();
|
||||
const callArgs = vi.mocked(aiService.aiService.enqueueFlyerProcessing).mock.calls[0];
|
||||
// Check the userProfile argument (3rd argument)
|
||||
expect(callArgs[2]?.user.user_id).toBe('auth-user-1');
|
||||
});
|
||||
|
||||
it('should pass user profile address to the job when authenticated user has an address', async () => {
|
||||
@@ -273,17 +254,35 @@ describe('AI Routes (/api/ai)', () => {
|
||||
basePath: '/api/ai',
|
||||
authenticatedUser: mockUserWithAddress,
|
||||
});
|
||||
|
||||
vi.mocked(aiService.aiService.enqueueFlyerProcessing).mockResolvedValue({ id: 'job-789' } as unknown as Job);
|
||||
|
||||
// Act
|
||||
await supertest(authenticatedApp)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.field('checksum', 'addr-checksum')
|
||||
.field('checksum', validChecksum)
|
||||
.attach('flyerFile', imagePath);
|
||||
|
||||
// Assert
|
||||
expect(vi.mocked(flyerQueue.add).mock.calls[0][1].userProfileAddress).toBe(
|
||||
'123 Pacific St, Anytown, BC, V8T 1A1, CA',
|
||||
);
|
||||
expect(aiService.aiService.enqueueFlyerProcessing).toHaveBeenCalled();
|
||||
// The service handles address extraction from profile, so we just verify the profile was passed
|
||||
const callArgs = vi.mocked(aiService.aiService.enqueueFlyerProcessing).mock.calls[0];
|
||||
expect(callArgs[2]?.address?.address_line_1).toBe('123 Pacific St');
|
||||
});
|
||||
|
||||
it('should clean up the uploaded file if validation fails (e.g., missing checksum)', async () => {
|
||||
// Spy on the unlink function to ensure it's called on error
|
||||
const unlinkSpy = vi.spyOn(fs.promises, 'unlink').mockResolvedValue(undefined);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/upload-and-process')
|
||||
.attach('flyerFile', imagePath); // No checksum field, will cause validation to throw
|
||||
|
||||
expect(response.status).toBe(400);
|
||||
// The validation error is now caught inside the route handler, which then calls cleanup.
|
||||
expect(unlinkSpy).toHaveBeenCalledTimes(1);
|
||||
|
||||
unlinkSpy.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
@@ -331,9 +330,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
flyer_id: 1,
|
||||
file_name: mockDataPayload.originalFileName,
|
||||
});
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined); // No duplicate
|
||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
||||
vi.mocked(mockedDb.adminRepo.logActivity).mockResolvedValue();
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(mockFlyer);
|
||||
|
||||
// Act
|
||||
const response = await supertest(app)
|
||||
@@ -344,7 +341,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
// Assert
|
||||
expect(response.status).toBe(201);
|
||||
expect(response.body.message).toBe('Flyer processed and saved successfully.');
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should return 400 if no flyer image is provided', async () => {
|
||||
@@ -356,8 +353,8 @@ describe('AI Routes (/api/ai)', () => {
|
||||
|
||||
it('should return 409 Conflict and delete the uploaded file if flyer checksum already exists', async () => {
|
||||
// Arrange
|
||||
const mockExistingFlyer = createMockFlyer({ flyer_id: 99 });
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(mockExistingFlyer); // Duplicate found
|
||||
const duplicateError = new aiService.DuplicateFlyerError('This flyer has already been processed.', 99);
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockRejectedValue(duplicateError);
|
||||
const unlinkSpy = vi.spyOn(fs.promises, 'unlink').mockResolvedValue(undefined);
|
||||
|
||||
// Act
|
||||
@@ -369,7 +366,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
// Assert
|
||||
expect(response.status).toBe(409);
|
||||
expect(response.body.message).toBe('This flyer has already been processed.');
|
||||
expect(mockedDb.createFlyerAndItems).not.toHaveBeenCalled();
|
||||
expect(mockedDb.createFlyerAndItems).not.toHaveBeenCalled(); // Should not be called if service throws
|
||||
// Assert that the file was deleted
|
||||
expect(unlinkSpy).toHaveBeenCalledTimes(1);
|
||||
// The filename is predictable in the test environment because of the multer config in ai.routes.ts
|
||||
@@ -384,12 +381,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
extractedData: { store_name: 'Partial Store' }, // no items key
|
||||
};
|
||||
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
const mockFlyer = createMockFlyer({
|
||||
flyer_id: 2,
|
||||
file_name: partialPayload.originalFileName,
|
||||
});
|
||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(createMockFlyer({ flyer_id: 2 }));
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/flyers/process')
|
||||
@@ -397,13 +389,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// verify the items array passed to DB was an empty array
|
||||
const callArgs = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0]?.[1];
|
||||
expect(callArgs).toBeDefined();
|
||||
expect(Array.isArray(callArgs)).toBe(true);
|
||||
// use non-null assertion for the runtime-checked variable so TypeScript is satisfied
|
||||
expect(callArgs!.length).toBe(0);
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should fallback to a safe store name when store_name is missing', async () => {
|
||||
@@ -413,12 +399,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
extractedData: { items: [] }, // store_name missing
|
||||
};
|
||||
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
const mockFlyer = createMockFlyer({
|
||||
flyer_id: 3,
|
||||
file_name: payloadNoStore.originalFileName,
|
||||
});
|
||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(createMockFlyer({ flyer_id: 3 }));
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/flyers/process')
|
||||
@@ -426,19 +407,11 @@ describe('AI Routes (/api/ai)', () => {
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// verify the flyerData.store_name passed to DB was the fallback string
|
||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
||||
expect(flyerDataArg.store_name).toContain('Unknown Store');
|
||||
// Also verify the warning was logged
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
'extractedData.store_name missing; using fallback store name to avoid DB constraint error.',
|
||||
);
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should handle a generic error during flyer creation', async () => {
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(mockedDb.createFlyerAndItems).mockRejectedValueOnce(
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockRejectedValueOnce(
|
||||
new Error('DB transaction failed'),
|
||||
);
|
||||
|
||||
@@ -461,8 +434,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
|
||||
beforeEach(() => {
|
||||
const mockFlyer = createMockFlyer({ flyer_id: 1 });
|
||||
vi.mocked(mockedDb.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(mockedDb.createFlyerAndItems).mockResolvedValue({ flyer: mockFlyer, items: [] });
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockResolvedValue(mockFlyer);
|
||||
});
|
||||
|
||||
it('should handle payload where "data" field is an object, not stringified JSON', async () => {
|
||||
@@ -472,7 +444,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should handle payload where extractedData is null', async () => {
|
||||
@@ -488,14 +460,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// Verify that extractedData was correctly defaulted to an empty object
|
||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
||||
expect(flyerDataArg.store_name).toContain('Unknown Store'); // Fallback should be used
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
{ bodyData: expect.any(Object) },
|
||||
'Missing extractedData in /api/ai/flyers/process payload.',
|
||||
);
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should handle payload where extractedData is a string', async () => {
|
||||
@@ -511,14 +476,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
// Verify that extractedData was correctly defaulted to an empty object
|
||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
||||
expect(flyerDataArg.store_name).toContain('Unknown Store'); // Fallback should be used
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
{ bodyData: expect.any(Object) },
|
||||
'Missing extractedData in /api/ai/flyers/process payload.',
|
||||
);
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should handle payload where extractedData is at the root of the body', async () => {
|
||||
@@ -532,9 +490,7 @@ describe('AI Routes (/api/ai)', () => {
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201); // This test was failing with 500, the fix is in ai.routes.ts
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
const flyerDataArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][0];
|
||||
expect(flyerDataArg.store_name).toBe('Root Store');
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should default item quantity to 1 if missing', async () => {
|
||||
@@ -553,9 +509,55 @@ describe('AI Routes (/api/ai)', () => {
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(201);
|
||||
expect(mockedDb.createFlyerAndItems).toHaveBeenCalledTimes(1);
|
||||
const itemsArg = vi.mocked(mockedDb.createFlyerAndItems).mock.calls[0][1];
|
||||
expect(itemsArg[0].quantity).toBe(1);
|
||||
expect(aiService.aiService.processLegacyFlyerUpload).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
});
|
||||
|
||||
describe('POST /flyers/process (Legacy Error Handling)', () => {
|
||||
const imagePath = path.resolve(__dirname, '../tests/assets/test-flyer-image.jpg');
|
||||
|
||||
it('should handle malformed JSON in data field and return 400', async () => {
|
||||
const malformedDataString = '{"checksum":'; // Invalid JSON
|
||||
|
||||
// Since the service parses the data, we mock it to throw a ValidationError when parsing fails
|
||||
// or when it detects the malformed input.
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockRejectedValue(new ValidationError([], 'Checksum is required.'));
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/flyers/process')
|
||||
.field('data', malformedDataString)
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
// The outer catch block should be hit, leading to empty parsed data.
|
||||
// The handler then fails the checksum validation.
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.message).toBe('Checksum is required.');
|
||||
// Note: The logging expectation was removed because if the service throws a ValidationError,
|
||||
// the route handler passes it to the global error handler, which might log differently or not as a "critical error during parsing" in the route itself.
|
||||
});
|
||||
|
||||
it('should return 400 if checksum is missing from legacy payload', async () => {
|
||||
const payloadWithoutChecksum = {
|
||||
originalFileName: 'flyer.jpg',
|
||||
extractedData: { store_name: 'Test Store', items: [] },
|
||||
};
|
||||
// Spy on fs.promises.unlink to verify file cleanup
|
||||
const unlinkSpy = vi.spyOn(fs.promises, 'unlink').mockResolvedValue(undefined);
|
||||
|
||||
// Mock the service to throw a ValidationError because the checksum is missing
|
||||
vi.mocked(aiService.aiService.processLegacyFlyerUpload).mockRejectedValue(new ValidationError([], 'Checksum is required.'));
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/flyers/process')
|
||||
.field('data', JSON.stringify(payloadWithoutChecksum))
|
||||
.attach('flyerImage', imagePath);
|
||||
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.message).toBe('Checksum is required.');
|
||||
// Ensure the uploaded file is cleaned up
|
||||
expect(unlinkSpy).toHaveBeenCalledTimes(1);
|
||||
|
||||
unlinkSpy.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
@@ -828,6 +830,39 @@ describe('AI Routes (/api/ai)', () => {
|
||||
expect(response.body.message).toBe('Maps API key invalid');
|
||||
});
|
||||
|
||||
it('POST /deep-dive should return 500 on a generic error', async () => {
|
||||
vi.mocked(mockLogger.info).mockImplementationOnce(() => {
|
||||
throw new Error('Deep dive logging failed');
|
||||
});
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/deep-dive')
|
||||
.send({ items: [{ name: 'test' }] });
|
||||
expect(response.status).toBe(500);
|
||||
expect(response.body.message).toBe('Deep dive logging failed');
|
||||
});
|
||||
|
||||
it('POST /search-web should return 500 on a generic error', async () => {
|
||||
vi.mocked(mockLogger.info).mockImplementationOnce(() => {
|
||||
throw new Error('Search web logging failed');
|
||||
});
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/search-web')
|
||||
.send({ query: 'test query' });
|
||||
expect(response.status).toBe(500);
|
||||
expect(response.body.message).toBe('Search web logging failed');
|
||||
});
|
||||
|
||||
it('POST /compare-prices should return 500 on a generic error', async () => {
|
||||
vi.mocked(mockLogger.info).mockImplementationOnce(() => {
|
||||
throw new Error('Compare prices logging failed');
|
||||
});
|
||||
const response = await supertest(app)
|
||||
.post('/api/ai/compare-prices')
|
||||
.send({ items: [{ name: 'Milk' }] });
|
||||
expect(response.status).toBe(500);
|
||||
expect(response.body.message).toBe('Compare prices logging failed');
|
||||
});
|
||||
|
||||
it('POST /quick-insights should return 400 if items are missing', async () => {
|
||||
const response = await supertest(app).post('/api/ai/quick-insights').send({});
|
||||
expect(response.status).toBe(400);
|
||||
|
||||
@@ -1,38 +1,32 @@
|
||||
// src/routes/ai.routes.ts
|
||||
import { Router, Request, Response, NextFunction } from 'express';
|
||||
import multer from 'multer';
|
||||
import path from 'path';
|
||||
import fs from 'node:fs';
|
||||
import { z } from 'zod';
|
||||
import passport from './passport.routes';
|
||||
import { optionalAuth } from './passport.routes';
|
||||
import * as db from '../services/db/index.db';
|
||||
import { createFlyerAndItems } from '../services/db/flyer.db';
|
||||
import * as aiService from '../services/aiService.server'; // Correctly import server-side AI service
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
import { sanitizeFilename } from '../utils/stringUtils';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { UserProfile, ExtractedCoreData, ExtractedFlyerItem } from '../types';
|
||||
import { flyerQueue } from '../services/queueService.server';
|
||||
import { aiService, DuplicateFlyerError } from '../services/aiService.server';
|
||||
import {
|
||||
createUploadMiddleware,
|
||||
handleMulterError,
|
||||
} from '../middleware/multer.middleware';
|
||||
import { logger } from '../services/logger.server'; // This was a duplicate, fixed.
|
||||
import { UserProfile } from '../types'; // This was a duplicate, fixed.
|
||||
import { validateRequest } from '../middleware/validation.middleware';
|
||||
import { requiredString } from '../utils/zodUtils';
|
||||
import { cleanupUploadedFile, cleanupUploadedFiles } from '../utils/fileUtils';
|
||||
import { monitoringService } from '../services/monitoringService.server';
|
||||
|
||||
const router = Router();
|
||||
|
||||
interface FlyerProcessPayload extends Partial<ExtractedCoreData> {
|
||||
checksum?: string;
|
||||
originalFileName?: string;
|
||||
extractedData?: Partial<ExtractedCoreData>;
|
||||
data?: FlyerProcessPayload; // For nested data structures
|
||||
}
|
||||
|
||||
// --- Zod Schemas for AI Routes (as per ADR-003) ---
|
||||
|
||||
const uploadAndProcessSchema = z.object({
|
||||
body: z.object({
|
||||
checksum: requiredString('File checksum is required.'),
|
||||
// Potential improvement: If checksum is always a specific format (e.g., SHA-256),
|
||||
// you could add `.length(64).regex(/^[a-f0-9]+$/)` for stricter validation.
|
||||
// Stricter validation for SHA-256 checksum. It must be a 64-character hexadecimal string.
|
||||
checksum: requiredString('File checksum is required.').pipe(
|
||||
z.string()
|
||||
.length(64, 'Checksum must be 64 characters long.')
|
||||
.regex(/^[a-f0-9]+$/, 'Checksum must be a valid hexadecimal string.'),
|
||||
),
|
||||
}),
|
||||
});
|
||||
|
||||
@@ -50,15 +44,6 @@ const errMsg = (e: unknown) => {
|
||||
return String(e || 'An unknown error occurred.');
|
||||
};
|
||||
|
||||
const cleanupUploadedFile = async (file?: Express.Multer.File) => {
|
||||
if (!file) return;
|
||||
try {
|
||||
await fs.promises.unlink(file.path);
|
||||
} catch (err) {
|
||||
// Ignore cleanup errors (e.g. file already deleted)
|
||||
}
|
||||
};
|
||||
|
||||
const cropAreaObjectSchema = z.object({
|
||||
x: z.number(),
|
||||
y: z.number(),
|
||||
@@ -87,7 +72,6 @@ const rescanAreaSchema = z.object({
|
||||
})
|
||||
.pipe(cropAreaObjectSchema), // Further validate the structure of the parsed object
|
||||
extractionType: z.enum(['store_name', 'dates', 'item_details'], {
|
||||
// This is the line with the error
|
||||
message: "extractionType must be one of 'store_name', 'dates', or 'item_details'.",
|
||||
}),
|
||||
}),
|
||||
@@ -95,13 +79,20 @@ const rescanAreaSchema = z.object({
|
||||
|
||||
const flyerItemForAnalysisSchema = z
|
||||
.object({
|
||||
item: z.string().nullish(),
|
||||
name: z.string().nullish(),
|
||||
// Sanitize item and name by trimming whitespace.
|
||||
// The transform ensures that null/undefined values are preserved
|
||||
// while trimming any actual string values.
|
||||
item: z.string().nullish().transform(val => (val ? val.trim() : val)),
|
||||
name: z.string().nullish().transform(val => (val ? val.trim() : val)),
|
||||
})
|
||||
// Using .passthrough() allows extra properties on the item object.
|
||||
// If the intent is to strictly enforce only 'item' and 'name' (and other known properties),
|
||||
// consider using .strict() instead for tighter security and data integrity.
|
||||
.passthrough()
|
||||
.refine(
|
||||
(data) =>
|
||||
(data.item && data.item.trim().length > 0) || (data.name && data.name.trim().length > 0),
|
||||
// After the transform, the values are already trimmed.
|
||||
(data.item && data.item.length > 0) || (data.name && data.name.length > 0),
|
||||
{
|
||||
message: "Item identifier is required (either 'item' or 'name').",
|
||||
},
|
||||
@@ -121,6 +112,8 @@ const comparePricesSchema = z.object({
|
||||
|
||||
const planTripSchema = z.object({
|
||||
body: z.object({
|
||||
// Consider if this array should be non-empty. If a trip plan requires at least one item,
|
||||
// you could add `.nonempty('At least one item is required to plan a trip.')`
|
||||
items: z.array(flyerItemForAnalysisSchema),
|
||||
store: z.object({ name: requiredString('Store name is required.') }),
|
||||
userLocation: z.object({
|
||||
@@ -148,40 +141,7 @@ const searchWebSchema = z.object({
|
||||
body: z.object({ query: requiredString('A search query is required.') }),
|
||||
});
|
||||
|
||||
// --- Multer Configuration for File Uploads ---
|
||||
const storagePath =
|
||||
process.env.STORAGE_PATH || '/var/www/flyer-crawler.projectium.com/flyer-images';
|
||||
|
||||
// Ensure the storage path exists at startup so multer can write files there.
|
||||
try {
|
||||
fs.mkdirSync(storagePath, { recursive: true });
|
||||
logger.debug(`AI upload storage path ready: ${storagePath}`);
|
||||
} catch (err) {
|
||||
logger.error(
|
||||
{ error: errMsg(err) },
|
||||
`Failed to create storage path (${storagePath}). File uploads may fail.`,
|
||||
);
|
||||
}
|
||||
const diskStorage = multer.diskStorage({
|
||||
destination: function (req, file, cb) {
|
||||
cb(null, storagePath);
|
||||
},
|
||||
filename: function (req, file, cb) {
|
||||
// If in a test environment, use a predictable filename for easy cleanup.
|
||||
if (process.env.NODE_ENV === 'test') {
|
||||
return cb(null, `${file.fieldname}-test-flyer-image.jpg`);
|
||||
} else {
|
||||
const uniqueSuffix = Date.now() + '-' + Math.round(Math.random() * 1e9);
|
||||
// Sanitize the original filename to remove spaces and special characters
|
||||
return cb(
|
||||
null,
|
||||
file.fieldname + '-' + uniqueSuffix + '-' + sanitizeFilename(file.originalname),
|
||||
);
|
||||
}
|
||||
},
|
||||
});
|
||||
|
||||
const uploadToDisk = multer({ storage: diskStorage });
|
||||
const uploadToDisk = createUploadMiddleware({ storageType: 'flyer' });
|
||||
|
||||
// Diagnostic middleware: log incoming AI route requests (headers and sizes)
|
||||
router.use((req: Request, res: Response, next: NextFunction) => {
|
||||
@@ -207,58 +167,29 @@ router.post(
|
||||
'/upload-and-process',
|
||||
optionalAuth,
|
||||
uploadToDisk.single('flyerFile'),
|
||||
validateRequest(uploadAndProcessSchema),
|
||||
// Validation is now handled inside the route to ensure file cleanup on failure.
|
||||
// validateRequest(uploadAndProcessSchema),
|
||||
async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
// Manually validate the request body. This will throw if validation fails.
|
||||
const { body } = uploadAndProcessSchema.parse({ body: req.body });
|
||||
|
||||
if (!req.file) {
|
||||
return res.status(400).json({ message: 'A flyer file (PDF or image) is required.' });
|
||||
}
|
||||
|
||||
logger.debug(
|
||||
{ filename: req.file.originalname, size: req.file.size, checksum: req.body.checksum },
|
||||
{ filename: req.file.originalname, size: req.file.size, checksum: body.checksum },
|
||||
'Handling /upload-and-process',
|
||||
);
|
||||
|
||||
const { checksum } = req.body;
|
||||
// Check for duplicate flyer using checksum before even creating a job
|
||||
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, req.log);
|
||||
if (existingFlyer) {
|
||||
logger.warn(`Duplicate flyer upload attempt blocked for checksum: ${checksum}`);
|
||||
// Use 409 Conflict for duplicates
|
||||
return res.status(409).json({
|
||||
message: 'This flyer has already been processed.',
|
||||
flyerId: existingFlyer.flyer_id,
|
||||
});
|
||||
}
|
||||
|
||||
const userProfile = req.user as UserProfile | undefined;
|
||||
// Construct a user address string from their profile if they are logged in.
|
||||
let userProfileAddress: string | undefined = undefined;
|
||||
if (userProfile?.address) {
|
||||
userProfileAddress = [
|
||||
userProfile.address.address_line_1,
|
||||
userProfile.address.address_line_2,
|
||||
userProfile.address.city,
|
||||
userProfile.address.province_state,
|
||||
userProfile.address.postal_code,
|
||||
userProfile.address.country,
|
||||
]
|
||||
.filter(Boolean)
|
||||
.join(', ');
|
||||
}
|
||||
|
||||
// Add job to the queue
|
||||
const job = await flyerQueue.add('process-flyer', {
|
||||
filePath: req.file.path,
|
||||
originalFileName: req.file.originalname,
|
||||
checksum: checksum,
|
||||
userId: userProfile?.user.user_id,
|
||||
submitterIp: req.ip, // Capture the submitter's IP address
|
||||
userProfileAddress: userProfileAddress, // Pass the user's profile address
|
||||
});
|
||||
|
||||
logger.info(
|
||||
`Enqueued flyer for processing. File: ${req.file.originalname}, Job ID: ${job.id}`,
|
||||
const job = await aiService.enqueueFlyerProcessing(
|
||||
req.file,
|
||||
body.checksum,
|
||||
userProfile,
|
||||
req.ip ?? 'unknown',
|
||||
req.log,
|
||||
);
|
||||
|
||||
// Respond immediately to the client with 202 Accepted
|
||||
@@ -267,6 +198,11 @@ router.post(
|
||||
jobId: job.id,
|
||||
});
|
||||
} catch (error) {
|
||||
await cleanupUploadedFile(req.file);
|
||||
if (error instanceof DuplicateFlyerError) {
|
||||
logger.warn(`Duplicate flyer upload attempt blocked for checksum: ${req.body?.checksum}`);
|
||||
return res.status(409).json({ message: error.message, flyerId: error.flyerId });
|
||||
}
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -283,18 +219,11 @@ router.get(
|
||||
const {
|
||||
params: { jobId },
|
||||
} = req as unknown as JobIdRequest;
|
||||
|
||||
try {
|
||||
const job = await flyerQueue.getJob(jobId);
|
||||
if (!job) {
|
||||
// Adhere to ADR-001 by throwing a specific error to be handled centrally.
|
||||
return res.status(404).json({ message: 'Job not found.' });
|
||||
}
|
||||
const state = await job.getState();
|
||||
const progress = job.progress;
|
||||
const returnValue = job.returnvalue;
|
||||
const failedReason = job.failedReason;
|
||||
logger.debug(`[API /ai/jobs] Status check for job ${jobId}: ${state}`);
|
||||
res.json({ id: job.id, state, progress, returnValue, failedReason });
|
||||
const jobStatus = await monitoringService.getFlyerJobStatus(jobId); // This was a duplicate, fixed.
|
||||
logger.debug(`[API /ai/jobs] Status check for job ${jobId}: ${jobStatus.state}`);
|
||||
res.json(jobStatus);
|
||||
} catch (error) {
|
||||
next(error);
|
||||
}
|
||||
@@ -316,184 +245,22 @@ router.post(
|
||||
return res.status(400).json({ message: 'Flyer image file is required.' });
|
||||
}
|
||||
|
||||
// Diagnostic & tolerant parsing for flyers/process
|
||||
logger.debug(
|
||||
{ keys: Object.keys(req.body || {}) },
|
||||
'[API /ai/flyers/process] Processing legacy upload',
|
||||
);
|
||||
logger.debug({ filePresent: !!req.file }, '[API /ai/flyers/process] file present:');
|
||||
|
||||
// Try several ways to obtain the payload so we are tolerant to client variations.
|
||||
let parsed: FlyerProcessPayload = {};
|
||||
let extractedData: Partial<ExtractedCoreData> | null | undefined = {};
|
||||
try {
|
||||
// If the client sent a top-level `data` field (stringified JSON), parse it.
|
||||
if (req.body && (req.body.data || req.body.extractedData)) {
|
||||
const raw = req.body.data ?? req.body.extractedData;
|
||||
logger.debug(
|
||||
{ type: typeof raw, length: raw?.length ?? 0 },
|
||||
'[API /ai/flyers/process] raw extractedData',
|
||||
);
|
||||
try {
|
||||
parsed = typeof raw === 'string' ? JSON.parse(raw) : raw;
|
||||
} catch (err) {
|
||||
logger.warn(
|
||||
{ error: errMsg(err) },
|
||||
'[API /ai/flyers/process] Failed to JSON.parse raw extractedData; falling back to direct assign',
|
||||
);
|
||||
parsed = (
|
||||
typeof raw === 'string' ? JSON.parse(String(raw).slice(0, 2000)) : raw
|
||||
) as FlyerProcessPayload;
|
||||
}
|
||||
// If parsed itself contains an `extractedData` field, use that, otherwise assume parsed is the extractedData
|
||||
extractedData = 'extractedData' in parsed ? parsed.extractedData : (parsed as Partial<ExtractedCoreData>);
|
||||
} else {
|
||||
// No explicit `data` field found. Attempt to interpret req.body as an object (Express may have parsed multipart fields differently).
|
||||
try {
|
||||
parsed = typeof req.body === 'string' ? JSON.parse(req.body) : req.body;
|
||||
} catch (err) {
|
||||
logger.warn(
|
||||
{ error: errMsg(err) },
|
||||
'[API /ai/flyers/process] Failed to JSON.parse req.body; using empty object',
|
||||
);
|
||||
parsed = (req.body as FlyerProcessPayload) || {};
|
||||
}
|
||||
// extractedData might be nested under `data` or `extractedData`, or the body itself may be the extracted data.
|
||||
if (parsed.data) {
|
||||
try {
|
||||
const inner = typeof parsed.data === 'string' ? JSON.parse(parsed.data) : parsed.data;
|
||||
extractedData = inner.extractedData ?? inner;
|
||||
} catch (err) {
|
||||
logger.warn(
|
||||
{ error: errMsg(err) },
|
||||
'[API /ai/flyers/process] Failed to parse parsed.data; falling back',
|
||||
);
|
||||
extractedData = parsed.data as unknown as Partial<ExtractedCoreData>;
|
||||
}
|
||||
} else if (parsed.extractedData) {
|
||||
extractedData = parsed.extractedData;
|
||||
} else {
|
||||
// Assume the body itself is the extracted data if it looks like it (has items or store_name keys)
|
||||
if ('items' in parsed || 'store_name' in parsed || 'valid_from' in parsed) {
|
||||
extractedData = parsed as Partial<ExtractedCoreData>;
|
||||
} else {
|
||||
extractedData = {};
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch (err) {
|
||||
logger.error(
|
||||
{ error: err },
|
||||
'[API /ai/flyers/process] Unexpected error while parsing request body',
|
||||
);
|
||||
parsed = {};
|
||||
extractedData = {};
|
||||
}
|
||||
|
||||
// Pull common metadata fields (checksum, originalFileName) from whichever shape we parsed.
|
||||
const checksum = parsed.checksum ?? parsed?.data?.checksum ?? '';
|
||||
|
||||
if (!checksum) {
|
||||
await cleanupUploadedFile(req.file);
|
||||
return res.status(400).json({ message: 'Checksum is required.' });
|
||||
}
|
||||
|
||||
const originalFileName =
|
||||
parsed.originalFileName ?? parsed?.data?.originalFileName ?? req.file.originalname;
|
||||
const userProfile = req.user as UserProfile | undefined;
|
||||
|
||||
// Validate extractedData to avoid database errors (e.g., null store_name)
|
||||
if (!extractedData || typeof extractedData !== 'object') {
|
||||
logger.warn(
|
||||
{ bodyData: parsed },
|
||||
'Missing extractedData in /api/ai/flyers/process payload.',
|
||||
);
|
||||
// Don't fail hard here; proceed with empty items and fallback store name so the upload can be saved for manual review.
|
||||
extractedData = {};
|
||||
}
|
||||
|
||||
// Transform the extracted items into the format required for database insertion.
|
||||
// This adds default values for fields like `view_count` and `click_count`
|
||||
// and makes this legacy endpoint consistent with the newer FlyerDataTransformer service.
|
||||
const rawItems = extractedData.items ?? [];
|
||||
const itemsArray = Array.isArray(rawItems)
|
||||
? rawItems
|
||||
: typeof rawItems === 'string'
|
||||
? JSON.parse(rawItems)
|
||||
: [];
|
||||
const itemsForDb = itemsArray.map((item: Partial<ExtractedFlyerItem>) => ({
|
||||
...item,
|
||||
master_item_id: item.master_item_id === null ? undefined : item.master_item_id,
|
||||
quantity: item.quantity ?? 1, // Default to 1 to satisfy DB constraint
|
||||
view_count: 0,
|
||||
click_count: 0,
|
||||
updated_at: new Date().toISOString(),
|
||||
}));
|
||||
|
||||
// Ensure we have a valid store name; the DB requires a non-null store name.
|
||||
const storeName =
|
||||
extractedData.store_name && String(extractedData.store_name).trim().length > 0
|
||||
? String(extractedData.store_name)
|
||||
: 'Unknown Store (auto)';
|
||||
if (storeName.startsWith('Unknown')) {
|
||||
logger.warn(
|
||||
'extractedData.store_name missing; using fallback store name to avoid DB constraint error.',
|
||||
);
|
||||
}
|
||||
|
||||
// 1. Check for duplicate flyer using checksum
|
||||
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, req.log);
|
||||
if (existingFlyer) {
|
||||
logger.warn(`Duplicate flyer upload attempt blocked for checksum: ${checksum}`);
|
||||
await cleanupUploadedFile(req.file);
|
||||
return res.status(409).json({ message: 'This flyer has already been processed.' });
|
||||
}
|
||||
|
||||
// Generate a 64x64 icon from the uploaded flyer image.
|
||||
const iconsDir = path.join(path.dirname(req.file.path), 'icons');
|
||||
const iconFileName = await generateFlyerIcon(req.file.path, iconsDir, req.log);
|
||||
const iconUrl = `/flyer-images/icons/${iconFileName}`;
|
||||
|
||||
// 2. Prepare flyer data for insertion
|
||||
const flyerData = {
|
||||
file_name: originalFileName,
|
||||
image_url: `/flyer-images/${req.file.filename}`, // Store the full URL path
|
||||
icon_url: iconUrl,
|
||||
checksum: checksum,
|
||||
// Use normalized store name (fallback applied above).
|
||||
store_name: storeName,
|
||||
valid_from: extractedData.valid_from ?? null,
|
||||
valid_to: extractedData.valid_to ?? null,
|
||||
store_address: extractedData.store_address ?? null,
|
||||
item_count: 0, // Set default to 0; the trigger will update it.
|
||||
uploaded_by: userProfile?.user.user_id, // Associate with user if logged in
|
||||
};
|
||||
|
||||
// 3. Create flyer and its items in a transaction
|
||||
const { flyer: newFlyer, items: newItems } = await createFlyerAndItems(
|
||||
flyerData,
|
||||
itemsForDb,
|
||||
req.log,
|
||||
);
|
||||
|
||||
logger.info(
|
||||
`Successfully processed and saved new flyer: ${newFlyer.file_name} (ID: ${newFlyer.flyer_id}) with ${newItems.length} items.`,
|
||||
);
|
||||
|
||||
// Log this significant event
|
||||
await db.adminRepo.logActivity(
|
||||
{
|
||||
userId: userProfile?.user.user_id,
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed a new flyer for ${flyerData.store_name}.`,
|
||||
details: { flyerId: newFlyer.flyer_id, storeName: flyerData.store_name },
|
||||
},
|
||||
const newFlyer = await aiService.processLegacyFlyerUpload(
|
||||
req.file,
|
||||
req.body,
|
||||
userProfile,
|
||||
req.log,
|
||||
);
|
||||
|
||||
res.status(201).json({ message: 'Flyer processed and saved successfully.', flyer: newFlyer });
|
||||
} catch (error) {
|
||||
await cleanupUploadedFile(req.file);
|
||||
if (error instanceof DuplicateFlyerError) {
|
||||
logger.warn(`Duplicate flyer upload attempt blocked.`);
|
||||
return res.status(409).json({ message: error.message, flyerId: error.flyerId });
|
||||
}
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
@@ -516,6 +283,8 @@ router.post(
|
||||
res.status(200).json({ is_flyer: true }); // Stubbed response
|
||||
} catch (error) {
|
||||
next(error);
|
||||
} finally {
|
||||
await cleanupUploadedFile(req.file);
|
||||
}
|
||||
},
|
||||
);
|
||||
@@ -533,6 +302,8 @@ router.post(
|
||||
res.status(200).json({ address: 'not identified' }); // Updated stubbed response
|
||||
} catch (error) {
|
||||
next(error);
|
||||
} finally {
|
||||
await cleanupUploadedFile(req.file);
|
||||
}
|
||||
},
|
||||
);
|
||||
@@ -550,6 +321,8 @@ router.post(
|
||||
res.status(200).json({ store_logo_base_64: null }); // Stubbed response
|
||||
} catch (error) {
|
||||
next(error);
|
||||
} finally {
|
||||
await cleanupUploadedFiles(req.files as Express.Multer.File[]);
|
||||
}
|
||||
},
|
||||
);
|
||||
@@ -626,7 +399,7 @@ router.post(
|
||||
try {
|
||||
const { items, store, userLocation } = req.body;
|
||||
logger.debug({ itemCount: items.length, storeName: store.name }, 'Trip planning requested.');
|
||||
const result = await aiService.aiService.planTripWithMaps(items, store, userLocation);
|
||||
const result = await aiService.planTripWithMaps(items, store, userLocation);
|
||||
res.status(200).json(result);
|
||||
} catch (error) {
|
||||
logger.error({ error: errMsg(error) }, 'Error in /api/ai/plan-trip endpoint:');
|
||||
@@ -686,7 +459,7 @@ router.post(
|
||||
'Rescan area requested',
|
||||
);
|
||||
|
||||
const result = await aiService.aiService.extractTextFromImageArea(
|
||||
const result = await aiService.extractTextFromImageArea(
|
||||
path,
|
||||
mimetype,
|
||||
cropArea,
|
||||
@@ -697,8 +470,13 @@ router.post(
|
||||
res.status(200).json(result);
|
||||
} catch (error) {
|
||||
next(error);
|
||||
} finally {
|
||||
await cleanupUploadedFile(req.file);
|
||||
}
|
||||
},
|
||||
);
|
||||
|
||||
/* Catches errors from multer (e.g., file size, file filter) */
|
||||
router.use(handleMulterError);
|
||||
|
||||
export default router;
|
||||
|
||||
@@ -2,13 +2,8 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import { Request, Response, NextFunction } from 'express';
|
||||
import cookieParser from 'cookie-parser';
|
||||
import * as bcrypt from 'bcrypt';
|
||||
import jwt from 'jsonwebtoken';
|
||||
import {
|
||||
createMockUserProfile,
|
||||
createMockUserWithPasswordHash,
|
||||
} from '../tests/utils/mockFactories';
|
||||
import cookieParser from 'cookie-parser'; // This was a duplicate, fixed.
|
||||
import { createMockUserProfile } from '../tests/utils/mockFactories';
|
||||
|
||||
// --- FIX: Hoist passport mocks to be available for vi.mock ---
|
||||
const passportMocks = vi.hoisted(() => {
|
||||
@@ -69,45 +64,20 @@ vi.mock('./passport.routes', () => ({
|
||||
optionalAuth: vi.fn((req: Request, res: Response, next: NextFunction) => next()),
|
||||
}));
|
||||
|
||||
// Mock the DB connection pool to control transactional behavior
|
||||
const { mockPool } = vi.hoisted(() => {
|
||||
const client = {
|
||||
query: vi.fn(),
|
||||
release: vi.fn(),
|
||||
};
|
||||
// Mock the authService, which is now the primary dependency of the routes.
|
||||
const { mockedAuthService } = vi.hoisted(() => {
|
||||
return {
|
||||
mockPool: {
|
||||
connect: vi.fn(() => Promise.resolve(client)),
|
||||
mockedAuthService: {
|
||||
registerAndLoginUser: vi.fn(),
|
||||
handleSuccessfulLogin: vi.fn(),
|
||||
resetPassword: vi.fn(),
|
||||
updatePassword: vi.fn(),
|
||||
refreshAccessToken: vi.fn(),
|
||||
logout: vi.fn(),
|
||||
},
|
||||
mockClient: client,
|
||||
};
|
||||
});
|
||||
// Mock the Service Layer directly.
|
||||
// We use async import inside the factory to properly hoist the UniqueConstraintError class usage.
|
||||
vi.mock('../services/db/index.db', async () => {
|
||||
const { UniqueConstraintError } = await import('../services/db/errors.db');
|
||||
return {
|
||||
userRepo: {
|
||||
findUserByEmail: vi.fn(),
|
||||
createUser: vi.fn(),
|
||||
saveRefreshToken: vi.fn(),
|
||||
createPasswordResetToken: vi.fn(),
|
||||
getValidResetTokens: vi.fn(),
|
||||
updateUserPassword: vi.fn(),
|
||||
deleteResetToken: vi.fn(),
|
||||
findUserByRefreshToken: vi.fn(),
|
||||
deleteRefreshToken: vi.fn(),
|
||||
},
|
||||
adminRepo: {
|
||||
logActivity: vi.fn(),
|
||||
},
|
||||
UniqueConstraintError: UniqueConstraintError,
|
||||
};
|
||||
});
|
||||
|
||||
vi.mock('../services/db/connection.db', () => ({
|
||||
getPool: () => mockPool,
|
||||
}));
|
||||
vi.mock('../services/authService', () => ({ authService: mockedAuthService }));
|
||||
|
||||
// Mock the logger
|
||||
vi.mock('../services/logger.server', async () => ({
|
||||
@@ -120,15 +90,8 @@ vi.mock('../services/emailService.server', () => ({
|
||||
sendPasswordResetEmail: vi.fn(),
|
||||
}));
|
||||
|
||||
// Mock bcrypt
|
||||
vi.mock('bcrypt', async (importOriginal) => {
|
||||
const actual = await importOriginal<typeof bcrypt>();
|
||||
return { ...actual, compare: vi.fn() };
|
||||
});
|
||||
|
||||
// Import the router AFTER mocks are established
|
||||
import authRouter from './auth.routes';
|
||||
import * as db from '../services/db/index.db'; // This was a duplicate, fixed.
|
||||
|
||||
import { UniqueConstraintError } from '../services/db/errors.db'; // Import actual class for instanceof checks
|
||||
|
||||
@@ -176,13 +139,11 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
user: { user_id: 'new-user-id', email: newUserEmail },
|
||||
full_name: 'Test User',
|
||||
});
|
||||
|
||||
// FIX: Mock the method on the imported singleton instance `userRepo` directly,
|
||||
// as this is what the route handler uses. Spying on the prototype does not
|
||||
// affect this already-created instance.
|
||||
vi.mocked(db.userRepo.createUser).mockResolvedValue(mockNewUser);
|
||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
||||
vi.mocked(db.adminRepo.logActivity).mockResolvedValue(undefined);
|
||||
mockedAuthService.registerAndLoginUser.mockResolvedValue({
|
||||
newUserProfile: mockNewUser,
|
||||
accessToken: 'new-access-token',
|
||||
refreshToken: 'new-refresh-token',
|
||||
});
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).post('/api/auth/register').send({
|
||||
@@ -190,22 +151,29 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
password: strongPassword,
|
||||
full_name: 'Test User',
|
||||
});
|
||||
|
||||
// Assert
|
||||
expect(response.status).toBe(201);
|
||||
expect(response.body.message).toBe('User registered successfully!');
|
||||
expect(response.body.userprofile.user.email).toBe(newUserEmail);
|
||||
expect(response.body.token).toBeTypeOf('string'); // This was a duplicate, fixed.
|
||||
expect(db.userRepo.createUser).toHaveBeenCalled();
|
||||
expect(mockedAuthService.registerAndLoginUser).toHaveBeenCalledWith(
|
||||
newUserEmail,
|
||||
strongPassword,
|
||||
'Test User',
|
||||
undefined, // avatar_url
|
||||
mockLogger,
|
||||
);
|
||||
});
|
||||
|
||||
it('should set a refresh token cookie on successful registration', async () => {
|
||||
const mockNewUser = createMockUserProfile({
|
||||
user: { user_id: 'new-user-id', email: 'cookie@test.com' },
|
||||
});
|
||||
vi.mocked(db.userRepo.createUser).mockResolvedValue(mockNewUser);
|
||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
||||
vi.mocked(db.adminRepo.logActivity).mockResolvedValue(undefined);
|
||||
mockedAuthService.registerAndLoginUser.mockResolvedValue({
|
||||
newUserProfile: mockNewUser,
|
||||
accessToken: 'new-access-token',
|
||||
refreshToken: 'new-refresh-token',
|
||||
});
|
||||
|
||||
const response = await supertest(app).post('/api/auth/register').send({
|
||||
email: 'cookie@test.com',
|
||||
@@ -235,15 +203,14 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
expect(errorMessages).toMatch(/Password is too weak/i);
|
||||
});
|
||||
|
||||
it('should reject registration if the email already exists', async () => {
|
||||
it('should reject registration if the auth service throws UniqueConstraintError', async () => {
|
||||
// Create an error object that includes the 'code' property for simulating a PG unique violation.
|
||||
// This is more type-safe than casting to 'any'.
|
||||
const dbError = new UniqueConstraintError(
|
||||
'User with that email already exists.',
|
||||
) as UniqueConstraintError & { code: string };
|
||||
dbError.code = '23505';
|
||||
|
||||
vi.mocked(db.userRepo.createUser).mockRejectedValue(dbError);
|
||||
mockedAuthService.registerAndLoginUser.mockRejectedValue(dbError);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/register')
|
||||
@@ -251,12 +218,11 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
|
||||
expect(response.status).toBe(409); // 409 Conflict
|
||||
expect(response.body.message).toBe('User with that email already exists.');
|
||||
expect(db.userRepo.createUser).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should return 500 if a generic database error occurs during registration', async () => {
|
||||
const dbError = new Error('DB connection lost');
|
||||
vi.mocked(db.userRepo.createUser).mockRejectedValue(dbError);
|
||||
mockedAuthService.registerAndLoginUser.mockRejectedValue(dbError);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/register')
|
||||
@@ -289,7 +255,10 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
it('should successfully log in a user and return a token and cookie', async () => {
|
||||
// Arrange:
|
||||
const loginCredentials = { email: 'test@test.com', password: 'password123' };
|
||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
||||
mockedAuthService.handleSuccessfulLogin.mockResolvedValue({
|
||||
accessToken: 'new-access-token',
|
||||
refreshToken: 'new-refresh-token',
|
||||
});
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
||||
@@ -309,25 +278,6 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
expect(response.headers['set-cookie']).toBeDefined();
|
||||
});
|
||||
|
||||
it('should contain the correct payload in the JWT token', async () => {
|
||||
// Arrange
|
||||
const loginCredentials = { email: 'payload.test@test.com', password: 'password123' };
|
||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
||||
|
||||
// Assert
|
||||
expect(response.status).toBe(200);
|
||||
const token = response.body.token;
|
||||
expect(token).toBeTypeOf('string');
|
||||
|
||||
const decodedPayload = jwt.decode(token) as { user_id: string; email: string; role: string };
|
||||
expect(decodedPayload.user_id).toBe('user-123');
|
||||
expect(decodedPayload.email).toBe(loginCredentials.email);
|
||||
expect(decodedPayload.role).toBe('user'); // Default role from mock factory
|
||||
});
|
||||
|
||||
it('should reject login for incorrect credentials', async () => {
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/login')
|
||||
@@ -359,7 +309,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
it('should return 500 if saving the refresh token fails', async () => {
|
||||
// Arrange:
|
||||
const loginCredentials = { email: 'test@test.com', password: 'password123' };
|
||||
vi.mocked(db.userRepo.saveRefreshToken).mockRejectedValue(new Error('DB write failed'));
|
||||
mockedAuthService.handleSuccessfulLogin.mockRejectedValue(new Error('DB write failed'));
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
||||
@@ -401,7 +351,10 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
password: 'password123',
|
||||
rememberMe: true,
|
||||
};
|
||||
vi.mocked(db.userRepo.saveRefreshToken).mockResolvedValue(undefined);
|
||||
mockedAuthService.handleSuccessfulLogin.mockResolvedValue({
|
||||
accessToken: 'remember-access-token',
|
||||
refreshToken: 'remember-refresh-token',
|
||||
});
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).post('/api/auth/login').send(loginCredentials);
|
||||
@@ -416,10 +369,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
describe('POST /forgot-password', () => {
|
||||
it('should send a reset link if the user exists', async () => {
|
||||
// Arrange
|
||||
vi.mocked(db.userRepo.findUserByEmail).mockResolvedValue(
|
||||
createMockUserWithPasswordHash({ user_id: 'user-123', email: 'test@test.com' }),
|
||||
);
|
||||
vi.mocked(db.userRepo.createPasswordResetToken).mockResolvedValue(undefined);
|
||||
mockedAuthService.resetPassword.mockResolvedValue('mock-reset-token');
|
||||
|
||||
// Act
|
||||
const response = await supertest(app)
|
||||
@@ -433,7 +383,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
});
|
||||
|
||||
it('should return a generic success message even if the user does not exist', async () => {
|
||||
vi.mocked(db.userRepo.findUserByEmail).mockResolvedValue(undefined);
|
||||
mockedAuthService.resetPassword.mockResolvedValue(undefined);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/forgot-password')
|
||||
@@ -444,7 +394,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
});
|
||||
|
||||
it('should return 500 if the database call fails', async () => {
|
||||
vi.mocked(db.userRepo.findUserByEmail).mockRejectedValue(new Error('DB connection failed'));
|
||||
mockedAuthService.resetPassword.mockRejectedValue(new Error('DB connection failed'));
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/forgot-password')
|
||||
.send({ email: 'any@test.com' });
|
||||
@@ -452,25 +402,6 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
expect(response.status).toBe(500);
|
||||
});
|
||||
|
||||
it('should still return 200 OK if the email service fails', async () => {
|
||||
// Arrange
|
||||
vi.mocked(db.userRepo.findUserByEmail).mockResolvedValue(
|
||||
createMockUserWithPasswordHash({ user_id: 'user-123', email: 'test@test.com' }),
|
||||
);
|
||||
vi.mocked(db.userRepo.createPasswordResetToken).mockResolvedValue(undefined);
|
||||
// Mock the email service to fail
|
||||
const { sendPasswordResetEmail } = await import('../services/emailService.server');
|
||||
vi.mocked(sendPasswordResetEmail).mockRejectedValue(new Error('SMTP server down'));
|
||||
|
||||
// Act
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/forgot-password')
|
||||
.send({ email: 'test@test.com' });
|
||||
|
||||
// Assert: The route should not fail even if the email does.
|
||||
expect(response.status).toBe(200);
|
||||
});
|
||||
|
||||
it('should return 400 for an invalid email format', async () => {
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/forgot-password')
|
||||
@@ -483,16 +414,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
|
||||
describe('POST /reset-password', () => {
|
||||
it('should reset the password with a valid token and strong password', async () => {
|
||||
const tokenRecord = {
|
||||
user_id: 'user-123',
|
||||
token_hash: 'hashed-token',
|
||||
expires_at: new Date(Date.now() + 3600000),
|
||||
};
|
||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([tokenRecord]); // This was a duplicate, fixed.
|
||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never); // Token matches
|
||||
vi.mocked(db.userRepo.updateUserPassword).mockResolvedValue(undefined);
|
||||
vi.mocked(db.userRepo.deleteResetToken).mockResolvedValue(undefined);
|
||||
vi.mocked(db.adminRepo.logActivity).mockResolvedValue(undefined);
|
||||
mockedAuthService.updatePassword.mockResolvedValue(true);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/reset-password')
|
||||
@@ -503,7 +425,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
});
|
||||
|
||||
it('should reject with an invalid or expired token', async () => {
|
||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([]); // No valid tokens found
|
||||
mockedAuthService.updatePassword.mockResolvedValue(null);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/reset-password')
|
||||
@@ -513,31 +435,8 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
expect(response.body.message).toBe('Invalid or expired password reset token.');
|
||||
});
|
||||
|
||||
it('should reject if token does not match any valid tokens in DB', async () => {
|
||||
const tokenRecord = {
|
||||
user_id: 'user-123',
|
||||
token_hash: 'hashed-token',
|
||||
expires_at: new Date(Date.now() + 3600000),
|
||||
};
|
||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([tokenRecord]);
|
||||
vi.mocked(bcrypt.compare).mockResolvedValue(false as never); // Token does not match
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/reset-password')
|
||||
.send({ token: 'wrong-token', newPassword: 'a-Very-Strong-Password-123!' });
|
||||
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.message).toBe('Invalid or expired password reset token.');
|
||||
});
|
||||
|
||||
it('should return 400 for a weak new password', async () => {
|
||||
const tokenRecord = {
|
||||
user_id: 'user-123',
|
||||
token_hash: 'hashed-token',
|
||||
expires_at: new Date(Date.now() + 3600000),
|
||||
};
|
||||
vi.mocked(db.userRepo.getValidResetTokens).mockResolvedValue([tokenRecord]);
|
||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never);
|
||||
// No need to mock the service here as validation runs first
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/reset-password')
|
||||
@@ -557,11 +456,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
|
||||
describe('POST /refresh-token', () => {
|
||||
it('should issue a new access token with a valid refresh token cookie', async () => {
|
||||
const mockUser = createMockUserWithPasswordHash({
|
||||
user_id: 'user-123',
|
||||
email: 'test@test.com',
|
||||
});
|
||||
vi.mocked(db.userRepo.findUserByRefreshToken).mockResolvedValue(mockUser);
|
||||
mockedAuthService.refreshAccessToken.mockResolvedValue({ accessToken: 'new-access-token' });
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/refresh-token')
|
||||
@@ -578,8 +473,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
});
|
||||
|
||||
it('should return 403 if refresh token is invalid', async () => {
|
||||
// Mock finding no user for this token, which should trigger the 403 logic
|
||||
vi.mocked(db.userRepo.findUserByRefreshToken).mockResolvedValue(undefined as any);
|
||||
mockedAuthService.refreshAccessToken.mockResolvedValue(null);
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/auth/refresh-token')
|
||||
@@ -590,7 +484,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
|
||||
it('should return 500 if the database call fails', async () => {
|
||||
// Arrange
|
||||
vi.mocked(db.userRepo.findUserByRefreshToken).mockRejectedValue(new Error('DB Error'));
|
||||
mockedAuthService.refreshAccessToken.mockRejectedValue(new Error('DB Error'));
|
||||
|
||||
// Act
|
||||
const response = await supertest(app)
|
||||
@@ -604,7 +498,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
describe('POST /logout', () => {
|
||||
it('should clear the refresh token cookie and return a success message', async () => {
|
||||
// Arrange
|
||||
vi.mocked(db.userRepo.deleteRefreshToken).mockResolvedValue(undefined);
|
||||
mockedAuthService.logout.mockResolvedValue(undefined);
|
||||
|
||||
// Act
|
||||
const response = await supertest(app)
|
||||
@@ -627,7 +521,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
it('should still return 200 OK even if deleting the refresh token from DB fails', async () => {
|
||||
// Arrange
|
||||
const dbError = new Error('DB connection lost');
|
||||
vi.mocked(db.userRepo.deleteRefreshToken).mockRejectedValue(dbError);
|
||||
mockedAuthService.logout.mockRejectedValue(dbError);
|
||||
const { logger } = await import('../services/logger.server');
|
||||
|
||||
// Act
|
||||
@@ -639,7 +533,7 @@ describe('Auth Routes (/api/auth)', () => {
|
||||
expect(response.status).toBe(200);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ error: dbError }),
|
||||
'Failed to delete refresh token from DB during logout.',
|
||||
'Logout token invalidation failed in background.',
|
||||
);
|
||||
});
|
||||
|
||||
|
||||
@@ -1,26 +1,18 @@
|
||||
// src/routes/auth.routes.ts
|
||||
import { Router, Request, Response, NextFunction } from 'express';
|
||||
import * as bcrypt from 'bcrypt';
|
||||
import { z } from 'zod';
|
||||
import jwt from 'jsonwebtoken';
|
||||
import crypto from 'crypto';
|
||||
import rateLimit from 'express-rate-limit';
|
||||
|
||||
import passport from './passport.routes';
|
||||
import { userRepo, adminRepo } from '../services/db/index.db';
|
||||
import { UniqueConstraintError } from '../services/db/errors.db';
|
||||
import { getPool } from '../services/db/connection.db';
|
||||
import { UniqueConstraintError } from '../services/db/errors.db'; // Import actual class for instanceof checks
|
||||
import { logger } from '../services/logger.server';
|
||||
import { sendPasswordResetEmail } from '../services/emailService.server';
|
||||
import { validateRequest } from '../middleware/validation.middleware';
|
||||
import type { UserProfile } from '../types';
|
||||
import { validatePasswordStrength } from '../utils/authUtils';
|
||||
import { requiredString } from '../utils/zodUtils';
|
||||
|
||||
import { authService } from '../services/authService';
|
||||
const router = Router();
|
||||
|
||||
const JWT_SECRET = process.env.JWT_SECRET!;
|
||||
|
||||
// Conditionally disable rate limiting for the test environment
|
||||
const isTestEnv = process.env.NODE_ENV === 'test';
|
||||
|
||||
@@ -31,7 +23,9 @@ const forgotPasswordLimiter = rateLimit({
|
||||
message: 'Too many password reset requests from this IP, please try again after 15 minutes.',
|
||||
standardHeaders: true,
|
||||
legacyHeaders: false,
|
||||
skip: () => isTestEnv, // Skip this middleware if in test environment
|
||||
// Do not skip in test environment so we can write integration tests for it.
|
||||
// The limiter uses an in-memory store by default, so counts are reset when the test server restarts.
|
||||
// skip: () => isTestEnv,
|
||||
});
|
||||
|
||||
const resetPasswordLimiter = rateLimit({
|
||||
@@ -45,21 +39,27 @@ const resetPasswordLimiter = rateLimit({
|
||||
|
||||
const registerSchema = z.object({
|
||||
body: z.object({
|
||||
email: z.string().email('A valid email is required.'),
|
||||
// Sanitize email by trimming and converting to lowercase.
|
||||
email: z.string().trim().toLowerCase().email('A valid email is required.'),
|
||||
password: z
|
||||
.string()
|
||||
.trim() // Prevent leading/trailing whitespace in passwords.
|
||||
.min(8, 'Password must be at least 8 characters long.')
|
||||
.superRefine((password, ctx) => {
|
||||
const strength = validatePasswordStrength(password);
|
||||
if (!strength.isValid) ctx.addIssue({ code: 'custom', message: strength.feedback });
|
||||
}),
|
||||
full_name: z.string().optional(),
|
||||
avatar_url: z.string().url().optional(),
|
||||
// Sanitize optional string inputs.
|
||||
full_name: z.string().trim().optional(),
|
||||
avatar_url: z.string().trim().url().optional(),
|
||||
}),
|
||||
});
|
||||
|
||||
const forgotPasswordSchema = z.object({
|
||||
body: z.object({ email: z.string().email('A valid email is required.') }),
|
||||
body: z.object({
|
||||
// Sanitize email by trimming and converting to lowercase.
|
||||
email: z.string().trim().toLowerCase().email('A valid email is required.'),
|
||||
}),
|
||||
});
|
||||
|
||||
const resetPasswordSchema = z.object({
|
||||
@@ -67,6 +67,7 @@ const resetPasswordSchema = z.object({
|
||||
token: requiredString('Token is required.'),
|
||||
newPassword: z
|
||||
.string()
|
||||
.trim() // Prevent leading/trailing whitespace in passwords.
|
||||
.min(8, 'Password must be at least 8 characters long.')
|
||||
.superRefine((password, ctx) => {
|
||||
const strength = validatePasswordStrength(password);
|
||||
@@ -88,39 +89,14 @@ router.post(
|
||||
} = req as unknown as RegisterRequest;
|
||||
|
||||
try {
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(password, saltRounds);
|
||||
logger.info(`Hashing password for new user: ${email}`);
|
||||
|
||||
// The createUser method in UserRepository now handles its own transaction.
|
||||
const newUser = await userRepo.createUser(
|
||||
const { newUserProfile, accessToken, refreshToken } = await authService.registerAndLoginUser(
|
||||
email,
|
||||
hashedPassword,
|
||||
{ full_name, avatar_url },
|
||||
password,
|
||||
full_name,
|
||||
avatar_url,
|
||||
req.log,
|
||||
);
|
||||
|
||||
const userEmail = newUser.user.email;
|
||||
const userId = newUser.user.user_id;
|
||||
logger.info(`Successfully created new user in DB: ${userEmail} (ID: ${userId})`);
|
||||
|
||||
// Use the new standardized logging function
|
||||
await adminRepo.logActivity(
|
||||
{
|
||||
userId: newUser.user.user_id,
|
||||
action: 'user_registered',
|
||||
displayText: `${userEmail} has registered.`,
|
||||
icon: 'user-plus',
|
||||
},
|
||||
req.log,
|
||||
);
|
||||
|
||||
const payload = { user_id: newUser.user.user_id, email: userEmail };
|
||||
const token = jwt.sign(payload, JWT_SECRET, { expiresIn: '1h' });
|
||||
|
||||
const refreshToken = crypto.randomBytes(64).toString('hex');
|
||||
await userRepo.saveRefreshToken(newUser.user.user_id, refreshToken, req.log);
|
||||
|
||||
res.cookie('refreshToken', refreshToken, {
|
||||
httpOnly: true,
|
||||
secure: process.env.NODE_ENV === 'production',
|
||||
@@ -128,7 +104,7 @@ router.post(
|
||||
});
|
||||
return res
|
||||
.status(201)
|
||||
.json({ message: 'User registered successfully!', userprofile: newUser, token });
|
||||
.json({ message: 'User registered successfully!', userprofile: newUserProfile, token: accessToken });
|
||||
} catch (error: unknown) {
|
||||
if (error instanceof UniqueConstraintError) {
|
||||
// If the email is a duplicate, return a 409 Conflict status.
|
||||
@@ -154,17 +130,6 @@ router.post('/login', (req: Request, res: Response, next: NextFunction) => {
|
||||
if (user) req.log.debug({ user }, '[API /login] Passport user object:'); // Log the user object passport returns
|
||||
if (user) req.log.info({ user }, '[API /login] Passport reported USER FOUND.');
|
||||
|
||||
try {
|
||||
const allUsersInDb = await getPool().query(
|
||||
'SELECT u.user_id, u.email, p.role FROM public.users u JOIN public.profiles p ON u.user_id = p.user_id',
|
||||
);
|
||||
req.log.debug('[API /login] Current users in DB from SERVER perspective:');
|
||||
console.table(allUsersInDb.rows);
|
||||
} catch (dbError) {
|
||||
req.log.error({ dbError }, '[API /login] Could not query users table for debugging.');
|
||||
}
|
||||
// --- END DEBUG LOGGING ---
|
||||
const { rememberMe } = req.body;
|
||||
if (err) {
|
||||
req.log.error(
|
||||
{ error: err },
|
||||
@@ -176,33 +141,24 @@ router.post('/login', (req: Request, res: Response, next: NextFunction) => {
|
||||
return res.status(401).json({ message: info.message || 'Login failed' });
|
||||
}
|
||||
|
||||
const userProfile = user as UserProfile;
|
||||
const payload = {
|
||||
user_id: userProfile.user.user_id,
|
||||
email: userProfile.user.email,
|
||||
role: userProfile.role,
|
||||
};
|
||||
const accessToken = jwt.sign(payload, JWT_SECRET, { expiresIn: '15m' });
|
||||
|
||||
try {
|
||||
const refreshToken = crypto.randomBytes(64).toString('hex');
|
||||
await userRepo.saveRefreshToken(userProfile.user.user_id, refreshToken, req.log);
|
||||
const { rememberMe } = req.body;
|
||||
const userProfile = user as UserProfile;
|
||||
const { accessToken, refreshToken } = await authService.handleSuccessfulLogin(userProfile, req.log);
|
||||
req.log.info(`JWT and refresh token issued for user: ${userProfile.user.email}`);
|
||||
|
||||
const cookieOptions = {
|
||||
httpOnly: true,
|
||||
secure: process.env.NODE_ENV === 'production',
|
||||
maxAge: rememberMe ? 30 * 24 * 60 * 60 * 1000 : undefined,
|
||||
maxAge: rememberMe ? 30 * 24 * 60 * 60 * 1000 : undefined, // 30 days
|
||||
};
|
||||
|
||||
res.cookie('refreshToken', refreshToken, cookieOptions);
|
||||
// Return the full user profile object on login to avoid a second fetch on the client.
|
||||
return res.json({ userprofile: userProfile, token: accessToken });
|
||||
} catch (tokenErr) {
|
||||
req.log.error(
|
||||
{ error: tokenErr },
|
||||
`Failed to save refresh token during login for user: ${userProfile.user.email}`,
|
||||
);
|
||||
const email = (user as UserProfile)?.user?.email || req.body.email;
|
||||
req.log.error({ error: tokenErr }, `Failed to process login for user: ${email}`);
|
||||
return next(tokenErr);
|
||||
}
|
||||
},
|
||||
@@ -221,38 +177,14 @@ router.post(
|
||||
} = req as unknown as ForgotPasswordRequest;
|
||||
|
||||
try {
|
||||
req.log.debug(`[API /forgot-password] Received request for email: ${email}`);
|
||||
const user = await userRepo.findUserByEmail(email, req.log);
|
||||
let token: string | undefined;
|
||||
req.log.debug(
|
||||
{ user: user ? { user_id: user.user_id, email: user.email } : 'NOT FOUND' },
|
||||
`[API /forgot-password] Database search result for ${email}:`,
|
||||
);
|
||||
|
||||
if (user) {
|
||||
token = crypto.randomBytes(32).toString('hex');
|
||||
const saltRounds = 10;
|
||||
const tokenHash = await bcrypt.hash(token, saltRounds);
|
||||
const expiresAt = new Date(Date.now() + 3600000); // 1 hour
|
||||
|
||||
await userRepo.createPasswordResetToken(user.user_id, tokenHash, expiresAt, req.log);
|
||||
|
||||
const resetLink = `${process.env.FRONTEND_URL}/reset-password/${token}`;
|
||||
|
||||
try {
|
||||
await sendPasswordResetEmail(email, resetLink, req.log);
|
||||
} catch (emailError) {
|
||||
req.log.error({ emailError }, `Email send failure during password reset for user`);
|
||||
}
|
||||
} else {
|
||||
req.log.warn(`Password reset requested for non-existent email: ${email}`);
|
||||
}
|
||||
// The service handles finding the user, creating the token, and sending the email.
|
||||
const token = await authService.resetPassword(email, req.log);
|
||||
|
||||
// For testability, return the token in the response only in the test environment.
|
||||
const responsePayload: { message: string; token?: string } = {
|
||||
message: 'If an account with that email exists, a password reset link has been sent.',
|
||||
};
|
||||
if (process.env.NODE_ENV === 'test' && user) responsePayload.token = token;
|
||||
if (process.env.NODE_ENV === 'test' && token) responsePayload.token = token;
|
||||
res.status(200).json(responsePayload);
|
||||
} catch (error) {
|
||||
req.log.error({ error }, `An error occurred during /forgot-password for email: ${email}`);
|
||||
@@ -273,38 +205,12 @@ router.post(
|
||||
} = req as unknown as ResetPasswordRequest;
|
||||
|
||||
try {
|
||||
const validTokens = await userRepo.getValidResetTokens(req.log);
|
||||
let tokenRecord;
|
||||
for (const record of validTokens) {
|
||||
const isMatch = await bcrypt.compare(token, record.token_hash);
|
||||
if (isMatch) {
|
||||
tokenRecord = record;
|
||||
break;
|
||||
}
|
||||
}
|
||||
const resetSuccessful = await authService.updatePassword(token, newPassword, req.log);
|
||||
|
||||
if (!tokenRecord) {
|
||||
if (!resetSuccessful) {
|
||||
return res.status(400).json({ message: 'Invalid or expired password reset token.' });
|
||||
}
|
||||
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
||||
|
||||
await userRepo.updateUserPassword(tokenRecord.user_id, hashedPassword, req.log);
|
||||
await userRepo.deleteResetToken(tokenRecord.token_hash, req.log);
|
||||
|
||||
// Log this security event after a successful password reset.
|
||||
await adminRepo.logActivity(
|
||||
{
|
||||
userId: tokenRecord.user_id,
|
||||
action: 'password_reset',
|
||||
displayText: `User ID ${tokenRecord.user_id} has reset their password.`,
|
||||
icon: 'key',
|
||||
details: { source_ip: req.ip ?? null },
|
||||
},
|
||||
req.log,
|
||||
);
|
||||
|
||||
res.status(200).json({ message: 'Password has been reset successfully.' });
|
||||
} catch (error) {
|
||||
req.log.error({ error }, `An error occurred during password reset.`);
|
||||
@@ -321,15 +227,11 @@ router.post('/refresh-token', async (req: Request, res: Response, next: NextFunc
|
||||
}
|
||||
|
||||
try {
|
||||
const user = await userRepo.findUserByRefreshToken(refreshToken, req.log);
|
||||
if (!user) {
|
||||
const result = await authService.refreshAccessToken(refreshToken, req.log);
|
||||
if (!result) {
|
||||
return res.status(403).json({ message: 'Invalid or expired refresh token.' });
|
||||
}
|
||||
|
||||
const payload = { user_id: user.user_id, email: user.email };
|
||||
const newAccessToken = jwt.sign(payload, JWT_SECRET, { expiresIn: '15m' });
|
||||
|
||||
res.json({ token: newAccessToken });
|
||||
res.json({ token: result.accessToken });
|
||||
} catch (error) {
|
||||
req.log.error({ error }, 'An error occurred during /refresh-token.');
|
||||
next(error);
|
||||
@@ -346,8 +248,8 @@ router.post('/logout', async (req: Request, res: Response) => {
|
||||
if (refreshToken) {
|
||||
// Invalidate the token in the database so it cannot be used again.
|
||||
// We don't need to wait for this to finish to respond to the user.
|
||||
userRepo.deleteRefreshToken(refreshToken, req.log).catch((err: Error) => {
|
||||
req.log.error({ error: err }, 'Failed to delete refresh token from DB during logout.');
|
||||
authService.logout(refreshToken, req.log).catch((err: Error) => {
|
||||
req.log.error({ error: err }, 'Logout token invalidation failed in background.');
|
||||
});
|
||||
}
|
||||
// Instruct the browser to clear the cookie by setting its expiration to the past.
|
||||
|
||||
@@ -1,11 +1,10 @@
|
||||
// src/routes/gamification.routes.ts
|
||||
import express, { NextFunction } from 'express';
|
||||
import { z } from 'zod';
|
||||
import passport, { isAdmin } from './passport.routes';
|
||||
import { gamificationRepo } from '../services/db/index.db';
|
||||
import passport, { isAdmin } from './passport.routes'; // Correctly imported
|
||||
import { gamificationService } from '../services/gamificationService';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { UserProfile } from '../types';
|
||||
import { ForeignKeyConstraintError } from '../services/db/errors.db';
|
||||
import { validateRequest } from '../middleware/validation.middleware';
|
||||
import { requiredString, optionalNumeric } from '../utils/zodUtils';
|
||||
|
||||
@@ -14,10 +13,12 @@ const adminGamificationRouter = express.Router(); // Create a new router for adm
|
||||
|
||||
// --- Zod Schemas for Gamification Routes (as per ADR-003) ---
|
||||
|
||||
const leaderboardQuerySchema = z.object({
|
||||
limit: optionalNumeric({ default: 10, integer: true, positive: true, max: 50 }),
|
||||
});
|
||||
|
||||
const leaderboardSchema = z.object({
|
||||
query: z.object({
|
||||
limit: optionalNumeric({ default: 10, integer: true, positive: true, max: 50 }),
|
||||
}),
|
||||
query: leaderboardQuerySchema,
|
||||
});
|
||||
|
||||
const awardAchievementSchema = z.object({
|
||||
@@ -35,7 +36,7 @@ const awardAchievementSchema = z.object({
|
||||
*/
|
||||
router.get('/', async (req, res, next: NextFunction) => {
|
||||
try {
|
||||
const achievements = await gamificationRepo.getAllAchievements(req.log);
|
||||
const achievements = await gamificationService.getAllAchievements(req.log);
|
||||
res.json(achievements);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching all achievements in /api/achievements:');
|
||||
@@ -51,14 +52,11 @@ router.get(
|
||||
'/leaderboard',
|
||||
validateRequest(leaderboardSchema),
|
||||
async (req, res, next: NextFunction): Promise<void> => {
|
||||
// Apply ADR-003 pattern for type safety.
|
||||
// Explicitly coerce query params to ensure numbers are passed to the repo,
|
||||
// as validateRequest might not replace req.query in all test environments.
|
||||
const query = req.query as unknown as { limit?: string };
|
||||
const limit = query.limit ? Number(query.limit) : 10;
|
||||
|
||||
try {
|
||||
const leaderboard = await gamificationRepo.getLeaderboard(limit, req.log);
|
||||
// The `validateRequest` middleware ensures `req.query` is valid.
|
||||
// We parse it here to apply Zod's coercions (string to number) and defaults.
|
||||
const { limit } = leaderboardQuerySchema.parse(req.query);
|
||||
const leaderboard = await gamificationService.getLeaderboard(limit!, req.log);
|
||||
res.json(leaderboard);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching leaderboard:');
|
||||
@@ -79,7 +77,7 @@ router.get(
|
||||
async (req, res, next: NextFunction): Promise<void> => {
|
||||
const userProfile = req.user as UserProfile;
|
||||
try {
|
||||
const userAchievements = await gamificationRepo.getUserAchievements(
|
||||
const userAchievements = await gamificationService.getUserAchievements(
|
||||
userProfile.user.user_id,
|
||||
req.log,
|
||||
);
|
||||
@@ -111,21 +109,13 @@ adminGamificationRouter.post(
|
||||
type AwardAchievementRequest = z.infer<typeof awardAchievementSchema>;
|
||||
const { body } = req as unknown as AwardAchievementRequest;
|
||||
try {
|
||||
await gamificationRepo.awardAchievement(body.userId, body.achievementName, req.log);
|
||||
await gamificationService.awardAchievement(body.userId, body.achievementName, req.log);
|
||||
res
|
||||
.status(200)
|
||||
.json({
|
||||
message: `Successfully awarded '${body.achievementName}' to user ${body.userId}.`,
|
||||
});
|
||||
} catch (error) {
|
||||
if (error instanceof ForeignKeyConstraintError) {
|
||||
res.status(400).json({ message: error.message });
|
||||
return;
|
||||
}
|
||||
logger.error(
|
||||
{ error, userId: body.userId, achievementName: body.achievementName },
|
||||
'Error awarding achievement via admin endpoint:',
|
||||
);
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
|
||||
@@ -164,11 +164,12 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect(response.body.message).toBe('DB connection failed'); // This is the message from the original error
|
||||
expect(response.body.stack).toBeDefined();
|
||||
expect(response.body.errorId).toEqual(expect.any(String));
|
||||
console.log('[DEBUG] health.routes.test.ts: Verifying logger.error for DB schema check failure');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
|
||||
@@ -186,7 +187,7 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect.objectContaining({
|
||||
err: expect.objectContaining({ message: 'DB connection failed' }),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
});
|
||||
@@ -220,7 +221,7 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
|
||||
@@ -239,7 +240,7 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
});
|
||||
@@ -300,7 +301,7 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
|
||||
@@ -321,7 +322,7 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect.objectContaining({
|
||||
err: expect.objectContaining({ message: 'Pool is not initialized' }),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
|
||||
@@ -336,11 +337,12 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect(response.body.message).toBe('Connection timed out');
|
||||
expect(response.body.stack).toBeDefined();
|
||||
expect(response.body.errorId).toEqual(expect.any(String));
|
||||
console.log('[DEBUG] health.routes.test.ts: Checking if logger.error was called with the correct pattern');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
|
||||
@@ -357,7 +359,7 @@ describe('Health Routes (/api/health)', () => {
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
}),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: \w+\)/),
|
||||
expect.stringMatching(/Unhandled API Error \(ID: [\w-]+\)/),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -39,7 +39,12 @@ router.get('/db-schema', validateRequest(emptySchema), async (req, res, next: Ne
|
||||
}
|
||||
return res.status(200).json({ success: true, message: 'All required database tables exist.' });
|
||||
} catch (error: unknown) {
|
||||
next(error);
|
||||
if (error instanceof Error) {
|
||||
return next(error);
|
||||
}
|
||||
const message =
|
||||
(error as any)?.message || 'An unknown error occurred during DB schema check.';
|
||||
return next(new Error(message));
|
||||
}
|
||||
});
|
||||
|
||||
@@ -48,7 +53,7 @@ router.get('/db-schema', validateRequest(emptySchema), async (req, res, next: Ne
|
||||
* This is important for features like file uploads.
|
||||
*/
|
||||
router.get('/storage', validateRequest(emptySchema), async (req, res, next: NextFunction) => {
|
||||
const storagePath = process.env.STORAGE_PATH || '/var/www/flyer-crawler.projectium.com/assets';
|
||||
const storagePath = process.env.STORAGE_PATH || '/var/www/flyer-crawler.projectium.com/flyer-images';
|
||||
try {
|
||||
await fs.access(storagePath, fs.constants.W_OK); // Use fs.promises
|
||||
return res
|
||||
@@ -88,7 +93,12 @@ router.get(
|
||||
.json({ success: false, message: `Pool may be under stress. ${message}` });
|
||||
}
|
||||
} catch (error: unknown) {
|
||||
next(error);
|
||||
if (error instanceof Error) {
|
||||
return next(error);
|
||||
}
|
||||
const message =
|
||||
(error as any)?.message || 'An unknown error occurred during DB pool check.';
|
||||
return next(new Error(message));
|
||||
}
|
||||
},
|
||||
);
|
||||
@@ -121,7 +131,12 @@ router.get(
|
||||
}
|
||||
throw new Error(`Unexpected Redis ping response: ${reply}`); // This will be caught below
|
||||
} catch (error: unknown) {
|
||||
next(error);
|
||||
if (error instanceof Error) {
|
||||
return next(error);
|
||||
}
|
||||
const message =
|
||||
(error as any)?.message || 'An unknown error occurred during Redis health check.';
|
||||
return next(new Error(message));
|
||||
}
|
||||
},
|
||||
);
|
||||
|
||||
@@ -260,6 +260,13 @@ const jwtOptions = {
|
||||
secretOrKey: JWT_SECRET,
|
||||
};
|
||||
|
||||
// --- DEBUG LOGGING FOR JWT SECRET ---
|
||||
if (!JWT_SECRET) {
|
||||
logger.fatal('[Passport] CRITICAL: JWT_SECRET is missing or empty in environment variables! JwtStrategy will fail.');
|
||||
} else {
|
||||
logger.info(`[Passport] JWT_SECRET loaded successfully (length: ${JWT_SECRET.length}).`);
|
||||
}
|
||||
|
||||
passport.use(
|
||||
new JwtStrategy(jwtOptions, async (jwt_payload, done) => {
|
||||
logger.debug(
|
||||
|
||||
@@ -28,10 +28,9 @@ router.get(
|
||||
validateRequest(mostFrequentSalesSchema),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
try {
|
||||
// Parse req.query to ensure coercion (string -> number) and defaults are applied.
|
||||
// Even though validateRequest checks validity, it may not mutate req.query with the parsed result.
|
||||
// The `validateRequest` middleware ensures `req.query` is valid.
|
||||
// We parse it here to apply Zod's coercions (string to number) and defaults.
|
||||
const { days, limit } = statsQuerySchema.parse(req.query);
|
||||
|
||||
const items = await db.adminRepo.getMostFrequentSaleItems(days!, limit!, req.log);
|
||||
res.json(items);
|
||||
} catch (error) {
|
||||
|
||||
@@ -1,26 +1,15 @@
|
||||
// src/routes/system.routes.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import systemRouter from './system.routes'; // This was a duplicate, fixed.
|
||||
import { exec, type ExecException, type ExecOptions } from 'child_process';
|
||||
import { geocodingService } from '../services/geocodingService.server';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
|
||||
// FIX: Use the simple factory pattern for child_process to avoid default export issues
|
||||
vi.mock('child_process', () => {
|
||||
const mockExec = vi.fn((command, callback) => {
|
||||
if (typeof callback === 'function') {
|
||||
callback(null, 'PM2 OK', '');
|
||||
}
|
||||
return { unref: () => {} };
|
||||
});
|
||||
|
||||
return {
|
||||
default: { exec: mockExec },
|
||||
exec: mockExec,
|
||||
};
|
||||
});
|
||||
|
||||
// 1. Mock the Service Layer
|
||||
// This decouples the route test from the service's implementation details.
|
||||
vi.mock('../services/systemService', () => ({
|
||||
systemService: {
|
||||
getPm2Status: vi.fn(),
|
||||
},
|
||||
}));
|
||||
// 2. Mock Geocoding
|
||||
vi.mock('../services/geocodingService.server', () => ({
|
||||
geocodingService: {
|
||||
@@ -39,44 +28,25 @@ vi.mock('../services/logger.server', () => ({
|
||||
},
|
||||
}));
|
||||
|
||||
// Import the router AFTER all mocks are defined to ensure systemService picks up the mocked util.promisify
|
||||
import { systemService } from '../services/systemService';
|
||||
import systemRouter from './system.routes';
|
||||
import { geocodingService } from '../services/geocodingService.server';
|
||||
|
||||
describe('System Routes (/api/system)', () => {
|
||||
const app = createTestApp({ router: systemRouter, basePath: '/api/system' });
|
||||
|
||||
beforeEach(() => {
|
||||
// We cast here to get type-safe access to mock functions like .mockImplementation
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
describe('GET /pm2-status', () => {
|
||||
it('should return success: true when pm2 process is online', async () => {
|
||||
// Arrange: Simulate a successful `pm2 describe` output for an online process.
|
||||
const pm2OnlineOutput = `
|
||||
┌─ PM2 info ────────────────┐
|
||||
│ status │ online │
|
||||
└───────────┴───────────┘
|
||||
`;
|
||||
|
||||
type ExecCallback = (error: ExecException | null, stdout: string, stderr: string) => void;
|
||||
|
||||
// A robust mock for `exec` that handles its multiple overloads.
|
||||
// This avoids the complex and error-prone `...args` signature.
|
||||
vi.mocked(exec).mockImplementation(
|
||||
(
|
||||
command: string,
|
||||
options?: ExecOptions | ExecCallback | null,
|
||||
callback?: ExecCallback | null,
|
||||
) => {
|
||||
// The actual callback can be the second or third argument.
|
||||
const actualCallback = (
|
||||
typeof options === 'function' ? options : callback
|
||||
) as ExecCallback;
|
||||
if (actualCallback) {
|
||||
actualCallback(null, pm2OnlineOutput, '');
|
||||
}
|
||||
// Return a minimal object that satisfies the ChildProcess type for .unref()
|
||||
return { unref: () => {} } as ReturnType<typeof exec>;
|
||||
},
|
||||
);
|
||||
vi.mocked(systemService.getPm2Status).mockResolvedValue({
|
||||
success: true,
|
||||
message: 'Application is online and running under PM2.',
|
||||
});
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).get('/api/system/pm2-status');
|
||||
@@ -90,28 +60,10 @@ describe('System Routes (/api/system)', () => {
|
||||
});
|
||||
|
||||
it('should return success: false when pm2 process is stopped or errored', async () => {
|
||||
const pm2StoppedOutput = `│ status │ stopped │`;
|
||||
|
||||
vi.mocked(exec).mockImplementation(
|
||||
(
|
||||
command: string,
|
||||
options?:
|
||||
| ExecOptions
|
||||
| ((error: ExecException | null, stdout: string, stderr: string) => void)
|
||||
| null,
|
||||
callback?: ((error: ExecException | null, stdout: string, stderr: string) => void) | null,
|
||||
) => {
|
||||
const actualCallback = (typeof options === 'function' ? options : callback) as (
|
||||
error: ExecException | null,
|
||||
stdout: string,
|
||||
stderr: string,
|
||||
) => void;
|
||||
if (actualCallback) {
|
||||
actualCallback(null, pm2StoppedOutput, '');
|
||||
}
|
||||
return { unref: () => {} } as ReturnType<typeof exec>;
|
||||
},
|
||||
);
|
||||
vi.mocked(systemService.getPm2Status).mockResolvedValue({
|
||||
success: false,
|
||||
message: 'Application process exists but is not online.',
|
||||
});
|
||||
|
||||
const response = await supertest(app).get('/api/system/pm2-status');
|
||||
|
||||
@@ -122,33 +74,10 @@ describe('System Routes (/api/system)', () => {
|
||||
|
||||
it('should return success: false when pm2 process does not exist', async () => {
|
||||
// Arrange: Simulate `pm2 describe` failing because the process isn't found.
|
||||
const processNotFoundOutput =
|
||||
"[PM2][ERROR] Process or Namespace flyer-crawler-api doesn't exist";
|
||||
const processNotFoundError = new Error(
|
||||
'Command failed: pm2 describe flyer-crawler-api',
|
||||
) as ExecException;
|
||||
processNotFoundError.code = 1;
|
||||
|
||||
vi.mocked(exec).mockImplementation(
|
||||
(
|
||||
command: string,
|
||||
options?:
|
||||
| ExecOptions
|
||||
| ((error: ExecException | null, stdout: string, stderr: string) => void)
|
||||
| null,
|
||||
callback?: ((error: ExecException | null, stdout: string, stderr: string) => void) | null,
|
||||
) => {
|
||||
const actualCallback = (typeof options === 'function' ? options : callback) as (
|
||||
error: ExecException | null,
|
||||
stdout: string,
|
||||
stderr: string,
|
||||
) => void;
|
||||
if (actualCallback) {
|
||||
actualCallback(processNotFoundError, processNotFoundOutput, '');
|
||||
}
|
||||
return { unref: () => {} } as ReturnType<typeof exec>;
|
||||
},
|
||||
);
|
||||
vi.mocked(systemService.getPm2Status).mockResolvedValue({
|
||||
success: false,
|
||||
message: 'Application process is not running under PM2.',
|
||||
});
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).get('/api/system/pm2-status');
|
||||
@@ -163,55 +92,17 @@ describe('System Routes (/api/system)', () => {
|
||||
|
||||
it('should return 500 if pm2 command produces stderr output', async () => {
|
||||
// Arrange: Simulate a successful exit code but with content in stderr.
|
||||
const stderrOutput = 'A non-fatal warning occurred.';
|
||||
|
||||
vi.mocked(exec).mockImplementation(
|
||||
(
|
||||
command: string,
|
||||
options?:
|
||||
| ExecOptions
|
||||
| ((error: ExecException | null, stdout: string, stderr: string) => void)
|
||||
| null,
|
||||
callback?: ((error: ExecException | null, stdout: string, stderr: string) => void) | null,
|
||||
) => {
|
||||
const actualCallback = (typeof options === 'function' ? options : callback) as (
|
||||
error: ExecException | null,
|
||||
stdout: string,
|
||||
stderr: string,
|
||||
) => void;
|
||||
if (actualCallback) {
|
||||
actualCallback(null, 'Some stdout', stderrOutput);
|
||||
}
|
||||
return { unref: () => {} } as ReturnType<typeof exec>;
|
||||
},
|
||||
);
|
||||
const serviceError = new Error('PM2 command produced an error: A non-fatal warning occurred.');
|
||||
vi.mocked(systemService.getPm2Status).mockRejectedValue(serviceError);
|
||||
|
||||
const response = await supertest(app).get('/api/system/pm2-status');
|
||||
expect(response.status).toBe(500);
|
||||
expect(response.body.message).toBe(`PM2 command produced an error: ${stderrOutput}`);
|
||||
expect(response.body.message).toBe(serviceError.message);
|
||||
});
|
||||
|
||||
it('should return 500 on a generic exec error', async () => {
|
||||
vi.mocked(exec).mockImplementation(
|
||||
(
|
||||
command: string,
|
||||
options?:
|
||||
| ExecOptions
|
||||
| ((error: ExecException | null, stdout: string, stderr: string) => void)
|
||||
| null,
|
||||
callback?: ((error: ExecException | null, stdout: string, stderr: string) => void) | null,
|
||||
) => {
|
||||
const actualCallback = (typeof options === 'function' ? options : callback) as (
|
||||
error: ExecException | null,
|
||||
stdout: string,
|
||||
stderr: string,
|
||||
) => void;
|
||||
if (actualCallback) {
|
||||
actualCallback(new Error('System error') as ExecException, '', 'stderr output');
|
||||
}
|
||||
return { unref: () => {} } as ReturnType<typeof exec>;
|
||||
},
|
||||
);
|
||||
const serviceError = new Error('System error');
|
||||
vi.mocked(systemService.getPm2Status).mockRejectedValue(serviceError);
|
||||
|
||||
// Act
|
||||
const response = await supertest(app).get('/api/system/pm2-status');
|
||||
|
||||
@@ -1,11 +1,11 @@
|
||||
// src/routes/system.routes.ts
|
||||
import { Router, Request, Response, NextFunction } from 'express';
|
||||
import { exec } from 'child_process';
|
||||
import { z } from 'zod';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { geocodingService } from '../services/geocodingService.server';
|
||||
import { validateRequest } from '../middleware/validation.middleware';
|
||||
import { z } from 'zod';
|
||||
import { requiredString } from '../utils/zodUtils';
|
||||
import { systemService } from '../services/systemService';
|
||||
|
||||
const router = Router();
|
||||
|
||||
@@ -25,39 +25,13 @@ const emptySchema = z.object({});
|
||||
router.get(
|
||||
'/pm2-status',
|
||||
validateRequest(emptySchema),
|
||||
(req: Request, res: Response, next: NextFunction) => {
|
||||
// The name 'flyer-crawler-api' comes from your ecosystem.config.cjs file.
|
||||
exec('pm2 describe flyer-crawler-api', (error, stdout, stderr) => {
|
||||
if (error) {
|
||||
// 'pm2 describe' exits with an error if the process is not found.
|
||||
// We can treat this as a "fail" status for our check.
|
||||
if (stdout && stdout.includes("doesn't exist")) {
|
||||
logger.warn('[API /pm2-status] PM2 process "flyer-crawler-api" not found.');
|
||||
return res.json({
|
||||
success: false,
|
||||
message: 'Application process is not running under PM2.',
|
||||
});
|
||||
}
|
||||
logger.error(
|
||||
{ error: stderr || error.message },
|
||||
'[API /pm2-status] Error executing pm2 describe:',
|
||||
);
|
||||
return next(error);
|
||||
}
|
||||
|
||||
// Check if there was output to stderr, even if the exit code was 0 (success).
|
||||
if (stderr && stderr.trim().length > 0) {
|
||||
logger.error({ stderr }, '[API /pm2-status] PM2 executed but produced stderr:');
|
||||
return next(new Error(`PM2 command produced an error: ${stderr}`));
|
||||
}
|
||||
|
||||
// If the command succeeds, we can parse stdout to check the status.
|
||||
const isOnline = /│ status\s+│ online\s+│/m.test(stdout);
|
||||
const message = isOnline
|
||||
? 'Application is online and running under PM2.'
|
||||
: 'Application process exists but is not online.';
|
||||
res.json({ success: isOnline, message });
|
||||
});
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
try {
|
||||
const status = await systemService.getPm2Status();
|
||||
res.json(status);
|
||||
} catch (error) {
|
||||
next(error);
|
||||
}
|
||||
},
|
||||
);
|
||||
|
||||
|
||||
@@ -1,8 +1,9 @@
|
||||
// src/routes/user.routes.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterAll } from 'vitest';
|
||||
import supertest from 'supertest';
|
||||
import express from 'express';
|
||||
import * as bcrypt from 'bcrypt';
|
||||
import path from 'path';
|
||||
import fs from 'node:fs/promises';
|
||||
import {
|
||||
createMockUserProfile,
|
||||
createMockMasterGroceryItem,
|
||||
@@ -16,10 +17,12 @@ import {
|
||||
createMockAddress,
|
||||
} from '../tests/utils/mockFactories';
|
||||
import { Appliance, Notification, DietaryRestriction } from '../types';
|
||||
import { ForeignKeyConstraintError, NotFoundError } from '../services/db/errors.db';
|
||||
import { ForeignKeyConstraintError, NotFoundError, ValidationError } from '../services/db/errors.db';
|
||||
import { createTestApp } from '../tests/utils/createTestApp';
|
||||
import { mockLogger } from '../tests/utils/mockLogger';
|
||||
import { cleanupFiles } from '../tests/utils/cleanupFiles';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { userService } from '../services/userService';
|
||||
|
||||
// 1. Mock the Service Layer directly.
|
||||
// The user.routes.ts file imports from '.../db/index.db'. We need to mock that module.
|
||||
@@ -28,9 +31,6 @@ vi.mock('../services/db/index.db', () => ({
|
||||
userRepo: {
|
||||
findUserProfileById: vi.fn(),
|
||||
updateUserProfile: vi.fn(),
|
||||
updateUserPassword: vi.fn(),
|
||||
findUserWithPasswordHashById: vi.fn(),
|
||||
deleteUserById: vi.fn(),
|
||||
updateUserPreferences: vi.fn(),
|
||||
},
|
||||
personalizationRepo: {
|
||||
@@ -69,22 +69,14 @@ vi.mock('../services/db/index.db', () => ({
|
||||
// Mock userService
|
||||
vi.mock('../services/userService', () => ({
|
||||
userService: {
|
||||
updateUserAvatar: vi.fn(),
|
||||
updateUserPassword: vi.fn(),
|
||||
deleteUserAccount: vi.fn(),
|
||||
getUserAddress: vi.fn(),
|
||||
upsertUserAddress: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// 2. Mock bcrypt.
|
||||
// We return an object that satisfies both default and named imports to be safe.
|
||||
vi.mock('bcrypt', () => {
|
||||
const hash = vi.fn();
|
||||
const compare = vi.fn();
|
||||
return {
|
||||
default: { hash, compare },
|
||||
hash,
|
||||
compare,
|
||||
};
|
||||
});
|
||||
|
||||
// Mock the logger
|
||||
vi.mock('../services/logger.server', async () => ({
|
||||
// Use async import to avoid hoisting issues with mockLogger
|
||||
@@ -93,7 +85,6 @@ vi.mock('../services/logger.server', async () => ({
|
||||
|
||||
// Import the router and other modules AFTER mocks are established
|
||||
import userRouter from './user.routes';
|
||||
import { userService } from '../services/userService'; // Import for checking calls
|
||||
// Import the mocked db module to control its functions in tests
|
||||
import * as db from '../services/db/index.db';
|
||||
|
||||
@@ -148,8 +139,8 @@ describe('User Routes (/api/users)', () => {
|
||||
|
||||
// Assert
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: mkdirError },
|
||||
'Failed to create avatar upload directory',
|
||||
{ error: mkdirError },
|
||||
'Failed to create multer storage directories on startup.',
|
||||
);
|
||||
vi.doUnmock('node:fs/promises'); // Clean up
|
||||
});
|
||||
@@ -177,6 +168,26 @@ describe('User Routes (/api/users)', () => {
|
||||
beforeEach(() => {
|
||||
// All tests in this block will use the authenticated app
|
||||
});
|
||||
|
||||
afterAll(async () => {
|
||||
// Safeguard to clean up any avatar files created during tests.
|
||||
const uploadDir = path.resolve(__dirname, '../../../uploads/avatars');
|
||||
try {
|
||||
const allFiles = await fs.readdir(uploadDir);
|
||||
// Files are named like 'avatar-user-123-timestamp.ext'
|
||||
const testFiles = allFiles
|
||||
.filter((f) => f.startsWith(`avatar-${mockUserProfile.user.user_id}`))
|
||||
.map((f) => path.join(uploadDir, f));
|
||||
|
||||
if (testFiles.length > 0) {
|
||||
await cleanupFiles(testFiles);
|
||||
}
|
||||
} catch (error) {
|
||||
if (error instanceof Error && (error as NodeJS.ErrnoException).code !== 'ENOENT') {
|
||||
console.error('Error during user routes test file cleanup:', error);
|
||||
}
|
||||
}
|
||||
});
|
||||
describe('GET /profile', () => {
|
||||
it('should return the full user profile', async () => {
|
||||
vi.mocked(db.userRepo.findUserProfileById).mockResolvedValue(mockUserProfile);
|
||||
@@ -598,20 +609,17 @@ describe('User Routes (/api/users)', () => {
|
||||
|
||||
describe('PUT /profile/password', () => {
|
||||
it('should update the password successfully with a strong password', async () => {
|
||||
vi.mocked(bcrypt.hash).mockResolvedValue('hashed-password' as never);
|
||||
vi.mocked(db.userRepo.updateUserPassword).mockResolvedValue(undefined);
|
||||
vi.mocked(userService.updateUserPassword).mockResolvedValue(undefined);
|
||||
const response = await supertest(app)
|
||||
.put('/api/users/profile/password')
|
||||
.send({ newPassword: 'a-Very-Strong-Password-456!' });
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.message).toBe('Password updated successfully.');
|
||||
});
|
||||
|
||||
it('should return 500 on a generic database error', async () => {
|
||||
const dbError = new Error('DB Connection Failed');
|
||||
vi.mocked(bcrypt.hash).mockResolvedValue('hashed-password' as never);
|
||||
vi.mocked(db.userRepo.updateUserPassword).mockRejectedValue(dbError);
|
||||
vi.mocked(userService.updateUserPassword).mockRejectedValue(dbError);
|
||||
const response = await supertest(app)
|
||||
.put('/api/users/profile/password')
|
||||
.send({ newPassword: 'a-Very-Strong-Password-456!' });
|
||||
@@ -623,7 +631,6 @@ describe('User Routes (/api/users)', () => {
|
||||
});
|
||||
|
||||
it('should return 400 for a weak password', async () => {
|
||||
// Use a password long enough to pass .min(8) but weak enough to fail strength check
|
||||
const response = await supertest(app)
|
||||
.put('/api/users/profile/password')
|
||||
.send({ newPassword: 'password123' });
|
||||
@@ -635,70 +642,38 @@ describe('User Routes (/api/users)', () => {
|
||||
|
||||
describe('DELETE /account', () => {
|
||||
it('should delete the account with the correct password', async () => {
|
||||
const userWithHash = createMockUserWithPasswordHash({
|
||||
...mockUserProfile.user,
|
||||
password_hash: 'hashed-password',
|
||||
});
|
||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithHash);
|
||||
vi.mocked(db.userRepo.deleteUserById).mockResolvedValue(undefined);
|
||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never);
|
||||
vi.mocked(userService.deleteUserAccount).mockResolvedValue(undefined);
|
||||
const response = await supertest(app)
|
||||
.delete('/api/users/account')
|
||||
.send({ password: 'correct-password' });
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.message).toBe('Account deleted successfully.');
|
||||
expect(userService.deleteUserAccount).toHaveBeenCalledWith('user-123', 'correct-password', expectLogger);
|
||||
});
|
||||
|
||||
it('should return 403 for an incorrect password', async () => {
|
||||
const userWithHash = createMockUserWithPasswordHash({
|
||||
...mockUserProfile.user,
|
||||
password_hash: 'hashed-password',
|
||||
});
|
||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithHash);
|
||||
vi.mocked(bcrypt.compare).mockResolvedValue(false as never);
|
||||
it('should return 400 for an incorrect password', async () => {
|
||||
vi.mocked(userService.deleteUserAccount).mockRejectedValue(new ValidationError([], 'Incorrect password.'));
|
||||
const response = await supertest(app)
|
||||
.delete('/api/users/account')
|
||||
.send({ password: 'wrong-password' });
|
||||
|
||||
expect(response.status).toBe(403);
|
||||
expect(response.status).toBe(400);
|
||||
expect(response.body.message).toBe('Incorrect password.');
|
||||
});
|
||||
|
||||
it('should return 404 if the user to delete is not found', async () => {
|
||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockRejectedValue(
|
||||
new NotFoundError('User not found or password not set.'),
|
||||
);
|
||||
const response = await supertest(app)
|
||||
.delete('/api/users/account')
|
||||
.send({ password: 'any-password' });
|
||||
expect(response.status).toBe(404);
|
||||
expect(response.body.message).toBe('User not found or password not set.');
|
||||
});
|
||||
|
||||
it('should return 404 if user is an OAuth user without a password', async () => {
|
||||
// Simulate an OAuth user who has no password_hash set.
|
||||
const userWithoutHash = createMockUserWithPasswordHash({
|
||||
...mockUserProfile.user,
|
||||
password_hash: null,
|
||||
});
|
||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithoutHash);
|
||||
vi.mocked(userService.deleteUserAccount).mockRejectedValue(new NotFoundError('User not found.'));
|
||||
|
||||
const response = await supertest(app)
|
||||
.delete('/api/users/account')
|
||||
.send({ password: 'any-password' });
|
||||
|
||||
expect(response.status).toBe(404);
|
||||
expect(response.body.message).toBe('User not found or password not set.');
|
||||
expect(response.body.message).toBe('User not found.');
|
||||
});
|
||||
|
||||
it('should return 500 on a generic database error', async () => {
|
||||
const userWithHash = createMockUserWithPasswordHash({
|
||||
...mockUserProfile.user,
|
||||
password_hash: 'hashed-password',
|
||||
});
|
||||
vi.mocked(db.userRepo.findUserWithPasswordHashById).mockResolvedValue(userWithHash);
|
||||
vi.mocked(bcrypt.compare).mockResolvedValue(true as never);
|
||||
vi.mocked(db.userRepo.deleteUserById).mockRejectedValue(new Error('DB Connection Failed'));
|
||||
vi.mocked(userService.deleteUserAccount).mockRejectedValue(new Error('DB Connection Failed'));
|
||||
const response = await supertest(app)
|
||||
.delete('/api/users/account')
|
||||
.send({ password: 'correct-password' });
|
||||
@@ -979,7 +954,7 @@ describe('User Routes (/api/users)', () => {
|
||||
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
||||
});
|
||||
const mockAddress = createMockAddress({ address_id: 1, address_line_1: '123 Main St' });
|
||||
vi.mocked(db.addressRepo.getAddressById).mockResolvedValue(mockAddress);
|
||||
vi.mocked(userService.getUserAddress).mockResolvedValue(mockAddress);
|
||||
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body).toEqual(mockAddress);
|
||||
@@ -991,7 +966,7 @@ describe('User Routes (/api/users)', () => {
|
||||
basePath,
|
||||
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
||||
});
|
||||
vi.mocked(db.addressRepo.getAddressById).mockRejectedValue(new Error('DB Error'));
|
||||
vi.mocked(userService.getUserAddress).mockRejectedValue(new Error('DB Error'));
|
||||
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
||||
expect(response.status).toBe(500);
|
||||
});
|
||||
@@ -1004,13 +979,10 @@ describe('User Routes (/api/users)', () => {
|
||||
});
|
||||
|
||||
it('GET /addresses/:addressId should return 403 if address does not belong to user', async () => {
|
||||
const appWithDifferentUser = createTestApp({
|
||||
router: userRouter,
|
||||
basePath,
|
||||
authenticatedUser: { ...mockUserProfile, address_id: 999 },
|
||||
});
|
||||
const response = await supertest(appWithDifferentUser).get('/api/users/addresses/1');
|
||||
expect(response.status).toBe(403);
|
||||
vi.mocked(userService.getUserAddress).mockRejectedValue(new ValidationError([], 'Forbidden'));
|
||||
const response = await supertest(app).get('/api/users/addresses/2'); // Requesting address 2
|
||||
expect(response.status).toBe(400); // ValidationError maps to 400 by default in the test error handler
|
||||
expect(response.body.message).toBe('Forbidden');
|
||||
});
|
||||
|
||||
it('GET /addresses/:addressId should return 404 if address not found', async () => {
|
||||
@@ -1019,7 +991,7 @@ describe('User Routes (/api/users)', () => {
|
||||
basePath,
|
||||
authenticatedUser: { ...mockUserProfile, address_id: 1 },
|
||||
});
|
||||
vi.mocked(db.addressRepo.getAddressById).mockRejectedValue(
|
||||
vi.mocked(userService.getUserAddress).mockRejectedValue(
|
||||
new NotFoundError('Address not found.'),
|
||||
);
|
||||
const response = await supertest(appWithUser).get('/api/users/addresses/1');
|
||||
@@ -1028,19 +1000,10 @@ describe('User Routes (/api/users)', () => {
|
||||
});
|
||||
|
||||
it('PUT /profile/address should call upsertAddress and updateUserProfile if needed', async () => {
|
||||
const appWithUser = createTestApp({
|
||||
router: userRouter,
|
||||
basePath,
|
||||
authenticatedUser: { ...mockUserProfile, address_id: null },
|
||||
}); // User has no address yet
|
||||
const addressData = { address_line_1: '123 New St' };
|
||||
vi.mocked(db.addressRepo.upsertAddress).mockResolvedValue(5); // New address ID is 5
|
||||
vi.mocked(db.userRepo.updateUserProfile).mockResolvedValue({
|
||||
...mockUserProfile,
|
||||
address_id: 5,
|
||||
});
|
||||
vi.mocked(userService.upsertUserAddress).mockResolvedValue(5);
|
||||
|
||||
const response = await supertest(appWithUser)
|
||||
const response = await supertest(app)
|
||||
.put('/api/users/profile/address')
|
||||
.send(addressData);
|
||||
|
||||
@@ -1072,11 +1035,11 @@ describe('User Routes (/api/users)', () => {
|
||||
|
||||
describe('POST /profile/avatar', () => {
|
||||
it('should upload an avatar and update the user profile', async () => {
|
||||
const mockUpdatedProfile = {
|
||||
const mockUpdatedProfile = createMockUserProfile({
|
||||
...mockUserProfile,
|
||||
avatar_url: '/uploads/avatars/new-avatar.png',
|
||||
};
|
||||
vi.mocked(db.userRepo.updateUserProfile).mockResolvedValue(mockUpdatedProfile);
|
||||
});
|
||||
vi.mocked(userService.updateUserAvatar).mockResolvedValue(mockUpdatedProfile);
|
||||
|
||||
// Create a dummy file path for supertest to attach
|
||||
const dummyImagePath = 'test-avatar.png';
|
||||
@@ -1086,17 +1049,17 @@ describe('User Routes (/api/users)', () => {
|
||||
.attach('avatar', Buffer.from('dummy-image-content'), dummyImagePath);
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.avatar_url).toContain('/uploads/avatars/');
|
||||
expect(db.userRepo.updateUserProfile).toHaveBeenCalledWith(
|
||||
expect(response.body.avatar_url).toContain('/uploads/avatars/'); // This was a duplicate, fixed.
|
||||
expect(userService.updateUserAvatar).toHaveBeenCalledWith(
|
||||
mockUserProfile.user.user_id,
|
||||
{ avatar_url: expect.any(String) },
|
||||
expect.any(Object),
|
||||
expectLogger,
|
||||
);
|
||||
});
|
||||
|
||||
it('should return 500 if updating the profile fails after upload', async () => {
|
||||
const dbError = new Error('DB Connection Failed');
|
||||
vi.mocked(db.userRepo.updateUserProfile).mockRejectedValue(dbError);
|
||||
vi.mocked(userService.updateUserAvatar).mockRejectedValue(dbError);
|
||||
const dummyImagePath = 'test-avatar.png';
|
||||
const response = await supertest(app)
|
||||
.post('/api/users/profile/avatar')
|
||||
@@ -1135,6 +1098,27 @@ describe('User Routes (/api/users)', () => {
|
||||
expect(response.body.message).toBe('No avatar file uploaded.');
|
||||
});
|
||||
|
||||
it('should clean up the uploaded file if updating the profile fails', async () => {
|
||||
// Spy on the unlink function to ensure it's called on error
|
||||
const unlinkSpy = vi.spyOn(fs, 'unlink').mockResolvedValue(undefined);
|
||||
|
||||
const dbError = new Error('DB Connection Failed');
|
||||
vi.mocked(userService.updateUserAvatar).mockRejectedValue(dbError);
|
||||
const dummyImagePath = 'test-avatar.png';
|
||||
|
||||
const response = await supertest(app)
|
||||
.post('/api/users/profile/avatar')
|
||||
.attach('avatar', Buffer.from('dummy-image-content'), dummyImagePath);
|
||||
|
||||
expect(response.status).toBe(500);
|
||||
// Verify that the cleanup function was called
|
||||
expect(unlinkSpy).toHaveBeenCalledTimes(1);
|
||||
// The filename is predictable because of the multer config in user.routes.ts
|
||||
expect(unlinkSpy).toHaveBeenCalledWith(expect.stringContaining('test-avatar.png'));
|
||||
|
||||
unlinkSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should return 400 for a non-numeric address ID', async () => {
|
||||
const response = await supertest(app).get('/api/users/addresses/abc');
|
||||
expect(response.status).toBe(400);
|
||||
|
||||
@@ -1,13 +1,14 @@
|
||||
// src/routes/user.routes.ts
|
||||
import express, { Request, Response, NextFunction } from 'express';
|
||||
import passport from './passport.routes';
|
||||
import multer from 'multer';
|
||||
import path from 'path';
|
||||
import fs from 'node:fs/promises';
|
||||
import * as bcrypt from 'bcrypt'; // This was a duplicate, fixed.
|
||||
import multer from 'multer'; // Keep for MulterError type check
|
||||
import { z } from 'zod';
|
||||
import { logger } from '../services/logger.server';
|
||||
import { UserProfile } from '../types';
|
||||
import {
|
||||
createUploadMiddleware,
|
||||
handleMulterError,
|
||||
} from '../middleware/multer.middleware';
|
||||
import { userService } from '../services/userService';
|
||||
import { ForeignKeyConstraintError } from '../services/db/errors.db';
|
||||
import { validateRequest } from '../middleware/validation.middleware';
|
||||
@@ -19,6 +20,7 @@ import {
|
||||
optionalBoolean,
|
||||
} from '../utils/zodUtils';
|
||||
import * as db from '../services/db/index.db';
|
||||
import { cleanupUploadedFile } from '../utils/fileUtils';
|
||||
|
||||
const router = express.Router();
|
||||
|
||||
@@ -34,6 +36,7 @@ const updatePasswordSchema = z.object({
|
||||
body: z.object({
|
||||
newPassword: z
|
||||
.string()
|
||||
.trim() // Trim whitespace from password input.
|
||||
.min(8, 'Password must be at least 8 characters long.')
|
||||
.superRefine((password, ctx) => {
|
||||
const strength = validatePasswordStrength(password);
|
||||
@@ -42,6 +45,9 @@ const updatePasswordSchema = z.object({
|
||||
}),
|
||||
});
|
||||
|
||||
// The `requiredString` utility (modified in `zodUtils.ts`) now handles trimming,
|
||||
// so no changes are needed here, but we are confirming that password trimming
|
||||
// is now implicitly handled for this schema.
|
||||
const deleteAccountSchema = z.object({
|
||||
body: z.object({ password: requiredString("Field 'password' is required.") }),
|
||||
});
|
||||
@@ -72,35 +78,10 @@ const emptySchema = z.object({});
|
||||
// Any request to a /api/users/* endpoint will now require a valid JWT.
|
||||
router.use(passport.authenticate('jwt', { session: false }));
|
||||
|
||||
// --- Multer Configuration for Avatar Uploads ---
|
||||
|
||||
// Ensure the directory for avatar uploads exists.
|
||||
const avatarUploadDir = path.join(process.cwd(), 'public', 'uploads', 'avatars');
|
||||
fs.mkdir(avatarUploadDir, { recursive: true }).catch((err) => {
|
||||
logger.error({ err }, 'Failed to create avatar upload directory');
|
||||
});
|
||||
|
||||
// Define multer storage configuration. The `req.user` object will be available
|
||||
// here because the passport middleware runs before this route handler.
|
||||
const avatarStorage = multer.diskStorage({
|
||||
destination: (req, file, cb) => cb(null, avatarUploadDir),
|
||||
filename: (req, file, cb) => {
|
||||
const uniqueSuffix = `${(req.user as UserProfile).user.user_id}-${Date.now()}${path.extname(file.originalname)}`;
|
||||
cb(null, uniqueSuffix);
|
||||
},
|
||||
});
|
||||
|
||||
const avatarUpload = multer({
|
||||
storage: avatarStorage,
|
||||
limits: { fileSize: 1 * 1024 * 1024 }, // 1MB file size limit
|
||||
fileFilter: (req, file, cb) => {
|
||||
if (file.mimetype.startsWith('image/')) {
|
||||
cb(null, true);
|
||||
} else {
|
||||
// Reject the file with a specific error
|
||||
cb(new Error('Only image files are allowed!'));
|
||||
}
|
||||
},
|
||||
const avatarUpload = createUploadMiddleware({
|
||||
storageType: 'avatar',
|
||||
fileSize: 1 * 1024 * 1024, // 1MB
|
||||
fileFilter: 'image',
|
||||
});
|
||||
|
||||
/**
|
||||
@@ -110,18 +91,17 @@ router.post(
|
||||
'/profile/avatar',
|
||||
avatarUpload.single('avatar'),
|
||||
async (req: Request, res: Response, next: NextFunction) => {
|
||||
// The try-catch block was already correct here.
|
||||
try {
|
||||
// The try-catch block was already correct here.
|
||||
// The `requireFileUpload` middleware is not used here, so we must check for `req.file`.
|
||||
if (!req.file) return res.status(400).json({ message: 'No avatar file uploaded.' });
|
||||
const userProfile = req.user as UserProfile;
|
||||
const avatarUrl = `/uploads/avatars/${req.file.filename}`;
|
||||
const updatedProfile = await db.userRepo.updateUserProfile(
|
||||
userProfile.user.user_id,
|
||||
{ avatar_url: avatarUrl },
|
||||
req.log,
|
||||
);
|
||||
const updatedProfile = await userService.updateUserAvatar(userProfile.user.user_id, req.file, req.log);
|
||||
res.json(updatedProfile);
|
||||
} catch (error) {
|
||||
// If an error occurs after the file has been uploaded (e.g., DB error),
|
||||
// we must clean up the orphaned file from the disk.
|
||||
await cleanupUploadedFile(req.file);
|
||||
logger.error({ error }, 'Error uploading avatar');
|
||||
next(error);
|
||||
}
|
||||
@@ -263,9 +243,7 @@ router.put(
|
||||
const { body } = req as unknown as UpdatePasswordRequest;
|
||||
|
||||
try {
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(body.newPassword, saltRounds);
|
||||
await db.userRepo.updateUserPassword(userProfile.user.user_id, hashedPassword, req.log);
|
||||
await userService.updateUserPassword(userProfile.user.user_id, body.newPassword, req.log);
|
||||
res.status(200).json({ message: 'Password updated successfully.' });
|
||||
} catch (error) {
|
||||
logger.error({ error }, `[ROUTE] PUT /api/users/profile/password - ERROR`);
|
||||
@@ -288,20 +266,7 @@ router.delete(
|
||||
const { body } = req as unknown as DeleteAccountRequest;
|
||||
|
||||
try {
|
||||
const userWithHash = await db.userRepo.findUserWithPasswordHashById(
|
||||
userProfile.user.user_id,
|
||||
req.log,
|
||||
);
|
||||
if (!userWithHash || !userWithHash.password_hash) {
|
||||
return res.status(404).json({ message: 'User not found or password not set.' });
|
||||
}
|
||||
|
||||
const isMatch = await bcrypt.compare(body.password, userWithHash.password_hash);
|
||||
if (!isMatch) {
|
||||
return res.status(403).json({ message: 'Incorrect password.' });
|
||||
}
|
||||
|
||||
await db.userRepo.deleteUserById(userProfile.user.user_id, req.log);
|
||||
await userService.deleteUserAccount(userProfile.user.user_id, body.password, req.log);
|
||||
res.status(200).json({ message: 'Account deleted successfully.' });
|
||||
} catch (error) {
|
||||
logger.error({ error }, `[ROUTE] DELETE /api/users/account - ERROR`);
|
||||
@@ -491,7 +456,11 @@ const addShoppingListItemSchema = shoppingListIdSchema.extend({
|
||||
body: z
|
||||
.object({
|
||||
masterItemId: z.number().int().positive().optional(),
|
||||
customItemName: z.string().min(1, 'customItemName cannot be empty if provided').optional(),
|
||||
customItemName: z
|
||||
.string()
|
||||
.trim()
|
||||
.min(1, 'customItemName cannot be empty if provided')
|
||||
.optional(),
|
||||
})
|
||||
.refine((data) => data.masterItemId || data.customItemName, {
|
||||
message: 'Either masterItemId or customItemName must be provided.',
|
||||
@@ -717,13 +686,7 @@ router.get(
|
||||
const { params } = req as unknown as GetAddressRequest;
|
||||
try {
|
||||
const addressId = params.addressId;
|
||||
// Security check: Ensure the requested addressId matches the one on the user's profile.
|
||||
if (userProfile.address_id !== addressId) {
|
||||
return res
|
||||
.status(403)
|
||||
.json({ message: 'Forbidden: You can only access your own address.' });
|
||||
}
|
||||
const address = await db.addressRepo.getAddressById(addressId, req.log); // This will throw NotFoundError if not found
|
||||
const address = await userService.getUserAddress(userProfile, addressId, req.log);
|
||||
res.json(address);
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'Error fetching user address');
|
||||
@@ -738,12 +701,12 @@ router.get(
|
||||
const updateUserAddressSchema = z.object({
|
||||
body: z
|
||||
.object({
|
||||
address_line_1: z.string().optional(),
|
||||
address_line_2: z.string().optional(),
|
||||
city: z.string().optional(),
|
||||
province_state: z.string().optional(),
|
||||
postal_code: z.string().optional(),
|
||||
country: z.string().optional(),
|
||||
address_line_1: z.string().trim().optional(),
|
||||
address_line_2: z.string().trim().optional(),
|
||||
city: z.string().trim().optional(),
|
||||
province_state: z.string().trim().optional(),
|
||||
postal_code: z.string().trim().optional(),
|
||||
country: z.string().trim().optional(),
|
||||
})
|
||||
.refine((data) => Object.keys(data).length > 0, {
|
||||
message: 'At least one address field must be provided.',
|
||||
@@ -803,13 +766,13 @@ router.delete(
|
||||
const updateRecipeSchema = recipeIdSchema.extend({
|
||||
body: z
|
||||
.object({
|
||||
name: z.string().optional(),
|
||||
description: z.string().optional(),
|
||||
instructions: z.string().optional(),
|
||||
name: z.string().trim().optional(),
|
||||
description: z.string().trim().optional(),
|
||||
instructions: z.string().trim().optional(),
|
||||
prep_time_minutes: z.number().int().optional(),
|
||||
cook_time_minutes: z.number().int().optional(),
|
||||
servings: z.number().int().optional(),
|
||||
photo_url: z.string().url().optional(),
|
||||
photo_url: z.string().trim().url().optional(),
|
||||
})
|
||||
.refine((data) => Object.keys(data).length > 0, { message: 'No fields provided to update.' }),
|
||||
});
|
||||
@@ -841,18 +804,7 @@ router.put(
|
||||
},
|
||||
);
|
||||
|
||||
// --- General Multer Error Handler ---
|
||||
// This should be placed after all routes that use multer.
|
||||
// It catches errors from `fileFilter` and other multer issues (e.g., file size limits).
|
||||
router.use((err: Error, req: Request, res: Response, next: NextFunction) => {
|
||||
if (err instanceof multer.MulterError) {
|
||||
// A Multer error occurred when uploading (e.g., file too large).
|
||||
return res.status(400).json({ message: `File upload error: ${err.message}` });
|
||||
} else if (err && err.message === 'Only image files are allowed!') {
|
||||
// A custom error from our fileFilter.
|
||||
return res.status(400).json({ message: err.message });
|
||||
}
|
||||
next(err); // Pass on to the next error handler if it's not a multer error we handle.
|
||||
});
|
||||
/* Catches errors from multer (e.g., file size, file filter) */
|
||||
router.use(handleMulterError);
|
||||
|
||||
export default router;
|
||||
|
||||
@@ -19,13 +19,15 @@ vi.mock('./logger.client', () => ({
|
||||
debug: vi.fn(),
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
// 2. Mock ./apiClient to simply pass calls through to the global fetch.
|
||||
vi.mock('./apiClient', async (importOriginal) => {
|
||||
return {
|
||||
apiFetch: (
|
||||
// This is the core logic we want to preserve: it calls the global fetch
|
||||
// which is then intercepted by MSW.
|
||||
const apiFetch = (
|
||||
url: string,
|
||||
options: RequestInit = {},
|
||||
apiOptions: import('./apiClient').ApiOptions = {},
|
||||
@@ -59,6 +61,26 @@ vi.mock('./apiClient', async (importOriginal) => {
|
||||
const request = new Request(fullUrl, options);
|
||||
console.log(`[apiFetch MOCK] Executing fetch for URL: ${request.url}.`);
|
||||
return fetch(request);
|
||||
};
|
||||
|
||||
return {
|
||||
// The original mock only had apiFetch. We need to add the helpers.
|
||||
apiFetch,
|
||||
|
||||
// These helpers are what aiApiClient.ts actually calls.
|
||||
// Their mock implementation should just call our mocked apiFetch.
|
||||
authedGet: (endpoint: string, options: import('./apiClient').ApiOptions = {}) => {
|
||||
return apiFetch(endpoint, { method: 'GET' }, options);
|
||||
},
|
||||
authedPost: <T>(endpoint: string, body: T, options: import('./apiClient').ApiOptions = {}) => {
|
||||
return apiFetch(
|
||||
endpoint,
|
||||
{ method: 'POST', headers: { 'Content-Type': 'application/json' }, body: JSON.stringify(body) },
|
||||
options,
|
||||
);
|
||||
},
|
||||
authedPostForm: (endpoint: string, formData: FormData, options: import('./apiClient').ApiOptions = {}) => {
|
||||
return apiFetch(endpoint, { method: 'POST', body: formData }, options);
|
||||
},
|
||||
// Add a mock for ApiOptions to satisfy the compiler
|
||||
ApiOptions: vi.fn(),
|
||||
@@ -178,6 +200,45 @@ describe('AI API Client (Network Mocking with MSW)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('uploadAndProcessFlyer error handling', () => {
|
||||
it('should throw a structured error with JSON body on non-ok response', async () => {
|
||||
const mockFile = new File(['content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
const checksum = 'checksum-abc-123';
|
||||
const errorBody = { message: 'Checksum already exists', flyerId: 99 };
|
||||
|
||||
server.use(
|
||||
http.post('http://localhost/api/ai/upload-and-process', () => {
|
||||
return HttpResponse.json(errorBody, { status: 409 });
|
||||
}),
|
||||
);
|
||||
|
||||
// The function now throws a structured object, not an Error instance.
|
||||
await expect(aiApiClient.uploadAndProcessFlyer(mockFile, checksum)).rejects.toEqual({
|
||||
status: 409,
|
||||
body: errorBody,
|
||||
});
|
||||
});
|
||||
|
||||
it('should throw a structured error with text body on non-ok, non-JSON response', async () => {
|
||||
const mockFile = new File(['content'], 'flyer.pdf', { type: 'application/pdf' });
|
||||
const checksum = 'checksum-abc-123';
|
||||
const errorText = 'Internal Server Error';
|
||||
|
||||
server.use(
|
||||
http.post('http://localhost/api/ai/upload-and-process', () => {
|
||||
return HttpResponse.text(errorText, { status: 500 });
|
||||
}),
|
||||
);
|
||||
|
||||
// The function now throws a structured object, not an Error instance.
|
||||
// The catch block in the implementation wraps the text in a message property.
|
||||
await expect(aiApiClient.uploadAndProcessFlyer(mockFile, checksum)).rejects.toEqual({
|
||||
status: 500,
|
||||
body: { message: errorText },
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('getJobStatus', () => {
|
||||
it('should send a GET request to the correct job status URL', async () => {
|
||||
const jobId = 'job-id-456';
|
||||
@@ -192,6 +253,82 @@ describe('AI API Client (Network Mocking with MSW)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('getJobStatus error handling', () => {
|
||||
const jobId = 'job-id-789';
|
||||
|
||||
it('should throw a JobFailedError if job state is "failed"', async () => {
|
||||
const failedStatus: aiApiClient.JobStatus = {
|
||||
id: jobId,
|
||||
state: 'failed',
|
||||
progress: { message: 'AI model exploded', errorCode: 'AI_ERROR' },
|
||||
returnValue: null,
|
||||
failedReason: 'Raw error from BullMQ',
|
||||
};
|
||||
|
||||
server.use(
|
||||
http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => {
|
||||
return HttpResponse.json(failedStatus);
|
||||
}),
|
||||
);
|
||||
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow(
|
||||
new aiApiClient.JobFailedError('AI model exploded', 'AI_ERROR'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should use failedReason for JobFailedError if progress message is missing', async () => {
|
||||
const failedStatus: aiApiClient.JobStatus = {
|
||||
id: jobId,
|
||||
state: 'failed',
|
||||
progress: null, // No progress object
|
||||
returnValue: null,
|
||||
failedReason: 'Raw error from BullMQ',
|
||||
};
|
||||
|
||||
server.use(
|
||||
http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => {
|
||||
return HttpResponse.json(failedStatus);
|
||||
}),
|
||||
);
|
||||
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow(
|
||||
new aiApiClient.JobFailedError('Raw error from BullMQ', 'UNKNOWN_ERROR'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error if the API response is not ok', async () => {
|
||||
const errorBody = { message: 'Job not found' };
|
||||
server.use(
|
||||
http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => {
|
||||
return HttpResponse.json(errorBody, { status: 404 });
|
||||
}),
|
||||
);
|
||||
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow('Job not found');
|
||||
});
|
||||
|
||||
it('should throw a specific error if a 200 OK response is not valid JSON', async () => {
|
||||
server.use(
|
||||
http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => {
|
||||
// A 200 OK response that is not JSON is a server-side contract violation.
|
||||
return HttpResponse.text('This should have been JSON', { status: 200 });
|
||||
}),
|
||||
);
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow(
|
||||
'Failed to parse job status from a successful API response.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error with status text if the non-ok API response is not valid JSON', async () => {
|
||||
server.use(
|
||||
http.get(`http://localhost/api/ai/jobs/${jobId}/status`, () => {
|
||||
return HttpResponse.text('Gateway Timeout', { status: 504, statusText: 'Gateway Timeout' });
|
||||
}),
|
||||
);
|
||||
await expect(aiApiClient.getJobStatus(jobId)).rejects.toThrow('Gateway Timeout');
|
||||
});
|
||||
});
|
||||
|
||||
describe('isImageAFlyer', () => {
|
||||
it('should construct FormData and send a POST request', async () => {
|
||||
const mockFile = new File(['dummy image content'], 'flyer.jpg', { type: 'image/jpeg' });
|
||||
|
||||
@@ -12,7 +12,7 @@ import type {
|
||||
GroundedResponse,
|
||||
} from '../types';
|
||||
import { logger } from './logger.client';
|
||||
import { apiFetch } from './apiClient';
|
||||
import { apiFetch, authedGet, authedPost, authedPostForm } from './apiClient';
|
||||
|
||||
/**
|
||||
* Uploads a flyer file to the backend to be processed asynchronously.
|
||||
@@ -33,21 +33,16 @@ export const uploadAndProcessFlyer = async (
|
||||
|
||||
logger.info(`[aiApiClient] Starting background processing for file: ${file.name}`);
|
||||
|
||||
const response = await apiFetch(
|
||||
'/ai/upload-and-process',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
const response = await authedPostForm('/ai/upload-and-process', formData, { tokenOverride });
|
||||
|
||||
if (!response.ok) {
|
||||
let errorBody;
|
||||
// Clone the response so we can read the body twice (once as JSON, and as text on failure).
|
||||
const clonedResponse = response.clone();
|
||||
try {
|
||||
errorBody = await response.json();
|
||||
} catch (e) {
|
||||
errorBody = { message: await response.text() };
|
||||
errorBody = { message: await clonedResponse.text() };
|
||||
}
|
||||
// Throw a structured error so the component can inspect the status and body
|
||||
throw { status: response.status, body: errorBody };
|
||||
@@ -63,12 +58,28 @@ export interface JobStatus {
|
||||
progress: {
|
||||
stages?: ProcessingStage[];
|
||||
estimatedTimeRemaining?: number;
|
||||
// The structured error payload from the backend worker
|
||||
errorCode?: string;
|
||||
message?: string;
|
||||
} | null;
|
||||
returnValue: {
|
||||
flyerId?: number;
|
||||
} | null;
|
||||
failedReason: string | null;
|
||||
failedReason: string | null; // The raw error string from BullMQ
|
||||
}
|
||||
|
||||
/**
|
||||
* Custom error class for job failures to make `catch` blocks more specific.
|
||||
* This allows the UI to easily distinguish between a job failure and a network error.
|
||||
*/
|
||||
export class JobFailedError extends Error {
|
||||
public errorCode: string;
|
||||
|
||||
constructor(message: string, errorCode: string) {
|
||||
super(message);
|
||||
this.name = 'JobFailedError';
|
||||
this.errorCode = errorCode;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -77,30 +88,59 @@ export interface JobStatus {
|
||||
* @param jobId The ID of the job to check.
|
||||
* @param tokenOverride Optional token for testing.
|
||||
* @returns A promise that resolves to the parsed job status object.
|
||||
* @throws An error if the network request fails or if the response is not valid JSON.
|
||||
* @throws A `JobFailedError` if the job has failed, or a generic `Error` for other issues.
|
||||
*/
|
||||
export const getJobStatus = async (
|
||||
jobId: string,
|
||||
tokenOverride?: string,
|
||||
): Promise<JobStatus> => {
|
||||
const response = await apiFetch(`/ai/jobs/${jobId}/status`, {}, { tokenOverride });
|
||||
const response = await authedGet(`/ai/jobs/${jobId}/status`, { tokenOverride });
|
||||
|
||||
// Handle non-OK responses first, as they might not have a JSON body.
|
||||
if (!response.ok) {
|
||||
let errorText = `API Error: ${response.status} ${response.statusText}`;
|
||||
let errorMessage = `API Error: ${response.status} ${response.statusText}`;
|
||||
try {
|
||||
const errorBody = await response.text();
|
||||
if (errorBody) errorText = `API Error ${response.status}: ${errorBody}`;
|
||||
// Try to get a more specific message from the body.
|
||||
const errorData = await response.json();
|
||||
if (errorData.message) {
|
||||
errorMessage = errorData.message;
|
||||
}
|
||||
} catch (e) {
|
||||
// ignore if reading body fails
|
||||
// The body was not JSON, which is fine for a server error page.
|
||||
// The default message is sufficient.
|
||||
logger.warn('getJobStatus received a non-JSON error response.', { status: response.status });
|
||||
}
|
||||
throw new Error(errorText);
|
||||
throw new Error(errorMessage);
|
||||
}
|
||||
|
||||
// If we get here, the response is OK (2xx). Now parse the body.
|
||||
try {
|
||||
return await response.json();
|
||||
const statusData: JobStatus = await response.json();
|
||||
|
||||
// If the job itself has failed, we should treat this as an error condition
|
||||
// for the polling logic by rejecting the promise. This will stop the polling loop.
|
||||
if (statusData.state === 'failed') {
|
||||
// The structured error payload is in the 'progress' object.
|
||||
const progress = statusData.progress;
|
||||
const userMessage =
|
||||
progress?.message || statusData.failedReason || 'Job failed with an unknown error.';
|
||||
const errorCode = progress?.errorCode || 'UNKNOWN_ERROR';
|
||||
|
||||
logger.error(`Job ${jobId} failed with code: ${errorCode}, message: ${userMessage}`);
|
||||
|
||||
// Throw a custom, structured error so the frontend can react to the errorCode.
|
||||
throw new JobFailedError(userMessage, errorCode);
|
||||
}
|
||||
|
||||
return statusData;
|
||||
} catch (error) {
|
||||
const rawText = await response.text();
|
||||
throw new Error(`Failed to parse JSON response from server. Body: ${rawText}`);
|
||||
// If it's the specific error we threw, just re-throw it.
|
||||
if (error instanceof JobFailedError) {
|
||||
throw error;
|
||||
}
|
||||
// This now primarily catches JSON parsing errors on an OK response, which is unexpected.
|
||||
logger.error('getJobStatus failed to parse a successful API response.', { error });
|
||||
throw new Error('Failed to parse job status from a successful API response.');
|
||||
}
|
||||
};
|
||||
|
||||
@@ -113,14 +153,7 @@ export const isImageAFlyer = (
|
||||
|
||||
// Use apiFetchWithAuth for FormData to let the browser set the correct Content-Type.
|
||||
// The URL must be relative, as the helper constructs the full path.
|
||||
return apiFetch(
|
||||
'/ai/check-flyer',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/check-flyer', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
export const extractAddressFromImage = (
|
||||
@@ -130,14 +163,7 @@ export const extractAddressFromImage = (
|
||||
const formData = new FormData();
|
||||
formData.append('image', imageFile);
|
||||
|
||||
return apiFetch(
|
||||
'/ai/extract-address',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/extract-address', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
export const extractLogoFromImage = (
|
||||
@@ -149,14 +175,7 @@ export const extractLogoFromImage = (
|
||||
formData.append('images', file);
|
||||
});
|
||||
|
||||
return apiFetch(
|
||||
'/ai/extract-logo',
|
||||
{
|
||||
method: 'POST',
|
||||
body: formData,
|
||||
},
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/extract-logo', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
export const getQuickInsights = (
|
||||
@@ -164,16 +183,7 @@ export const getQuickInsights = (
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
return apiFetch(
|
||||
'/ai/quick-insights',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/quick-insights', { items }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
export const getDeepDiveAnalysis = (
|
||||
@@ -181,16 +191,7 @@ export const getDeepDiveAnalysis = (
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
return apiFetch(
|
||||
'/ai/deep-dive',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/deep-dive', { items }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
export const searchWeb = (
|
||||
@@ -198,16 +199,7 @@ export const searchWeb = (
|
||||
signal?: AbortSignal,
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
return apiFetch(
|
||||
'/ai/search-web',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ query }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/search-web', { query }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
// ============================================================================
|
||||
@@ -222,15 +214,7 @@ export const planTripWithMaps = async (
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
logger.debug('Stub: planTripWithMaps called with location:', { userLocation });
|
||||
return apiFetch(
|
||||
'/ai/plan-trip',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items, store, userLocation }),
|
||||
},
|
||||
{ signal, tokenOverride },
|
||||
);
|
||||
return authedPost('/ai/plan-trip', { items, store, userLocation }, { signal, tokenOverride });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -244,16 +228,7 @@ export const generateImageFromText = (
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
logger.debug('Stub: generateImageFromText called with prompt:', { prompt });
|
||||
return apiFetch(
|
||||
'/ai/generate-image',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ prompt }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/generate-image', { prompt }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -267,16 +242,7 @@ export const generateSpeechFromText = (
|
||||
tokenOverride?: string,
|
||||
): Promise<Response> => {
|
||||
logger.debug('Stub: generateSpeechFromText called with text:', { text });
|
||||
return apiFetch(
|
||||
'/ai/generate-speech',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ text }),
|
||||
signal,
|
||||
},
|
||||
{ tokenOverride, signal },
|
||||
);
|
||||
return authedPost('/ai/generate-speech', { text }, { tokenOverride, signal });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -329,11 +295,7 @@ export const rescanImageArea = (
|
||||
formData.append('cropArea', JSON.stringify(cropArea));
|
||||
formData.append('extractionType', extractionType);
|
||||
|
||||
return apiFetch(
|
||||
'/ai/rescan-area',
|
||||
{ method: 'POST', body: formData },
|
||||
{ tokenOverride },
|
||||
);
|
||||
return authedPostForm('/ai/rescan-area', formData, { tokenOverride });
|
||||
};
|
||||
|
||||
/**
|
||||
@@ -347,12 +309,5 @@ export const compareWatchedItemPrices = (
|
||||
): Promise<Response> => {
|
||||
// Use the apiFetch wrapper for consistency with other API calls in this file.
|
||||
// This centralizes token handling and base URL logic.
|
||||
return apiFetch(
|
||||
'/ai/compare-prices',
|
||||
{
|
||||
method: 'POST',
|
||||
headers: { 'Content-Type': 'application/json' },
|
||||
body: JSON.stringify({ items: watchedItems }),
|
||||
},
|
||||
{ signal },
|
||||
)};
|
||||
return authedPost('/ai/compare-prices', { items: watchedItems }, { signal });
|
||||
};
|
||||
|
||||
@@ -1,11 +1,17 @@
|
||||
// src/services/aiService.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, afterEach } from 'vitest';
|
||||
import { describe, it, expect, vi, beforeEach, afterEach, type Mock } from 'vitest';
|
||||
import { createMockLogger } from '../tests/utils/mockLogger';
|
||||
import type { Logger } from 'pino';
|
||||
import type { MasterGroceryItem } from '../types';
|
||||
import type { FlyerStatus, MasterGroceryItem, UserProfile } from '../types';
|
||||
// Import the class, not the singleton instance, so we can instantiate it with mocks.
|
||||
import { AIService } from './aiService.server';
|
||||
import {
|
||||
AIService,
|
||||
AiFlyerDataSchema,
|
||||
aiService as aiServiceSingleton,
|
||||
DuplicateFlyerError,
|
||||
} from './aiService.server';
|
||||
import { createMockMasterGroceryItem } from '../tests/utils/mockFactories';
|
||||
import { ValidationError } from './db/errors.db';
|
||||
|
||||
// Mock the logger to prevent the real pino instance from being created, which causes issues with 'pino-pretty' in tests.
|
||||
vi.mock('./logger.server', () => ({
|
||||
@@ -45,6 +51,55 @@ vi.mock('@google/genai', () => {
|
||||
};
|
||||
});
|
||||
|
||||
// --- New Mocks for Database and Queue ---
|
||||
vi.mock('./db/index.db', () => ({
|
||||
flyerRepo: {
|
||||
findFlyerByChecksum: vi.fn(),
|
||||
},
|
||||
adminRepo: {
|
||||
logActivity: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./queueService.server', () => ({
|
||||
flyerQueue: {
|
||||
add: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
vi.mock('./db/flyer.db', () => ({
|
||||
createFlyerAndItems: vi.fn(),
|
||||
}));
|
||||
|
||||
vi.mock('../utils/imageProcessor', () => ({
|
||||
generateFlyerIcon: vi.fn(),
|
||||
}));
|
||||
|
||||
// Import mocked modules to assert on them
|
||||
import * as dbModule from './db/index.db';
|
||||
import { flyerQueue } from './queueService.server';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
|
||||
// Define a mock interface that closely resembles the actual Flyer type for testing purposes.
|
||||
// This helps ensure type safety in mocks without relying on 'any'.
|
||||
interface MockFlyer {
|
||||
flyer_id: number;
|
||||
file_name: string;
|
||||
image_url: string;
|
||||
icon_url: string;
|
||||
checksum: string;
|
||||
store_name: string;
|
||||
valid_from: string | null;
|
||||
valid_to: string | null;
|
||||
store_address: string | null;
|
||||
item_count: number;
|
||||
status: FlyerStatus;
|
||||
uploaded_by: string | null | undefined;
|
||||
created_at: string;
|
||||
updated_at: string;
|
||||
}
|
||||
|
||||
describe('AI Service (Server)', () => {
|
||||
// Create mock dependencies that will be injected into the service
|
||||
const mockAiClient = { generateContent: vi.fn() };
|
||||
@@ -65,6 +120,25 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('AiFlyerDataSchema', () => {
|
||||
it('should fail validation if store_name is null or empty, covering requiredString', () => {
|
||||
const dataWithNull = { store_name: null, items: [] };
|
||||
const dataWithEmpty = { store_name: '', items: [] };
|
||||
const resultNull = AiFlyerDataSchema.safeParse(dataWithNull);
|
||||
const resultEmpty = AiFlyerDataSchema.safeParse(dataWithEmpty);
|
||||
|
||||
expect(resultNull.success).toBe(false);
|
||||
if (!resultNull.success) {
|
||||
expect(resultNull.error.issues[0].message).toBe('Store name cannot be empty');
|
||||
}
|
||||
|
||||
expect(resultEmpty.success).toBe(false);
|
||||
if (!resultEmpty.success) {
|
||||
expect(resultEmpty.error.issues[0].message).toBe('Store name cannot be empty');
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
describe('Constructor', () => {
|
||||
const originalEnv = process.env;
|
||||
|
||||
@@ -148,7 +222,7 @@ describe('AI Service (Server)', () => {
|
||||
await adapter.generateContent(request);
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledWith({
|
||||
model: 'gemini-2.5-flash',
|
||||
model: 'gemini-3-flash-preview',
|
||||
...request,
|
||||
});
|
||||
});
|
||||
@@ -202,21 +276,22 @@ describe('AI Service (Server)', () => {
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(2);
|
||||
|
||||
// Check first call
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, {
|
||||
model: 'gemini-2.5-flash',
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, { // The first model in the list is now 'gemini-3-flash-preview'
|
||||
model: 'gemini-3-flash-preview',
|
||||
...request,
|
||||
});
|
||||
|
||||
// Check second call
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, {
|
||||
model: 'gemini-3-flash',
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, { // The second model in the list is 'gemini-2.5-flash'
|
||||
model: 'gemini-2.5-flash',
|
||||
...request,
|
||||
});
|
||||
|
||||
// Check that a warning was logged
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
// The warning should be for the model that failed ('gemini-3-flash-preview'), not the next one.
|
||||
expect.stringContaining(
|
||||
"Model 'gemini-2.5-flash' failed due to quota/rate limit. Trying next model.",
|
||||
"Model 'gemini-3-flash-preview' failed due to quota/rate limit. Trying next model.",
|
||||
),
|
||||
);
|
||||
});
|
||||
@@ -239,8 +314,8 @@ describe('AI Service (Server)', () => {
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(1);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ error: nonRetriableError },
|
||||
`[AIService Adapter] Model 'gemini-2.5-flash' failed with a non-retriable error.`,
|
||||
{ error: nonRetriableError }, // The first model in the list is now 'gemini-3-flash-preview'
|
||||
`[AIService Adapter] Model 'gemini-3-flash-preview' failed with a non-retriable error.`,
|
||||
);
|
||||
});
|
||||
|
||||
@@ -267,15 +342,15 @@ describe('AI Service (Server)', () => {
|
||||
);
|
||||
|
||||
expect(mockGenerateContent).toHaveBeenCalledTimes(3);
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, {
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(1, { // The first model in the list is now 'gemini-3-flash-preview'
|
||||
model: 'gemini-3-flash-preview',
|
||||
...request,
|
||||
});
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, { // The second model in the list is 'gemini-2.5-flash'
|
||||
model: 'gemini-2.5-flash',
|
||||
...request,
|
||||
});
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(2, {
|
||||
model: 'gemini-3-flash',
|
||||
...request,
|
||||
});
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(3, {
|
||||
expect(mockGenerateContent).toHaveBeenNthCalledWith(3, { // The third model in the list is 'gemini-2.5-flash-lite'
|
||||
model: 'gemini-2.5-flash-lite',
|
||||
...request,
|
||||
});
|
||||
@@ -577,40 +652,6 @@ describe('AI Service (Server)', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('_normalizeExtractedItems (private method)', () => {
|
||||
it('should replace null or undefined fields with default values', () => {
|
||||
const rawItems: {
|
||||
item: string;
|
||||
price_display: null;
|
||||
quantity: undefined;
|
||||
category_name: null;
|
||||
master_item_id: null;
|
||||
}[] = [
|
||||
{
|
||||
item: 'Test',
|
||||
price_display: null,
|
||||
quantity: undefined,
|
||||
category_name: null,
|
||||
master_item_id: null,
|
||||
},
|
||||
];
|
||||
const [normalized] = (
|
||||
aiServiceInstance as unknown as {
|
||||
_normalizeExtractedItems: (items: typeof rawItems) => {
|
||||
price_display: string;
|
||||
quantity: string;
|
||||
category_name: string;
|
||||
master_item_id: undefined;
|
||||
}[];
|
||||
}
|
||||
)._normalizeExtractedItems(rawItems);
|
||||
expect(normalized.price_display).toBe('');
|
||||
expect(normalized.quantity).toBe('');
|
||||
expect(normalized.category_name).toBe('Other/Miscellaneous');
|
||||
expect(normalized.master_item_id).toBeUndefined();
|
||||
});
|
||||
});
|
||||
|
||||
describe('extractTextFromImageArea', () => {
|
||||
it('should call sharp to crop the image and call the AI with the correct prompt', async () => {
|
||||
console.log("TEST START: 'should call sharp to crop...'");
|
||||
@@ -706,4 +747,315 @@ describe('AI Service (Server)', () => {
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('planTripWithMaps', () => {
|
||||
const mockUserLocation: GeolocationCoordinates = {
|
||||
latitude: 45,
|
||||
longitude: -75,
|
||||
accuracy: 10,
|
||||
altitude: null,
|
||||
altitudeAccuracy: null,
|
||||
heading: null,
|
||||
speed: null,
|
||||
toJSON: () => ({}),
|
||||
};
|
||||
const mockStore = { name: 'Test Store' };
|
||||
|
||||
it('should throw a "feature disabled" error', async () => {
|
||||
// This test verifies the current implementation which has the feature disabled.
|
||||
await expect(
|
||||
aiServiceInstance.planTripWithMaps([], mockStore, mockUserLocation, mockLoggerInstance),
|
||||
).rejects.toThrow("The 'planTripWithMaps' feature is currently disabled due to API costs.");
|
||||
|
||||
// Also verify that the warning is logged
|
||||
expect(mockLoggerInstance.warn).toHaveBeenCalledWith(
|
||||
'[AIService] planTripWithMaps called, but feature is disabled. Throwing error.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('enqueueFlyerProcessing', () => {
|
||||
const mockFile = {
|
||||
path: '/tmp/test.pdf',
|
||||
originalname: 'test.pdf',
|
||||
} as Express.Multer.File;
|
||||
const mockProfile = {
|
||||
user: { user_id: 'user123' },
|
||||
address: {
|
||||
address_line_1: '123 St',
|
||||
city: 'City',
|
||||
country: 'Country', // This was a duplicate, fixed.
|
||||
},
|
||||
} as UserProfile;
|
||||
|
||||
it('should throw DuplicateFlyerError if flyer already exists', async () => {
|
||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue({ flyer_id: 99 } as any);
|
||||
|
||||
await expect(
|
||||
aiServiceInstance.enqueueFlyerProcessing(
|
||||
mockFile,
|
||||
'checksum123',
|
||||
mockProfile,
|
||||
'127.0.0.1',
|
||||
mockLoggerInstance,
|
||||
),
|
||||
).rejects.toThrow(DuplicateFlyerError);
|
||||
});
|
||||
|
||||
it('should enqueue job with user address if profile exists', async () => {
|
||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job123' } as any);
|
||||
|
||||
const result = await aiServiceInstance.enqueueFlyerProcessing(
|
||||
mockFile,
|
||||
'checksum123',
|
||||
mockProfile,
|
||||
'127.0.0.1',
|
||||
mockLoggerInstance,
|
||||
);
|
||||
|
||||
expect(flyerQueue.add).toHaveBeenCalledWith('process-flyer', {
|
||||
filePath: mockFile.path,
|
||||
originalFileName: mockFile.originalname,
|
||||
checksum: 'checksum123',
|
||||
userId: 'user123',
|
||||
submitterIp: '127.0.0.1',
|
||||
userProfileAddress: '123 St, City, Country', // Partial address match based on filter(Boolean)
|
||||
});
|
||||
expect(result.id).toBe('job123');
|
||||
});
|
||||
|
||||
it('should enqueue job without address if profile is missing', async () => {
|
||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(flyerQueue.add).mockResolvedValue({ id: 'job456' } as any);
|
||||
|
||||
await aiServiceInstance.enqueueFlyerProcessing(
|
||||
mockFile,
|
||||
'checksum123',
|
||||
undefined, // No profile
|
||||
'127.0.0.1',
|
||||
mockLoggerInstance,
|
||||
);
|
||||
|
||||
expect(flyerQueue.add).toHaveBeenCalledWith(
|
||||
'process-flyer',
|
||||
expect.objectContaining({
|
||||
userId: undefined,
|
||||
userProfileAddress: undefined,
|
||||
}),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('processLegacyFlyerUpload', () => {
|
||||
const mockFile = {
|
||||
path: '/tmp/upload.jpg',
|
||||
filename: 'upload.jpg',
|
||||
originalname: 'orig.jpg',
|
||||
} as Express.Multer.File; // This was a duplicate, fixed.
|
||||
const mockProfile = { user: { user_id: 'u1' } } as UserProfile;
|
||||
|
||||
beforeEach(() => {
|
||||
// Default success mocks
|
||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue(undefined);
|
||||
vi.mocked(generateFlyerIcon).mockResolvedValue('icon.jpg');
|
||||
vi.mocked(createFlyerAndItems).mockResolvedValue({
|
||||
flyer: {
|
||||
flyer_id: 100,
|
||||
file_name: 'orig.jpg',
|
||||
image_url: '/flyer-images/upload.jpg',
|
||||
icon_url: '/flyer-images/icons/icon.jpg',
|
||||
checksum: 'mock-checksum-123',
|
||||
store_name: 'Mock Store',
|
||||
valid_from: null,
|
||||
valid_to: null,
|
||||
store_address: null,
|
||||
item_count: 0,
|
||||
status: 'processed',
|
||||
uploaded_by: 'u1',
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
} as MockFlyer, // Use the more specific MockFlyer type
|
||||
items: [],
|
||||
});
|
||||
});
|
||||
|
||||
it('should throw ValidationError if checksum is missing', async () => {
|
||||
const body = { data: JSON.stringify({}) }; // No checksum
|
||||
await expect(
|
||||
aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
),
|
||||
).rejects.toThrow(ValidationError);
|
||||
});
|
||||
|
||||
it('should throw DuplicateFlyerError if checksum exists', async () => {
|
||||
vi.mocked(dbModule.flyerRepo.findFlyerByChecksum).mockResolvedValue({ flyer_id: 55 } as any);
|
||||
const body = { checksum: 'dup-sum' };
|
||||
|
||||
await expect(
|
||||
aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
),
|
||||
).rejects.toThrow(DuplicateFlyerError);
|
||||
});
|
||||
|
||||
it('should parse "data" string property containing extractedData', async () => {
|
||||
const payload = {
|
||||
checksum: 'abc',
|
||||
originalFileName: 'test.jpg',
|
||||
extractedData: {
|
||||
store_name: 'My Store',
|
||||
items: [{ item: 'Milk', price_in_cents: 200 }],
|
||||
},
|
||||
};
|
||||
const body = { data: JSON.stringify(payload) };
|
||||
|
||||
await aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
);
|
||||
|
||||
expect(createFlyerAndItems).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
store_name: 'My Store',
|
||||
checksum: 'abc',
|
||||
}),
|
||||
expect.arrayContaining([expect.objectContaining({ item: 'Milk' })]),
|
||||
mockLoggerInstance,
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle direct object body with extractedData', async () => {
|
||||
const body = {
|
||||
checksum: 'xyz',
|
||||
extractedData: {
|
||||
store_name: 'Direct Store',
|
||||
valid_from: '2023-01-01',
|
||||
},
|
||||
};
|
||||
|
||||
await aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
);
|
||||
|
||||
expect(createFlyerAndItems).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
store_name: 'Direct Store',
|
||||
valid_from: '2023-01-01',
|
||||
}),
|
||||
[], // No items
|
||||
mockLoggerInstance,
|
||||
);
|
||||
});
|
||||
|
||||
it('should fallback for missing store name and normalize items', async () => {
|
||||
const body = {
|
||||
checksum: 'fallback',
|
||||
extractedData: {
|
||||
// store_name missing
|
||||
items: [{ item: 'Bread' }], // minimal item
|
||||
},
|
||||
};
|
||||
|
||||
await aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
);
|
||||
|
||||
expect(createFlyerAndItems).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
store_name: 'Unknown Store (auto)',
|
||||
}),
|
||||
expect.arrayContaining([
|
||||
expect.objectContaining({
|
||||
item: 'Bread',
|
||||
quantity: 1, // Default
|
||||
view_count: 0,
|
||||
}),
|
||||
]),
|
||||
mockLoggerInstance,
|
||||
);
|
||||
expect(mockLoggerInstance.warn).toHaveBeenCalledWith(
|
||||
expect.stringContaining('extractedData.store_name missing'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should log activity and return the new flyer', async () => {
|
||||
const body = { checksum: 'act', extractedData: { store_name: 'Act Store' } };
|
||||
const result = await aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
);
|
||||
|
||||
expect(result).toHaveProperty('flyer_id', 100);
|
||||
expect(dbModule.adminRepo.logActivity).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
action: 'flyer_processed',
|
||||
userId: 'u1',
|
||||
}),
|
||||
mockLoggerInstance,
|
||||
);
|
||||
});
|
||||
|
||||
it('should catch JSON parsing errors in _parseLegacyPayload and log warning (errMsg coverage)', async () => {
|
||||
// Sending a body where 'data' is a malformed JSON string to trigger the catch block in _parseLegacyPayload
|
||||
const body = { data: '{ "malformed": json ' };
|
||||
|
||||
// This will eventually throw ValidationError because checksum won't be found
|
||||
await expect(
|
||||
aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
),
|
||||
).rejects.toThrow(ValidationError);
|
||||
|
||||
// Verify that the error was caught and logged using errMsg logic
|
||||
expect(mockLoggerInstance.warn).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ error: expect.any(String) }),
|
||||
'[AIService] Failed to parse nested "data" property string.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle body as a string', async () => {
|
||||
const payload = { checksum: 'str-body', extractedData: { store_name: 'String Body' } };
|
||||
const body = JSON.stringify(payload);
|
||||
|
||||
await aiServiceInstance.processLegacyFlyerUpload(
|
||||
mockFile,
|
||||
body,
|
||||
mockProfile,
|
||||
mockLoggerInstance,
|
||||
);
|
||||
|
||||
expect(createFlyerAndItems).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ checksum: 'str-body' }),
|
||||
expect.anything(),
|
||||
mockLoggerInstance,
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Singleton Export', () => {
|
||||
it('should export a singleton instance of AIService', () => {
|
||||
expect(aiServiceSingleton).toBeInstanceOf(AIService);
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -10,7 +10,23 @@ import fsPromises from 'node:fs/promises';
|
||||
import type { Logger } from 'pino';
|
||||
import { z } from 'zod';
|
||||
import { pRateLimit } from 'p-ratelimit';
|
||||
import type { FlyerItem, MasterGroceryItem, ExtractedFlyerItem } from '../types';
|
||||
import type {
|
||||
FlyerItem,
|
||||
MasterGroceryItem,
|
||||
ExtractedFlyerItem,
|
||||
UserProfile,
|
||||
ExtractedCoreData,
|
||||
FlyerInsert,
|
||||
Flyer,
|
||||
} from '../types';
|
||||
import { FlyerProcessingError } from './processingErrors';
|
||||
import * as db from './db/index.db';
|
||||
import { flyerQueue } from './queueService.server';
|
||||
import type { Job } from 'bullmq';
|
||||
import { createFlyerAndItems } from './db/flyer.db';
|
||||
import { generateFlyerIcon } from '../utils/imageProcessor';
|
||||
import path from 'path';
|
||||
import { ValidationError } from './db/errors.db';
|
||||
|
||||
// Helper for consistent required string validation (handles missing/null/empty)
|
||||
const requiredString = (message: string) =>
|
||||
@@ -34,6 +50,21 @@ export const AiFlyerDataSchema = z.object({
|
||||
items: z.array(ExtractedFlyerItemSchema),
|
||||
});
|
||||
|
||||
interface FlyerProcessPayload extends Partial<ExtractedCoreData> {
|
||||
checksum?: string;
|
||||
originalFileName?: string;
|
||||
extractedData?: Partial<ExtractedCoreData>;
|
||||
data?: FlyerProcessPayload; // For nested data structures
|
||||
}
|
||||
|
||||
// Helper to safely extract an error message from unknown `catch` values.
|
||||
const errMsg = (e: unknown) => {
|
||||
if (e instanceof Error) return e.message;
|
||||
if (typeof e === 'object' && e !== null && 'message' in e)
|
||||
return String((e as { message: unknown }).message);
|
||||
return String(e || 'An unknown error occurred.');
|
||||
};
|
||||
|
||||
/**
|
||||
* Defines the contract for a file system utility. This interface allows for
|
||||
* dependency injection, making the AIService testable without hitting the real file system.
|
||||
@@ -67,12 +98,21 @@ type RawFlyerItem = {
|
||||
master_item_id?: number | null | undefined;
|
||||
};
|
||||
|
||||
export class DuplicateFlyerError extends FlyerProcessingError {
|
||||
constructor(message: string, public flyerId: number) {
|
||||
super(message, 'DUPLICATE_FLYER', message);
|
||||
}
|
||||
}
|
||||
|
||||
export class AIService {
|
||||
private aiClient: IAiClient;
|
||||
private fs: IFileSystem;
|
||||
private rateLimiter: <T>(fn: () => Promise<T>) => Promise<T>;
|
||||
private logger: Logger;
|
||||
private readonly models = ['gemini-2.5-flash', 'gemini-3-flash', 'gemini-2.5-flash-lite'];
|
||||
// The fallback list is ordered by preference (speed/cost vs. power).
|
||||
// We try the fastest models first, then the more powerful 'pro' model as a high-quality fallback,
|
||||
// and finally the 'lite' model as a last resort.
|
||||
private readonly models = [ 'gemini-3-flash-preview', 'gemini-2.5-flash', 'gemini-2.5-flash-lite'];
|
||||
|
||||
constructor(logger: Logger, aiClient?: IAiClient, fs?: IFileSystem) {
|
||||
this.logger = logger;
|
||||
@@ -193,7 +233,8 @@ export class AIService {
|
||||
errorMessage.includes('quota') ||
|
||||
errorMessage.includes('429') || // HTTP 429 Too Many Requests
|
||||
errorMessage.includes('resource_exhausted') || // Make case-insensitive
|
||||
errorMessage.includes('model is overloaded')
|
||||
errorMessage.includes('model is overloaded') ||
|
||||
errorMessage.includes('not found') // Also retry if model is not found (e.g., regional availability or API version issue)
|
||||
) {
|
||||
this.logger.warn(
|
||||
`[AIService Adapter] Model '${modelName}' failed due to quota/rate limit. Trying next model. Error: ${errorMessage}`,
|
||||
@@ -551,6 +592,11 @@ export class AIService {
|
||||
private _normalizeExtractedItems(items: RawFlyerItem[]): ExtractedFlyerItem[] {
|
||||
return items.map((item: RawFlyerItem) => ({
|
||||
...item,
|
||||
// Ensure 'item' is always a string, defaulting to 'Unknown Item' if null/undefined.
|
||||
item:
|
||||
item.item === null || item.item === undefined || String(item.item).trim() === ''
|
||||
? 'Unknown Item'
|
||||
: String(item.item),
|
||||
price_display:
|
||||
item.price_display === null || item.price_display === undefined
|
||||
? ''
|
||||
@@ -685,6 +731,168 @@ export class AIService {
|
||||
}
|
||||
*/
|
||||
}
|
||||
|
||||
|
||||
async enqueueFlyerProcessing(
|
||||
file: Express.Multer.File,
|
||||
checksum: string,
|
||||
userProfile: UserProfile | undefined,
|
||||
submitterIp: string,
|
||||
logger: Logger,
|
||||
): Promise<Job> {
|
||||
// 1. Check for duplicate flyer
|
||||
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, logger);
|
||||
if (existingFlyer) {
|
||||
// Throw a specific error for the route to handle
|
||||
throw new DuplicateFlyerError(
|
||||
'This flyer has already been processed.',
|
||||
existingFlyer.flyer_id,
|
||||
);
|
||||
}
|
||||
|
||||
// 2. Construct user address string
|
||||
let userProfileAddress: string | undefined = undefined;
|
||||
if (userProfile?.address) {
|
||||
userProfileAddress = [
|
||||
userProfile.address.address_line_1,
|
||||
userProfile.address.address_line_2,
|
||||
userProfile.address.city,
|
||||
userProfile.address.province_state,
|
||||
userProfile.address.postal_code,
|
||||
userProfile.address.country,
|
||||
]
|
||||
.filter(Boolean)
|
||||
.join(', ');
|
||||
}
|
||||
|
||||
// 3. Add job to the queue
|
||||
const job = await flyerQueue.add('process-flyer', {
|
||||
filePath: file.path,
|
||||
originalFileName: file.originalname,
|
||||
checksum: checksum,
|
||||
userId: userProfile?.user.user_id,
|
||||
submitterIp: submitterIp,
|
||||
userProfileAddress: userProfileAddress,
|
||||
});
|
||||
|
||||
logger.info(
|
||||
`Enqueued flyer for processing. File: ${file.originalname}, Job ID: ${job.id}`,
|
||||
);
|
||||
|
||||
return job;
|
||||
}
|
||||
|
||||
private _parseLegacyPayload(
|
||||
body: any,
|
||||
logger: Logger,
|
||||
): { parsed: FlyerProcessPayload; extractedData: Partial<ExtractedCoreData> | null | undefined } {
|
||||
let parsed: FlyerProcessPayload = {};
|
||||
|
||||
try {
|
||||
parsed = typeof body === 'string' ? JSON.parse(body) : body || {};
|
||||
} catch (e) {
|
||||
logger.warn({ error: errMsg(e) }, '[AIService] Failed to parse top-level request body string.');
|
||||
return { parsed: {}, extractedData: {} };
|
||||
}
|
||||
|
||||
// If the real payload is nested inside a 'data' property (which could be a string),
|
||||
// we parse it out but keep the original `parsed` object for top-level properties like checksum.
|
||||
let potentialPayload: FlyerProcessPayload = parsed;
|
||||
if (parsed.data) {
|
||||
if (typeof parsed.data === 'string') {
|
||||
try {
|
||||
potentialPayload = JSON.parse(parsed.data);
|
||||
} catch (e) {
|
||||
logger.warn({ error: errMsg(e) }, '[AIService] Failed to parse nested "data" property string.');
|
||||
}
|
||||
} else if (typeof parsed.data === 'object') {
|
||||
potentialPayload = parsed.data;
|
||||
}
|
||||
}
|
||||
|
||||
// The extracted data is either in an `extractedData` key or is the payload itself.
|
||||
const extractedData = potentialPayload.extractedData ?? potentialPayload;
|
||||
|
||||
// Merge for checksum lookup: properties in the outer `parsed` object (like a top-level checksum)
|
||||
// take precedence over any same-named properties inside `potentialPayload`.
|
||||
const finalParsed = { ...potentialPayload, ...parsed };
|
||||
|
||||
return { parsed: finalParsed, extractedData };
|
||||
}
|
||||
|
||||
async processLegacyFlyerUpload(
|
||||
file: Express.Multer.File,
|
||||
body: any,
|
||||
userProfile: UserProfile | undefined,
|
||||
logger: Logger,
|
||||
): Promise<Flyer> {
|
||||
const { parsed, extractedData: initialExtractedData } = this._parseLegacyPayload(body, logger);
|
||||
let extractedData = initialExtractedData;
|
||||
|
||||
const checksum = parsed.checksum ?? parsed?.data?.checksum ?? '';
|
||||
if (!checksum) {
|
||||
throw new ValidationError([], 'Checksum is required.');
|
||||
}
|
||||
|
||||
const existingFlyer = await db.flyerRepo.findFlyerByChecksum(checksum, logger);
|
||||
if (existingFlyer) {
|
||||
throw new DuplicateFlyerError('This flyer has already been processed.', existingFlyer.flyer_id);
|
||||
}
|
||||
|
||||
const originalFileName = parsed.originalFileName ?? parsed?.data?.originalFileName ?? file.originalname;
|
||||
|
||||
if (!extractedData || typeof extractedData !== 'object') {
|
||||
logger.warn({ bodyData: parsed }, 'Missing extractedData in legacy payload.');
|
||||
extractedData = {};
|
||||
}
|
||||
|
||||
const rawItems = extractedData.items ?? [];
|
||||
const itemsArray = Array.isArray(rawItems) ? rawItems : typeof rawItems === 'string' ? JSON.parse(rawItems) : [];
|
||||
const itemsForDb = itemsArray.map((item: Partial<ExtractedFlyerItem>) => ({
|
||||
...item,
|
||||
master_item_id: item.master_item_id === null ? undefined : item.master_item_id,
|
||||
quantity: item.quantity ?? 1,
|
||||
view_count: 0,
|
||||
click_count: 0,
|
||||
updated_at: new Date().toISOString(),
|
||||
}));
|
||||
|
||||
const storeName = extractedData.store_name && String(extractedData.store_name).trim().length > 0 ? String(extractedData.store_name) : 'Unknown Store (auto)';
|
||||
if (storeName.startsWith('Unknown')) {
|
||||
logger.warn('extractedData.store_name missing; using fallback store name.');
|
||||
}
|
||||
|
||||
const iconsDir = path.join(path.dirname(file.path), 'icons');
|
||||
const iconFileName = await generateFlyerIcon(file.path, iconsDir, logger);
|
||||
const iconUrl = `/flyer-images/icons/${iconFileName}`;
|
||||
|
||||
const flyerData: FlyerInsert = {
|
||||
file_name: originalFileName,
|
||||
image_url: `/flyer-images/${file.filename}`,
|
||||
icon_url: iconUrl,
|
||||
checksum: checksum,
|
||||
store_name: storeName,
|
||||
valid_from: extractedData.valid_from ?? null,
|
||||
valid_to: extractedData.valid_to ?? null,
|
||||
store_address: extractedData.store_address ?? null,
|
||||
item_count: 0,
|
||||
status: 'needs_review',
|
||||
uploaded_by: userProfile?.user.user_id,
|
||||
};
|
||||
|
||||
const { flyer: newFlyer, items: newItems } = await createFlyerAndItems(flyerData, itemsForDb, logger);
|
||||
|
||||
logger.info(`Successfully processed legacy flyer: ${newFlyer.file_name} (ID: ${newFlyer.flyer_id}) with ${newItems.length} items.`);
|
||||
|
||||
await db.adminRepo.logActivity({
|
||||
userId: userProfile?.user.user_id,
|
||||
action: 'flyer_processed',
|
||||
displayText: `Processed a new flyer for ${flyerData.store_name}.`,
|
||||
details: { flyerId: newFlyer.flyer_id, storeName: flyerData.store_name },
|
||||
}, logger);
|
||||
|
||||
return newFlyer;
|
||||
}
|
||||
}
|
||||
|
||||
// Export a singleton instance of the service for use throughout the application.
|
||||
|
||||
153
src/services/analyticsService.server.test.ts
Normal file
153
src/services/analyticsService.server.test.ts
Normal file
@@ -0,0 +1,153 @@
|
||||
// src/services/analyticsService.server.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, afterEach } from 'vitest';
|
||||
import { AnalyticsService } from './analyticsService.server';
|
||||
import { logger } from './logger.server';
|
||||
import type { Job } from 'bullmq';
|
||||
import type { AnalyticsJobData, WeeklyAnalyticsJobData } from '../types/job-data';
|
||||
|
||||
// Mock logger
|
||||
vi.mock('./logger.server', () => ({
|
||||
logger: {
|
||||
child: vi.fn(),
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
describe('AnalyticsService', () => {
|
||||
let service: AnalyticsService;
|
||||
let mockLoggerInstance: any;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
vi.useFakeTimers();
|
||||
|
||||
// Setup mock logger instance returned by child()
|
||||
mockLoggerInstance = {
|
||||
info: vi.fn(),
|
||||
error: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
};
|
||||
vi.mocked(logger.child).mockReturnValue(mockLoggerInstance);
|
||||
|
||||
service = new AnalyticsService();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
vi.useRealTimers();
|
||||
});
|
||||
|
||||
const createMockJob = <T>(data: T): Job<T> =>
|
||||
({
|
||||
id: 'job-123',
|
||||
name: 'analytics-job',
|
||||
data,
|
||||
attemptsMade: 1,
|
||||
updateProgress: vi.fn(),
|
||||
} as unknown as Job<T>);
|
||||
|
||||
describe('processDailyReportJob', () => {
|
||||
it('should process successfully', async () => {
|
||||
const job = createMockJob<AnalyticsJobData>({ reportDate: '2023-10-27' } as AnalyticsJobData);
|
||||
|
||||
const promise = service.processDailyReportJob(job);
|
||||
|
||||
// Fast-forward time to bypass the 10s delay
|
||||
await vi.advanceTimersByTimeAsync(10000);
|
||||
|
||||
const result = await promise;
|
||||
|
||||
expect(result).toEqual({ status: 'success', reportDate: '2023-10-27' });
|
||||
expect(logger.child).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
jobId: 'job-123',
|
||||
reportDate: '2023-10-27',
|
||||
}),
|
||||
);
|
||||
expect(mockLoggerInstance.info).toHaveBeenCalledWith('Picked up daily analytics job.');
|
||||
expect(mockLoggerInstance.info).toHaveBeenCalledWith(
|
||||
'Successfully generated report for 2023-10-27.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle failure when reportDate is FAIL', async () => {
|
||||
const job = createMockJob<AnalyticsJobData>({ reportDate: 'FAIL' } as AnalyticsJobData);
|
||||
|
||||
const promise = service.processDailyReportJob(job);
|
||||
|
||||
await expect(promise).rejects.toThrow('This is a test failure for the analytics job.');
|
||||
|
||||
expect(mockLoggerInstance.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
attemptsMade: 1,
|
||||
}),
|
||||
'Daily analytics job failed.',
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('processWeeklyReportJob', () => {
|
||||
it('should process successfully', async () => {
|
||||
const job = createMockJob<WeeklyAnalyticsJobData>({
|
||||
reportYear: 2023,
|
||||
reportWeek: 43,
|
||||
} as WeeklyAnalyticsJobData);
|
||||
|
||||
const promise = service.processWeeklyReportJob(job);
|
||||
|
||||
await vi.advanceTimersByTimeAsync(30000);
|
||||
|
||||
const result = await promise;
|
||||
|
||||
expect(result).toEqual({ status: 'success', reportYear: 2023, reportWeek: 43 });
|
||||
expect(logger.child).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
jobId: 'job-123',
|
||||
reportYear: 2023,
|
||||
reportWeek: 43,
|
||||
}),
|
||||
);
|
||||
expect(mockLoggerInstance.info).toHaveBeenCalledWith('Picked up weekly analytics job.');
|
||||
expect(mockLoggerInstance.info).toHaveBeenCalledWith(
|
||||
'Successfully generated weekly report for week 43, 2023.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle errors during processing', async () => {
|
||||
const job = createMockJob<WeeklyAnalyticsJobData>({
|
||||
reportYear: 2023,
|
||||
reportWeek: 43,
|
||||
} as WeeklyAnalyticsJobData);
|
||||
|
||||
// Make the second info call throw to simulate an error inside the try block
|
||||
mockLoggerInstance.info
|
||||
.mockImplementationOnce(() => {}) // "Picked up..."
|
||||
.mockImplementationOnce(() => {
|
||||
throw new Error('Processing failed');
|
||||
}); // "Successfully generated..."
|
||||
|
||||
// Get the promise from the service method.
|
||||
const promise = service.processWeeklyReportJob(job);
|
||||
|
||||
// Capture the expectation promise BEFORE triggering the rejection.
|
||||
const expectation = expect(promise).rejects.toThrow('Processing failed');
|
||||
|
||||
// Advance timers to trigger the part of the code that throws.
|
||||
await vi.advanceTimersByTimeAsync(30000);
|
||||
|
||||
// Await the expectation to ensure assertions ran.
|
||||
await expectation;
|
||||
|
||||
// Verify the side effect (error logging) after the rejection is confirmed.
|
||||
expect(mockLoggerInstance.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
err: expect.any(Error),
|
||||
attemptsMade: 1,
|
||||
}),
|
||||
'Weekly analytics job failed.',
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
79
src/services/analyticsService.server.ts
Normal file
79
src/services/analyticsService.server.ts
Normal file
@@ -0,0 +1,79 @@
|
||||
// src/services/analyticsService.server.ts
|
||||
import type { Job } from 'bullmq';
|
||||
import { logger as globalLogger } from './logger.server';
|
||||
import type { AnalyticsJobData, WeeklyAnalyticsJobData } from '../types/job-data';
|
||||
|
||||
/**
|
||||
* A service class to encapsulate business logic for analytics-related background jobs.
|
||||
*/
|
||||
export class AnalyticsService {
|
||||
/**
|
||||
* Processes a job to generate a daily analytics report.
|
||||
* This is currently a mock implementation.
|
||||
* @param job The BullMQ job object.
|
||||
*/
|
||||
async processDailyReportJob(job: Job<AnalyticsJobData>) {
|
||||
const { reportDate } = job.data;
|
||||
const logger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
reportDate,
|
||||
});
|
||||
|
||||
logger.info(`Picked up daily analytics job.`);
|
||||
|
||||
try {
|
||||
// This is mock logic, but we keep it in the service
|
||||
if (reportDate === 'FAIL') {
|
||||
throw new Error('This is a test failure for the analytics job.');
|
||||
}
|
||||
// Simulate work
|
||||
await new Promise((resolve) => setTimeout(resolve, 10000));
|
||||
logger.info(`Successfully generated report for ${reportDate}.`);
|
||||
return { status: 'success', reportDate };
|
||||
} catch (error) {
|
||||
const wrappedError = error instanceof Error ? error : new Error(String(error));
|
||||
logger.error(
|
||||
{
|
||||
err: wrappedError,
|
||||
attemptsMade: job.attemptsMade,
|
||||
},
|
||||
`Daily analytics job failed.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Processes a job to generate a weekly analytics report.
|
||||
* This is currently a mock implementation.
|
||||
* @param job The BullMQ job object.
|
||||
*/
|
||||
async processWeeklyReportJob(job: Job<WeeklyAnalyticsJobData>) {
|
||||
const { reportYear, reportWeek } = job.data;
|
||||
const logger = globalLogger.child({
|
||||
jobId: job.id,
|
||||
jobName: job.name,
|
||||
reportYear,
|
||||
reportWeek,
|
||||
});
|
||||
|
||||
logger.info(`Picked up weekly analytics job.`);
|
||||
|
||||
try {
|
||||
// Mock logic
|
||||
await new Promise((resolve) => setTimeout(resolve, 30000));
|
||||
logger.info(`Successfully generated weekly report for week ${reportWeek}, ${reportYear}.`);
|
||||
return { status: 'success', reportYear, reportWeek };
|
||||
} catch (error) {
|
||||
const wrappedError = error instanceof Error ? error : new Error(String(error));
|
||||
logger.error(
|
||||
{ err: wrappedError, attemptsMade: job.attemptsMade },
|
||||
`Weekly analytics job failed.`,
|
||||
);
|
||||
throw wrappedError;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
export const analyticsService = new AnalyticsService();
|
||||
@@ -7,6 +7,17 @@ import { http, HttpResponse } from 'msw';
|
||||
vi.unmock('./apiClient');
|
||||
|
||||
import * as apiClient from './apiClient';
|
||||
import {
|
||||
createMockAddressPayload,
|
||||
createMockBudget,
|
||||
createMockLoginPayload,
|
||||
createMockProfileUpdatePayload,
|
||||
createMockRecipeCommentPayload,
|
||||
createMockRegisterUserPayload,
|
||||
createMockSearchQueryPayload,
|
||||
createMockShoppingListItemPayload,
|
||||
createMockWatchedItemPayload,
|
||||
} from '../tests/utils/mockFactories';
|
||||
|
||||
// Mock the logger to keep test output clean and verifiable.
|
||||
vi.mock('./logger', () => ({
|
||||
@@ -229,33 +240,6 @@ describe('API Client', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('Analytics API Functions', () => {
|
||||
it('trackFlyerItemInteraction should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
|
||||
// We can now await this properly because we added 'return' in apiClient.ts
|
||||
await apiClient.trackFlyerItemInteraction(123, 'click');
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
});
|
||||
|
||||
it('logSearchQuery should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
|
||||
await apiClient.logSearchQuery({
|
||||
query_text: 'test',
|
||||
result_count: 0,
|
||||
was_successful: false,
|
||||
});
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
});
|
||||
});
|
||||
|
||||
describe('apiFetch (with FormData)', () => {
|
||||
it('should handle FormData correctly by not setting Content-Type', async () => {
|
||||
localStorage.setItem('authToken', 'form-data-token');
|
||||
@@ -317,10 +301,11 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('addWatchedItem should send a POST request with the correct body', async () => {
|
||||
await apiClient.addWatchedItem('Apples', 'Produce');
|
||||
const watchedItemData = createMockWatchedItemPayload({ itemName: 'Apples', category: 'Produce' });
|
||||
await apiClient.addWatchedItem(watchedItemData.itemName, watchedItemData.category);
|
||||
|
||||
expect(capturedUrl?.pathname).toBe('/api/users/watched-items');
|
||||
expect(capturedBody).toEqual({ itemName: 'Apples', category: 'Produce' });
|
||||
expect(capturedBody).toEqual(watchedItemData);
|
||||
});
|
||||
|
||||
it('removeWatchedItem should send a DELETE request to the correct URL', async () => {
|
||||
@@ -337,12 +322,12 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('createBudget should send a POST request with budget data', async () => {
|
||||
const budgetData = {
|
||||
const budgetData = createMockBudget({
|
||||
name: 'Groceries',
|
||||
amount_cents: 50000,
|
||||
period: 'monthly' as const,
|
||||
period: 'monthly',
|
||||
start_date: '2024-01-01',
|
||||
};
|
||||
});
|
||||
await apiClient.createBudget(budgetData);
|
||||
|
||||
expect(capturedUrl?.pathname).toBe('/api/budgets');
|
||||
@@ -461,7 +446,7 @@ describe('API Client', () => {
|
||||
|
||||
it('addShoppingListItem should send a POST request with item data', async () => {
|
||||
const listId = 42;
|
||||
const itemData = { customItemName: 'Paper Towels' };
|
||||
const itemData = createMockShoppingListItemPayload({ customItemName: 'Paper Towels' });
|
||||
await apiClient.addShoppingListItem(listId, itemData);
|
||||
|
||||
expect(capturedUrl?.pathname).toBe(`/api/users/shopping-lists/${listId}/items`);
|
||||
@@ -547,7 +532,7 @@ describe('API Client', () => {
|
||||
|
||||
it('addRecipeComment should send a POST request with content and optional parentId', async () => {
|
||||
const recipeId = 456;
|
||||
const commentData = { content: 'This is a reply', parentCommentId: 789 };
|
||||
const commentData = createMockRecipeCommentPayload({ content: 'This is a reply', parentCommentId: 789 });
|
||||
await apiClient.addRecipeComment(recipeId, commentData.content, commentData.parentCommentId);
|
||||
expect(capturedUrl?.pathname).toBe(`/api/recipes/${recipeId}/comments`);
|
||||
expect(capturedBody).toEqual(commentData);
|
||||
@@ -563,7 +548,7 @@ describe('API Client', () => {
|
||||
describe('User Profile and Settings API Functions', () => {
|
||||
it('updateUserProfile should send a PUT request with profile data', async () => {
|
||||
localStorage.setItem('authToken', 'user-settings-token');
|
||||
const profileData = { full_name: 'John Doe' };
|
||||
const profileData = createMockProfileUpdatePayload({ full_name: 'John Doe' });
|
||||
await apiClient.updateUserProfile(profileData, { tokenOverride: 'override-token' });
|
||||
expect(capturedUrl?.pathname).toBe('/api/users/profile');
|
||||
expect(capturedBody).toEqual(profileData);
|
||||
@@ -619,14 +604,14 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('registerUser should send a POST request with user data', async () => {
|
||||
await apiClient.registerUser('test@example.com', 'password123', 'Test User');
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/register');
|
||||
expect(capturedBody).toEqual({
|
||||
const userData = createMockRegisterUserPayload({
|
||||
email: 'test@example.com',
|
||||
password: 'password123',
|
||||
full_name: 'Test User',
|
||||
avatar_url: undefined,
|
||||
});
|
||||
await apiClient.registerUser(userData.email, userData.password, userData.full_name);
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/register');
|
||||
expect(capturedBody).toEqual(userData);
|
||||
});
|
||||
|
||||
it('deleteUserAccount should send a DELETE request with the confirmation password', async () => {
|
||||
@@ -654,7 +639,7 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('updateUserAddress should send a PUT request with address data', async () => {
|
||||
const addressData = { address_line_1: '123 Main St', city: 'Anytown' };
|
||||
const addressData = createMockAddressPayload({ address_line_1: '123 Main St', city: 'Anytown' });
|
||||
await apiClient.updateUserAddress(addressData);
|
||||
expect(capturedUrl?.pathname).toBe('/api/users/profile/address');
|
||||
expect(capturedBody).toEqual(addressData);
|
||||
@@ -890,6 +875,11 @@ describe('API Client', () => {
|
||||
expect(capturedUrl?.pathname).toBe('/api/admin/corrections');
|
||||
});
|
||||
|
||||
it('getFlyersForReview should call the correct endpoint', async () => {
|
||||
await apiClient.getFlyersForReview();
|
||||
expect(capturedUrl?.pathname).toBe('/api/admin/review/flyers');
|
||||
});
|
||||
|
||||
it('rejectCorrection should send a POST request to the correct URL', async () => {
|
||||
const correctionId = 46;
|
||||
await apiClient.rejectCorrection(correctionId);
|
||||
@@ -942,53 +932,49 @@ describe('API Client', () => {
|
||||
});
|
||||
|
||||
it('logSearchQuery should send a POST request with query data', async () => {
|
||||
const queryData = { query_text: 'apples', result_count: 10, was_successful: true };
|
||||
const queryData = createMockSearchQueryPayload({ query_text: 'apples', result_count: 10, was_successful: true });
|
||||
await apiClient.logSearchQuery(queryData);
|
||||
expect(capturedUrl?.pathname).toBe('/api/search/log');
|
||||
expect(capturedBody).toEqual(queryData);
|
||||
});
|
||||
|
||||
it('trackFlyerItemInteraction should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
const { logger } = await import('./logger.client');
|
||||
|
||||
// We can now await this properly because we added 'return' in apiClient.ts
|
||||
await apiClient.trackFlyerItemInteraction(123, 'click');
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to track flyer item interaction', {
|
||||
error: apiError,
|
||||
});
|
||||
});
|
||||
|
||||
it('logSearchQuery should log a warning on failure', async () => {
|
||||
const { logger } = await import('./logger.client');
|
||||
const apiError = new Error('Network failed');
|
||||
vi.mocked(global.fetch).mockRejectedValue(apiError);
|
||||
const { logger } = await import('./logger.client');
|
||||
|
||||
await apiClient.logSearchQuery({
|
||||
const queryData = createMockSearchQueryPayload({
|
||||
query_text: 'test',
|
||||
result_count: 0,
|
||||
was_successful: false,
|
||||
});
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
|
||||
await apiClient.logSearchQuery(queryData);
|
||||
expect(logger.warn).toHaveBeenCalledWith('Failed to log search query', { error: apiError });
|
||||
});
|
||||
});
|
||||
|
||||
describe('Authentication API Functions', () => {
|
||||
it('loginUser should send a POST request with credentials', async () => {
|
||||
await apiClient.loginUser('test@example.com', 'password123', true);
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/login');
|
||||
expect(capturedBody).toEqual({
|
||||
const loginData = createMockLoginPayload({
|
||||
email: 'test@example.com',
|
||||
password: 'password123',
|
||||
rememberMe: true,
|
||||
});
|
||||
await apiClient.loginUser(loginData.email, loginData.password, loginData.rememberMe);
|
||||
expect(capturedUrl?.pathname).toBe('/api/auth/login');
|
||||
expect(capturedBody).toEqual(loginData);
|
||||
});
|
||||
});
|
||||
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
339
src/services/authService.test.ts
Normal file
339
src/services/authService.test.ts
Normal file
@@ -0,0 +1,339 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import type { UserProfile } from '../types';
|
||||
import type * as jsonwebtoken from 'jsonwebtoken';
|
||||
|
||||
describe('AuthService', () => {
|
||||
let authService: typeof import('./authService').authService;
|
||||
let bcrypt: typeof import('bcrypt');
|
||||
let jwt: typeof jsonwebtoken & { default: typeof jsonwebtoken };
|
||||
let userRepo: typeof import('./db/index.db').userRepo;
|
||||
let adminRepo: typeof import('./db/index.db').adminRepo;
|
||||
let logger: typeof import('./logger.server').logger;
|
||||
let sendPasswordResetEmail: typeof import('./emailService.server').sendPasswordResetEmail;
|
||||
let UniqueConstraintError: typeof import('./db/errors.db').UniqueConstraintError;
|
||||
|
||||
const reqLog = {}; // Mock request logger object
|
||||
const mockUser = {
|
||||
user_id: 'user-123',
|
||||
email: 'test@example.com',
|
||||
password_hash: 'hashed-password',
|
||||
};
|
||||
const mockUserProfile: UserProfile = {
|
||||
user: mockUser,
|
||||
role: 'user',
|
||||
} as unknown as UserProfile;
|
||||
|
||||
beforeEach(async () => {
|
||||
vi.clearAllMocks();
|
||||
vi.resetModules();
|
||||
|
||||
// Set environment variables before any modules are imported
|
||||
process.env.JWT_SECRET = 'test-secret';
|
||||
process.env.FRONTEND_URL = 'http://localhost:3000';
|
||||
|
||||
// Mock all dependencies before dynamically importing the service
|
||||
// Core modules like bcrypt, jsonwebtoken, and crypto are now mocked globally in tests-setup-unit.ts
|
||||
vi.mock('bcrypt');
|
||||
vi.mock('./db/index.db', () => ({
|
||||
userRepo: {
|
||||
createUser: vi.fn(),
|
||||
saveRefreshToken: vi.fn(),
|
||||
findUserByEmail: vi.fn(),
|
||||
createPasswordResetToken: vi.fn(),
|
||||
getValidResetTokens: vi.fn(),
|
||||
updateUserPassword: vi.fn(),
|
||||
deleteResetToken: vi.fn(),
|
||||
findUserByRefreshToken: vi.fn(),
|
||||
findUserProfileById: vi.fn(),
|
||||
deleteRefreshToken: vi.fn(),
|
||||
},
|
||||
adminRepo: {
|
||||
logActivity: vi.fn(),
|
||||
},
|
||||
}));
|
||||
vi.mock('./logger.server', () => ({
|
||||
logger: { info: vi.fn(), error: vi.fn(), warn: vi.fn(), debug: vi.fn() },
|
||||
}));
|
||||
vi.mock('./emailService.server', () => ({
|
||||
sendPasswordResetEmail: vi.fn(),
|
||||
}));
|
||||
vi.mock('./db/connection.db', () => ({ getPool: vi.fn() }));
|
||||
vi.mock('../utils/authUtils', () => ({ validatePasswordStrength: vi.fn() }));
|
||||
|
||||
// Dynamically import modules to get the mocked versions and the service instance
|
||||
authService = (await import('./authService')).authService;
|
||||
bcrypt = await import('bcrypt');
|
||||
jwt = (await import('jsonwebtoken')) as typeof jwt;
|
||||
const dbModule = await import('./db/index.db');
|
||||
userRepo = dbModule.userRepo;
|
||||
adminRepo = dbModule.adminRepo;
|
||||
logger = (await import('./logger.server')).logger;
|
||||
sendPasswordResetEmail = (await import('./emailService.server')).sendPasswordResetEmail;
|
||||
UniqueConstraintError = (await import('./db/errors.db')).UniqueConstraintError;
|
||||
});
|
||||
|
||||
describe('registerUser', () => {
|
||||
it('should successfully register a new user', async () => {
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
vi.mocked(userRepo.createUser).mockResolvedValue(mockUserProfile);
|
||||
|
||||
const result = await authService.registerUser(
|
||||
'test@example.com',
|
||||
'password123',
|
||||
'Test User',
|
||||
undefined,
|
||||
reqLog,
|
||||
);
|
||||
|
||||
expect(bcrypt.hash).toHaveBeenCalledWith('password123', 10);
|
||||
expect(userRepo.createUser).toHaveBeenCalledWith(
|
||||
'test@example.com',
|
||||
'hashed-password',
|
||||
{ full_name: 'Test User', avatar_url: undefined },
|
||||
reqLog,
|
||||
);
|
||||
expect(adminRepo.logActivity).toHaveBeenCalledWith(
|
||||
expect.objectContaining({
|
||||
action: 'user_registered',
|
||||
userId: 'user-123',
|
||||
}),
|
||||
reqLog,
|
||||
);
|
||||
expect(result).toEqual(mockUserProfile);
|
||||
});
|
||||
|
||||
it('should throw UniqueConstraintError if email already exists', async () => {
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
const error = new UniqueConstraintError('Email exists');
|
||||
vi.mocked(userRepo.createUser).mockRejectedValue(error);
|
||||
|
||||
await expect(
|
||||
authService.registerUser('test@example.com', 'password123', undefined, undefined, reqLog),
|
||||
).rejects.toThrow(UniqueConstraintError);
|
||||
|
||||
expect(logger.error).not.toHaveBeenCalled(); // Should not log expected unique constraint errors as system errors
|
||||
});
|
||||
|
||||
it('should log and throw other errors', async () => {
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
const error = new Error('Database failed');
|
||||
vi.mocked(userRepo.createUser).mockRejectedValue(error);
|
||||
|
||||
await expect(
|
||||
authService.registerUser('test@example.com', 'password123', undefined, undefined, reqLog),
|
||||
).rejects.toThrow('Database failed');
|
||||
|
||||
expect(logger.error).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
describe('registerAndLoginUser', () => {
|
||||
it('should register user and return tokens', async () => {
|
||||
// Mock registerUser logic (since we can't easily spy on the same class instance method without prototype spying, we rely on the underlying calls)
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-password');
|
||||
vi.mocked(userRepo.createUser).mockResolvedValue(mockUserProfile);
|
||||
// FIX: The global mock for jsonwebtoken provides a `default` export.
|
||||
// The code under test (`authService`) uses `import jwt from 'jsonwebtoken'`, so it gets the default export.
|
||||
// We must mock `jwt.default.sign` to affect the code under test.
|
||||
vi.mocked(jwt.default.sign).mockImplementation(() => 'access-token');
|
||||
|
||||
const result = await authService.registerAndLoginUser(
|
||||
'test@example.com',
|
||||
'password123',
|
||||
'Test User',
|
||||
undefined,
|
||||
reqLog,
|
||||
);
|
||||
|
||||
expect(result).toEqual({
|
||||
newUserProfile: mockUserProfile,
|
||||
accessToken: 'access-token',
|
||||
refreshToken: 'mocked_random_id',
|
||||
});
|
||||
expect(userRepo.saveRefreshToken).toHaveBeenCalledWith(
|
||||
'user-123',
|
||||
'mocked_random_id',
|
||||
reqLog,
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('generateAuthTokens', () => {
|
||||
it('should generate access and refresh tokens', () => {
|
||||
// FIX: The global mock for jsonwebtoken provides a `default` export.
|
||||
// The code under test (`authService`) uses `import jwt from 'jsonwebtoken'`, so it gets the default export.
|
||||
// We must mock `jwt.default.sign` to affect the code under test.
|
||||
vi.mocked(jwt.default.sign).mockImplementation(() => 'access-token');
|
||||
|
||||
const result = authService.generateAuthTokens(mockUserProfile);
|
||||
|
||||
expect(vi.mocked(jwt.default.sign)).toHaveBeenCalledWith(
|
||||
{
|
||||
user_id: 'user-123',
|
||||
email: 'test@example.com',
|
||||
role: 'user',
|
||||
},
|
||||
'test-secret',
|
||||
{ expiresIn: '15m' },
|
||||
);
|
||||
expect(result).toEqual({
|
||||
accessToken: 'access-token',
|
||||
refreshToken: 'mocked_random_id',
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('saveRefreshToken', () => {
|
||||
it('should save refresh token to db', async () => {
|
||||
await authService.saveRefreshToken('user-123', 'token', reqLog);
|
||||
expect(userRepo.saveRefreshToken).toHaveBeenCalledWith('user-123', 'token', reqLog);
|
||||
});
|
||||
|
||||
it('should log and throw error on failure', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(userRepo.saveRefreshToken).mockRejectedValue(error);
|
||||
|
||||
await expect(authService.saveRefreshToken('user-123', 'token', reqLog)).rejects.toThrow(
|
||||
'DB Error',
|
||||
);
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ error }),
|
||||
expect.stringContaining('Failed to save refresh token'),
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('resetPassword', () => {
|
||||
it('should process password reset for existing user', async () => {
|
||||
vi.mocked(userRepo.findUserByEmail).mockResolvedValue(mockUser as any);
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'hashed-token');
|
||||
|
||||
const result = await authService.resetPassword('test@example.com', reqLog);
|
||||
|
||||
expect(userRepo.createPasswordResetToken).toHaveBeenCalledWith(
|
||||
'user-123',
|
||||
'hashed-token',
|
||||
expect.any(Date),
|
||||
reqLog,
|
||||
);
|
||||
expect(sendPasswordResetEmail).toHaveBeenCalledWith(
|
||||
'test@example.com',
|
||||
expect.stringContaining('/reset-password/mocked_random_id'),
|
||||
reqLog,
|
||||
);
|
||||
expect(result).toBe('mocked_random_id');
|
||||
});
|
||||
|
||||
it('should log warning and return undefined for non-existent user', async () => {
|
||||
vi.mocked(userRepo.findUserByEmail).mockResolvedValue(undefined);
|
||||
|
||||
const result = await authService.resetPassword('unknown@example.com', reqLog);
|
||||
|
||||
expect(logger.warn).toHaveBeenCalledWith(
|
||||
expect.stringContaining('Password reset requested for non-existent email'),
|
||||
);
|
||||
expect(sendPasswordResetEmail).not.toHaveBeenCalled();
|
||||
expect(result).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should log error and throw on failure', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(userRepo.findUserByEmail).mockRejectedValue(error);
|
||||
|
||||
await expect(authService.resetPassword('test@example.com', reqLog)).rejects.toThrow(
|
||||
'DB Error',
|
||||
);
|
||||
expect(logger.error).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
describe('updatePassword', () => {
|
||||
it('should update password if token is valid', async () => {
|
||||
const mockTokenRecord = {
|
||||
user_id: 'user-123',
|
||||
token_hash: 'hashed-token',
|
||||
};
|
||||
vi.mocked(userRepo.getValidResetTokens).mockResolvedValue([mockTokenRecord] as any);
|
||||
vi.mocked(bcrypt.compare).mockImplementation(async () => true); // Match found
|
||||
vi.mocked(bcrypt.hash).mockImplementation(async () => 'new-hashed-password');
|
||||
|
||||
const result = await authService.updatePassword('valid-token', 'newPassword', reqLog);
|
||||
|
||||
expect(userRepo.updateUserPassword).toHaveBeenCalledWith(
|
||||
'user-123',
|
||||
'new-hashed-password',
|
||||
reqLog,
|
||||
);
|
||||
expect(userRepo.deleteResetToken).toHaveBeenCalledWith('hashed-token', reqLog);
|
||||
expect(adminRepo.logActivity).toHaveBeenCalledWith(
|
||||
expect.objectContaining({ action: 'password_reset' }),
|
||||
reqLog,
|
||||
);
|
||||
expect(result).toBe(true);
|
||||
});
|
||||
|
||||
it('should return null if token is invalid or not found', async () => {
|
||||
vi.mocked(userRepo.getValidResetTokens).mockResolvedValue([]);
|
||||
|
||||
const result = await authService.updatePassword('invalid-token', 'newPassword', reqLog);
|
||||
|
||||
expect(userRepo.updateUserPassword).not.toHaveBeenCalled();
|
||||
expect(result).toBeNull();
|
||||
});
|
||||
});
|
||||
|
||||
describe('getUserByRefreshToken', () => {
|
||||
it('should return user profile if token exists', async () => {
|
||||
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue({ user_id: 'user-123' } as any);
|
||||
vi.mocked(userRepo.findUserProfileById).mockResolvedValue(mockUserProfile);
|
||||
|
||||
const result = await authService.getUserByRefreshToken('valid-token', reqLog);
|
||||
|
||||
expect(result).toEqual(mockUserProfile);
|
||||
});
|
||||
|
||||
it('should return null if token not found', async () => {
|
||||
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue(undefined);
|
||||
|
||||
const result = await authService.getUserByRefreshToken('invalid-token', reqLog);
|
||||
|
||||
expect(result).toBeNull();
|
||||
});
|
||||
});
|
||||
|
||||
describe('logout', () => {
|
||||
it('should delete refresh token', async () => {
|
||||
await authService.logout('token', reqLog);
|
||||
expect(userRepo.deleteRefreshToken).toHaveBeenCalledWith('token', reqLog);
|
||||
});
|
||||
|
||||
it('should log and throw on error', async () => {
|
||||
const error = new Error('DB Error');
|
||||
vi.mocked(userRepo.deleteRefreshToken).mockRejectedValue(error);
|
||||
|
||||
await expect(authService.logout('token', reqLog)).rejects.toThrow('DB Error');
|
||||
expect(logger.error).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
describe('refreshAccessToken', () => {
|
||||
it('should return new access token if user found', async () => {
|
||||
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue({ user_id: 'user-123' } as any);
|
||||
vi.mocked(userRepo.findUserProfileById).mockResolvedValue(mockUserProfile);
|
||||
// FIX: The global mock for jsonwebtoken provides a `default` export.
|
||||
// The code under test (`authService`) uses `import jwt from 'jsonwebtoken'`, so it gets the default export.
|
||||
// We must mock `jwt.default.sign` to affect the code under test.
|
||||
vi.mocked(jwt.default.sign).mockImplementation(() => 'new-access-token');
|
||||
|
||||
const result = await authService.refreshAccessToken('valid-token', reqLog);
|
||||
|
||||
expect(result).toEqual({ accessToken: 'new-access-token' });
|
||||
});
|
||||
|
||||
it('should return null if user not found', async () => {
|
||||
vi.mocked(userRepo.findUserByRefreshToken).mockResolvedValue(undefined);
|
||||
const result = await authService.refreshAccessToken('invalid-token', reqLog);
|
||||
expect(result).toBeNull();
|
||||
});
|
||||
});
|
||||
});
|
||||
221
src/services/authService.ts
Normal file
221
src/services/authService.ts
Normal file
@@ -0,0 +1,221 @@
|
||||
// src/services/authService.ts
|
||||
import * as bcrypt from 'bcrypt';
|
||||
import jwt from 'jsonwebtoken';
|
||||
import crypto from 'crypto';
|
||||
import { userRepo, adminRepo } from './db/index.db';
|
||||
import { UniqueConstraintError } from './db/errors.db';
|
||||
import { getPool } from './db/connection.db';
|
||||
import { logger } from './logger.server';
|
||||
import { sendPasswordResetEmail } from './emailService.server';
|
||||
import type { UserProfile } from '../types';
|
||||
import { validatePasswordStrength } from '../utils/authUtils';
|
||||
|
||||
const JWT_SECRET = process.env.JWT_SECRET!;
|
||||
|
||||
class AuthService {
|
||||
async registerUser(
|
||||
email: string,
|
||||
password: string,
|
||||
fullName: string | undefined,
|
||||
avatarUrl: string | undefined,
|
||||
reqLog: any,
|
||||
) {
|
||||
try {
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(password, saltRounds);
|
||||
logger.info(`Hashing password for new user: ${email}`);
|
||||
|
||||
// The createUser method in UserRepository now handles its own transaction.
|
||||
const newUser = await userRepo.createUser(
|
||||
email,
|
||||
hashedPassword,
|
||||
{ full_name: fullName, avatar_url: avatarUrl },
|
||||
reqLog,
|
||||
);
|
||||
|
||||
const userEmail = newUser.user.email;
|
||||
const userId = newUser.user.user_id;
|
||||
logger.info(`Successfully created new user in DB: ${userEmail} (ID: ${userId})`);
|
||||
|
||||
// Use the new standardized logging function
|
||||
await adminRepo.logActivity(
|
||||
{
|
||||
userId: newUser.user.user_id,
|
||||
action: 'user_registered',
|
||||
displayText: `${userEmail} has registered.`,
|
||||
icon: 'user-plus',
|
||||
},
|
||||
reqLog,
|
||||
);
|
||||
|
||||
return newUser;
|
||||
} catch (error: unknown) {
|
||||
if (error instanceof UniqueConstraintError) {
|
||||
// If the email is a duplicate, return a 409 Conflict status.
|
||||
throw error;
|
||||
}
|
||||
logger.error({ error }, `User registration route failed for email: ${email}.`);
|
||||
// Pass the error to the centralized handler
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
async registerAndLoginUser(
|
||||
email: string,
|
||||
password: string,
|
||||
fullName: string | undefined,
|
||||
avatarUrl: string | undefined,
|
||||
reqLog: any,
|
||||
): Promise<{ newUserProfile: UserProfile; accessToken: string; refreshToken: string }> {
|
||||
const newUserProfile = await this.registerUser(
|
||||
email,
|
||||
password,
|
||||
fullName,
|
||||
avatarUrl,
|
||||
reqLog,
|
||||
);
|
||||
const { accessToken, refreshToken } = await this.handleSuccessfulLogin(newUserProfile, reqLog);
|
||||
return { newUserProfile, accessToken, refreshToken };
|
||||
}
|
||||
|
||||
generateAuthTokens(userProfile: UserProfile) {
|
||||
const payload = {
|
||||
user_id: userProfile.user.user_id,
|
||||
email: userProfile.user.email,
|
||||
role: userProfile.role,
|
||||
};
|
||||
const accessToken = jwt.sign(payload, JWT_SECRET, { expiresIn: '15m' });
|
||||
const refreshToken = crypto.randomBytes(64).toString('hex');
|
||||
|
||||
return { accessToken, refreshToken };
|
||||
}
|
||||
|
||||
async saveRefreshToken(userId: string, refreshToken: string, reqLog: any) {
|
||||
try {
|
||||
await userRepo.saveRefreshToken(userId, refreshToken, reqLog);
|
||||
} catch (tokenErr) {
|
||||
logger.error(
|
||||
{ error: tokenErr },
|
||||
`Failed to save refresh token during login for user: ${userId}`,
|
||||
);
|
||||
throw tokenErr;
|
||||
}
|
||||
}
|
||||
|
||||
async handleSuccessfulLogin(userProfile: UserProfile, reqLog: any) {
|
||||
const { accessToken, refreshToken } = this.generateAuthTokens(userProfile);
|
||||
await this.saveRefreshToken(userProfile.user.user_id, refreshToken, reqLog);
|
||||
return { accessToken, refreshToken };
|
||||
}
|
||||
|
||||
async resetPassword(email: string, reqLog: any) {
|
||||
try {
|
||||
logger.debug(`[API /forgot-password] Received request for email: ${email}`);
|
||||
const user = await userRepo.findUserByEmail(email, reqLog);
|
||||
let token: string | undefined;
|
||||
logger.debug(
|
||||
{ user: user ? { user_id: user.user_id, email: user.email } : 'NOT FOUND' },
|
||||
`[API /forgot-password] Database search result for ${email}:`,
|
||||
);
|
||||
|
||||
if (user) {
|
||||
token = crypto.randomBytes(32).toString('hex');
|
||||
const saltRounds = 10;
|
||||
const tokenHash = await bcrypt.hash(token, saltRounds);
|
||||
const expiresAt = new Date(Date.now() + 3600000); // 1 hour
|
||||
|
||||
await userRepo.createPasswordResetToken(user.user_id, tokenHash, expiresAt, reqLog);
|
||||
|
||||
const resetLink = `${process.env.FRONTEND_URL}/reset-password/${token}`;
|
||||
|
||||
try {
|
||||
await sendPasswordResetEmail(email, resetLink, reqLog);
|
||||
} catch (emailError) {
|
||||
logger.error({ emailError }, `Email send failure during password reset for user`);
|
||||
}
|
||||
} else {
|
||||
logger.warn(`Password reset requested for non-existent email: ${email}`);
|
||||
}
|
||||
|
||||
return token;
|
||||
} catch (error) {
|
||||
logger.error({ error }, `An error occurred during /forgot-password for email: ${email}`);
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
async updatePassword(token: string, newPassword: string, reqLog: any) {
|
||||
try {
|
||||
const validTokens = await userRepo.getValidResetTokens(reqLog);
|
||||
let tokenRecord;
|
||||
for (const record of validTokens) {
|
||||
const isMatch = await bcrypt.compare(token, record.token_hash);
|
||||
if (isMatch) {
|
||||
tokenRecord = record;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (!tokenRecord) {
|
||||
return null;
|
||||
}
|
||||
|
||||
const saltRounds = 10;
|
||||
const hashedPassword = await bcrypt.hash(newPassword, saltRounds);
|
||||
|
||||
await userRepo.updateUserPassword(tokenRecord.user_id, hashedPassword, reqLog);
|
||||
await userRepo.deleteResetToken(tokenRecord.token_hash, reqLog);
|
||||
|
||||
// Log this security event after a successful password reset.
|
||||
await adminRepo.logActivity(
|
||||
{
|
||||
userId: tokenRecord.user_id,
|
||||
action: 'password_reset',
|
||||
displayText: `User ID ${tokenRecord.user_id} has reset their password.`,
|
||||
icon: 'key',
|
||||
details: { source_ip: null },
|
||||
},
|
||||
reqLog,
|
||||
);
|
||||
|
||||
return true;
|
||||
} catch (error) {
|
||||
logger.error({ error }, `An error occurred during password reset.`);
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
async getUserByRefreshToken(refreshToken: string, reqLog: any) {
|
||||
try {
|
||||
const basicUser = await userRepo.findUserByRefreshToken(refreshToken, reqLog);
|
||||
if (!basicUser) {
|
||||
return null;
|
||||
}
|
||||
const userProfile = await userRepo.findUserProfileById(basicUser.user_id, reqLog);
|
||||
return userProfile;
|
||||
} catch (error) {
|
||||
logger.error({ error }, 'An error occurred during /refresh-token.');
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
async logout(refreshToken: string, reqLog: any) {
|
||||
try {
|
||||
await userRepo.deleteRefreshToken(refreshToken, reqLog);
|
||||
} catch (err: any) {
|
||||
logger.error({ error: err }, 'Failed to delete refresh token from DB during logout.');
|
||||
throw err;
|
||||
}
|
||||
}
|
||||
|
||||
async refreshAccessToken(refreshToken: string, reqLog: any): Promise<{ accessToken: string } | null> {
|
||||
const user = await this.getUserByRefreshToken(refreshToken, reqLog);
|
||||
if (!user) {
|
||||
return null;
|
||||
}
|
||||
const { accessToken } = this.generateAuthTokens(user);
|
||||
return { accessToken };
|
||||
}
|
||||
}
|
||||
|
||||
export const authService = new AuthService();
|
||||
@@ -335,8 +335,14 @@ describe('Background Job Service', () => {
|
||||
// Use fake timers to control promise resolution
|
||||
vi.useFakeTimers();
|
||||
|
||||
// Create a controllable promise
|
||||
let resolveRun!: () => void;
|
||||
const runPromise = new Promise<void>((resolve) => {
|
||||
resolveRun = resolve;
|
||||
});
|
||||
|
||||
// Make the first call hang indefinitely
|
||||
vi.mocked(mockBackgroundJobService.runDailyDealCheck).mockReturnValue(new Promise(() => {}));
|
||||
vi.mocked(mockBackgroundJobService.runDailyDealCheck).mockReturnValue(runPromise);
|
||||
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
@@ -352,12 +358,60 @@ describe('Background Job Service', () => {
|
||||
// Trigger it a second time immediately
|
||||
const secondCall = dailyDealCheckCallback();
|
||||
|
||||
// Resolve the first call so the test can finish
|
||||
resolveRun();
|
||||
|
||||
await Promise.all([firstCall, secondCall]);
|
||||
|
||||
// The service method should only have been called once
|
||||
expect(mockBackgroundJobService.runDailyDealCheck).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should handle unhandled rejections in the daily deal check cron wrapper', async () => {
|
||||
// Use fake timers to control promise resolution
|
||||
vi.useFakeTimers();
|
||||
|
||||
// Create a controllable promise
|
||||
let resolveRun!: () => void;
|
||||
const runPromise = new Promise<void>((resolve) => {
|
||||
resolveRun = resolve;
|
||||
});
|
||||
|
||||
// Make the first call hang indefinitely to keep the lock active
|
||||
vi.mocked(mockBackgroundJobService.runDailyDealCheck).mockReturnValue(runPromise);
|
||||
|
||||
// Make logger.warn throw an error. This is outside the main try/catch in the cron job.
|
||||
const warnError = new Error('Logger warn failed');
|
||||
vi.mocked(globalMockLogger.warn).mockImplementationOnce(() => {
|
||||
throw warnError;
|
||||
});
|
||||
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
mockAnalyticsQueue,
|
||||
mockWeeklyAnalyticsQueue,
|
||||
mockTokenCleanupQueue,
|
||||
globalMockLogger,
|
||||
);
|
||||
const dailyDealCheckCallback = mockCronSchedule.mock.calls[0][1];
|
||||
|
||||
// Trigger the job once, it will hang and set the lock. Then trigger it a second time
|
||||
// to enter the `if (isDailyDealCheckRunning)` block and call the throwing logger.warn.
|
||||
const firstCall = dailyDealCheckCallback();
|
||||
const secondCall = dailyDealCheckCallback();
|
||||
|
||||
// Resolve the first call so the test can finish
|
||||
resolveRun();
|
||||
|
||||
await Promise.allSettled([firstCall, secondCall]);
|
||||
|
||||
// The outer catch block should have been called with the error from logger.warn
|
||||
expect(globalMockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: warnError },
|
||||
'[BackgroundJob] Unhandled rejection in daily deal check cron wrapper.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should enqueue an analytics job when the second cron job function is executed', async () => {
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
@@ -421,6 +475,31 @@ describe('Background Job Service', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle unhandled rejections in the analytics report cron wrapper', async () => {
|
||||
const infoError = new Error('Logger info failed');
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
mockAnalyticsQueue,
|
||||
mockWeeklyAnalyticsQueue,
|
||||
mockTokenCleanupQueue,
|
||||
globalMockLogger,
|
||||
);
|
||||
|
||||
// Make logger.info throw, which is outside the try/catch in the cron job.
|
||||
const infoSpy = vi.spyOn(globalMockLogger, 'info').mockImplementation(() => {
|
||||
throw infoError;
|
||||
});
|
||||
|
||||
const analyticsJobCallback = mockCronSchedule.mock.calls[1][1];
|
||||
await analyticsJobCallback();
|
||||
|
||||
expect(globalMockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: infoError }, // The implementation uses `err` key here
|
||||
'[BackgroundJob] Unhandled rejection in analytics report cron wrapper.',
|
||||
);
|
||||
infoSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should enqueue a weekly analytics job when the third cron job function is executed', async () => {
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
@@ -483,6 +562,30 @@ describe('Background Job Service', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle unhandled rejections in the weekly analytics report cron wrapper', async () => {
|
||||
const infoError = new Error('Logger info failed');
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
mockAnalyticsQueue,
|
||||
mockWeeklyAnalyticsQueue,
|
||||
mockTokenCleanupQueue,
|
||||
globalMockLogger,
|
||||
);
|
||||
|
||||
const infoSpy = vi.spyOn(globalMockLogger, 'info').mockImplementation(() => {
|
||||
throw infoError;
|
||||
});
|
||||
|
||||
const weeklyAnalyticsJobCallback = mockCronSchedule.mock.calls[2][1];
|
||||
await weeklyAnalyticsJobCallback();
|
||||
|
||||
expect(globalMockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: infoError },
|
||||
'[BackgroundJob] Unhandled rejection in weekly analytics report cron wrapper.',
|
||||
);
|
||||
infoSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should enqueue a token cleanup job when the fourth cron job function is executed', async () => {
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
@@ -542,6 +645,30 @@ describe('Background Job Service', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle unhandled rejections in the token cleanup cron wrapper', async () => {
|
||||
const infoError = new Error('Logger info failed');
|
||||
startBackgroundJobs(
|
||||
mockBackgroundJobService,
|
||||
mockAnalyticsQueue,
|
||||
mockWeeklyAnalyticsQueue,
|
||||
mockTokenCleanupQueue,
|
||||
globalMockLogger,
|
||||
);
|
||||
|
||||
const infoSpy = vi.spyOn(globalMockLogger, 'info').mockImplementation(() => {
|
||||
throw infoError;
|
||||
});
|
||||
|
||||
const tokenCleanupCallback = mockCronSchedule.mock.calls[3][1];
|
||||
await tokenCleanupCallback();
|
||||
|
||||
expect(globalMockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: infoError },
|
||||
'[BackgroundJob] Unhandled rejection in token cleanup cron wrapper.',
|
||||
);
|
||||
infoSpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should log a critical error if scheduling fails', () => {
|
||||
mockCronSchedule.mockImplementation(() => {
|
||||
throw new Error('Scheduling failed');
|
||||
|
||||
@@ -7,6 +7,7 @@ import { getSimpleWeekAndYear } from '../utils/dateUtils';
|
||||
// Import types for repositories from their source files
|
||||
import type { PersonalizationRepository } from './db/personalization.db';
|
||||
import type { NotificationRepository } from './db/notification.db';
|
||||
import { analyticsQueue, weeklyAnalyticsQueue } from './queueService.server';
|
||||
|
||||
interface EmailJobData {
|
||||
to: string;
|
||||
@@ -23,6 +24,24 @@ export class BackgroundJobService {
|
||||
private logger: Logger,
|
||||
) {}
|
||||
|
||||
public async triggerAnalyticsReport(): Promise<string> {
|
||||
const reportDate = new Date().toISOString().split('T')[0]; // YYYY-MM-DD
|
||||
const jobId = `manual-report-${reportDate}-${Date.now()}`;
|
||||
const job = await analyticsQueue.add('generate-daily-report', { reportDate }, { jobId });
|
||||
return job.id!;
|
||||
}
|
||||
|
||||
public async triggerWeeklyAnalyticsReport(): Promise<string> {
|
||||
const { year: reportYear, week: reportWeek } = getSimpleWeekAndYear();
|
||||
const jobId = `manual-weekly-report-${reportYear}-${reportWeek}-${Date.now()}`;
|
||||
const job = await weeklyAnalyticsQueue.add(
|
||||
'generate-weekly-report',
|
||||
{ reportYear, reportWeek },
|
||||
{ jobId },
|
||||
);
|
||||
return job.id!;
|
||||
}
|
||||
|
||||
/**
|
||||
* Prepares the data for an email notification job based on a user's deals.
|
||||
* @param user The user to whom the email will be sent.
|
||||
@@ -212,7 +231,7 @@ export function startBackgroundJobs(
|
||||
})().catch((error: unknown) => {
|
||||
// This catch is for unhandled promise rejections from the async wrapper itself.
|
||||
logger.error(
|
||||
{ error },
|
||||
{ err: error },
|
||||
'[BackgroundJob] Unhandled rejection in daily deal check cron wrapper.',
|
||||
);
|
||||
isDailyDealCheckRunning = false;
|
||||
|
||||
51
src/services/brandService.test.ts
Normal file
51
src/services/brandService.test.ts
Normal file
@@ -0,0 +1,51 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { brandService } from './brandService';
|
||||
import * as db from './db/index.db';
|
||||
import type { Logger } from 'pino';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('./db/index.db', () => ({
|
||||
adminRepo: {
|
||||
updateBrandLogo: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
describe('BrandService', () => {
|
||||
const mockLogger = {} as Logger;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
describe('updateBrandLogo', () => {
|
||||
it('should update brand logo and return the new URL', async () => {
|
||||
const brandId = 123;
|
||||
const mockFile = {
|
||||
filename: 'test-logo.jpg',
|
||||
} as Express.Multer.File;
|
||||
|
||||
vi.mocked(db.adminRepo.updateBrandLogo).mockResolvedValue(undefined);
|
||||
|
||||
const result = await brandService.updateBrandLogo(brandId, mockFile, mockLogger);
|
||||
|
||||
expect(result).toBe('/flyer-images/test-logo.jpg');
|
||||
expect(db.adminRepo.updateBrandLogo).toHaveBeenCalledWith(
|
||||
brandId,
|
||||
'/flyer-images/test-logo.jpg',
|
||||
mockLogger,
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw error if database update fails', async () => {
|
||||
const brandId = 123;
|
||||
const mockFile = {
|
||||
filename: 'test-logo.jpg',
|
||||
} as Express.Multer.File;
|
||||
const dbError = new Error('DB Error');
|
||||
|
||||
vi.mocked(db.adminRepo.updateBrandLogo).mockRejectedValue(dbError);
|
||||
|
||||
await expect(brandService.updateBrandLogo(brandId, mockFile, mockLogger)).rejects.toThrow('DB Error');
|
||||
});
|
||||
});
|
||||
});
|
||||
13
src/services/brandService.ts
Normal file
13
src/services/brandService.ts
Normal file
@@ -0,0 +1,13 @@
|
||||
// src/services/brandService.ts
|
||||
import * as db from './db/index.db';
|
||||
import type { Logger } from 'pino';
|
||||
|
||||
class BrandService {
|
||||
async updateBrandLogo(brandId: number, file: Express.Multer.File, logger: Logger): Promise<string> {
|
||||
const logoUrl = `/flyer-images/${file.filename}`;
|
||||
await db.adminRepo.updateBrandLogo(brandId, logoUrl, logger);
|
||||
return logoUrl;
|
||||
}
|
||||
}
|
||||
|
||||
export const brandService = new BrandService();
|
||||
@@ -1,14 +1,9 @@
|
||||
// src/services/db/address.db.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import type { Pool } from 'pg';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import { AddressRepository } from './address.db';
|
||||
import type { Address } from '../../types';
|
||||
import { UniqueConstraintError, NotFoundError } from './errors.db';
|
||||
|
||||
// Un-mock the module we are testing
|
||||
vi.unmock('./address.db');
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('../logger.server', () => ({
|
||||
logger: { info: vi.fn(), warn: vi.fn(), error: vi.fn(), debug: vi.fn() },
|
||||
@@ -17,10 +12,13 @@ import { logger as mockLogger } from '../logger.server';
|
||||
|
||||
describe('Address DB Service', () => {
|
||||
let addressRepo: AddressRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
addressRepo = new AddressRepository(mockPoolInstance as unknown as Pool);
|
||||
addressRepo = new AddressRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getAddressById', () => {
|
||||
@@ -35,19 +33,19 @@ describe('Address DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
};
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockAddress] });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockAddress], rowCount: 1 });
|
||||
|
||||
const result = await addressRepo.getAddressById(1, mockLogger);
|
||||
|
||||
expect(result).toEqual(mockAddress);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.addresses WHERE address_id = $1',
|
||||
[1],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if no address is found', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(addressRepo.getAddressById(999, mockLogger)).rejects.toThrow(NotFoundError);
|
||||
await expect(addressRepo.getAddressById(999, mockLogger)).rejects.toThrow(
|
||||
'Address with ID 999 not found.',
|
||||
@@ -56,7 +54,7 @@ describe('Address DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.getAddressById(1, mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve address.',
|
||||
@@ -71,12 +69,12 @@ describe('Address DB Service', () => {
|
||||
describe('upsertAddress', () => {
|
||||
it('should INSERT a new address when no address_id is provided', async () => {
|
||||
const newAddressData = { address_line_1: '456 New Ave', city: 'Newville' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ address_id: 2 }] });
|
||||
mockDb.query.mockResolvedValue({ rows: [{ address_id: 2 }] });
|
||||
|
||||
const result = await addressRepo.upsertAddress(newAddressData, mockLogger);
|
||||
|
||||
expect(result).toBe(2);
|
||||
const [query, values] = mockPoolInstance.query.mock.calls[0];
|
||||
const [query, values] = mockDb.query.mock.calls[0];
|
||||
expect(query).toContain('INSERT INTO public.addresses');
|
||||
expect(query).toContain('ON CONFLICT (address_id) DO UPDATE');
|
||||
expect(values).toEqual(['456 New Ave', 'Newville']);
|
||||
@@ -84,64 +82,47 @@ describe('Address DB Service', () => {
|
||||
|
||||
it('should UPDATE an existing address when an address_id is provided', async () => {
|
||||
const existingAddressData = { address_id: 1, address_line_1: '789 Old Rd', city: 'Oldtown' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [{ address_id: 1 }] });
|
||||
mockDb.query.mockResolvedValue({ rows: [{ address_id: 1 }] });
|
||||
|
||||
const result = await addressRepo.upsertAddress(existingAddressData, mockLogger);
|
||||
|
||||
expect(result).toBe(1);
|
||||
const [query, values] = mockPoolInstance.query.mock.calls[0];
|
||||
const [query, values] = mockDb.query.mock.calls[0];
|
||||
expect(query).toContain('INSERT INTO public.addresses');
|
||||
expect(query).toContain('ON CONFLICT (address_id) DO UPDATE');
|
||||
// The values array should now include the address_id at the beginning
|
||||
expect(values).toEqual([1, '789 Old Rd', 'Oldtown']);
|
||||
});
|
||||
|
||||
it('should throw a generic error on INSERT failure', async () => {
|
||||
const newAddressData = { address_line_1: '456 New Ave', city: 'Newville' };
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
it('should throw UniqueConstraintError on unique constraint violation', async () => {
|
||||
const addressData = { address_line_1: '123 Duplicate St' };
|
||||
const dbError = new Error('duplicate key value violates unique constraint');
|
||||
(dbError as any).code = '23505';
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.upsertAddress(newAddressData, mockLogger)).rejects.toThrow(
|
||||
'Failed to upsert address.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: newAddressData },
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error on UPDATE failure', async () => {
|
||||
const existingAddressData = { address_id: 1, address_line_1: '789 Old Rd', city: 'Oldtown' };
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.upsertAddress(existingAddressData, mockLogger)).rejects.toThrow(
|
||||
'Failed to upsert address.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: existingAddressData },
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw UniqueConstraintError on duplicate address insert', async () => {
|
||||
const newAddressData = { address_line_1: '123 Main St', city: 'Anytown' };
|
||||
const dbError = new Error('duplicate key value violates unique constraint') as Error & {
|
||||
code: string;
|
||||
};
|
||||
dbError.code = '23505';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.upsertAddress(newAddressData, mockLogger)).rejects.toThrow(
|
||||
await expect(addressRepo.upsertAddress(addressData, mockLogger)).rejects.toThrow(
|
||||
UniqueConstraintError,
|
||||
);
|
||||
await expect(addressRepo.upsertAddress(newAddressData, mockLogger)).rejects.toThrow(
|
||||
await expect(addressRepo.upsertAddress(addressData, mockLogger)).rejects.toThrow(
|
||||
'An identical address already exists.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: newAddressData },
|
||||
{ err: dbError, address: addressData },
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error if the database query fails for other reasons', async () => {
|
||||
const addressData = { address_line_1: '789 Failure Rd' };
|
||||
const dbError = new Error('DB Connection Error');
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(addressRepo.upsertAddress(addressData, mockLogger)).rejects.toThrow(
|
||||
'Failed to upsert address.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, address: addressData },
|
||||
'Database error in upsertAddress',
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -6,9 +6,11 @@ import { UniqueConstraintError, NotFoundError } from './errors.db';
|
||||
import { Address } from '../../types';
|
||||
|
||||
export class AddressRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -1,14 +1,14 @@
|
||||
// src/services/db/admin.db.test.ts
|
||||
import { describe, it, expect, vi, beforeEach, Mock } from 'vitest';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import type { Pool, PoolClient } from 'pg';
|
||||
import { ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
||||
import { AdminRepository } from './admin.db';
|
||||
import type { SuggestedCorrection, AdminUserView, Profile } from '../../types';
|
||||
import type { SuggestedCorrection, AdminUserView, Profile, Flyer } from '../../types';
|
||||
import {
|
||||
createMockSuggestedCorrection,
|
||||
createMockAdminUserView,
|
||||
createMockProfile,
|
||||
createMockFlyer,
|
||||
} from '../../tests/utils/mockFactories';
|
||||
// Un-mock the module we are testing
|
||||
vi.unmock('./admin.db');
|
||||
@@ -33,6 +33,9 @@ import { withTransaction } from './connection.db';
|
||||
|
||||
describe('Admin DB Service', () => {
|
||||
let adminRepo: AdminRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset the global mock's call history before each test.
|
||||
@@ -43,8 +46,8 @@ describe('Admin DB Service', () => {
|
||||
const mockClient = { query: vi.fn() };
|
||||
return callback(mockClient as unknown as PoolClient);
|
||||
});
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
adminRepo = new AdminRepository(mockPoolInstance as unknown as Pool);
|
||||
// Instantiate the repository with the minimal mock db for each test
|
||||
adminRepo = new AdminRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getSuggestedCorrections', () => {
|
||||
@@ -52,11 +55,11 @@ describe('Admin DB Service', () => {
|
||||
const mockCorrections: SuggestedCorrection[] = [
|
||||
createMockSuggestedCorrection({ suggested_correction_id: 1 }),
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockCorrections });
|
||||
mockDb.query.mockResolvedValue({ rows: mockCorrections });
|
||||
|
||||
const result = await adminRepo.getSuggestedCorrections(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.suggested_corrections sc'),
|
||||
);
|
||||
expect(result).toEqual(mockCorrections);
|
||||
@@ -64,7 +67,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getSuggestedCorrections(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve suggested corrections.',
|
||||
);
|
||||
@@ -77,10 +80,10 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('approveCorrection', () => {
|
||||
it('should call the approve_correction database function', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] }); // Mock the function call
|
||||
mockDb.query.mockResolvedValue({ rows: [] }); // Mock the function call
|
||||
await adminRepo.approveCorrection(123, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT public.approve_correction($1)',
|
||||
[123],
|
||||
);
|
||||
@@ -88,7 +91,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database function fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.approveCorrection(123, mockLogger)).rejects.toThrow(
|
||||
'Failed to approve correction.',
|
||||
);
|
||||
@@ -101,17 +104,17 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('rejectCorrection', () => {
|
||||
it('should update the correction status to rejected', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 1 });
|
||||
await adminRepo.rejectCorrection(123, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining("UPDATE public.suggested_corrections SET status = 'rejected'"),
|
||||
[123],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if the correction is not found or not pending', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0 });
|
||||
await expect(adminRepo.rejectCorrection(123, mockLogger)).rejects.toThrow(NotFoundError);
|
||||
await expect(adminRepo.rejectCorrection(123, mockLogger)).rejects.toThrow(
|
||||
"Correction with ID 123 not found or not in 'pending' state.",
|
||||
@@ -119,7 +122,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
mockDb.query.mockRejectedValue(new Error('DB Error'));
|
||||
await expect(adminRepo.rejectCorrection(123, mockLogger)).rejects.toThrow(
|
||||
'Failed to reject correction.',
|
||||
);
|
||||
@@ -136,11 +139,11 @@ describe('Admin DB Service', () => {
|
||||
suggested_correction_id: 1,
|
||||
suggested_value: '300',
|
||||
});
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockCorrection], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockCorrection], rowCount: 1 });
|
||||
|
||||
const result = await adminRepo.updateSuggestedCorrection(1, '300', mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.suggested_corrections SET suggested_value = $1'),
|
||||
['300', 1],
|
||||
);
|
||||
@@ -148,7 +151,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the correction is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(
|
||||
adminRepo.updateSuggestedCorrection(999, 'new value', mockLogger),
|
||||
).rejects.toThrow(NotFoundError);
|
||||
@@ -158,7 +161,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
mockDb.query.mockRejectedValue(new Error('DB Error'));
|
||||
await expect(adminRepo.updateSuggestedCorrection(1, 'new value', mockLogger)).rejects.toThrow(
|
||||
'Failed to update suggested correction.',
|
||||
);
|
||||
@@ -172,7 +175,7 @@ describe('Admin DB Service', () => {
|
||||
describe('getApplicationStats', () => {
|
||||
it('should execute 5 parallel count queries and return the aggregated stats', async () => {
|
||||
// Mock responses for each of the 5 parallel queries
|
||||
mockPoolInstance.query
|
||||
mockDb.query
|
||||
.mockResolvedValueOnce({ rows: [{ count: '10' }] }) // flyerCount
|
||||
.mockResolvedValueOnce({ rows: [{ count: '20' }] }) // userCount
|
||||
.mockResolvedValueOnce({ rows: [{ count: '300' }] }) // flyerItemCount
|
||||
@@ -182,7 +185,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
const stats = await adminRepo.getApplicationStats(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(6);
|
||||
expect(mockDb.query).toHaveBeenCalledTimes(6);
|
||||
expect(stats).toEqual({
|
||||
flyerCount: 10,
|
||||
userCount: 20,
|
||||
@@ -195,7 +198,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if one of the parallel queries fails', async () => {
|
||||
// Mock one query to succeed and another to fail
|
||||
mockPoolInstance.query
|
||||
mockDb.query
|
||||
.mockResolvedValueOnce({ rows: [{ count: '10' }] })
|
||||
.mockRejectedValueOnce(new Error('DB Read Error'));
|
||||
|
||||
@@ -211,11 +214,11 @@ describe('Admin DB Service', () => {
|
||||
describe('getDailyStatsForLast30Days', () => {
|
||||
it('should execute the correct query to get daily stats', async () => {
|
||||
const mockStats = [{ date: '2023-01-01', new_users: 5, new_flyers: 2 }];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockStats });
|
||||
mockDb.query.mockResolvedValue({ rows: mockStats });
|
||||
|
||||
const result = await adminRepo.getDailyStatsForLast30Days(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('WITH date_series AS'),
|
||||
);
|
||||
expect(result).toEqual(mockStats);
|
||||
@@ -223,7 +226,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getDailyStatsForLast30Days(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve daily statistics.',
|
||||
);
|
||||
@@ -236,18 +239,18 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('logActivity', () => {
|
||||
it('should insert a new activity log entry', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
const logData = { userId: 'user-123', action: 'test_action', displayText: 'Test activity' };
|
||||
await adminRepo.logActivity(logData, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO public.activity_log'),
|
||||
[logData.userId, logData.action, logData.displayText, null, null],
|
||||
);
|
||||
});
|
||||
|
||||
it('should not throw an error if the database query fails (non-critical)', async () => {
|
||||
mockPoolInstance.query.mockRejectedValue(new Error('DB Error'));
|
||||
mockDb.query.mockRejectedValue(new Error('DB Error'));
|
||||
const logData = { action: 'test_action', displayText: 'Test activity' };
|
||||
await expect(adminRepo.logActivity(logData, mockLogger)).resolves.toBeUndefined();
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
@@ -259,9 +262,9 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('getMostFrequentSaleItems', () => {
|
||||
it('should call the correct database function', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.getMostFrequentSaleItems(30, 10, mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.flyer_items fi'),
|
||||
[30, 10],
|
||||
);
|
||||
@@ -269,7 +272,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getMostFrequentSaleItems(30, 10, mockLogger)).rejects.toThrow(
|
||||
'Failed to get most frequent sale items.',
|
||||
);
|
||||
@@ -283,9 +286,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateRecipeCommentStatus', () => {
|
||||
it('should update the comment status and return the updated comment', async () => {
|
||||
const mockComment = { comment_id: 1, status: 'hidden' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockComment], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockComment], rowCount: 1 });
|
||||
const result = await adminRepo.updateRecipeCommentStatus(1, 'hidden', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.recipe_comments'),
|
||||
['hidden', 1],
|
||||
);
|
||||
@@ -293,7 +296,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the comment is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateRecipeCommentStatus(999, 'hidden', mockLogger)).rejects.toThrow(
|
||||
'Recipe comment with ID 999 not found.',
|
||||
);
|
||||
@@ -301,7 +304,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateRecipeCommentStatus(1, 'hidden', mockLogger)).rejects.toThrow(
|
||||
'Failed to update recipe comment status.',
|
||||
);
|
||||
@@ -314,16 +317,16 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('getUnmatchedFlyerItems', () => {
|
||||
it('should execute the correct query to get unmatched items', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.getUnmatchedFlyerItems(mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.unmatched_flyer_items ufi'),
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getUnmatchedFlyerItems(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve unmatched flyer items.',
|
||||
);
|
||||
@@ -337,9 +340,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateRecipeStatus', () => {
|
||||
it('should update the recipe status and return the updated recipe', async () => {
|
||||
const mockRecipe = { recipe_id: 1, status: 'public' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockRecipe], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockRecipe], rowCount: 1 });
|
||||
const result = await adminRepo.updateRecipeStatus(1, 'public', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.recipes'),
|
||||
['public', 1],
|
||||
);
|
||||
@@ -347,7 +350,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the recipe is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateRecipeStatus(999, 'public', mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -358,7 +361,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateRecipeStatus(1, 'public', mockLogger)).rejects.toThrow(
|
||||
'Failed to update recipe status.',
|
||||
);
|
||||
@@ -437,16 +440,16 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('ignoreUnmatchedFlyerItem', () => {
|
||||
it('should update the status of an unmatched item to "ignored"', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 1 });
|
||||
await adminRepo.ignoreUnmatchedFlyerItem(1, mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
"UPDATE public.unmatched_flyer_items SET status = 'ignored' WHERE unmatched_flyer_item_id = $1 AND status = 'pending'",
|
||||
[1],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if the unmatched item is not found or not pending', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0 });
|
||||
await expect(adminRepo.ignoreUnmatchedFlyerItem(999, mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -457,11 +460,11 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.ignoreUnmatchedFlyerItem(1, mockLogger)).rejects.toThrow(
|
||||
'Failed to ignore unmatched flyer item.',
|
||||
);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining("UPDATE public.unmatched_flyer_items SET status = 'ignored'"),
|
||||
[1],
|
||||
);
|
||||
@@ -474,7 +477,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('resetFailedLoginAttempts', () => {
|
||||
it('should execute a specific UPDATE query to reset attempts and log login details', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.resetFailedLoginAttempts('user-123', '127.0.0.1', mockLogger);
|
||||
|
||||
// Use a regular expression to match the SQL query while ignoring whitespace differences.
|
||||
@@ -482,7 +485,7 @@ describe('Admin DB Service', () => {
|
||||
const expectedQueryRegex =
|
||||
/UPDATE\s+public\.users\s+SET\s+failed_login_attempts\s*=\s*0,\s*last_failed_login\s*=\s*NULL,\s*last_login_ip\s*=\s*\$2,\s*last_login_at\s*=\s*NOW\(\)\s+WHERE\s+user_id\s*=\s*\$1\s+AND\s+failed_login_attempts\s*>\s*0/;
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
// The test now verifies the full structure of the query.
|
||||
expect.stringMatching(expectedQueryRegex),
|
||||
['user-123', '127.0.0.1'],
|
||||
@@ -491,7 +494,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should not throw an error if the database query fails (non-critical)', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
adminRepo.resetFailedLoginAttempts('user-123', '127.0.0.1', mockLogger),
|
||||
).resolves.toBeUndefined();
|
||||
@@ -506,21 +509,21 @@ describe('Admin DB Service', () => {
|
||||
describe('incrementFailedLoginAttempts', () => {
|
||||
it('should execute an UPDATE query and return the new attempt count', async () => {
|
||||
// Mock the DB to return the new count
|
||||
mockPoolInstance.query.mockResolvedValue({
|
||||
mockDb.query.mockResolvedValue({
|
||||
rows: [{ failed_login_attempts: 3 }],
|
||||
rowCount: 1,
|
||||
});
|
||||
const newCount = await adminRepo.incrementFailedLoginAttempts('user-123', mockLogger);
|
||||
|
||||
expect(newCount).toBe(3);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('RETURNING failed_login_attempts'),
|
||||
['user-123'],
|
||||
);
|
||||
});
|
||||
|
||||
it('should return 0 if the user is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
const newCount = await adminRepo.incrementFailedLoginAttempts('user-not-found', mockLogger);
|
||||
expect(newCount).toBe(0);
|
||||
expect(mockLogger.warn).toHaveBeenCalledWith(
|
||||
@@ -531,7 +534,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should return -1 if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
const newCount = await adminRepo.incrementFailedLoginAttempts('user-123', mockLogger);
|
||||
|
||||
expect(newCount).toBe(-1);
|
||||
@@ -544,16 +547,16 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('updateBrandLogo', () => {
|
||||
it('should execute an UPDATE query for the brand logo', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.updateBrandLogo(1, '/logo.png', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'UPDATE public.brands SET logo_url = $1 WHERE brand_id = $2',
|
||||
['/logo.png', 1],
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw NotFoundError if the brand is not found', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0 });
|
||||
await expect(adminRepo.updateBrandLogo(999, '/logo.png', mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -564,11 +567,11 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateBrandLogo(1, '/logo.png', mockLogger)).rejects.toThrow(
|
||||
'Failed to update brand logo in database.',
|
||||
);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.brands SET logo_url'),
|
||||
['/logo.png', 1],
|
||||
);
|
||||
@@ -582,9 +585,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateReceiptStatus', () => {
|
||||
it('should update the receipt status and return the updated receipt', async () => {
|
||||
const mockReceipt = { receipt_id: 1, status: 'completed' };
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockReceipt], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockReceipt], rowCount: 1 });
|
||||
const result = await adminRepo.updateReceiptStatus(1, 'completed', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.receipts'),
|
||||
['completed', 1],
|
||||
);
|
||||
@@ -592,7 +595,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the receipt is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateReceiptStatus(999, 'completed', mockLogger)).rejects.toThrow(
|
||||
NotFoundError,
|
||||
);
|
||||
@@ -603,7 +606,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateReceiptStatus(1, 'completed', mockLogger)).rejects.toThrow(
|
||||
'Failed to update receipt status.',
|
||||
);
|
||||
@@ -616,9 +619,9 @@ describe('Admin DB Service', () => {
|
||||
|
||||
describe('getActivityLog', () => {
|
||||
it('should call the get_activity_log database function', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
await adminRepo.getActivityLog(50, 0, mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.get_activity_log($1, $2)',
|
||||
[50, 0],
|
||||
);
|
||||
@@ -626,7 +629,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getActivityLog(50, 0, mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve activity log.',
|
||||
);
|
||||
@@ -642,9 +645,9 @@ describe('Admin DB Service', () => {
|
||||
const mockUsers: AdminUserView[] = [
|
||||
createMockAdminUserView({ user_id: '1', email: 'test@test.com' }),
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockUsers });
|
||||
mockDb.query.mockResolvedValue({ rows: mockUsers });
|
||||
const result = await adminRepo.getAllUsers(mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.users u JOIN public.profiles p'),
|
||||
);
|
||||
expect(result).toEqual(mockUsers);
|
||||
@@ -652,7 +655,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.getAllUsers(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve all users.',
|
||||
);
|
||||
@@ -666,9 +669,9 @@ describe('Admin DB Service', () => {
|
||||
describe('updateUserRole', () => {
|
||||
it('should update the user role and return the updated user', async () => {
|
||||
const mockProfile: Profile = createMockProfile({ role: 'admin' });
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockProfile], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockProfile], rowCount: 1 });
|
||||
const result = await adminRepo.updateUserRole('1', 'admin', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'UPDATE public.profiles SET role = $1 WHERE user_id = $2 RETURNING *',
|
||||
['admin', '1'],
|
||||
);
|
||||
@@ -676,7 +679,7 @@ describe('Admin DB Service', () => {
|
||||
});
|
||||
|
||||
it('should throw an error if the user is not found (rowCount is 0)', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 0, rows: [] });
|
||||
await expect(adminRepo.updateUserRole('999', 'admin', mockLogger)).rejects.toThrow(
|
||||
'User with ID 999 not found.',
|
||||
);
|
||||
@@ -684,7 +687,7 @@ describe('Admin DB Service', () => {
|
||||
|
||||
it('should re-throw a generic error if the database query fails for other reasons', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(adminRepo.updateUserRole('1', 'admin', mockLogger)).rejects.toThrow('DB Error');
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, userId: '1', role: 'admin' },
|
||||
@@ -697,7 +700,7 @@ describe('Admin DB Service', () => {
|
||||
const dbError = new Error('violates foreign key constraint');
|
||||
// Create a more specific type for the error object to avoid using 'any'
|
||||
(dbError as Error & { code: string }).code = '23503';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(
|
||||
adminRepo.updateUserRole('non-existent-user', 'admin', mockLogger),
|
||||
@@ -710,4 +713,28 @@ describe('Admin DB Service', () => {
|
||||
'Database error in updateUserRole',
|
||||
);
|
||||
});
|
||||
|
||||
describe('getFlyersForReview', () => {
|
||||
it('should retrieve flyers with "needs_review" status', async () => {
|
||||
const mockFlyers: Flyer[] = [createMockFlyer({ status: 'needs_review' })];
|
||||
mockDb.query.mockResolvedValue({ rows: mockFlyers });
|
||||
|
||||
const result = await adminRepo.getFlyersForReview(mockLogger);
|
||||
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining("WHERE f.status = 'needs_review'"),
|
||||
);
|
||||
expect(result).toEqual(mockFlyers);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(adminRepo.getFlyersForReview(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve flyers for review.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith({ err: dbError }, 'Database error in getFlyersForReview');
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
@@ -13,12 +13,15 @@ import {
|
||||
Receipt,
|
||||
AdminUserView,
|
||||
Profile,
|
||||
Flyer,
|
||||
} from '../../types';
|
||||
|
||||
export class AdminRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
@@ -612,4 +615,32 @@ export class AdminRepository {
|
||||
throw error; // Re-throw to be handled by the route
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Retrieves all flyers that have been flagged with a 'needs_review' status.
|
||||
* @param logger The logger instance.
|
||||
* @returns A promise that resolves to an array of Flyer objects.
|
||||
*/
|
||||
async getFlyersForReview(logger: Logger): Promise<Flyer[]> {
|
||||
try {
|
||||
const query = `
|
||||
SELECT
|
||||
f.*,
|
||||
json_build_object(
|
||||
'store_id', s.store_id,
|
||||
'name', s.name,
|
||||
'logo_url', s.logo_url
|
||||
) as store
|
||||
FROM public.flyers f
|
||||
LEFT JOIN public.stores s ON f.store_id = s.store_id
|
||||
WHERE f.status = 'needs_review'
|
||||
ORDER BY f.created_at DESC;
|
||||
`;
|
||||
const res = await this.db.query<Flyer>(query);
|
||||
return res.rows;
|
||||
} catch (error) {
|
||||
logger.error({ err: error }, 'Database error in getFlyersForReview');
|
||||
throw new Error('Failed to retrieve flyers for review.');
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -7,7 +7,6 @@ vi.unmock('./budget.db');
|
||||
|
||||
import { BudgetRepository } from './budget.db';
|
||||
import type { Pool, PoolClient } from 'pg';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import type { Budget, SpendingByCategory } from '../../types';
|
||||
|
||||
// Mock the logger to prevent console output during tests
|
||||
@@ -42,11 +41,14 @@ import { withTransaction } from './connection.db';
|
||||
|
||||
describe('Budget DB Service', () => {
|
||||
let budgetRepo: BudgetRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
budgetRepo = new BudgetRepository(mockPoolInstance as unknown as Pool);
|
||||
// Instantiate the repository with the minimal mock db for each test
|
||||
budgetRepo = new BudgetRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getBudgetsForUser', () => {
|
||||
@@ -63,11 +65,11 @@ describe('Budget DB Service', () => {
|
||||
updated_at: new Date().toISOString(),
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockBudgets });
|
||||
mockDb.query.mockResolvedValue({ rows: mockBudgets });
|
||||
|
||||
const result = await budgetRepo.getBudgetsForUser('user-123', mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.budgets WHERE user_id = $1 ORDER BY start_date DESC',
|
||||
['user-123'],
|
||||
);
|
||||
@@ -75,15 +77,15 @@ describe('Budget DB Service', () => {
|
||||
});
|
||||
|
||||
it('should return an empty array if the user has no budgets', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
const result = await budgetRepo.getBudgetsForUser('user-123', mockLogger);
|
||||
expect(result).toEqual([]);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(expect.any(String), ['user-123']);
|
||||
expect(mockDb.query).toHaveBeenCalledWith(expect.any(String), ['user-123']);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(budgetRepo.getBudgetsForUser('user-123', mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve budgets.',
|
||||
);
|
||||
@@ -236,11 +238,11 @@ describe('Budget DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
updated_at: new Date().toISOString(),
|
||||
};
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [mockUpdatedBudget], rowCount: 1 });
|
||||
mockDb.query.mockResolvedValue({ rows: [mockUpdatedBudget], rowCount: 1 });
|
||||
|
||||
const result = await budgetRepo.updateBudget(1, 'user-123', budgetUpdates, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('UPDATE public.budgets SET'),
|
||||
[budgetUpdates.name, budgetUpdates.amount_cents, undefined, undefined, 1, 'user-123'],
|
||||
);
|
||||
@@ -249,7 +251,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if no rows are updated', async () => {
|
||||
// Arrange: Mock the query to return 0 rows affected
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
|
||||
await expect(
|
||||
budgetRepo.updateBudget(999, 'user-123', { name: 'Fail' }, mockLogger),
|
||||
@@ -258,7 +260,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
budgetRepo.updateBudget(1, 'user-123', { name: 'Fail' }, mockLogger),
|
||||
).rejects.toThrow('Failed to update budget.');
|
||||
@@ -271,9 +273,9 @@ describe('Budget DB Service', () => {
|
||||
|
||||
describe('deleteBudget', () => {
|
||||
it('should execute a DELETE query with user ownership check', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rowCount: 1, command: 'DELETE', rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rowCount: 1, command: 'DELETE', rows: [] });
|
||||
await budgetRepo.deleteBudget(1, 'user-123', mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'DELETE FROM public.budgets WHERE budget_id = $1 AND user_id = $2',
|
||||
[1, 'user-123'],
|
||||
);
|
||||
@@ -281,7 +283,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if no rows are deleted', async () => {
|
||||
// Arrange: Mock the query to return 0 rows affected
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
mockDb.query.mockResolvedValue({ rows: [], rowCount: 0 });
|
||||
|
||||
await expect(budgetRepo.deleteBudget(999, 'user-123', mockLogger)).rejects.toThrow(
|
||||
'Budget not found or user does not have permission to delete.',
|
||||
@@ -290,7 +292,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(budgetRepo.deleteBudget(1, 'user-123', mockLogger)).rejects.toThrow(
|
||||
'Failed to delete budget.',
|
||||
);
|
||||
@@ -306,7 +308,7 @@ describe('Budget DB Service', () => {
|
||||
const mockSpendingData: SpendingByCategory[] = [
|
||||
{ category_id: 1, category_name: 'Produce', total_spent_cents: 12345 },
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockSpendingData });
|
||||
mockDb.query.mockResolvedValue({ rows: mockSpendingData });
|
||||
|
||||
const result = await budgetRepo.getSpendingByCategory(
|
||||
'user-123',
|
||||
@@ -315,7 +317,7 @@ describe('Budget DB Service', () => {
|
||||
mockLogger,
|
||||
);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.get_spending_by_category($1, $2, $3)',
|
||||
['user-123', '2024-01-01', '2024-01-31'],
|
||||
);
|
||||
@@ -323,7 +325,7 @@ describe('Budget DB Service', () => {
|
||||
});
|
||||
|
||||
it('should return an empty array if there is no spending data', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
const result = await budgetRepo.getSpendingByCategory(
|
||||
'user-123',
|
||||
'2024-01-01',
|
||||
@@ -335,7 +337,7 @@ describe('Budget DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
budgetRepo.getSpendingByCategory('user-123', '2024-01-01', '2024-01-31', mockLogger),
|
||||
).rejects.toThrow('Failed to get spending analysis.');
|
||||
|
||||
@@ -7,9 +7,11 @@ import type { Budget, SpendingByCategory } from '../../types';
|
||||
import { GamificationRepository } from './gamification.db';
|
||||
|
||||
export class BudgetRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -1,9 +1,7 @@
|
||||
// src/services/db/deals.db.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import { DealsRepository } from './deals.db';
|
||||
import type { WatchedItemDeal } from '../../types';
|
||||
import type { Pool } from 'pg';
|
||||
|
||||
// Un-mock the module we are testing to ensure we use the real implementation.
|
||||
vi.unmock('./deals.db');
|
||||
@@ -22,11 +20,17 @@ import { logger as mockLogger } from '../logger.server';
|
||||
describe('Deals DB Service', () => {
|
||||
// Import the Pool type to use for casting the mock instance.
|
||||
let dealsRepo: DealsRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn()
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
dealsRepo = new DealsRepository(mockPoolInstance as unknown as Pool);
|
||||
|
||||
mockDb.query.mockReset()
|
||||
|
||||
// Instantiate the repository with the minimal mock db for each test
|
||||
dealsRepo = new DealsRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('findBestPricesForWatchedItems', () => {
|
||||
@@ -50,14 +54,14 @@ describe('Deals DB Service', () => {
|
||||
valid_to: '2025-12-24',
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockDeals });
|
||||
mockDb.query.mockResolvedValue({ rows: mockDeals });
|
||||
|
||||
// Act
|
||||
const result = await dealsRepo.findBestPricesForWatchedItems('user-123', mockLogger);
|
||||
|
||||
// Assert
|
||||
expect(result).toEqual(mockDeals);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM flyer_items fi'),
|
||||
['user-123'],
|
||||
);
|
||||
@@ -68,7 +72,7 @@ describe('Deals DB Service', () => {
|
||||
});
|
||||
|
||||
it('should return an empty array if no deals are found', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
mockDb.query.mockResolvedValue({ rows: [] });
|
||||
|
||||
const result = await dealsRepo.findBestPricesForWatchedItems(
|
||||
'user-with-no-deals',
|
||||
@@ -80,7 +84,7 @@ describe('Deals DB Service', () => {
|
||||
|
||||
it('should re-throw the error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Connection Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(dealsRepo.findBestPricesForWatchedItems('user-1', mockLogger)).rejects.toThrow(
|
||||
dbError,
|
||||
|
||||
@@ -6,9 +6,11 @@ import type { Logger } from 'pino';
|
||||
import { logger as globalLogger } from '../logger.server';
|
||||
|
||||
export class DealsRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -40,8 +40,7 @@ describe('Flyer DB Service', () => {
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
// In a transaction, `pool.connect()` returns a client. That client has a `release` method.
|
||||
//In a transaction, `pool.connect()` returns a client. That client has a `release` method.
|
||||
// For these tests, we simulate this by having `connect` resolve to the pool instance itself,
|
||||
// and we ensure the `release` method is mocked on that instance.
|
||||
const mockClient = { ...mockPoolInstance, release: vi.fn() } as unknown as PoolClient;
|
||||
@@ -56,7 +55,7 @@ describe('Flyer DB Service', () => {
|
||||
const result = await flyerRepo.findOrCreateStore('Existing Store', mockLogger);
|
||||
expect(result).toBe(1);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'SELECT store_id FROM public.stores WHERE name = $1',
|
||||
expect.stringContaining('SELECT store_id FROM public.stores WHERE name = $1'),
|
||||
['Existing Store'],
|
||||
);
|
||||
});
|
||||
@@ -64,11 +63,11 @@ describe('Flyer DB Service', () => {
|
||||
it('should create a new store if it does not exist', async () => {
|
||||
mockPoolInstance.query
|
||||
.mockResolvedValueOnce({ rows: [] }) // First SELECT finds nothing
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] }); // INSERT returns new ID
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 2 }] })
|
||||
const result = await flyerRepo.findOrCreateStore('New Store', mockLogger);
|
||||
expect(result).toBe(2);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
'INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id',
|
||||
expect.stringContaining('INSERT INTO public.stores (name) VALUES ($1) RETURNING store_id'),
|
||||
['New Store'],
|
||||
);
|
||||
});
|
||||
@@ -83,11 +82,11 @@ describe('Flyer DB Service', () => {
|
||||
.mockResolvedValueOnce({ rows: [{ store_id: 3 }] }); // Second SELECT finds the store
|
||||
|
||||
const result = await flyerRepo.findOrCreateStore('Racy Store', mockLogger);
|
||||
expect(result).toBe(3);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(3);
|
||||
});
|
||||
expect(result).toBe(3);
|
||||
//expect(mockDb.query).toHaveBeenCalledTimes(3);
|
||||
});
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
await expect(flyerRepo.findOrCreateStore('Any Store', mockLogger)).rejects.toThrow(
|
||||
@@ -129,6 +128,7 @@ describe('Flyer DB Service', () => {
|
||||
valid_from: '2024-01-01',
|
||||
valid_to: '2024-01-07',
|
||||
store_address: '123 Test St',
|
||||
status: 'processed',
|
||||
item_count: 10,
|
||||
uploaded_by: 'user-1',
|
||||
};
|
||||
@@ -139,7 +139,7 @@ describe('Flyer DB Service', () => {
|
||||
|
||||
expect(result).toEqual(mockFlyer);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(1);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('INSERT INTO flyers'),
|
||||
[
|
||||
'test.jpg',
|
||||
@@ -150,6 +150,7 @@ describe('Flyer DB Service', () => {
|
||||
'2024-01-01',
|
||||
'2024-01-07',
|
||||
'123 Test St',
|
||||
'processed',
|
||||
10,
|
||||
'user-1',
|
||||
],
|
||||
|
||||
@@ -13,9 +13,11 @@ import type {
|
||||
} from '../../types';
|
||||
|
||||
export class FlyerRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
@@ -78,10 +80,10 @@ export class FlyerRepository {
|
||||
try {
|
||||
const query = `
|
||||
INSERT INTO flyers (
|
||||
file_name, image_url, icon_url, checksum, store_id, valid_from, valid_to,
|
||||
store_address, item_count, uploaded_by
|
||||
file_name, image_url, icon_url, checksum, store_id, valid_from, valid_to, store_address,
|
||||
status, item_count, uploaded_by
|
||||
)
|
||||
VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10)
|
||||
VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11)
|
||||
RETURNING *;
|
||||
`;
|
||||
const values = [
|
||||
@@ -93,8 +95,9 @@ export class FlyerRepository {
|
||||
flyerData.valid_from, // $6
|
||||
flyerData.valid_to, // $7
|
||||
flyerData.store_address, // $8
|
||||
flyerData.item_count, // $9
|
||||
flyerData.uploaded_by, // $10
|
||||
flyerData.status, // $9
|
||||
flyerData.item_count, // $10
|
||||
flyerData.uploaded_by, // $11
|
||||
];
|
||||
|
||||
const result = await this.db.query<Flyer>(query, values);
|
||||
|
||||
@@ -22,14 +22,18 @@ import { logger as mockLogger } from '../logger.server';
|
||||
|
||||
describe('Gamification DB Service', () => {
|
||||
let gamificationRepo: GamificationRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset the global mock's call history before each test.
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
gamificationRepo = new GamificationRepository(mockPoolInstance as unknown as Pool);
|
||||
});
|
||||
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
gamificationRepo = new GamificationRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('getAllAchievements', () => {
|
||||
it('should execute the correct SELECT query and return achievements', async () => {
|
||||
const mockAchievements: Achievement[] = [
|
||||
@@ -42,11 +46,11 @@ describe('Gamification DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockAchievements });
|
||||
mockDb.query.mockResolvedValue({ rows: mockAchievements });
|
||||
|
||||
const result = await gamificationRepo.getAllAchievements(mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT * FROM public.achievements ORDER BY points_value ASC, name ASC',
|
||||
);
|
||||
expect(result).toEqual(mockAchievements);
|
||||
@@ -54,7 +58,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(gamificationRepo.getAllAchievements(mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve achievements.',
|
||||
);
|
||||
@@ -79,11 +83,11 @@ describe('Gamification DB Service', () => {
|
||||
created_at: new Date().toISOString(),
|
||||
},
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockUserAchievements });
|
||||
mockDb.query.mockResolvedValue({ rows: mockUserAchievements });
|
||||
|
||||
const result = await gamificationRepo.getUserAchievements('user-123', mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('FROM public.user_achievements ua'),
|
||||
['user-123'],
|
||||
);
|
||||
@@ -92,7 +96,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(gamificationRepo.getUserAchievements('user-123', mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve user achievements.',
|
||||
);
|
||||
@@ -105,10 +109,10 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
describe('awardAchievement', () => {
|
||||
it('should call the award_achievement database function with the correct parameters', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] }); // The function returns void
|
||||
mockDb.query.mockResolvedValue({ rows: [] }); // The function returns void
|
||||
await gamificationRepo.awardAchievement('user-123', 'Test Achievement', mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
'SELECT public.award_achievement($1, $2)',
|
||||
['user-123', 'Test Achievement'],
|
||||
);
|
||||
@@ -117,7 +121,7 @@ describe('Gamification DB Service', () => {
|
||||
it('should throw ForeignKeyConstraintError if user or achievement does not exist', async () => {
|
||||
const dbError = new Error('violates foreign key constraint');
|
||||
(dbError as Error & { code: string }).code = '23503';
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
gamificationRepo.awardAchievement(
|
||||
'non-existent-user',
|
||||
@@ -133,7 +137,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(
|
||||
gamificationRepo.awardAchievement('user-123', 'Test Achievement', mockLogger),
|
||||
).rejects.toThrow('Failed to award achievement.');
|
||||
@@ -147,15 +151,14 @@ describe('Gamification DB Service', () => {
|
||||
describe('getLeaderboard', () => {
|
||||
it('should execute the correct SELECT query with a LIMIT and return leaderboard users', async () => {
|
||||
const mockLeaderboard: LeaderboardUser[] = [
|
||||
{ user_id: 'user-1', full_name: 'User One', avatar_url: null, points: 500, rank: '1' },
|
||||
{ user_id: 'user-2', full_name: 'User Two', avatar_url: null, points: 450, rank: '2' },
|
||||
{ user_id: 'user-1', full_name: 'User One', avatar_url: null, points: 500, rank: '1' },
|
||||
{ user_id: 'user-2', full_name: 'User Two', avatar_url: null, points: 450, rank: '2' }
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockLeaderboard });
|
||||
mockDb.query.mockResolvedValue({ rows: mockLeaderboard });
|
||||
|
||||
const result = await gamificationRepo.getLeaderboard(10, mockLogger);
|
||||
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledTimes(1);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect(mockDb.query).toHaveBeenCalledTimes(1);
|
||||
expect(mockDb.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('RANK() OVER (ORDER BY points DESC)'),
|
||||
[10],
|
||||
);
|
||||
@@ -164,7 +167,7 @@ describe('Gamification DB Service', () => {
|
||||
|
||||
it('should throw an error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
mockDb.query.mockRejectedValue(dbError);
|
||||
await expect(gamificationRepo.getLeaderboard(10, mockLogger)).rejects.toThrow(
|
||||
'Failed to retrieve leaderboard.',
|
||||
);
|
||||
|
||||
@@ -6,9 +6,11 @@ import type { Logger } from 'pino';
|
||||
import { Achievement, UserAchievement, LeaderboardUser } from '../../types';
|
||||
|
||||
export class GamificationRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -2,7 +2,6 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import type { Pool } from 'pg';
|
||||
|
||||
// Un-mock the module we are testing to ensure we use the real implementation.
|
||||
vi.unmock('./notification.db');
|
||||
|
||||
import { NotificationRepository } from './notification.db';
|
||||
@@ -11,6 +10,7 @@ import { ForeignKeyConstraintError, NotFoundError } from './errors.db';
|
||||
import type { Notification } from '../../types';
|
||||
import { createMockNotification } from '../../tests/utils/mockFactories';
|
||||
|
||||
|
||||
// Mock the logger to prevent console output during tests
|
||||
vi.mock('../logger.server', () => ({
|
||||
logger: {
|
||||
@@ -24,10 +24,14 @@ import { logger as mockLogger } from '../logger.server';
|
||||
|
||||
describe('Notification DB Service', () => {
|
||||
let notificationRepo: NotificationRepository;
|
||||
const mockDb = {
|
||||
query: vi.fn(),
|
||||
};
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Instantiate the repository with the mock pool for each test
|
||||
|
||||
notificationRepo = new NotificationRepository(mockPoolInstance as unknown as Pool);
|
||||
});
|
||||
|
||||
|
||||
@@ -6,9 +6,11 @@ import type { Logger } from 'pino';
|
||||
import type { Notification } from '../../types';
|
||||
|
||||
export class NotificationRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
@@ -16,9 +16,11 @@ import {
|
||||
} from '../../types';
|
||||
|
||||
export class PersonalizationRepository {
|
||||
private db: Pool | PoolClient;
|
||||
// The repository only needs an object with a `query` method, matching the Pool/PoolClient interface.
|
||||
// Using `Pick` makes this dependency explicit and simplifies testing by reducing the mock surface.
|
||||
private db: Pick<Pool | PoolClient, 'query'>;
|
||||
|
||||
constructor(db: Pool | PoolClient = getPool()) {
|
||||
constructor(db: Pick<Pool | PoolClient, 'query'> = getPool()) {
|
||||
this.db = db;
|
||||
}
|
||||
|
||||
|
||||
96
src/services/db/price.db.test.ts
Normal file
96
src/services/db/price.db.test.ts
Normal file
@@ -0,0 +1,96 @@
|
||||
// src/services/db/price.db.test.ts
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { mockPoolInstance } from '../../tests/setup/tests-setup-unit';
|
||||
import { getPool } from './connection.db';
|
||||
import { priceRepo } from './price.db';
|
||||
import type { PriceHistoryData } from '../../types';
|
||||
|
||||
// Un-mock the module we are testing to ensure we use the real implementation.
|
||||
vi.unmock('./price.db');
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('./connection.db', () => ({
|
||||
getPool: vi.fn(),
|
||||
}));
|
||||
|
||||
vi.mock('../logger.server', () => ({
|
||||
logger: {
|
||||
info: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
error: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
},
|
||||
}));
|
||||
|
||||
import { logger as mockLogger } from '../logger.server';
|
||||
|
||||
describe('Price DB Service', () => {
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
// Make getPool return our mock instance for each test
|
||||
vi.mocked(getPool).mockReturnValue(mockPoolInstance as any);
|
||||
});
|
||||
|
||||
describe('getPriceHistory', () => {
|
||||
it('should return an empty array if masterItemIds is empty and not query the db', async () => {
|
||||
const result = await priceRepo.getPriceHistory([], mockLogger);
|
||||
expect(result).toEqual([]);
|
||||
expect(mockPoolInstance.query).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should execute the correct query with default limit and offset', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
await priceRepo.getPriceHistory([1, 2], mockLogger);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('LIMIT $2 OFFSET $3'),
|
||||
[[1, 2], 1000, 0],
|
||||
);
|
||||
});
|
||||
|
||||
it('should execute the correct query with provided limit and offset', async () => {
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: [] });
|
||||
await priceRepo.getPriceHistory([1, 2], mockLogger, 50, 10);
|
||||
expect(mockPoolInstance.query).toHaveBeenCalledWith(
|
||||
expect.stringContaining('LIMIT $2 OFFSET $3'),
|
||||
[[1, 2], 50, 10],
|
||||
);
|
||||
});
|
||||
|
||||
it('should return price history data on success', async () => {
|
||||
const mockHistory: PriceHistoryData[] = [
|
||||
{ master_item_id: 1, price_in_cents: 199, date: '2024-01-01' },
|
||||
{ master_item_id: 1, price_in_cents: 209, date: '2024-01-08' },
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockHistory });
|
||||
|
||||
const result = await priceRepo.getPriceHistory([1], mockLogger);
|
||||
expect(result).toEqual(mockHistory);
|
||||
});
|
||||
|
||||
it('should log the result count on success', async () => {
|
||||
const mockHistory: PriceHistoryData[] = [
|
||||
{ master_item_id: 1, price_in_cents: 199, date: '2024-01-01' },
|
||||
];
|
||||
mockPoolInstance.query.mockResolvedValue({ rows: mockHistory });
|
||||
|
||||
await priceRepo.getPriceHistory([1], mockLogger, 50, 10);
|
||||
expect(mockLogger.debug).toHaveBeenCalledWith(
|
||||
{ count: 1, itemIds: 1, limit: 50, offset: 10 },
|
||||
'Fetched price history from database.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should throw a generic error if the database query fails', async () => {
|
||||
const dbError = new Error('DB Connection Error');
|
||||
mockPoolInstance.query.mockRejectedValue(dbError);
|
||||
|
||||
await expect(priceRepo.getPriceHistory([1], mockLogger, 50, 10)).rejects.toThrow(
|
||||
'Failed to retrieve price history.',
|
||||
);
|
||||
expect(mockLogger.error).toHaveBeenCalledWith(
|
||||
{ err: dbError, masterItemIds: [1], limit: 50, offset: 10 },
|
||||
'Database error in getPriceHistory',
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -43,11 +43,19 @@ export const priceRepo = {
|
||||
LIMIT $2 OFFSET $3;
|
||||
`;
|
||||
|
||||
const result = await getPool().query(query, [masterItemIds, limit, offset]);
|
||||
logger.debug(
|
||||
{ count: result.rows.length, itemIds: masterItemIds.length, limit, offset },
|
||||
'Fetched price history from database.',
|
||||
);
|
||||
return result.rows;
|
||||
try {
|
||||
const result = await getPool().query(query, [masterItemIds, limit, offset]);
|
||||
logger.debug(
|
||||
{ count: result.rows.length, itemIds: masterItemIds.length, limit, offset },
|
||||
'Fetched price history from database.',
|
||||
);
|
||||
return result.rows;
|
||||
} catch (error) {
|
||||
logger.error(
|
||||
{ err: error, masterItemIds, limit, offset },
|
||||
'Database error in getPriceHistory',
|
||||
);
|
||||
throw new Error('Failed to retrieve price history.');
|
||||
}
|
||||
},
|
||||
};
|
||||
@@ -29,6 +29,7 @@ vi.mock('./logger.server', () => ({
|
||||
info: vi.fn(),
|
||||
debug: vi.fn(),
|
||||
error: vi.fn(),
|
||||
child: vi.fn().mockReturnThis(),
|
||||
},
|
||||
}));
|
||||
|
||||
@@ -37,10 +38,13 @@ import {
|
||||
sendPasswordResetEmail,
|
||||
sendWelcomeEmail,
|
||||
sendDealNotificationEmail,
|
||||
processEmailJob,
|
||||
} from './emailService.server';
|
||||
import type { WatchedItemDeal } from '../types';
|
||||
import { createMockWatchedItemDeal } from '../tests/utils/mockFactories';
|
||||
import { logger } from './logger.server';
|
||||
import type { Job } from 'bullmq';
|
||||
import type { EmailJobData } from '../types/job-data';
|
||||
|
||||
describe('Email Service (Server)', () => {
|
||||
beforeEach(async () => {
|
||||
@@ -219,4 +223,51 @@ describe('Email Service (Server)', () => {
|
||||
);
|
||||
});
|
||||
});
|
||||
|
||||
describe('processEmailJob', () => {
|
||||
const mockJobData: EmailJobData = {
|
||||
to: 'job@example.com',
|
||||
subject: 'Job Email',
|
||||
html: '<p>Job</p>',
|
||||
text: 'Job',
|
||||
};
|
||||
|
||||
const createMockJob = (data: EmailJobData): Job<EmailJobData> =>
|
||||
({
|
||||
id: 'job-123',
|
||||
name: 'email-job',
|
||||
data,
|
||||
attemptsMade: 1,
|
||||
} as unknown as Job<EmailJobData>);
|
||||
|
||||
it('should call sendMail with job data and log success', async () => {
|
||||
const job = createMockJob(mockJobData);
|
||||
mocks.sendMail.mockResolvedValue({ messageId: 'job-test-id' });
|
||||
|
||||
await processEmailJob(job);
|
||||
|
||||
expect(mocks.sendMail).toHaveBeenCalledTimes(1);
|
||||
const mailOptions = mocks.sendMail.mock.calls[0][0];
|
||||
expect(mailOptions.to).toBe(mockJobData.to);
|
||||
expect(mailOptions.subject).toBe(mockJobData.subject);
|
||||
expect(logger.info).toHaveBeenCalledWith('Picked up email job.');
|
||||
expect(logger.info).toHaveBeenCalledWith(
|
||||
{ to: 'job@example.com', subject: 'Job Email', messageId: 'job-test-id' },
|
||||
'Email sent successfully.',
|
||||
);
|
||||
});
|
||||
|
||||
it('should log an error and re-throw if sendMail fails', async () => {
|
||||
const job = createMockJob(mockJobData);
|
||||
const emailError = new Error('SMTP Connection Failed');
|
||||
mocks.sendMail.mockRejectedValue(emailError);
|
||||
|
||||
await expect(processEmailJob(job)).rejects.toThrow(emailError);
|
||||
|
||||
expect(logger.error).toHaveBeenCalledWith(
|
||||
{ err: emailError, jobData: mockJobData, attemptsMade: 1 },
|
||||
'Email job failed.',
|
||||
);
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user